11
Machine Learning Part 3 – Transformers and Large Language Models
In this chapter, we will cover the currently best-performing techniques in natural language processing (NLP) – transformers and pretrained models. We will discuss the concepts behind transformers and include examples of using transformers and large language models (LLMs) for text classification. The code for this chapter will be based on the TensorFlow/Keras Python libraries and the cloud services provided by OpenAI.
The topics covered in this chapter are important because although transformers and LLMs are only a few years old, they have become state-of-the-art for many different types of NLP applications. In fact, LLM systems such as ChatGPT have been widely covered in the ...
Get Natural Language Understanding with Python now with the O’Reilly learning platform.
O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.