Book description
Master the art of training vision and large language models with conceptual fundaments and industry-expert guidance. Learn about AWS services and design patterns, with relevant coding examples
Key Features
- Learn to develop, train, tune, and apply foundation models with optimized end-to-end pipelines
- Explore large-scale distributed training for models and datasets with AWS and SageMaker examples
- Evaluate, deploy, and operationalize your custom models with bias detection and pipeline monitoring
Book Description
Foundation models have forever changed machine learning. From BERT to ChatGPT, CLIP to Stable Diffusion, when billions of parameters are combined with large datasets and hundreds to thousands of GPUs, the result is nothing short of record-breaking. The recommendations, advice, and code samples in this book will help you pretrain and fine-tune your own foundation models from scratch on AWS and Amazon SageMaker, while applying them to hundreds of use cases across your organization.
With advice from seasoned AWS and machine learning expert Emily Webber, this book helps you learn everything you need to go from project ideation to dataset preparation, training, evaluation, and deployment for large language, vision, and multimodal models. With step-by-step explanations of essential concepts and practical examples, you’ll go from mastering the concept of pretraining to preparing your dataset and model, configuring your environment, training, fine-tuning, evaluating, deploying, and optimizing your foundation models.
You will learn how to apply the scaling laws to distributing your model and dataset over multiple GPUs, remove bias, achieve high throughput, and build deployment pipelines.
By the end of this book, you’ll be well equipped to embark on your own project to pretrain and fine-tune the foundation models of the future.
What you will learn
- Find the right use cases and datasets for pretraining and fine-tuning
- Prepare for large-scale training with custom accelerators and GPUs
- Configure environments on AWS and SageMaker to maximize performance
- Select hyperparameters based on your model and constraints
- Distribute your model and dataset using many types of parallelism
- Avoid pitfalls with job restarts, intermittent health checks, and more
- Evaluate your model with quantitative and qualitative insights
- Deploy your models with runtime improvements and monitoring pipelines
Who this book is for
If you’re a machine learning researcher or enthusiast who wants to start a foundation modelling project, this book is for you. Applied scientists, data scientists, machine learning engineers, solution architects, product managers, and students will all benefit from this book. Intermediate Python is a must, along with introductory concepts of cloud computing. A strong understanding of deep learning fundamentals is needed, while advanced topics will be explained. The content covers advanced machine learning and cloud techniques, explaining them in an actionable, easy-to-understand way.
Table of contents
- Pretrain Vision and Large Language Models in Python
- Foreword
- Contributors
- About the author
- Acknowledgment
- About the reviewer
- Preface
- Part 1: Before Pretraining
- Chapter 1: An Introduction to Pretraining Foundation Models
- Chapter 2: Dataset Preparation: Part One
- Chapter 3: Model Preparation
- Part 2: Configure Your Environment
- Chapter 4: Containers and Accelerators on the Cloud
- Chapter 5: Distribution Fundamentals
-
Chapter 6: Dataset Preparation: Part Two, the Data Loader
- Introducing the data loader in Python
- Building and testing your own data loader – a case study from Stable Diffusion
- Creating embeddings – tokenizers and other key steps for smart features
- Optimizing your data pipeline on Amazon SageMaker
- Transforming deep learning datasets at scale on AWS
- Summary
- References
- Part 3: Train Your Model
- Chapter 7: Finding the Right Hyperparameters
- Chapter 8: Large-Scale Training on SageMaker
-
Chapter 9: Advanced Training Concepts
- Evaluating and improving throughput
- Using Flash Attention to speed up your training runs
- Speeding up your jobs with compilation
- Amazon SageMaker Training Compiler and Neo
- Running compiled models on Amazon’s Trainium and Inferentia custom hardware
- Solving for an optimal training time
- Summary
- References
- Part 4: Evaluate Your Model
- Chapter 10: Fine-Tuning and Evaluating
- Chapter 11: Detecting, Mitigating, and Monitoring Bias
- Chapter 12: How to Deploy Your Model
- Part 5: Deploy Your Model
- Chapter 13: Prompt Engineering
- Chapter 14: MLOps for Vision and Language
- Chapter 15: Future Trends in Pretraining Foundation Models
- Index
- Other Books You May Enjoy
Product information
- Title: Pretrain Vision and Large Language Models in Python
- Author(s):
- Release date: May 2023
- Publisher(s): Packt Publishing
- ISBN: 9781804618257
You might also like
book
Natural Language Understanding with Python
Build advanced NLU systems by utilizing NLP libraries such as NLTK, SpaCy, BERT, and OpenAI; ML …
book
Deep Learning with Python, Second Edition
Printed in full color! Unlock the groundbreaking advances of deep learning with this extensively revised new …
video
Deep Learning with TensorFlow, Keras, and PyTorch
7+ Hours of Video Instruction An intuitive, application-focused introduction to deep learning and TensorFlow, Keras, and …
video
Quick Guide to ChatGPT, Embeddings, and Other Large Language Models (LLMs)
9+ Hours of Video Instruction Learn how to use and launch large language models (LLMs) like …