12
Model Training Optimizations
Before serving pre-trained machine learning models, which we will discuss extensively in Chapter 13, Operationalizing PyTorch Models into Production, we need to train them. In Chapters 2 to 6, we saw the vast expanse of increasingly complex deep learning model architectures. Such gigantic models often have millions and even billions of parameters. The recent (at the time of writing) Pathways Language Model (PaLM) can have up to 540 billion parameters, for example using backpropagation to tune these many parameters requires enormous amounts of memory and compute power. And even then, model training can take days to finish.
In this chapter, we will explore ways of speeding up the model training process by distributing ...
Get Mastering PyTorch - Second Edition now with the O’Reilly learning platform.
O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.