Skip to Content
Distributed Machine Learning with Python
book

Distributed Machine Learning with Python

by Guanhua Wang
April 2022
Intermediate to advanced
284 pages
5h 53m
English
Packt Publishing
Content preview from Distributed Machine Learning with Python

Chapter 4: Bottlenecks and Solutions

Using the code we designed in Chapter 3, Building a Data Parallel Training and Serving Pipeline, we can build data parallel training and serving pipelines using either the parameter server or the All-Reduce paradigm. Similar to what we did in Chapter 3, Building a Data Parallel Training and Serving Pipeline, in this chapter, we will focus on the more widely used All-Reduce paradigm.

In this chapter, we will discuss the shortcomings in the current data parallel training and serving pipelines. For practical system bottleneck discussions, we will make the following assumptions:

  • We use homogenous accelerators for all our model training nodes.
  • Compared to CPU memory (that is, main memory), the on-device memory ...
Become an O’Reilly member and get unlimited access to this title plus top books and audiobooks from O’Reilly and nearly 200 top publishers, thousands of courses curated by job role, 150+ live events each month,
and much more.
Start your free trial

You might also like

Interpretable Machine Learning with Python

Interpretable Machine Learning with Python

Serg Masís
Distributed Computing with Python

Distributed Computing with Python

Francesco Pierfederici

Publisher Resources

ISBN: 9781801815697Supplemental Content