Advanced model deployments with TensorFlow Serving

Video description

TensorFlow Serving is one of the cornerstones in the TensorFlow ecosystem. It has eased the deployment of machine learning models tremendously and led to an acceleration of model deployments. Unfortunately, machine learning engineers aren’t familiar with the details of TensorFlow Serving, and they’re missing out on significant performance increases.

Hannes Hapke (SAP ConcurLabs) provides a brief introduction to TensorFlow Serving, then leads a deep dive into advanced settings and use cases. He introduces advanced concepts and implementation suggestions to increase the performance of the TensorFlow Serving setup, which includes an introduction to how clients can request model meta-information from the model server, an overview of model optimization options for optimal prediction throughput, an introduction to batching requests to improve the throughput performance, an example implementation to support model A/B testing, and an overview of monitoring your TensorFlow Serving setup.

Prerequisite knowledge

  • A basic understanding of Docker functionality and how HTTP requests work
  • General knowledge of machine learning (useful but not required)

What you'll learn

  • Learn how to increase the TensorFlow Serving inference performance, increase the inference response time by tweaking the request payload, and run TensorFlow Serving with NVIDIA's TensorRT for further performance improvements
  • Discover how to configure batch requests in TensorFlow Serving and how to configure TensorFlow Serving to provide A/B Testing capabilities

This session is from the 2019 O'Reilly TensorFlow World Conference in Santa Clara, CA.

Product information

  • Title: Advanced model deployments with TensorFlow Serving
  • Author(s): Hannes Hapke
  • Release date: February 2020
  • Publisher(s): O'Reilly Media, Inc.
  • ISBN: 0636920372585