
Speaker "Chris Fregly" Details Back


-
Name
Chris Fregly
-
Company
Amazon Web Services
-
Designation
Developer Advocate
Topic
Using AWS SageMaker, Kubernetes, and PipelineAI for High Performance, Hybrid-Cloud Distributed TensorFlow Model Training and Serving with GPUs.
Abstract
In this talk, I will demonstrate how to train, optimize, and serve distributed machine learning models across various environments including the following:
1) Local Laptop
2) Kubernetes Cluster (Running Anywhere)
3) AWS's New SageMaker Service
I'll also present some post-training model-optimization techniques to improve model serving performance for TensorFlow running on GPUs. These techniques include 16-bit model training, neural network layer fusing, and 8-bit weight quantization.
Lastly, I'll discuss alternate runtimes for TensorFlow on GPUs including and TensorFlow Lite and Nvidia's TensorRT.