Overview
Dive into a 38-minute technical deep dive on TensorFlow Lite, presented by Software Engineer Jared Duke from the TensorFlow team. Explore how TensorFlow Lite enables deployment of machine learning models on mobile and IoT devices. Learn about the differences between TensorFlow and TensorFlow Lite, the interpreter, acceleration techniques using delegates, model conversion in Python, and inference in Java. Discover selective registration processes in Bazel and C++, performance optimization strategies, benchmarking on Android, and inference with NNAPI and GPU passthrough. Gain insights into fast execution methods, post-training quantization techniques for model conversion, and available documentation resources. Get a glimpse of the TensorFlow Lite roadmap and participate in a Q&A session to deepen your understanding of this powerful tool for on-device machine learning.
Syllabus
Intro
Outline
Why on device?
TensorFlow vs TensorFlow Lite
Why not TFMobile?
What is TensorFlow Lite?
Interpreter
Acceleration (Delegates)
Model conversion (Python)
Inference (Java)
Selective Registration (Bazel)
Selective Registration (C++)
Performance
Benchmarking (Android)
Inference w/ NNAPI
Inference w/ GPU passthrough
Fast execution
Model conversion w/ Post-Training Quant (Hybrid)
Optimization
Model conversion w/ Post-Training Quant (Full)
Documentation
Model repository
TensorFlow Lite Roadmap
Questions?
Taught by
TensorFlow