TensorRT for Tensorflow (Free Coursera Project)
Coursera has opened for free the Optimize TensorFlow Models For Deployment with TensorRT project.
Basics
Goal: Speed up Tensorflow model inference performance.
Time length: 1.5-2 hours
Prerequisite: experience with pre-trained Tensorflow models.
Reference: Accelerating Inference in TensorFlow with TensorRT User Guide
My rating: ★★★★
Procs
- Nice explanatory videos.
- You have to type a lot, but it does encourage you to understand things better (didactic trick).
- Comparison to Nvidia’s alternative:
- Seems to explain better
- Colab instead fully pre-configured dedicated virtual machine
- Toy database instead of the ‘real’ one
- Headache of outdated Colab configuration
- Certificate
Notes
-
Making things work require some fixes, since the configuration is outdated. The following installation procedure is to be used (Nvidia doc):
%%bash sudo apt install python3-libnvinfer python3 -m pip install --upgrade tensorrt
The result may also be verified by
import tensorrt print(tensorrt.__version__)
-
INT8
conversion has taken me about 10 minutes on Colab. The videos show that Colab worked twice faster a few years ago. - The certification exam is trivial.
- Supplementary video and presentation on the subject by Nvidia
Enjoy Reading This Article?
Here are some more articles you might like to read next: