Deployment of Machine Learning Models
- Description
- Curriculum
- FAQ
- Reviews
This course is for AI and ML Engineers, Practitioners and Researchers who already built an awesome Deep Learning model, and they have a great idea for an app. But they discovered that it is not straight forward to deploy their model in a production App. Another example, say you want to build a robot that uses the Camera sensor to perceive the surrounding environment, build a map of it and eventually navigate it. Here also you discover that you still have a long Journey to go after your model is already performing great on your training machine. Finally, Software Engineers, who have their primary job is to build a working system or an app, often find themselves in a situation where they need to integrate an AI model in their software, which happens a lot today with the expansion of AI applications. They might get this model from a research team in their firm or company, or even use an API or pre-trained model on the internet to do their task.
We cover all those deployment scenarios, covering the journey from working trained model to an optimized deployed model. Our focus will be on CV deployment mainly. We cover Mobile deployment like on Android devices, Edge deployment on Embedded boards like Rasperry Pi, and Browser deployment where your AI model is running in the browser like Chrome, Edge, Safari or any other browser. Also, we cover server deployment scenarios, which are often found in highly scalable apps and systems with millions of users, and also in industrial scenarios like AI visual inspection in factories.
While the course is mostly practical, focusing on “How” things are done and the best way of doing it, we cover also some theoretical parts about the “what” and “why” those techniques are used.
This requires sometimes to understand new types of convolution operations that are optimized for speed and memory, or understanding some model compression techniques that makes them suitable for Embedded and Edge deployments, which was not in scope during building the initial model that was already performing great.
-
6Model compression overviewVideo lesson
-
7PruningVideo lesson
-
8DistillationVideo lesson
-
9QuantizationVideo lesson
-
10Compression pipelinesVideo lesson
-
11Convolution FactorizationVideo lesson
-
121x1 ConvolutionVideo lesson
-
13Group convolution and 1x1 group convolutionVideo lesson
-
14Depth-wise separable convolutionVideo lesson
-
15ConDenseNetVideo lesson
-
16ResNet, ResNext and 1x1 skip connectionsVideo lesson
-
17InceptionNet and 1x1 bottleneckVideo lesson
-
18SqueezeNetVideo lesson
-
19MobileNet v1Video lesson
-
20ShuffleNetVideo lesson
-
21MobileNet v2Video lesson
-
22Cheat Sheet: Optimized pre-trained models vs Convolution typesVideo lesson
-
23NASNetVideo lesson
-
24MobileNet v3Video lesson
-
25EfficientNetVideo lesson
-
26TFLite overview and converter processVideo lesson
-
27Rasperry PiVideo lesson
-
28TFLite + Rasperry PiVideo lesson
-
29Intel MovidiusVideo lesson
-
30TFLite + AndroidVideo lesson
-
31Android Image Classification App on Static imagesVideo lesson
-
32Android Image Classification App on Camera FeedVideo lesson
-
33Android Object Detection App on Camera feedVideo lesson
-
34Browser based deployment using TFJSVideo lesson
-
35Device deployment wrap upVideo lesson
-
36Client server overview and model optionsVideo lesson
-
37Cloud based deployment and TFHubVideo lesson
-
38Object detection use case with TF-API Blackbox Model GardenVideo lesson
-
39TF-API Whitebox scenario: Fine tune on custom data for Masks DetectionVideo lesson
-
40TF-API From scratch whitebox Yolov5 and DETR on SF Street Signs DetectionVideo lesson
-
41Serving qualitiesVideo lesson
-
42Serving landscapeVideo lesson
-
43FlaskVideo lesson
-
44Django on AWS EC2Video lesson
-
45TFServing Native DeploymentVideo lesson
-
46DockerVideo lesson
-
47TFServing + Docker DeploymentVideo lesson
External Links May Contain Affiliate Links read more