Course Outline

Preparing Machine Learning Models for Deployment

  • Packaging models with Docker
  • Exporting models from TensorFlow and PyTorch
  • Versioning and storage considerations

Model Serving on Kubernetes

  • Overview of inference servers
  • Deploying TensorFlow Serving and TorchServe
  • Setting up model endpoints

Inference Optimization Techniques

  • Batching strategies
  • Concurrent request handling
  • Latency and throughput tuning

Autoscaling ML Workloads

  • Horizontal Pod Autoscaler (HPA)
  • Vertical Pod Autoscaler (VPA)
  • Kubernetes Event-Driven Autoscaling (KEDA)

GPU Provisioning and Resource Management

  • Configuring GPU nodes
  • NVIDIA device plugin overview
  • Resource requests and limits for ML workloads

Model Rollout and Release Strategies

  • Blue/green deployments
  • Canary rollout patterns
  • A/B testing for model evaluation

Monitoring and Observability for ML in Production

  • Metrics for inference workloads
  • Logging and tracing practices
  • Dashboards and alerting

Security and Reliability Considerations

  • Securing model endpoints
  • Network policies and access control
  • Ensuring high availability

Summary and Next Steps

Requirements

  • An understanding of containerized application workflows
  • Experience with Python-based machine learning models
  • Familiarity with Kubernetes fundamentals

Audience

  • ML engineers
  • DevOps engineers
  • Platform engineering teams
 14 Hours

Custom Corporate Training

Training solutions designed exclusively for businesses.

  • Customized Content: We adapt the syllabus and practical exercises to the real goals and needs of your project.
  • Flexible Schedule: Dates and times adapted to your team's agenda.
  • Format: Online (live), In-company (at your offices), or Hybrid.
Investment

Price per private group, online live training, starting from 3200 € + VAT*

Contact us for an exact quote and to hear our latest promotions

Testimonials (5)

Upcoming Courses

Related Categories