Thank you for sending your enquiry! One of our team members will contact you shortly.
Thank you for sending your booking! One of our team members will contact you shortly.
Course Outline
Preparing Machine Learning Models for Deployment
- Packaging models with Docker to ensure consistent and reproducible environments for government operations.
- Exporting models from TensorFlow and PyTorch to facilitate seamless integration into existing systems for government use.
- Evaluating versioning and storage considerations to maintain model integrity and traceability in public sector applications.
Model Serving on Kubernetes
- Overview of inference servers to support scalable and efficient model serving for government operations.
- Deploying TensorFlow Serving and TorchServe to enhance performance and reliability in governmental IT environments.
- Setting up model endpoints to ensure secure and accessible deployment for public sector use.
Inference Optimization Techniques
- Batching strategies to optimize resource utilization and reduce latency in government applications.
- Concurrent request handling to improve throughput and response times for high-demand services.
- Latency and throughput tuning to meet performance standards required by public sector operations.
Autoscaling ML Workloads
- Horizontal Pod Autoscaler (HPA) for dynamic scaling based on demand, ensuring efficient resource allocation in government environments.
- Vertical Pod Autoscaler (VPA) to automatically adjust the resources allocated to pods, enhancing operational efficiency for government workloads.
- Kubernetes Event-Driven Autoscaling (KEDA) to enable more granular and event-based scaling for government applications.
GPU Provisioning and Resource Management
- Configuring GPU nodes to support computationally intensive machine learning tasks in public sector environments.
- NVIDIA device plugin overview to ensure optimal utilization of GPU resources for government workloads.
- Resource requests and limits for ML workloads to prevent resource contention and ensure stable operations for government applications.
Model Rollout and Release Strategies
- Blue/green deployments to minimize downtime and risk during model updates in public sector systems.
- Canary rollout patterns to gradually introduce new models and monitor performance in a controlled manner for government use.
- A/B testing for model evaluation to ensure the effectiveness and reliability of machine learning solutions in governmental applications.
Monitoring and Observability for ML in Production
- Metrics for inference workloads to provide insights into model performance and resource usage for government operations.
- Logging and tracing practices to facilitate troubleshooting and enhance transparency in public sector IT systems.
- Dashboards and alerting to enable real-time monitoring and proactive management of machine learning services for government use.
Security and Reliability Considerations
- Securing model endpoints to protect sensitive data and maintain the integrity of government operations.
- Network policies and access control to enforce strict security measures in governmental IT environments.
- Ensuring high availability to provide reliable and continuous service for critical public sector applications.
Summary and Next Steps
Requirements
- An understanding of containerized application workflows for government
- Experience with Python-based machine learning models
- Familiarity with Kubernetes fundamentals
Audience
- Machine Learning Engineers
- DevOps Engineers
- Platform Engineering Teams
14 Hours
Testimonials (3)
About the microservices and how to maintenance kubernetes
Yufri Isnaini Rochmat Maulana - Bank Indonesia
Course - Advanced Platform Engineering: Scaling with Microservices and Kubernetes
How trainer deliver knowledge so effectively
Vu Thoai Le - Reply Polska sp. z o. o.
Course - Certified Kubernetes Administrator (CKA) - exam preparation
The knowledge and the patience from the trainer to answer to our questions.