From DevOps to MLOps: Scaling ML models to 2 Million+ requests per day

The challenge with Machine Learning (ML) models is productionizing. It requires data ingestion, data preparation, model training, model deployment, and monitoring.Adopting MLOps practices is similar to DevOps practices. In MLOps, the workload changes, but some core principles like automation, continuous integration/continuous deployment (CI/CD), and monitoring. Taking DevOps practices, I will discuss the similarities and differences in adopting MLOps practices.In this talk, Chinmay takes a production use case to scale ML models to 2 million+ daily requests. It leverages Google Cloud's (GCP) infrastructure to use its GPU and other services. This talk will help you draw similarities between DevOps and MLOps as a DevOps practitioner and help you learn how to run Machine Learning models at the production scale with best practices.

Related Talks

Integrate WireMock into LocalStack for End-to-End Local Testing

In this live session, Brian from LocalStack will demonstrate the WireMock extension for LocalStack, showing how developers can achieve end-to-end local testing by combining AWS service emulation with external API mocking. Brian will walk through real-world use cases, demonstrate the integration in action, and explain how this unified approach simplifies testing complex cloud applications that depend on both AWS services and third-party APIs.

Watch recording
Watch recording
Simulate Microservices, Cloud Services, and Everything Else with WireMock & LocalStack

In this live session, WireMock CTO Tom Akehurst will introduce hybrid API simulation (local + cloud) with WireMock Runner. Tom will explain why we built Runner, how developers are using it today, and how it fits into modern dev and test workflows - such as simulating APIs during testing, prototyping, and AI-native development.

Watch recording
Watch recording
Getting started with the LocalStack Model Context Protocol (MCP) Server

What if your AI coding assistant could not only write infrastructure code, but also deploy it, test it, and fix issues automatically — all on your local machine? That's exactly what the LocalStack MCP Server makes possible.In this session, we'll introduce the LocalStack Model Context Protocol (MCP) Server, a new tool that lets AI agents manage your entire local cloud development lifecycle through a conversational interface. You'll learn:What MCP is and why it's a game-changer for AI-assisted developmentHow the LocalStack MCP Server turns manual cloud tasks into automated workflowsHow to set up and configure the server with your favorite AI editor (Cursor, VS Code, etc.)Real-world demos: deploying CDK apps, analyzing logs, running chaos tests, managing state with Cloud Pods, and more.Through hands-on examples, we'll walk through a complete workflow where an AI agent deploys a serverless application, verifies resources, troubleshoots issues, and tests resilience, all without leaving the conversation.If you've ever wished your AI assistant could do more than just generate code, this talk will show you what's possible when agents can actually manage your local cloud environment.

Watch recording
Watch recording

Launch yourself in the world of local cloud development

Try for free
Try for free
Talk to Sales
Talk to Sales