Creating self-healing software systems via effective usage of telemetry data and AI agents

Modern software systems operate in complex, dynamic environments where failures are inevitable. Traditional monitoring and manual incident response are no longer sufficient to ensure resilience or customer satisfaction. This talk explores how to design and implement self-healing software systems by combining telemetry data with an AI-driven agentic approach. We’ll start by examining how high-quality telemetry forms the foundation for detecting anomalies and predicting failures. Next, we’ll show how modern GenAI (LLMs) can transform this telemetry into actionable insights for AI agents that interpret data, pinpoint root causes, and apply automated fixes. Through a practical, real-world example, you’ll see how telemetry and AI work together to create adaptive feedback loops that continuously improve system reliability, while freeing engineers from repetitive operational tasks.

Related Talks

Integrate WireMock into LocalStack for End-to-End Local Testing

In this live session, Brian from LocalStack will demonstrate the WireMock extension for LocalStack, showing how developers can achieve end-to-end local testing by combining AWS service emulation with external API mocking. Brian will walk through real-world use cases, demonstrate the integration in action, and explain how this unified approach simplifies testing complex cloud applications that depend on both AWS services and third-party APIs.

Watch recording
Watch recording
Simulate Microservices, Cloud Services, and Everything Else with WireMock & LocalStack

In this live session, WireMock CTO Tom Akehurst will introduce hybrid API simulation (local + cloud) with WireMock Runner. Tom will explain why we built Runner, how developers are using it today, and how it fits into modern dev and test workflows - such as simulating APIs during testing, prototyping, and AI-native development.

Watch recording
Watch recording
Getting started with the LocalStack Model Context Protocol (MCP) Server

What if your AI coding assistant could not only write infrastructure code, but also deploy it, test it, and fix issues automatically — all on your local machine? That's exactly what the LocalStack MCP Server makes possible.In this session, we'll introduce the LocalStack Model Context Protocol (MCP) Server, a new tool that lets AI agents manage your entire local cloud development lifecycle through a conversational interface. You'll learn:What MCP is and why it's a game-changer for AI-assisted developmentHow the LocalStack MCP Server turns manual cloud tasks into automated workflowsHow to set up and configure the server with your favorite AI editor (Cursor, VS Code, etc.)Real-world demos: deploying CDK apps, analyzing logs, running chaos tests, managing state with Cloud Pods, and more.Through hands-on examples, we'll walk through a complete workflow where an AI agent deploys a serverless application, verifies resources, troubleshoots issues, and tests resilience, all without leaving the conversation.If you've ever wished your AI assistant could do more than just generate code, this talk will show you what's possible when agents can actually manage your local cloud environment.

Watch recording
Watch recording

Launch yourself in the world of local cloud development

Try for free
Try for free
Talk to Sales
Talk to Sales