Everything You Need To Know About Agent Observability — Danny Gollapalli and Ben Hylak, Raindrop

AI Engineer · Intermediate ·🤖 AI Agents & Automation ·2h ago
Agent failures do not look like normal software failures. In this workshop, the Raindrop team breaks down what it actually takes to monitor production agents, from explicit signals like tool errors, latency, and cost to fuzzier signals like user frustration, refusals, task failure, and capability gaps. The session covers how to move beyond evals toward real production observability, how to use classifiers, regex, and experiments to catch regressions, and how to instrument self-diagnostics so agents can report their own failures and strange behavior. If you're running agents in production, this is a practical framework for understanding what is going wrong and how to catch it early. Speaker info: - https://x.com/benhylak - https://www.linkedin.com/in/benhylak/ - Danny Gollapalli
Watch on YouTube ↗ (saves to browser)
Sign in to unlock AI tutor explanation · ⚡30

Related AI Lessons

Pay.sh pays for an API call. Hashlock settles a trade. Know the difference.
Understand the difference between Pay.sh and Hashlock in the context of API calls and trade settlements on the Solana blockchain
Dev.to AI
6 Protocols for Agent Infrastructure — Trust Score, Deployment, SLA, Identity, Compliance
Learn 6 essential protocols for building robust agent infrastructure, ensuring trust, deployment, and compliance in autonomous workflows
Dev.to AI
Why an MCP Security Library Beats a Security Proxy
Learn why an MCP Security Library is a better approach than a Security Proxy for securing AI agents, and how to implement it
Dev.to AI
How We Got Listed on the Official MCP Registry — A Server Builder Checklist
Get listed on the MCP Registry with a step-by-step checklist for MCP server builders to increase discoverability for AI agents
Dev.to AI
Up next
New Way Now: Endeavor Health puts an AI-first lens on pathology, improving efficiency by up to 40%
Google Cloud
Watch →