<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/9729bee88cde4da58d30e8d05737c824&quot; frameborder=&quot;0&quot; width=&quot;1280&quot; height=&quot;960&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>960</height><width>1280</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>960</thumbnail_height><thumbnail_width>1280</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/9729bee88cde4da58d30e8d05737c824-9b2d9bd92eba3523.gif</thumbnail_url><duration>160.193</duration><title>Transforming AI Troubleshooting with Monte Carlo: A Case Study ✈️</title><description>In this video, I discuss how Monte Carlo&apos;s troubleshooting agent can dramatically improve our ability to diagnose AI failures in production, which often manifest as incorrect outputs rather than obvious errors. For instance, we observed a drop in recommendation validity from 99.2% to 94.6%, leading to over 2,000 invalid recommendations due to stale flight data. The agent quickly traced the issue back to a failing airflow DAG that had not updated the data for six hours. I emphasize that the root cause was not an AI problem but rather stale source data, and I outline a remediation playbook that includes restarting the DAG and enhancing our data freshness checks. I encourage everyone to be proactive in ensuring our data pipelines are functioning optimally to maintain the quality of our AI outputs.</description></oembed>