<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/2f85e69bec7041a9ba7f33f333656a51&quot; frameborder=&quot;0&quot; width=&quot;1110&quot; height=&quot;832&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>832</height><width>1110</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>832</thumbnail_height><thumbnail_width>1110</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/2f85e69bec7041a9ba7f33f333656a51-ebd40feb3353ed05.gif</thumbnail_url><duration>283.998</duration><title>Patronus AI&apos;s Percival Debugging a Toy Claims Agent</title><description>In this video, I walk you through how to use our Patronus platform to trace, evaluate, and enhance the performance of your GenAI applications, specifically in the context of developing a claims processing agent. We cover the setup process, including importing necessary packages, defining tools, and running experiments to compare our agent&apos;s outputs against a golden dataset. I highlight how to investigate traces for errors and utilize our agent judge for automatic insights into performance issues. I encourage you to take these insights back to your development workflow to refine your prompts and improve your agent&apos;s effectiveness. Finally, I suggest rerunning the experiment after making adjustments to see the improvements in action.</description></oembed>