<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/a19c698d8f4c42289a48cc069e126e48&quot; frameborder=&quot;0&quot; width=&quot;1920&quot; height=&quot;1440&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>1440</height><width>1920</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>1440</thumbnail_height><thumbnail_width>1920</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/a19c698d8f4c42289a48cc069e126e48-c6c10e2a6bda0e61.gif</thumbnail_url><duration>181.334</duration><title>Agent Testing Framework Project Walkthrough</title><description>Hello, I am Aniket, and I walked you through my Agent Testing Framework project. I booted up the code and explained how main.py calls runner.py to loop through 20 JSON test cases, send each test case to the agent, evaluate performance using LLM scoring plus a small rule base for refusal, length, and prompt injection, then generate a report with pass and fail. In the run, 14 passed and 6 failed, for a 70 percent pass rate, including adversarial cases like requests to reveal hidden policies. There was no action requested from viewers.</description></oembed>