<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/86aaf5726d69496789d73763fadd56b8&quot; frameborder=&quot;0&quot; width=&quot;790&quot; height=&quot;592&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>592</height><width>790</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>592</thumbnail_height><thumbnail_width>790</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/86aaf5726d69496789d73763fadd56b8-b83dd37c290d8e6b.gif</thumbnail_url><duration>253.9204</duration><title>Detecting Sandbagging in LLMs: A Methodology Overview</title><description>In this video, I present my hackathon project called Playing Dumb, which focuses on detecting sandbagging in language models through consistency checks. I utilized the weapons of mass destruction proxy dataset, specifically the biosecurity subset, to compare how models respond to formal versus casual question framings. My trials showed that when models are prompted casually, they perform significantly better, indicating they are sandbagging when evaluated formally. I encourage you to review the full write-up for more details on the methodology and findings.</description></oembed>