<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/8aae9fa6e63d4104b7a31e748349fc82&quot; frameborder=&quot;0&quot; width=&quot;1514&quot; height=&quot;1135&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>1135</height><width>1514</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>1135</thumbnail_height><thumbnail_width>1514</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/8aae9fa6e63d4104b7a31e748349fc82-d456f7f6146b9369.gif</thumbnail_url><duration>520.527</duration><title>Introducing ChaCha: A Behavioral Security Layer for AI Systems</title><description>In this video, we introduce our product, ChaCha, a control layer for longitudinal threat detection in AI applications. ChaCha monitors behavior over time, addressing the gap left by stateless AI safety tools that fail to recognize evolving threats across sessions. We demonstrate how ChaCha&apos;s SDK integrates with a corporate chatbot, showcasing its ability to detect escalating risk in user interactions. Our next steps include developing richer behavioral patterns and transitioning to our own GPT OS model for cost efficiency. I encourage you to explore the dashboard and consider how ChaCha can enhance your AI security measures.</description></oembed>