<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/e0f3ae24367841b58b31b2d53a222e5d&quot; frameborder=&quot;0&quot; width=&quot;1658&quot; height=&quot;1243&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>1243</height><width>1658</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>1243</thumbnail_height><thumbnail_width>1658</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/e0f3ae24367841b58b31b2d53a222e5d-87297abb995b6cc5.gif</thumbnail_url><duration>286.684</duration><title>ShopEase Voice Agent Implementation Overview 🚀</title><description>Hi, this is my submission for the Cosmo Take-Home Engineering assignment, showcasing a PipeCat implementation. I&apos;m utilizing Grok LLAMA 3.18b for low-latency LLM inference and DeepGram for both text-to-speech and speech-to-text, with a SQLite database running locally. During the demo, I’ll be testing order updates and refund policy inquiries while streaming latency metrics in real-time. I also have retry logic in place for handling rate limits, and I&apos;m randomizing inputs to demonstrate the system&apos;s capabilities. Please take a look at the K8 image specs in the repo for deployment details and let me know your thoughts!</description></oembed>