<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/9d307bc4f47040668cc92c50b9711800&quot; frameborder=&quot;0&quot; width=&quot;1728&quot; height=&quot;1296&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>1296</height><width>1728</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>1296</thumbnail_height><thumbnail_width>1728</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/9d307bc4f47040668cc92c50b9711800-37b14da704fd9fb8.gif</thumbnail_url><duration>154.84</duration><title>Privacy-First Document Q&amp;amp;A on Cloudflare</title><description>Hi everyone, I built Walltrack for small law firms and healthcare clinics that need AI documentation search but cannot send confidential files to OpenAI or other third party providers. I explain why the default RAG stack is not enough and that self hosted setups require heavy GPU and can cost about $6 million for enterprise platforms. Walltrack runs entirely on Cloudflare, with no data leaving the Cloudflare network, using products like Workers, Workers AI, Vectorize, R2, D1, AI Gateway, KV, and Secrets. I show uploading a PDF, chunking and embedding it, then asking a question and getting answers from Llama 3.1. We also log everything for compliance and auditing, and there is no action requested from viewers.</description></oembed>