<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/ed55d0b4c60f4cdaa4d67b9c6a2f155a&quot; frameborder=&quot;0&quot; width=&quot;1920&quot; height=&quot;1440&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>1440</height><width>1920</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>1440</thumbnail_height><thumbnail_width>1920</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/ed55d0b4c60f4cdaa4d67b9c6a2f155a-e8c1f84d4f07d681.gif</thumbnail_url><duration>314.593</duration><title>Document QA Chatbot Architecture Walkthrough ⚡</title><description>Hi, this is Karthik. I built an AAPowerDrive chatbot where users upload documents and ask questions, and the system retrieves relevant chunks using a Pinecone vector database and answers with an LLM in a retrieval architecture. I use FastAPI for the backend, Gradio for the UI, and deploy on AWS EC2, with lightweight Lightbit embeddings for AWS free tier. It splits documents into chunks, creates embeddings, and stores them, then returns answers with content scores and the document source. No action was requested from viewers.</description></oembed>