{"type":"video","version":"1.0","html":"<iframe src=\"https://www.loom.com/embed/00791498f1d84e4ba6d7476bd2e1442f\" frameborder=\"0\" width=\"1108\" height=\"831\" webkitallowfullscreen mozallowfullscreen allowfullscreen></iframe>","height":831,"width":1108,"provider_name":"Loom","provider_url":"https://www.loom.com","thumbnail_height":831,"thumbnail_width":1108,"thumbnail_url":"https://cdn.loom.com/sessions/thumbnails/00791498f1d84e4ba6d7476bd2e1442f-3120c50efa263701.gif","duration":235.489,"title":"Setting Up OpenCode with LightLLM: A Quick Demo 🚀","description":"In this quick demo, I walk you through setting up OpenCode with LightLLM, highlighting my local LightLLM proxy and the three Azure GPT models I've exposed. After installing OpenCode, I show you how to configure it at a global level or project level based on your needs. I emphasize the importance of defining the provider configuration, particularly for LightLLM, and how to set the base URL to point to localhost. I also guide you on listing the modules and using aliases for easy reference in OpenCode. Please follow along and enter your API key to see the specified modules."}