<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/0b4bb303513d44a1bc5d01d2c22f1a85&quot; frameborder=&quot;0&quot; width=&quot;2208&quot; height=&quot;1656&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>1656</height><width>2208</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>1656</thumbnail_height><thumbnail_width>2208</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/0b4bb303513d44a1bc5d01d2c22f1a85-7bc72666732139e2.gif</thumbnail_url><duration>376.111</duration><title>Felafax -- building AI infra for non-NVIDIA GPUs</title><description>Hey, we are twin brothers who worked on ML Infra at Google and Meta the last 5 years and we built a new AI stack for fine-tuning and serving LLMs. Our platform works both on non-NVIDIA chipsets like TPUs, Trainium and AMD GPUs, and also works on NVIDIA GPUs.</description></oembed>