<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/a808c1c6a5034a5cb857b697ed5a8a77&quot; frameborder=&quot;0&quot; width=&quot;1386&quot; height=&quot;1039&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>1039</height><width>1386</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>1039</thumbnail_height><thumbnail_width>1386</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/a808c1c6a5034a5cb857b697ed5a8a77-00001.gif</thumbnail_url><duration>366.9</duration><title>Server-Sent Events (SSE) support demonstrated based on OpenAI&apos;s Codex (Co-Pilot) code-generation model with streaming mode on</title><description></description></oembed>