<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/6addf8673d4f440ca20964519f5f1047&quot; frameborder=&quot;0&quot; width=&quot;1416&quot; height=&quot;1062&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>1062</height><width>1416</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>1062</thumbnail_height><thumbnail_width>1416</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/6addf8673d4f440ca20964519f5f1047-add2fc53b7c8c085.gif</thumbnail_url><duration>129.3543</duration><title>Testing Prompts Against Different Models</title><description>In this video, I discuss the importance of evaluating prompts across different AI models to determine which performs best. I demonstrate how to set up an evaluation using an AI Chef prompt, showing how to duplicate and modify models, specifically using GPT-4 and GPT-5. I emphasize that while we can change prompts inline, overriding the model for evaluation is a straightforward approach. I encourage you to experiment with various models and prompts to see how they perform. Please take the time to run these evaluations and share your findings.</description></oembed>