<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/bfdb15672ff04dcf83472d41e9cbbfe8&quot; frameborder=&quot;0&quot; width=&quot;1728&quot; height=&quot;1296&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>1296</height><width>1728</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>1296</thumbnail_height><thumbnail_width>1728</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/bfdb15672ff04dcf83472d41e9cbbfe8-4f82363d30421a18.gif</thumbnail_url><duration>222.336</duration><title>Pipelines demo video</title><description>In this video, I walk you through how to effectively use pipelines in PromptHub, focusing on our product feedback classifier. I demonstrate how the pipeline integrates evaluations, datasets, and raw versioning to track performance during merge requests. Each time a merge request is made, the pipeline runs evaluations on a dataset of five test cases to ensure we maintain performance standards. I also highlight the importance of running guardrails to prevent issues like including secrets or profanity in prompts. Please make sure to implement these practices in your workflow to enhance our product&apos;s reliability.</description></oembed>