<?xml version="1.0" encoding="UTF-8"?><oembed><type>video</type><version>1.0</version><html>&lt;iframe src=&quot;https://www.loom.com/embed/7474d8c38f9c4c5bb09225761aef0c0d&quot; frameborder=&quot;0&quot; width=&quot;1680&quot; height=&quot;1260&quot; webkitallowfullscreen mozallowfullscreen allowfullscreen&gt;&lt;/iframe&gt;</html><height>1260</height><width>1680</width><provider_name>Loom</provider_name><provider_url>https://www.loom.com</provider_url><thumbnail_height>1260</thumbnail_height><thumbnail_width>1680</thumbnail_width><thumbnail_url>https://cdn.loom.com/sessions/thumbnails/7474d8c38f9c4c5bb09225761aef0c0d-ee949423959aa363.gif</thumbnail_url><duration>358.639</duration><title>7-Foundations of Probability in Deep Learning 📊</title><description>In this video, I discuss the fundamental concepts of probability as they relate to deep learning, covering topics such as sample spaces, conditional probability, and Bayes&apos; theorem. I also explain the significance of random variables, expected value, variance, and key distributions like Bernoulli and Gaussian. Additionally, I delve into information theory, maximum likelihood estimation, and the role of sampling methods in deep learning. I encourage you to implement Softmax, CrossEntropyLoss, and DropoutLayer from scratch as practical exercises to reinforce your understanding. Thank you for your attention!</description></oembed>