{"type":"video","version":"1.0","html":"<iframe src=\"https://www.loom.com/embed/b8a6d28cc1694121ac3869079d76085b\" frameborder=\"0\" width=\"1920\" height=\"1440\" webkitallowfullscreen mozallowfullscreen allowfullscreen></iframe>","height":1440,"width":1920,"provider_name":"Loom","provider_url":"https://www.loom.com","thumbnail_height":1440,"thumbnail_width":1920,"thumbnail_url":"https://cdn.loom.com/sessions/thumbnails/b8a6d28cc1694121ac3869079d76085b-1acb4262bccddfe9.gif","duration":849.908,"title":"AI Hallucinations Are Now an Epidemic in the Legal Field","description":"Tony DeSimone shares that AI hallucinations are becoming a serious and growing problem in the legal field, with prosecutors and law firms recently facing sanctions and costly consequences for submitting court filings containing false or AI-generated information. He explains that hallucinations aren’t rare mistakes, but a built-in risk of generative AI, and he breaks down the three main factors behind them to help you understand the best ways to avoid them."}