😵‍💫 Why Language Models Hallucinate

In this episode, we delve into why language models "hallucinate," generating plausible yet incorrect information instead of admitting uncertainty. We'll explore how these overconfident falsehoods arise from the statistical objectives minimized during pretraining and are further reinforced by current evaluation methods that reward guessing over expressing doubt. Join us as we uncover the socio-technical factors behind this persistent problem and discuss proposed solutions to foster more trustworthy AI systems.

Om Podcasten

> Building the future of products with AI-powered innovation. < Build Wiz AI Show is your go-to podcast for transforming the latest and most interesting papers, articles, and blogs about AI into an easy-to-digest audio format. Using NotebookLM, we break down complex ideas into engaging discussions, making AI knowledge more accessible. Have a resource you’d love to hear in podcast form? Send us the link, and we might feature it in an upcoming episode! 🚀🎙️