"I love hallucinations. I really do, because there is a creativity to it." This provocative statement from Chris Hay, a Distinguished Engineer, encapsulates a central, often counterintuitive, theme of the recent Mixture of Experts podcast from IBM Think. The episode, hosted by Tim Hwang, brought together Hay, Senior Research Scientist Skyler Speakman, and Kate Soule, Director of Technical Product Management for Granite, to dissect the nuanced realities of artificial intelligence, moving beyond simplistic narratives of AI as either infallible or fundamentally flawed. Their discussion explored the origins of large language model inaccuracies, revisited a prominent prediction about AI's impact on coding, and delved into the evolving landscape of the AI-driven job market and the burgeoning era of micro-models.
A core insight emerging from the discussion centered on the very nature of AI hallucinations. The panel unpacked a recent OpenAI paper suggesting that these inaccuracies are not merely inherent flaws but are significantly influenced by the current training and reward functions. Kate Soule articulated this succinctly, explaining that models are "always rewarded more if they guess... than if they say I don't know." This fundamental incentive structure, driven by binary evaluation metrics, pushes models to generate plausible but incorrect information rather than admitting uncertainty. Chris Hay further elaborated on this, highlighting how the shift towards reinforcement learning in post-training has created an "eval nightmare land," where models are rewarded for "getting it right" and penalized for saying "I don't know." This environment inadvertently exacerbates the hallucination problem by prioritizing a definitive, even if incorrect, answer over a cautious, accurate one. Skyler Speakman reinforced this, noting that the conventional wisdom that increased accuracy would naturally reduce hallucinations is being challenged. The issue, he suggested, lies not just in accuracy, but in the models' ability to assess the "feasibility" or "reasonableness" of their own statements, a capability not adequately captured by current evaluation methods.
