"The choice of AI model is use case-specific." This concise statement from Martin Keen, Master Inventor at IBM, cuts through the current hype surrounding large language models (LLMs) to provide a clear, actionable framework for enterprise leaders navigating the sprawling landscape of generative AI. Keen’s presentation, aimed at demystifying the differences between Small Language Models (SLMs), LLMs, and Frontier Models (FMs), provides sharp analysis on how to strategically match model capability, cost, and governance requirements to specific business needs.
Keen began by positioning LLMs—the models most people associate with AI, boasting tens of billions of parameters—as generalists. These are the broad, open-source workhorses capable of sophisticated, back-and-forth conversations and complex reasoning across multiple domains. However, they are computationally expensive, typically requiring cloud or SaaS environments due to their large GPU memory and processing needs. In contrast, SLMs, with fewer than 10 billion parameters, are presented not as inferior versions, but as efficient specialists. Keen points out that a well-tuned SLM can often "match or even beat these bigger models at focused tasks," operating faster and cheaper. Frontier Models (FMs), the most capable tier, are defined by their sheer scale—often exceeding hundreds of billions of parameters—and deep tool integration, making them the best choice for highly complex, multi-step tasks.
