Foundation models like GPT-5 and Claude Sonnet 4 have redefined AI, but their broad capabilities often falter when it comes to practical, nuanced tasks. A new startup, Osmosis AI, is emerging from stealth with $6.3 million in seed funding, co-led by CRV and Audacious Ventures, to tackle this very problem. They claim their approach, centered on 'Reinforcement fine-tuning', can make AI agents "better, faster, and cheaper" than their foundation model counterparts.
Osmosis's own research highlights the shortcomings of generalized models. In a recent analysis of leading closed and open-source models using Anthropic's Model Context Protocol (MCP) for tool integration, they found significant "unforced errors." Models frequently failed to use necessary tools, or conversely, invoked irrelevant ones, leading to degraded performance. GPT-5, despite generally performing best, saw notable success rate drops when presented with only relevant tools versus a broader set. The study even noted GPT-5's tendency to unnecessarily call `get_me` (a Slack tool) in general knowledge queries, suggesting a post-training skew.
