Daily AI deployments are a rarity, with just 7% of organizations achieving this benchmark, according to the CNCF's 2025 Annual Survey. For traditional software, this pace would signal a crisis, prompting urgent investigations into bottlenecks and automation. Yet, for AI, this sluggishness is often accepted.
While AI models are complex and data science workflows differ from software engineering, the data suggests a deeper problem. A combined 93% of organizations deploy AI models only occasionally or somewhere in between. This points to a significant gap in delivery infrastructure, where practices enabling rapid application delivery – CI/CD automation, GitOps, and observability – are largely absent for AI.
The Delivery Infrastructure Gap
AI model serving places unique stresses on delivery systems. DORA's 2024 analysis revealed that higher AI adoption correlates with lower engineering performance, with increased AI usage linked to reduced throughput and system stability.
Traditional CI/CD pipelines, optimized for stateless applications, falter under the demands of model serving. Google Cloud's MLOps documentation highlights the mismatch: AI models require statistical validation on holdout datasets, a fundamentally different gate than the unit and integration tests used for software. Models are also prone to failure due to data distribution shifts or environmental changes, necessitating continuous monitoring and retraining that most existing delivery systems are not designed to handle.
This disconnect often results in data scientists handing over models as opaque artifacts, leaving engineering teams to struggle with deployment without understanding model-specific needs. Model serving diverges from application deployment in three key areas:
