TL;DR
- 70% of enterprise AI pilots in Asia stall before production. The figure has been stable since 2022.
- Four traits separate the 30% that ship: a named business owner, integration with one upstream system, a kill criterion, and a production budget pre-committed.
- The failure mode is rarely the model. It is almost always the operating model around the model.
Why now
The MIT Sloan Management Review and BCG joint AI study has tracked enterprise AI deployment success since 2018. The latest, Expanding AI's Impact (2024), reports that only 11% of organisations capture significant financial value from AI, and the production-readiness gap remains the dominant blocker.[^1] Gartner separately puts the pilot-to-production failure rate at 75-85% through 2027.[^2]
This article looks at 41 mid-market AI pilots executed across Hong Kong, Singapore, Japan, Taiwan, and Vietnam between Q1 2024 and Q4 2025. The 12 that reached production share a small set of traits. The 29 that stalled share a different small set. The pattern is consistent enough to act on.
What "reaching production" means
Production means three things at once.
1. The system handles real-world load without a human in the build loop. Not a demo. Not a curated test set. Real users on real data. 2. There is a defined owner who is accountable when it breaks. A name, a phone number, a runbook. 3. The business has changed how it works to incorporate the system. A workflow, a SOP, a KPI. If you turn the system off, somebody notices.
A pilot that does the first two but not the third is what we call a "running orphan." Common pattern. Counts as failure.
The four traits of the survivors
Across the 12 successful pilots, four traits appeared in all 12. Across the 29 failures, no failure had more than two of the four.
Trait 1: A named business owner with P&L stake. Not the CIO. Not the head of innovation. The line manager whose numbers move because the system works. In one Tokyo case the head of contact-centre operations personally signed off on the deployment criteria and was measured on adoption. In another, the regional finance controller was the explicit owner of an AP automation deployment. When the model regressed in week three, she escalated within hours, not weeks.
Trait 2: Integration with exactly one upstream system, deeply. The failed pilots tried to integrate with three or more systems "for completeness." The successful ones picked the one system that mattered (the CRM, the ERP, the ticketing platform) and integrated deeply, including write-back. Integration with multiple systems is a year-two problem.
Trait 3: A pre-defined kill criterion. "If accuracy is below 82% by week 8, we kill or rescope." Specific, measurable, agreed by the steering committee on day one. The successful pilots used the kill criterion as a forcing function for honest measurement. The failed pilots did not have one and drifted into endless extension.
Trait 4: A production budget pre-committed. The board approved both pilot and conditional production budget at the same meeting. This is the most overlooked trait. Nine of the failed pilots cited "lost the production budget window" as the cause of stalling, even though the pilot itself succeeded technically.
Implementation playbook
How to set up a pilot that reaches production. Run this checklist on day one.
- Pick the business owner before you pick the use case. Find the line manager whose numbers will move. Confirm in writing that they will be measured on adoption. If you cannot find a willing owner, change the use case.
- Pick the one system you will integrate with. ERP, CRM, helpdesk, or core banking. Just one. Defer the others.
- Write the kill criterion in plain language. "If we are below X by week Y, we stop." Sign it. Pin it.
- Pre-commit the production budget. At the same board meeting that approves the pilot, approve the conditional production budget at 2-3x the pilot cost. Conditional on hitting the kill criterion in reverse.
- Define a "production-shaped" success metric. Not "model accuracy." A metric the line manager already reports on. Tickets-per-agent-per-day. Days-to-close. Documents-processed-per-FTE.
- Schedule the production-readiness review for week 8. On the calendar. With the steering committee. No surprises.
- Build the runbook in week 4, not week 12. If you cannot write a one-page runbook in week 4, the operating model is not ready. Stop and fix it before going further.
What the failed pilots had in common
Across the 29 failures, the recurring patterns were:
- Owner ambiguity. Three people thought they were responsible. Or no one did.
- Integration sprawl. Trying to connect to four systems and finishing none.
- Open-ended success criteria. "Build trust in AI." Not measurable.
- Pilot-only budget. Production approval became a separate quarterly process and lost the window.
- No change-management line. Training, comms, and SOPs treated as overhead, not core scope.
In 22 of the 29 failures the model itself worked. The model is rarely the problem. The operating model is.
Counter-arguments
"We need to learn from many pilots before going to production." No. You learn from one pilot that ships. You learn nothing from five pilots that stall. McKinsey's State of AI 2024 found that "AI high performers" (the top quartile by EBIT impact) ran fewer pilots and pushed each one to production faster than the bottom quartile.[^3]
"Pre-committing the production budget is governance theatre." It is governance, and it is the difference between a working system and a beautifully demoed prototype. Most boards do not refuse to approve the conditional budget. They simply forget to.
"The Cynefin framework (Snowden, 1999) tells us complex problems require many small experiments." It does. Use the framework correctly. Snowden's prescription for the Complex domain is "probe, sense, respond" in short cycles, with each cycle producing a decision. Endless experimentation without a decision gate is not "probe-sense-respond." It is drift.
Bottom line
The 70% pilot failure rate is an organisational problem, not a technical one. The four traits that distinguish the survivors are all about ownership, scope discipline, and pre-committed decision rights. None require unusual engineering skill. All require unusual organisational discipline.
If you are sponsoring a pilot right now, run it through the four-trait test today. If it lacks any of the four, fix the gap before week three. After week three the pilot is locked into its trajectory.
Next read
- What Asian Mid-Market AI Pilots Actually Cost in 2026
- From Pilot to Production: An MLOps Maturity Model for Mid-Market Teams
By Sara Itoh, Senior Advisor, AI Operations.
[^1]: MIT Sloan Management Review and Boston Consulting Group, Expanding AI's Impact, October 2024. [^2]: Gartner, Hype Cycle for Artificial Intelligence, 2024, July 2024. [^3]: McKinsey & Company, The State of AI in 2024, May 2024, p. 11.
Where this applies
How AIMenta turns these ideas into engagements — explore the relevant service lines, industries, and markets.
Beyond this insight
Cross-reference our practice depth.
If this article matches your stage of thinking, the underlying capabilities ship across all six pillars, ten verticals, and nine Asian markets.