Why you need an Anthropic AI operations manager (not just a consultant)
Consultants get you started with Claude. Operations managers keep it running. With 95% of AI pilots failing after launch, the real work begins after implementation - monitoring usage, optimizing costs, preventing drift, and maintaining the systems that actually deliver ROI.

Key takeaways
- 95% of AI pilots die after consultants leave - MIT research shows implementations fail without ongoing operational management
- Daily tasks compound into major costs - unmanaged Claude usage can spiral from minimal to substantial monthly costs without monitoring
- Operations delivers significant cost savings - dedicated teams achieve better ROI through usage optimization and configuration management
- Build vs buy changes everything - external vendor implementations succeed majority of the time vs minority for internal builds
- Need operational support for Claude? Let's discuss your requirements.
MIT found 95% of generative AI pilots fail. Not because the technology doesn’t work. They fail six months after the consultants leave, when nobody’s watching usage costs, updating prompts, or fixing integrations that quietly broke.
I watched a client’s Claude implementation go from hero to zero in four months. The consultants delivered a beautiful solution. Everyone celebrated. Then reality hit - daily token costs tripled, prompt accuracy substantially degraded, and nobody knew how to fix it. They needed operations, not more consulting.
The implementation illusion
Here’s what every AI consultant sells you: three months of implementation, a handover document, and a prayer that your team figures out the rest.
They’ll set up Claude, build some workflows, maybe train your team for a day. Then they disappear, leaving you with a sophisticated system that starts degrading immediately.
Gartner predicts significant percentage of agentic AI projects will be canceled by 2027 due to escalating costs and unclear business value. The pattern is predictable - initial success, gradual degradation, then abandonment.
The real issue? The gap between implementation and value realization. Enterprise AI initiatives achieve minimal ROI while requiring significant capital investment. That’s not technology failure. That’s operations failure.
Daily operations reality
After spending months building Tallyfy’s AI capabilities, I learned what actually keeps AI systems alive. It’s not the big strategic decisions. It’s the boring daily work nobody talks about.
Morning monitoring routine (30 minutes daily): Token usage spikes tell you when something’s wrong. I caught a recursive loop burning through significant token costs before 9am. Without daily monitoring, that’s substantial monthly waste. Grafana’s Anthropic integration tracks this automatically, but someone needs to watch the dashboards.
Prompt library maintenance (2 hours weekly): Prompts degrade. What worked perfectly last month starts hallucinating. Anthropic’s Claude Sonnet 4.5 behaves differently than 4.0. Each model update requires prompt adjustments. Miss this, and accuracy drops substantially within weeks.
Cost optimization reviews (4 hours weekly): Claude pricing varies significantly per million tokens depending on the model. Most companies use Opus for everything when Haiku would handle majority of tasks. That’s paying premium pricing when budget-friendly options would work. One client saved significant monthly costs just by routing requests to the right model.
The governance gap
88% of AI pilots fail to reach production. The survivors have one thing in common - someone owns the operational governance.
Compliance requirements multiply: The EU AI Act requires transparency audits and risk assessments. ISO 42001 establishes management standards. NIST AI RMF mandates risk frameworks. Each regulation needs documentation, monitoring, and reporting. Without dedicated operations, compliance becomes impossible.
Security monitoring never stops: Every model update introduces new vulnerabilities. Prompt injection attacks evolve weekly. Data leakage patterns change. Someone needs to track Anthropic’s security updates and implement them immediately.
Team access management grows complex: Enterprise Claude plans offer SSO, SCIM, and role-based permissions. But someone needs to configure them, update them when people leave, and audit usage patterns. I’ve seen companies paying for 50 Claude licenses when 12 people actually use it.
ROI patterns that work
Leading companies attribute substantial portion of operating profits to AI. The difference? They treat AI as operations, not projects.
The Carbon model works: Carbon processes 150,000 loan applications monthly through DataRobot, with dedicated teams managing the operations. They save an entire end-to-end process by having operations managers handle the infrastructure while business teams focus on strategy.
Infrastructure savings compound: Companies achieve significant infrastructure cost savings through better configurations and workflow orchestration. But only with dedicated operations teams. Part-time resources fail - fully staffed AI teams boost success rates meaningfully.
Time savings multiply: Red Hat Cloud customers report majority time savings for developers through automated infrastructure. Data scientists save significant portion through standardized workflows. But these savings require operational management to maintain the systems.
Build vs buy reality and the operations evolution
Here’s what MIT’s research revealed: purchasing AI tools and building partnerships succeeds majority of the time. Internal builds succeed minority of the time.
The math is brutal. You need someone who understands Claude’s API changes, monitors 15,000+ AI operations roles worth of collective knowledge, and can implement fixes within hours, not weeks.
Operations managers track hundreds of configuration parameters. They know when Anthropic changes rate limits, how to optimize context windows, and which prompts cause token explosions. This isn’t knowledge you build internally.
We’re watching the birth of AIops - what happened when DevOps met machine learning. The MLOps market will experience substantial growth from current levels.
This isn’t about managing servers anymore. It’s about managing intelligence systems that drift, degrade, and evolve continuously.
Without downtime management, companies lose hundreds of millions annually from digital system failures. AI systems are even more fragile - a misconfigured Claude implementation can burn through extreme per-minute costs in token usage.
The companies succeeding with Claude aren’t the ones with the best initial implementations. They’re the ones with dedicated operations managers preventing the 46% failure rate that kills most AI initiatives within six months.
Your consultants are gone. Your Claude implementation is drifting. Every day without operations management costs you money and degrades performance. The question isn’t whether you need an operations manager - it’s how much drift you can afford before you get one.
About the Author
Amit Kothari is an experienced consultant, advisor, and educator specializing in AI and operations. With 25+ years of experience and as the founder of Tallyfy (raised $3.6m), he helps mid-size companies identify, plan, and implement practical AI solutions that actually work. Originally British and now based in St. Louis, MO, Amit combines deep technical expertise with real-world business understanding.
Disclaimer: The content in this article represents personal opinions based on extensive research and practical experience. While every effort has been made to ensure accuracy through data analysis and source verification, this should not be considered professional advice. Always consult with qualified professionals for decisions specific to your situation.