Uber Burned Its Annual AI Budget in Four Months
Ricardo Argüello — April 15, 2026
CEO & Founder
General summary
Uber's CTO Praveen Neppalli Naga told his company this week that Claude Code and other AI coding tools had already consumed the company's entire 2026 AI budget just months into the year. His quote: 'I'm back to the drawing board, because the budget I thought I would need is blown away already.' This isn't an adoption story. It's a structural governance story.
- Uber's Claude Code adoption moved from 32% of engineers in December 2024 to 63% by February 2026 — that isn't a rollout, it's pull from the bottom up
- 11% of Uber's pull requests are opened by agents, uReview analyzes 90% of 65,000 weekly diffs, and 75% of its AI review comments are marked helpful
- The real cost driver isn't seats, it's tokens: Gartner reports agentic workflows burn 5x to 30x more tokens per task than a standard chatbot
- Anthropic's Claude Code hit $2.5B in annualized revenue in February, up from $1B in November 2025 — the fastest enterprise software ramp in history
- 91% of C-suite executives plan to raise their agentic AI spend in 2026 and CFO estimates are running 500% to 1,000% off reality according to Gartner's own finance VP
Picture hiring 5,000 new engineers and handing each one a corporate card with no ceiling. Individually, daily spend looks fine. Four months in, your CFO opens the statement and realizes consumption isn't linear — it compounds. That is exactly where Uber is today. The AI worked too well, and the budget was never built to measure pull.
AI-generated summary
Anissa Gardizy, a reporter at The Information, surfaced a quote on Monday from Uber CTO Praveen Neppalli Naga that every engineering leader planning 2026 AI budgets should stop and read:
I’m back to the drawing board, because the budget I thought I would need is blown away already.
Uber opened Claude Code access to its engineers in December 2024. Back then, 32% picked it up. By February of this year, 63% were on it. By April, Praveen told Laura Bratton at The Information that the annual AI budget was already torched.
That isn’t a rollout story. That’s a pull story.
The curve isn’t what finance planned for
When a company moves from 32% to 63% internal adoption in fourteen months, the platform team isn’t pushing. The engineering team is pulling. Aakash Gupta broke down the rest of the numbers: 92% of Uber’s developers now use AI agents monthly. Between 65% and 72% of the code that gets written inside the IDE is already AI-generated. 11% of pull requests are opened by agents, not humans. Uber’s internal AI review system, uReview, touches more than 90% of the 65,000 diffs the company ships every week. 75% of the comments uReview posts get marked as helpful.
These are production numbers, not pilot numbers.
Which means Uber can’t walk it back. The engineering flow has absorbed the tool. If somebody cut Claude Code tomorrow, developer velocity would drop in a way the board would see immediately. If nobody cuts it, margin drops quietly. Pick one of those two problems — you’re not actually getting to skip both.
The real unit isn’t a seat
Most enterprises still budget software the way they budget SaaS: cost per seat times headcount times twelve months. That mental model breaks in the first week of agentic workflows.
Claude Code for enterprise teams runs $100 to $200 per developer per month on the Premium plan, with consumption billed at API rates on top of the seat. The community-reported average daily burn lands around $6 per developer. Multiply that across 5,000 engineers and you get $30K in daily base spend before you open the rest of the invoice. That number assumes linear usage. AI isn’t linear.
Agents spawn sub-agents. Each sub-agent carries its own context window. A single pull request can trigger a chain that runs for thirty minutes, consuming tokens with no human watching. Drop that inside a CI/CD loop and the spend detonates on every commit.
Gartner measured it in March: an agentic workflow uses between 5x and 30x more tokens per task than a standard chatbot. AnalyticsWeek named the failure mode “Agentic Resource Exhaustion”: one agent in a semantic loop can incinerate thousands of dollars in a single afternoon. Their aggregated estimate across the Fortune 500 this year sits near $400 million in unplanned cloud spend attributable to agentic AI.
Jason Calacanis wrote it out in public a few weeks back: his team’s agents were running $300/day. Annualized, that’s over $100K per agent, operating at a fraction of capacity. This isn’t a Fortune 500 problem exclusively. The people paying attention earliest are narrating the alarm in real time.
Uber can’t get off the train
When you see a bill of this size, the reflex is obvious: pause, renegotiate, roll back. Uber doesn’t have that option.
11% of your PRs opened by agents and 90% of your diffs reviewed by an internal AI system isn’t a layer you can switch off without breaking the development cycle. And you can’t break the development cycle when you’re competing with Lyft across rider-driver matching, dynamic pricing, and real-time logistics across 70+ countries. Uber built its operational moat on software. AI now lives inside that software.
That leaves one lever that still works: govern consumption, don’t cut it.
The CFO and CTO show up to the meeting with different numbers
Annual budgeting was built for predictable SaaS and headcount. Two line items with linear, explainable unit costs. Agentic AI is neither.
CIO.com reported Gartner’s finance practice VP saying CFOs don’t really know what AI costs — estimates are running off by 500% to 1,000%. Glean audited enterprise AI budgets and found most underestimate true total cost of ownership by 40% to 60%. McKinsey had already quantified the other side: for every dollar a company puts into model development, they typically spend three on change management, training, and workflow redesign.
Stack the three numbers and you get a consistent picture. Companies aren’t budgeting AI badly out of carelessness. They’re budgeting AI badly because the financial model they inherited was never designed to measure exponential consumption.
The part that doesn’t get repeated enough: per the 2026 State of FinOps report, 98% of organizations now actively manage AI spend. Two years ago that number was 31%. And at the same time, 91% of C-suite executives plan to increase agentic AI budgets this year. Nobody wants to spend less. Everyone is realizing the governance work nobody did when the contract got signed is the actual project.
Praveen is budgeting from that spot right now. Most CTOs I talk to are one quarter behind him.
Taste debt’s financial sibling
I wrote yesterday about taste debt — the quality bill that accrues when you pull the human out of the loop too early. Today’s bill reads differently on paper. The underlying mechanism is identical.
Taste debt shows up in brand, decisions, and customer complaints. Uber’s budget overrun shows up in P&L, board decks, and quarterly conversations with the CFO. Two faces of the same liability: silent accrual that compounds while nobody is looking, visible only when someone arrives to collect all at once.
The amortization method is the same too. You don’t stop using agents. You install a governance layer that measures, limits, pauses, and demands human sign-off where the cost of an error — either quality or financial — outweighs the cost of the review.
Team OS is the name we use for that layer. Same layer for both liabilities.
Three moves every CTO should have on next week’s agenda
After the Uber news, any technology leader scaling agents has three items that belong on the next steering committee agenda. I’m listing them in order of effort.
Per-workflow visibility. Not per seat, not per team. Per agent, per execution. If your AI dashboard only tells you how many licenses got activated, it tells you nothing. Step one is tokenized accounting: every agent, every run, every context window, tagged. Without that, you can’t set limits because you can’t see where the money leaks.
Budgets with circuit breakers. Cap per agent, cap per team, cap per workflow. Automatic circuit breakers when an execution crosses a threshold without a result. An agent in a semantic loop burning $3,000 in an afternoon shouldn’t be possible without a human-in-the-loop alarm somewhere in the middle. FinOps-for-AI platforms already exist. Adoption is slow because companies assumed the cost was predictable.
Quality gates that double as cost gates. Where error cost is high, a human signs. Where error cost is low, the agent runs free. That boundary doesn’t get drawn in a quarterly workshop. It gets instrumented. If a workflow wants to spend more than X dollars or modify more than Y files, it needs approval before continuing. That isn’t friction. That’s insurance.
None of the three slows productivity. All three make productivity measurable, and therefore budgetable.
What won’t work
Three moves I expect to see in the next six months that aren’t going to solve this:
Freezing AI usage. This is the finance team’s reflex when the invoice lands. It’s structurally impossible once the flow has absorbed the tool. Freezing AI at Uber today is freezing a meaningful share of the development cycle. No CTO signs that.
Renegotiating the enterprise contract. Useful at the margin. But the problem isn’t the vendor’s price sheet. It’s the internal consumption pattern without governance. Even at zero cost per token, an agent in a loop still steals machine time and context.
Hoping the next model generation will be cheaper. It will be cheaper per token. It will be more expensive per task. When a model gets more capable, teams point it at longer, more complex, more multi-step work. Total cost per useful task goes up, not down. That’s Jevons’ paradox applied to tokens, and it’s already visible in the pricing data.
What IQ Source does with this
The Uber story is the most visible version of a problem any company scaling agents without consumption instruments is going to hit. Fortune 500 or a fifty-person startup: the token math doesn’t discriminate by headcount.
Our job is installing those instruments. Not an abstract framework. A concrete method for answering four questions every CTO should be able to walk into the next CFO conversation with: how much is being spent, across which workflows, what ROI each one is returning, and where the circuit breakers sit.
We implement it with Team OS — the same governance layer we use for the taste problem, applied to the consumption problem. The two liabilities live in the same flow. Treating them as separate projects is a diagnostic error.
The advantage of installing this before your company ends up in The Information is that early governance compounds into a real moat. The companies that figure out token cost optimization first are going to hold a structural edge over every competitor still running annual budget cycles against exponential adoption curves.
Uber won’t be the last company to have this conversation. The question is whether you have it before the board meeting, or during it.
Before the next board meeting, know the number. That’s where we start.
Frequently Asked Questions
Uber's CTO Praveen Neppalli Naga confirmed to The Information that Claude Code and related AI coding tools maxed out Uber's full 2026 AI budget in the first months of the year. Internal Claude Code adoption went from 32% to 63% in fourteen months. It matters because it exposes that the enterprise AI bottleneck is no longer the technology, it's the governance of consumption.
Claude Code for enterprise teams runs $100 to $200 per developer per month on the Premium plan, with usage billed at API rates on top of the seat. The reported daily burn averages around $6 per developer, but it scales non-linearly when agents invoke sub-agents or run inside CI/CD loops. That non-linearity is what broke Uber's budget math.
AI coding agent costs grow faster than SaaS because the billing model is per-token consumption, not per seat. Gartner reports an agentic workflow burns 5 to 30 times more tokens per task than a standard chatbot. When an agent spawns sub-agents that each hold their own context windows, spend compounds inside a single execution, and a CI/CD loop can drain a monthly budget in days.
A CTO governs AI consumption without slowing productivity with three layers: per-workflow visibility (not per-seat), per-team and per-agent budgets with automatic circuit breakers, and quality gates that also function as cost gates. Pulling AI back isn't an option once the workflow has absorbed the tool. Governance has to live inside the flow, not on top of it.
Related Articles
Taste Debt: The Real Cost of Removing Yourself From AI
Peter Steinberger named the real failure mode of agentic workflows: pulling yourself out too early. The bill that shows up later I call taste debt.
Google's $180B: The Enterprise Signal Nobody Reads
Google went from $30B to $180B in AI CapEx. No keynote needed. For enterprise buyers evaluating AI vendors, that number tells more than any product demo.