Uber CTO shows how Claude Code can blow up AI budgets
Uber’s chief technology officer has revealed that the ride‑hailing giant’s rapid rollout of AI coding tools, particularly Anthropic’s Claude Code, has forced the company to revisit its 2026 artificial‑intelligence budget after spending more than planned in just a few months. Praveen Neppalli Naga told insiders that engineering teams’ adoption of Claude Code and similar assistants has surged far beyond internal forecasts, with many engineers now integrating these models into daily workflows to generate, review, and refine backend code. The move has delivered speed and productivity gains, but also exposed the steep cost of scaling large‑language‑model usage across thousands of developers.
According to reports drawing on company disclosures and internal commentary, Uber’s AI and software‑development spending is embedded within a broader R&D envelope that reached 3.4 billion dollars in the prior fiscal year. Within that envelope, the portion allocated specifically to AI tools and platform development has been largely consumed in early 2026, driven by token‑based usage fees tied to Claude Code and other coding‑focused models. Engineers were given access to Claude Code in bulk late last year, and usage quickly ramped up, with data indicating that AI‑written or AI‑assisted code now accounts for a rising share of live updates to Uber’s backend systems. This has turned what was initially seen as a supplementary tool into a core, recurring cost center.
Naga described the situation as a “back to the drawing board” moment, emphasizing that Uber did not anticipate the pace at which engineers would embrace AI‑driven coding assistants. He noted that Claude Code has become the dominant tool within the company’s expanding AI toolkit, outpacing other coding‑focused models whose growth has stabilized. The CTO also signaled that Uber is now evaluating additional coding assistants, including OpenAI’s Codex, as it seeks to balance productivity, reliability, and cost control. For the company, the episode underscores the tension between unlocking faster engineering throughput and containing the financial risk of runaway AI‑compute bills.
Analysts point out that Uber’s experience reflects a broader pattern in the tech sector, where once‑modest AI experiments quickly morph into enterprise‑scale deployments with outsized token‑usage profiles. When thousands of engineers regularly invoke models to generate boilerplate, refactor legacy systems, or scan codebases for vulnerabilities, even “small” per‑request fees compound into multimillion‑dollar quarterly bills. The CTO’s remarks imply that Uber will likely introduce more granular cost monitoring, usage quotas, and policy guardrails for AI coding tools, while still preserving their role in accelerating development cycles. For other large employers, the episode serves as a cautionary signal that AI budgeting must now treat language models as permanent infrastructure, not one‑off experiments.
-
11:40
-
11:20
-
11:00
-
10:42
-
10:19
-
09:51
-
09:28
-
09:21
-
09:01
-
08:40
-
08:20
-
08:00
-
07:42
-
07:19
-
07:00
-
16:21
-
15:59
-
15:40
-
15:20
-
14:59
-
14:40
-
14:17
-
13:59
-
13:43
-
13:23
-
13:04
-
12:15
-
12:00