All News

Tesla Ends Dojo Project and Recasts AI Strategy

Tesla has quietly shuttered Dojo, its custom AI supercomputer program, and disbanded the team behind it. Once billed as critical to Full Self-Driving and Optimus robot training, Dojo lost internal momentum as Tesla shifted investment toward Cortex and a $16.5B AI6 chip deal with Samsung. The closure highlights risks of bespoke silicon projects and the rising appeal of partnership-based compute strategies.

Published September 2, 2025 at 01:13 PM EDT in Artificial Intelligence (AI)

Tesla has officially pulled the plug on Dojo, the in-house AI supercomputer program that promised to be the backbone of the company’s Full Self-Driving (FSD) and Optimus robot ambitions. In mid‑August 2025 Tesla disbanded the Dojo team and closed the project, marking a stark reversal after years of public evangelism by Elon Musk.

What Dojo was meant to achieve

Dojo was Tesla’s custom training supercomputer designed around proprietary D1 (and planned D2) chips to train vision‑only neural networks at massive scale. The goal: convert billions of miles of car video into models that could enable true autonomy and power humanoid robots — while reducing reliance on expensive third‑party GPUs.

Why Tesla shut Dojo down

The end came after a strategic pivot: Tesla inked a $16.5 billion deal for AI6 chips from Samsung and doubled down on Cortex, a large GPU training cluster in Austin. Musk called Dojo 2 an “evolutionary dead end,” arguing convergence around AI6 made continued Dojo investment unjustifiable.

Talent flight and organizational friction accelerated the decision. Dojo’s leader departed and dozens of engineers left to found a startup, undercutting the institutional knowledge required to complete a bespoke silicon program.

What this shift signals

Dojo’s closure is less an accusation of failure than a lesson in trade‑offs. Building vertical, custom compute can yield advantages in latency and cost, but it is capital‑intensive, dependent on rare talent, and vulnerable to ecosystem lock‑in. Partnering for chips and expanding large GPU clusters can be faster and less risky when market access and software compatibility matter.

Industry players should ask practical questions: Is custom silicon required to meet product SLAs? Can software and tooling be adapted cost‑effectively? How will talent and IP be retained if direction changes?

  • Cost and procurement risk: buying economies of scale vs. building bespoke wares
  • Software portability: most AI stacks assume GPUs, so custom chips may demand costly rewrites
  • Talent and IP retention: leaving a project stranded can fragment teams and create startups that compete for the same niche

How organizations should respond

Leaders building AI infrastructure can take Boeing’s or Apple’s lesson: align hardware bets to clear product milestones, keep software stacks portable, and model both upside and downside scenarios. Hybrid strategies — combining partner chips, cloud bursts, and selective custom silicon for unique workloads — often hit the best balance.

For policy makers and enterprise buyers, Dojo’s story underlines the value of transparent procurement, ecosystem compatibility, and contingency plans when critical compute is concentrated in a single internal program.

QuarkyByte’s approach is to quantify these trade‑offs for technical and business leaders: map cost vs. throughput curves, simulate training timelines under different chipsets, and design organizational safeguards so projects survive strategic pivots. Dojo is a reminder that bold engineering bets work best when they’re measured against market realities.

Tesla will still fund a supercomputer in Buffalo, but not Dojo as originally conceived. Whether the company now accelerates its robotaxi and humanoid ambitions on Samsung AI6 chips and Cortex clusters — or faces new challenges — remains an unfolding test of how AI strategy and execution intersect.

Keep Reading

View All
The Future of Business is AI

AI Tools Built for Agencies That Move Fast.

Let QuarkyByte model the trade-offs between in-house silicon and partner-led AI stacks, simulate cost vs. performance for training fleets, and design retention plans to protect hard-won talent. Connect with our analysts to map an optimized compute roadmap tied to measurable product milestones.