Amazon's Custom AI Chips Now Generate Over $20 Billion a Year — and Jassy Hints They Could Be Sold Externally, Challenging NVIDIA
Mubboo Editorial Team
April 12, 2026 · 4 min read
Buried inside Andy Jassy's annual shareholder letter — the one everyone noticed for the $15 billion AWS AI revenue disclosure — was a second number that may matter more in the long run: Amazon's custom chip business is now generating over $20 billion in annualized revenue, growing at triple-digit rates. And Jassy hinted that Amazon may start selling those chips to third parties.
The chip business nobody expected
Amazon designs three custom silicon lines in-house. Graviton handles general compute workloads. Trainium powers AI training. Nitro manages security and networking. Together, they crossed $20 billion in annualized revenue as of Jassy's April 9 letter — doubling from the $10 billion milestone Amazon reported earlier in 2026 and growing at triple-digit year-over-year rates.
Demand is outstripping supply. Jassy disclosed that two large, unnamed customers attempted to purchase all of Amazon's available Graviton capacity for 2026. Amazon refused, choosing to keep the chips available across its broader customer base rather than lock them into exclusive deals. But the demand signal prompted a notable comment: "There's so much demand for our chips that it's quite possible we'll sell racks of them to third parties in the future," Jassy wrote.
If Amazon follows through, it enters direct competition with NVIDIA in the AI chip market — not as a startup with unproven silicon, but as a company already running these chips at hyperscale inside the world's largest cloud provider.
Why does this matter for the AI ecosystem?
NVIDIA dominates AI chip supply today. Amazon, Google with its TPU chips, and Microsoft with its Maia accelerators are all developing alternatives, but NVIDIA still commands the majority of AI training and inference compute. Amazon's position is different from a typical challenger because it already operates its custom chips at massive scale inside AWS. Customers do not need to build new infrastructure or rearchitect their workloads — they access Trainium and Graviton through the same AWS services they already use.
The usage numbers back up the demand story. Amazon Bedrock, the managed AI service that runs foundation models on AWS infrastructure including Trainium chips, processed more tokens in Q1 2026 than in all prior periods combined. Inference volumes were nearly doubling month-over-month through March. That growth feeds directly into custom chip utilization — every token processed on Trainium is a token not processed on NVIDIA hardware.
The AI chip market is shifting from single-supplier dependency toward a multi-vendor ecosystem. Amazon selling chips externally would accelerate that shift by giving companies outside AWS access to silicon that has already been validated at production scale.
The financial picture behind the bet
Amazon is investing approximately $200 billion in capital expenditure in 2026, the majority directed at AI infrastructure. The spending is visible in the company's cash flow: 2025 revenue grew 12 percent to $717 billion and operating income rose 17 percent to $80 billion, but free cash flow dropped from $38 billion to $11 billion as capex surged. The stock rose roughly 5 percent on the day of the letter's release.
Jassy addressed the spending directly: "We're not investing on a hunch." The custom chip business generating $20 billion in measurable revenue — not projections, not addressable market estimates — is the strongest evidence that Amazon's infrastructure bets are converting into real returns. The question is no longer whether Amazon can build competitive AI chips. It is whether it will sell them to the companies currently buying from NVIDIA.
Mubboo's Take
Amazon's custom chip business generating $20 billion in revenue is a signal that the cost of AI inference is headed down. When multiple companies — Amazon, Google, and eventually their third-party customers — compete on AI chip supply, the price of running AI models falls for everyone. For consumer platforms like Mubboo that depend on AI for content production, product comparison, and travel planning, cheaper inference means doing more with the same budget. The AI chip wars are not a hardware story. They are a story about how fast AI-powered consumer services become affordable enough to operate at scale.
Mubboo Editorial Team
The Mubboo Editorial Team covers the latest in AI, consumer technology, e-commerce, and travel.