Skip to content Skip to sidebar Skip to footer

Amazon (AMZN) is rolling out a new house chip, and this one is aimed squarely at the silicon royalty that has been renting out its throne. Trainium3, unveiled this week at AWS re:Invent 2025 in Las Vegas, is Amazon Web Services’ first 3‑nanometer AI accelerator and the company’s boldest attempt yet to loosen Nvidia’s grip and crowd into the same frame as Google’s custom tensor silicon.

A 3‑Nanometer Shot at the Throne

Trainium3 is built on a 3‑nanometer process, shrinking transistors and swelling Amazon’s ambitions in one go. AWS says the chip delivers up to 4.4 times higher performance, nearly four times the memory bandwidth, and roughly four times better performance per watt than its Trainium2 predecessor, the sort of generational leap that suggests the marketing department did not have to strain for superlatives. Each chip can deliver several petaflops of FP8 compute and connect into large clusters, targeting the trillion‑parameter models that have become table stakes in modern AI.

The energy story is pitched almost as aggressively as the speed story. Trainium3 UltraServer systems are said to be roughly 40% more energy efficient than the prior generation, translating into more output tokens per megawatt and giving CIOs something to say in both the earnings call and the sustainability report. In practice, AWS is advertising up to 50% lower costs for training and inference versus older setups, hoping that cloud bills, not spec sheets, will ultimately win the argument.

UltraServers: Racks as Status Symbols

The new silicon arrives pre‑packaged in Amazon EC2 Trn3 UltraServers, dense AI boxes that are quickly becoming the new status symbol of hyperscale computing. A fully configured UltraServer can stitch together up to 144 Trainium3 chips, adding up to more than 360 petaflops of FP8 compute and over 20 terabytes of HBM3e memory in a single integrated system—enough horsepower to train frontier‑scale models or, at minimum, to impress the procurement committee.

For customers with ambitions that do not fit neatly inside one rack, AWS is extending things to UltraClusters 3.0, which can connect thousands of these UltraServers and, in theory, scale to on the order of a million Trainium‑class chips. That kind of bigness is aimed at customers training multimodal, trillion‑token models or serving real‑time inference to millions of users, but it also conveniently reinforces the idea that the only real limit to AI scale is how much of it a customer is willing to rent.

Pointed at Nvidia, Smiling at Google

Publicly, AWS insists it is still the Switzerland of accelerators, promising to keep buying “very significant” quantities of chips from Nvidia (NVDA), AMD, and Intel (INTC). Privately—and increasingly, not so privately—Trainium3 is clearly designed to be the in‑house alternative when customers start asking whether they really need to pay top‑shelf prices for every last GPU in the rack. On Amazon Bedrock, the company’s managed foundation model service, Trainium3 is already billed as the fastest accelerator, delivering up to three times the performance of Trainium2 and over five times the output tokens per megawatt, an efficiency pitch that lands squarely in Nvidia territory,

Google (GOOG), with its TPU line, gets pulled into the same comparison set, whether it likes it or not. Where Google positions TPUs as the in‑house muscle of its own cloud, Amazon is effectively arguing that it can do the same—with chips tuned for its own stack from Neuron to Bedrock, but sold as neutral ground for everyone else’s models. In this framing, customers are invited to think less about whose logo is stamped on the die and more about the price‑performance curve of training their next conversational agent or video generator.

From Frontier Labs to Regular IT

AWS is careful to point out that Trainium has already graduated from curiosity to revenue line, with Trainium2 described as a multibillion‑dollar business largely driven by a small set of very large customers. Trainium3 is supposed to broaden that base, moving from a handful of headline accounts—think frontier‑model labs and deep‑pocketed platforms—into a wider population of enterprises looking to run generative AI without handing over their entire capital budget to GPUs.

To smooth that path, Amazon is pitching Trainium3 not just as a cloud instance but as a building block for “AI factories,” dedicated infrastructure that can be dropped into customers’ existing data centers and wired back into AWS services. The result is a narrative in which AI hardware, once a specialized toy for research labs, becomes a configurable line item that IT departments can choose much like storage tiers or database engines—albeit with more petaflops and fewer refunds.

The Quiet Hedge on the Future

The long game is visible in the fine print. On the software side, AWS Neuron is expanding support for Trainium3, with deeper integration into popular frameworks like PyTorch, a bid to ensure developers are not locked out by the learning curve. On the hardware side, AWS is already previewing a Trainium4 roadmap, including plans to support Nvidia’s own NVLink Fusion interconnect, a reminder that in the cloud business, hedging is not just prudent—it is policy.

Whether Trainium3 can “crack” Nvidia’s grip on AI compute is an open question, but that may not be the only test that matters. If Amazon can persuade enough customers that custom silicon, wrapped in UltraServers and wired into Bedrock, delivers materially better economics than a wall of third‑party GPUs, it will have quietly shifted one of the most profitable parts of the AI stack back under its own roof—one 3‑nanometer sliver at a time.

The Sources…

  1. https://www.aboutamazon.com/news/aws/aws-re-invent-2025-ai-news-updates
  2. https://www.tipranks.com/news/amazon-amzn-takes-on-nvidia-and-google-with-new-trainium3-ai-chip
  3. https://www.moomoo.com/news/post/62341913/amazon-unveils-major-release-challenging-google-and-nvidia-ai-chip
  4. https://aws.amazon.com/ec2/instance-types/trn3/
  5. https://www.aboutamazon.com/news/aws/trainium-3-ultraserver-faster-ai-training-lower-cost
  6. https://technologymagazine.com/news/trainium3-new-aws-chip-promises-4x-performance-boost
  7. https://datacentremagazine.com/news/trainium3-new-aws-chip-promises-4x-performance-boost
  8. https://www.aboutamazon.com/news/aws/aws-data-centers-ai-factories
  9. https://www.constellationr.com/blog-news/insights/aws-reinvent-2025-feature-trainium3-scaling-secure-ai-agents
  10. https://aws.amazon.com/about-aws/whats-new/2025/12/amazon-ec2-trn3-ultraservers
  11. https://www.datacenterknowledge.com/data-center-chips/aws-launches-tranium3-chip-to-challenge-nvidia-ai-dominance
  12. https://awsdocs-neuron.readthedocs-hosted.com/en/latest/about-neuron/whats-new.html
  13. https://www.aa.com.tr/en/science-technology/amazon-introduces-its-new-ai-chip-trainium3/3760363
  14. https://theaieconomy.substack.com/p/aws-trainium3-ultraservers-trainium4-preview
  15. https://finance.yahoo.com/news/amazons-ai-chip-strike-trainium3-183152867.html
  16. https://techcrunch.com/2025/12/02/amazon-releases-an-impressive-new-ai-chip-and-teases-a-nvidia-friendly-roadmap/
  17. https://www.datacenterdynamics.com/en/news/aws-makes-trainium3-ultraservers-generally-available/
  18. https://www.businesswire.com/news/home/20251201046353/en/Trainium3-UltraServers-Now-Available-Enabling-Customers-to-Train-and-Deploy-AI-Models-Faster-at-Lower-Cost
  19. https://finance.yahoo.com/news/trainium3-ultraservers-now-available-enabling-183000469.html
  20. https://www.youtube.com/watch?v=4y3pMGIS6DU
Your Guide To Staying Informed In The Markets

Subscribe For Free Email Updates Access To Exclusive Research

Vista Partners — © 2025 — Vista Partners LLC (“Vista”) is a Registered Investment Advisor in the State of California. Vista is not licensed as a broker, broker-dealer, market maker, investment banker, or underwriter in any jurisdiction. By viewing this website and all of its pages, you agree to our terms. Read the full disclaimer here