Jensen Huang Says Nvidia’s New Vera Rubin Chips Are in ‘Full Production’

0
11


Nvidia CEO Jensen Huang says that the company’s next-generation AI superchip platform, Vera Rubin, is on schedule to begin arriving to customers later this year. “Today, I can tell you that Vera Rubin is in full production,” Huang said during a press event on Monday at the annual CES technology trade show in Las Vegas.

Rubin will cut the cost of running AI models to about one-tenth of Nvidia’s current leading chip system, Blackwell, the company told analysts and journalists during a call on Sunday. Nvidia also said Rubin can train certain large models using roughly one-fourth as many chips as Blackwell requires. Taken together, those gains could make advanced AI systems significantly cheaper to operate and make it harder for Nvidia’s customers to justify moving away from its hardware.

Nvidia said on the call that two of its existing partners, Microsoft and CoreWeave, will be among the first companies to begin offering services powered by Rubin chips later this year. Two major AI data centers that Microsoft is currently building in Georgia and Wisconsin will eventually include thousands of Rubin chips, Nvidia added. Some of Nvidia’s partners have already started running their next-generation AI models on early Rubin systems, the company said.

The semiconductor giant also said it’s working with Red Hat, which makes open source enterprise software for banks, automakers, airlines, and government agencies, to offer more products that will run on the new Rubin chip system.

Nvidia’s latest chip platform is named after Vera Rubin, an American astronomer who reshaped how scientists understand the properties of galaxies. The system includes six different chips, including the Rubin GPU and an Vera CPU, both of which are built using Taiwan Semiconductor Manufacturing Company’s 3 nanometer fabrication process and the most advanced bandwidth memory technology currently available. Nvidia’s sixth-generation interconnect and switching technologies link the various chips together.

Each part of this chip system is “completely revolutionary and the best of its kind,” Huang proclaimed during the company’s CES press conference.

Nvidia has been developing the Rubin system for years, and Huang first announced the chips were coming during a keynote speech in 2024. Last year, the company said that systems built on Rubin would begin arriving in the second half of 2026.

It’s unclear exactly what Nvidia means by saying that Vera Rubin is in “full production.” Typically, production for chips this advanced—which Nvidia is building with its longtime partner TSMC—starts at low volume while the chips go through testing and validation and ramps up at a later stage.

“This CES announcement around Rubin is to tell investors, ‘We’re on track,’” says Austin Lyons, an analyst at Creative Strategists and author of the semiconductor industry newsletter Chipstrat. There were rumors on Wall Street that the Rubin GPU was running behind schedule, Lyons says, so Nvidia is now pushing back by saying it has cleared key development and testing steps, and it’s confident Rubin is still on course to begin scaling up production in the second half of 2026.

In 2024, Nvidia had to delay delivery of its then-new Blackwell chips due to a design flaw that caused them to overheat when they were connected together in server racks. Shipments for Blackwell were back on schedule by the middle of 2025.

As the AI industry rapidly expands, software companies and cloud service providers have had to fiercely compete for access to Nvidia’s newest GPUs. Demand will likely be just as high for Rubin. But some firms are also hedging their bets by investing in their own custom chip designs. OpenAI, for example, has said it is working with Broadcom to build bespoke silicon for its next generation of AI models. These partnerships highlight a longer-term risk for Nvidia: Customers that design their own chips can gain a level of control over their hardware that the company doesn’t offer.

But Lyons says today’s announcements demonstrate how Nvidia is evolving beyond merely offering GPUs to becoming a “full AI system architect, spanning compute, networking, memory hierarchy, storage, and software orchestration.” Even as hyperscalers pour money into custom silicon, he adds, Nvidia’s tightly integrated platform “is getting harder to displace.”

LEAVE A REPLY

Please enter your comment!
Please enter your name here