5 Sources
5 Sources
[1]
Nvidia Offers First Look at Vera Rubin Superchip
Nvidia CEO Jensen Huang (Credit: Kent Nishimura/Bloomberg via Getty Images) Don't miss out on our latest stories. Add PCMag as a preferred source on Google. While batting away deepfakes, Nvidia's GTC DC show showed off the superchip platform that will power next-generation data centers: Vera Rubin. This single-board system that combines an 88-core Vera CPU with twin Rubin GPUs will be the backbone of Nvidia's 2026 strategy to continue powering the AI revolution, and its next trillion dollars in valuation. "This is the next-generation Rubin," CEO Jensen Huang said at GTC. "While we are shipping GB300, we are preparing Rubin to be in production this time next year, maybe slightly earlier. This is just an incredibly beautiful computer," with 100 Petaflops of FP4 performance for AI. Although Nvidia is facing some skepticism over its performance claims of late, there's no denying that Vera Rubin will be the most powerful "Superchip" ever made when it hits mass production in 2026. It's built using next-generation architectures for Nvidia's CPU and GPU designs, and more than doubles the number of CPU cores over the current-generation GB200 Grace Blackwell flagship. Each Rubin GPU has access to 288GB of HBM4 memory, with the GPUs and CPUs linked together using an enhanced NVLINK-C2C with increased throughput, as per VideoCardz. All of that equates to a huge increase in performance over existing Grace Blackwell systems that Nvidia has been hyping so much in its big AI data center deals this year. The full NVL144 system will be able to offer 3.6 Exaflops FP4 inference and 1.2 Exaflops FP8 in training workloads. That works out to around 3.3 times faster than Nvidia's GB300 NVL72 platform. As if that wasn't enough, Nvidia teased what's coming after what's coming next, in the form of Rubin Ultra NVL576. Slated to debut sometime in the first of 2027, the super-sized version of Vera Rubin would have more GPUs, a terabyte of HBM4e memory, and up to four times the performance of the standard Vera Rubin configuration. Nvidia is bringing graphics card FOMO to the data center space in dramatic fashion. Not only is it promising something really exciting just next year, but get your investor wallets ready because there's going to be even more capable and expensive hardware coming after that. The question is, how long can this gravy train keep rolling?
[2]
Nvidia reveals Vera Rubin Superchip for the first time -- incredibly compact board features 88-core Vera CPU, two Rubin GPUs, and 8 SOCAMM modules
At its GTC keynote in DC on Tuesday, Nvidia unveiled its next-generation Vera Rubin Superchip, comprising two Rubin GPUs for AI and HPC as well as its custom 88-core Vera CPU. All three components will be in production this time next year, Nvidia says. "This is the next generation Rubin," said Jensen Huang, chief executive of Nvidia, at GTC. "While we are shipping GB300, we are preparing Rubin to be in production this time next year, maybe slightly earlier. [...] This is just an incredibly beautiful computer. So, this is amazing, this is 100 PetaFLOPS [of FP4 performance for AI]." Indeed, Nvidia's Superchips tend to look more like a motherboard (on an extremely thick PCB) rather than a 'chip' as they carry a general-purpose custom CPU and two high-performance compute GPUs for AI and HPC workloads. The Vera Rubin Superchip is not an exception, and the board carries Nvidia's next-generation 88-core Vera CPU surrounded by SOCAMM2 memory modules carrying LPDDR memory and two Rubin GPUs covered with two large rectangular aluminum heat spreaders. Markings on the Rubin GPU say that they were packaged in Taiwan on the 38 week of 2025, which is late September, something that proves that the company has been playing with the new processor for some time now. The size of the heatspreader is about the same size as the heatspreader of Blackwell processors, so we cannot figure out the exact size of GPU packaging or die sizes of compute chiplets. Meanwhile, the Vera CPU does not seem to be monolithic as it has visible internal seams, implying that we are dealing with a multi-chiplet design. A picture of the board that Nvidia demonstrated once again reveals that each Rubin GPU is comprised of two compute chiplets, eight HBM4 memory stacks, and one or two I/O chiplets. Interestingly, but this time around, Nvidia demonstrated the Vera CPU with a very distinct I/O chiplet located next to it. Also, the image shows green features coming from the I/O pads of the CPU die, the purpose of which is unknown. Perhaps, some of Vera's I/O capabilities are enabled by external chiplets that are located beneath the CPU itself. Of course, we are speculating, but there is definitely an intrigue with the Vera processor. Interestingly, the Vera Rubin Superchip board no longer has industry-standard slots for cabled connectors. Instead, there are two NVLink backplane connectors on top to connect GPUs to the NVLink switch, enabling scale-up scalability within a rack and three connectors on the bottom edge for power, PCIe, CXL, and so on. In general, Nvidia's Vera Rubin Superchip board looks quite baked, so expect the unit to ship sometime in late 2026 and get deployed by early 2027.
[3]
Nvidia's new six-trillion transistor Vera Rubin 'superchip' for AI makes the 92-billion transistor RTX 5090 gaming GPU look positively puny
Nvidia's ever-optimistic CEO, Jensen Huang, has revealed more details of the company's next-gen AI platform at the same GTC event where the company's quantum computing plans were outlined. He calls it the Vera Rubin "superchip" and claims it contains no fewer than six trillion transistors. Yup, six trillion. In very simple terms, that means the Vera Rubin superchip sports 60 times the number of transistors of an Nvidia RTX 5090 gaming GPU. And that is a completely bananas comparison. What's more, the Vera Rubin superchip is capable of 100 petaflops of raw compute. And that is exactly 100 times more compute than Nvidia claims for its neat little DGX Spark AI box. Oh, and it packs 88 custom Arm cores suporting 176 software threads for conventional CPU compute, too. On a side note, that Arm CPU bit (which alone is known as Vera) is interesting from a PC perspective as you might expect its custom designed CPU cores, which notably support multi-threading, could turn up in a PC processor in the future. Watch this space. At this point you may be wondering how any of this is possible. Six trillion transistors, 100 PF of compute, 88 CPU cores, all in one chip? OK, Vera Rubin is expected to upgrade to TSMC's latest N3 process node. But the N5-derived node that Nvidia currently uses for its GPUs tops out at a little over 100 billion transistors for a single, monolithic chip. N3 is more dense than N5, for sure. But not anything like enough to go from about 100 billion to six trillion transistors. Actually, it turns out that Vera Rubin isn't just more than one chip, it's more than one chip package. Huang held up an engineering sample of Vera Rubin and what's clear to see are the three main chip packages, one for the CPU and two GPU packages. In a later part of the presentation, it becomes clear that each GPU package actually contains two GPUs. So that's four GPUs total, plus the CPU. A close up on the Vera CPU package also reveals it to be a chiplet design, further elevating the overall transistor count headroom. That said, each Rubin AI GPU die is monolithic and described as reticle-sized, which means it's as big as TSMC's production process will allow. But even if you were generous and assumed that TSMC could squeeze 200 billion transistors into a single monolithic chip using N3, you'd need 30 such chips to get to that six trillion figure. Actually, Nvidia also revealed a "Rubin Ultra" GPU package with four Rubin GPU dies, so this stuff is just going to scale and scale. But, still, it's unclear how Nvidia is doing its transistor count. I'd have to assume that everything, including all the memory for the GPUs and CPU, plus the supporting chips on the Vera Rubin motherboard, is thrown in. However Nvidia is coming up with the numbers, the Vera Rubin AI board (as opposed to "chip" as Huang is calling it), is an unfathomably complex thing. Indeed, Huang says it's 100 times faster than the Nvidia equivalent product from nine years ago, the DGX One, which Huang says he delivered to what must then have been a nascent OpenAI. The rest, as they say, is history. And trillions and trillions of transistors. And trillions of dollars for Nvidia. For the record, Huang says Vera Rubin goes into production "this time next year, maybe even slightly earlier" and it probably won't be cheap. So get saving up.
[4]
NVIDIA Vera Rubin Superchip Unveiled at GTC 2025 - 88-Core Arm CPU and Dual Rubin GPUs
At the GTC event in Washington D.C., NVIDIA gave the world its first look at the new Vera Rubin Superchip system -- a massive leap in design for high-performance computing and AI workloads. This setup combines two Rubin GPUs with a single Vera CPU inside one package, forming what NVIDIA calls a "symbiotic" processor architecture. The Vera CPU packs 88 custom-built Arm-based cores with 176 threads, giving it serious muscle for data-heavy workloads. Together, the two Rubin GPUs deliver about 100 PetaFLOPS of FP4 performance -- that's 50 PetaFLOPS each -- and are aimed at large-scale AI training and inference tasks. Each Rubin GPU looks like a twin-chiplet design, using two giant dies of around 830 mmΒ² apiece and eight HBM4 memory stacks. That adds up to roughly 288 GB of HBM4 per GPU, or 576 GB in total for the full Superchip. On top of that, the board also carries SOCAMM2 LPDDR5X memory modules, providing around 1.5 TB of extra low-latency system RAM per Vera CPU. NVLink has been upgraded to an incredible 1.8 TB/s bandwidth, allowing massive data flow between CPU and GPU -- essential for today's enormous AI models. NVIDIA built this system as the heart of its future NVL-class data center hardware. The smaller NVL144 configuration will use 144 GPUs to reach around 3.6 ExaFLOPS of FP4 inference performance and about 1.2 ExaFLOPS for FP8 training. That setup combines roughly 13 TB/s of memory bandwidth and 75 TB of system memory. For the biggest data centers, the upcoming Rubin Ultra NVL576 design scales even further, with total compute pushing close to 15 ExaFLOPS FP4 and hundreds of terabytes of fast memory. Production of the Vera Rubin Superchip is expected to begin in 2026, with wider rollout planned for 2027. Engineering samples are already being tested in labs, and early data suggests NVIDIA is taking a fully in-house approach with the Vera CPU instead of licensing Arm's Neoverse cores as it did with the Grace chips. Source: NVIDIA
[5]
NVIDIA Shows Next-Gen Vera Rubin Superchip For The First Time, Two Massive GPUs Primed For Production Next Year
NVIDIA has shown off its next-gen Vera Rubin Superchip for the first time at GTC in Washington, primed to spark the next wave of AI. At GTC October 2025, NVIDIA's CEO Jensen Huang showcased the next-gen Vera Rubin Superchip. This is the first time that we are seeing an actual sample of the motherboard, or Superchip as NVIDIA loves to call it, featuring the Vera CPU and two massive Rubin GPUs. The motherboard also hosts lots of LPDDR system memory, which will be combined with HBM4 memory featured on the Rubin GPUs. In terms of specifications, the NVIDIA Vera Rubin NVL144 platform will utilize two new chips. The Rubin GPU will make use of two Reticle-sized chips, with up to 50 PFLOPs of FP4 performance and 288 GB of next-gen HBM4 memory. These chips will be equipped alongside an 88-core Vera CPU with a custom Arm architecture, 176 threads, and up to 1.8 TB/s of NVLINK-C2C interconnect. In terms of performance scaling, the NVIDIA Vera Rubin NVL144 platform will feature 3.6 Exaflops of FP4 inference and 1.2 Exaflops of FP8 Training capabilities, a 3.3x increase over GB300 NVL72, 13 TB/s of HBM4 memory with 75 TB of fast memory, a 60% uplift over GB300 and 2x the NVLINK and CX9 capabilities, rated at up to 260 TB/s and 28.8 TB/s, respectively. The second platform will be arriving in the second half of 2027 and will be called Rubin Ultra. This platform will scale the NVL system from 144 to 576. The architecture for the CPU remains the same, but the Rubin Ultra GPU will feature four reticle-sized chips, offering up to 100 PFLOPS of FP4 and a total HBM4e capacity of 1 TB scattered across 16 HBM sites. In terms of performance scaling, the NVIDIA Rubin Ultra NVL576 platform will feature 15 Exaflops of FP4 inference and 5 Exaflops of FP8 Training capabilities, a 14x increase over GB300 NVL72, 4.6 PB/s of HBM4 memory with 365 TB of fast memory, a 8x uplift over GB300 and 12x the NVLINK and 8x the CX9 capabilities, rated at up to 1.5 PB/s and 115.2 TB/s, respectively.
Share
Share
Copy Link
Nvidia reveals its next-generation Vera Rubin Superchip at GTC, featuring 88-core Vera CPU, dual Rubin GPUs, and 100 PetaFLOPS performance. The system promises 3.3x performance increase over current GB300 platforms and targets 2026 production.
Nvidia has officially revealed its next-generation Vera Rubin Superchip at the GTC conference in Washington D.C., marking a significant leap in AI computing capabilities. CEO Jensen Huang presented the first physical sample of what the company calls an "incredibly beautiful computer," featuring an unprecedented six trillion transistors and delivering 100 PetaFLOPS of FP4 performance for AI workloads
1
.
Source: PC Magazine
The Vera Rubin Superchip represents a dramatic evolution from traditional chip designs, functioning more like a sophisticated motherboard that combines multiple processing units into a single package. This "symbiotic" processor architecture integrates an 88-core Vera CPU with dual Rubin GPUs, creating a unified system optimized for large-scale AI training and inference tasks
4
.The Vera CPU utilizes custom-built Arm-based cores supporting 176 software threads, representing a significant departure from Nvidia's previous Grace processors that licensed Arm's Neoverse cores. Early analysis suggests this marks Nvidia's move toward a fully in-house CPU design approach
4
. The CPU package appears to feature a multi-chiplet design with visible internal seams and distinct I/O chiplets2
.
Source: Tom's Hardware
Each Rubin GPU incorporates two compute chiplets with dies measuring approximately 830mmΒ² each, representing reticle-sized chips that maximize TSMC's N3 process capabilities. The GPUs feature eight HBM4 memory stacks each, providing 288GB of high-bandwidth memory per GPU for a total of 576GB across the dual-GPU configuration
4
. Additionally, the system incorporates SOCAMM2 LPDDR5X memory modules, delivering approximately 1.5TB of low-latency system RAM4
.The enhanced NVLink-C2C interconnect achieves 1.8TB/s bandwidth, enabling massive data flow between CPU and GPU components essential for processing today's enormous AI models
4
. The board design eliminates traditional cabled connectors in favor of two NVLink backplane connectors for GPU-to-switch connectivity and three bottom-edge connectors for power, PCIe, and CXL connections2
.The NVL144 system configuration utilizing 144 GPUs will deliver 3.6 ExaFLOPS of FP4 inference performance and 1.2 ExaFLOPS for FP8 training workloads, representing a 3.3x performance increase over Nvidia's current GB300 NVL72 platform
1
. This configuration combines 13TB/s of memory bandwidth with 75TB of system memory, providing a 60% uplift in memory capabilities5
.Related Stories
Nvidia expects to begin production of the Vera Rubin Superchip in late 2026, with broader deployment anticipated in early 2027. Engineering samples are already undergoing testing in laboratories, with markings indicating GPU packaging occurred in Taiwan during the 38th week of 2025
2
.Looking beyond the standard Vera Rubin configuration, Nvidia has announced the Rubin Ultra NVL576 platform scheduled for the first half of 2027. This enhanced system will scale from 144 to 576 GPUs, featuring four reticle-sized chips per GPU package and delivering up to 100 PFLOPS of FP4 performance with 1TB of HBM4e memory capacity
5
. The Rubin Ultra configuration promises 15 ExaFLOPS of FP4 inference and 5 ExaFLOPS of FP8 training capabilities, representing a 14x increase over current GB300 platforms5
.
Source: Guru3D.com
Summarized by
Navi
[1]
[2]
[4]
19 Mar 2025β’Technology

10 Jun 2025β’Technology

23 Aug 2025β’Technology
