ads header

Breaking News

NVIDIA Rubin: The Next Frontier of AI Supercomputing

 

NVIDIA Rubin: The Next Frontier of AI Supercomputing

At CES 2026, NVIDIA officially launched the Rubin platform, the successor to the Blackwell architecture. Named after the legendary astronomer Vera Rubin, this platform isn't just a new chip—it’s a massive leap in how we build and scale artificial intelligence.

If you thought Blackwell was fast, Rubin is designed to take us into the era of "Agentic AI" and trillion-parameter models with efficiency we haven't seen before.


What Makes Rubin Different?

The Rubin platform is built on a philosophy NVIDIA calls "Extreme Co-design." Instead of just focusing on the GPU, NVIDIA has integrated six distinct chips into a single, unified AI supercomputer architecture:

  1. Rubin GPU: Features 5th Gen Tensor Cores and 288GB of HBM4 memory.

  2. Vera CPU: A custom Arm-based processor with 88 "Olympus" cores, optimized for AI reasoning.

  3. NVLink 6 Switch: Doubles the interconnect speed to 3.6TB/s per GPU.

  4. ConnectX-9 SuperNIC: Provides 1.6Tb/s networking for massive data movement.

  5. BlueField-4 DPU: Acts as the "operating system" of the data center, handling security and storage.

  6. Spectrum-6 Ethernet Switch: Dramatically improves power efficiency for large-scale clusters.

The Stats: Rubin vs. Blackwell

The performance jumps are staggering. When compared to the previous Blackwell generation, the Rubin platform delivers:

  • 10x Lower Cost: Massive reduction in the cost per token for inference.

  • 4x Faster Training: Can train Mixture-of-Experts (MoE) models using four times fewer GPUs.

  • 50 Petaflops: Huge compute power for AI inferencing (NVFP4).

  • Higher Efficiency: The new Vera Rubin NVL72 rack system uses 5x less power for networking tasks.

Focus on Agentic AI

One of the most exciting reveals was the NVIDIA Inference Context Memory Storage Platform. This is designed specifically for "Agentic AI"—AI agents that need long-term memory and the ability to reason through complex, multi-step tasks. By moving memory management to the hardware level, Rubin allows these agents to "think" much faster and remember more context than ever before.


When Can We Get It?

NVIDIA confirmed that the Rubin platform is currently in full production. Partners like AWS, Google Cloud, and Microsoft are expected to start offering Rubin-based systems in the second half of 2026.

The Bottom Line: Rubin marks NVIDIA’s shift from being a "chip maker" to a "data center architect." It is the foundation for the next decade of AI evolution.

Read the official announcement here: NVIDIA Kicks Off the Next Generation of AI With Rubin


 

No comments

  What is Codex Security? Codex Security is an autonomous application security agent powered by OpenAI's frontier models (including GPT-...