Particle.news
Download on the App Store

Nvidia Unveils Vera Rubin at CES, Accelerating AI Server Roadmap to H2 2026

The full-stack platform debuts with aggressive, company-reported efficiency gains to steer hyperscalers toward Nvidia’s blueprint.

Overview

  • Microsoft, Google Cloud, AWS, Oracle, and CoreWeave said they plan to offer Rubin-based systems.
  • Nvidia claims Rubin cuts inference cost by 10x, reduces required training GPUs by 75%, and delivers up to 5x faster inference and 3.5x faster training versus Blackwell using HBM4, a redesigned Transformer Engine, adaptive compression, and next-gen NVLink.
  • The company says systems will be available in the second half of 2026, with analysts expecting the meaningful revenue impact as hyperscaler orders convert to shipments late in the year.
  • The early CES reveal was framed as a response to fresh accelerator announcements from AMD and Intel, with AMD showcasing its Helios rack-scale competitor.
  • Nvidia is positioning Rubin to power long-context, simulation-heavy “physical AI” workloads in areas such as autonomy, robotics, factories, digital twins, defense, healthcare, and logistics.