Overview
- Microsoft, Google Cloud, AWS, Oracle, and CoreWeave said they plan to offer Rubin-based systems.
- Nvidia claims Rubin cuts inference cost by 10x, reduces required training GPUs by 75%, and delivers up to 5x faster inference and 3.5x faster training versus Blackwell using HBM4, a redesigned Transformer Engine, adaptive compression, and next-gen NVLink.
- The company says systems will be available in the second half of 2026, with analysts expecting the meaningful revenue impact as hyperscaler orders convert to shipments late in the year.
- The early CES reveal was framed as a response to fresh accelerator announcements from AMD and Intel, with AMD showcasing its Helios rack-scale competitor.
- Nvidia is positioning Rubin to power long-context, simulation-heavy “physical AI” workloads in areas such as autonomy, robotics, factories, digital twins, defense, healthcare, and logistics.