NVIDIA의 전략은 확실하네요. CES 2026 Keynote 핵심 정리 | 자율주행 '플랫폼‘을 팔고 Vera-Rubin GPU, Cosmos와 DLSS 4.5까지
Video

NVIDIA의 전략은 확실하네요. CES 2026 Keynote 핵심 정리 | 자율주행 '플랫폼‘을 팔고 Vera-Rubin GPU, Cosmos와 DLSS 4.5까지

안될공학 - IT 테크 신기술
2026.01.11
·YouTube·by 이호민
#AI#Autonomous Driving#CES#CES 2026#GPU#HBM4#Nvidia#Supercomputer

Key Points

  • 1NVIDIA unveiled the VeraRubin AI supercomputer platform, integrating a new Vera CPU, Rubin GPU with HBM4, and advanced networking components like MV Link 6 and DPUs, designed to eliminate bottlenecks and optimize large-scale AI factory operations.
  • 2Key technological advancements include HBM4 with significantly increased bandwidth and capacity (288GB, 22 TB/s), a doubled MV Link speed, and dedicated DPUs for efficient data and memory hierarchy management to accelerate AI model training and inference.
  • 3The announcement also highlighted AlphaMayo, an open model providing data and simulation tools to accelerate autonomous driving development, alongside the Drive Hyperion platform for Level 2 to Level 4 autonomy, reinforcing NVIDIA's focus on physical AI and integrated solutions.

The paper details several key announcements and technological advancements from NVIDIA, primarily focusing on the Rubin platform for AI supercomputing, advancements in autonomous driving, the latest iteration of DLSS, and the concept of Physical AI.

The core of the discussion revolves around the VeraRubin GPU and the broader Rubin AI supercomputer platform, which is presented not merely as a chip but as a platform integrating six core components to form AI supercomputers:

  1. Vera CPU: An 88-core CPU (American Olympus core) with 1.5TB memory, utilizing MV C2C for 1.8TB data transfer, designed to efficiently orchestrate GPU workloads and minimize bottlenecks.
  2. Rubin GPU: The flagship GPU featuring HBM4, highlighting a significant leap in memory technology. SK Hynix's 16-layer HBM4 with 2048 channels substantially increases bandwidth. The Rubin GPU is specialized for agent-type workloads, long contexts, and efficient matrix/tensor operations for both learning and inference. It achieves 50 petaflops (MVFP4 precision), leveraging FP4 (4-bit floating-point) arithmetic to maximize computational density, a decision justified by the accelerating pace of AI model development. The HBM4 configuration boasts 288GB of capacity (from eight stacks), and a bandwidth of 2.2 TB/s. This increased capacity and bandwidth are crucial for storing large KV caches and managing attention-related memory traffic, allowing for the processing of larger models and longer contexts directly on the GPU, ultimately aiming to reduce training costs per token.
  3. MV Link 6 Switch: NVIDIA's proprietary inter-chip communication standard, upgraded to 3.6 TB/s bidirectional bandwidth, doubling the previous MV Link 5's 1.8 TB/s. This ensures high-speed data transfer between CPUs and GPUs, crucial for large-scale AI operations.
  4. Connect X9 NIC: A Network Interface Card supporting an 800GB Ethernet port with Palm OS, facilitating high-speed data transfer between servers and across racks in data centers.
  5. Bluefield 4 DPU: A Data Processing Unit designed to offload tasks from the CPU and GPU. It manages KV cache offloading to storage (e.g., SSDs), and handles network and security functions, allowing the main computing units to focus on core AI tasks. This hierarchy optimizes memory utilization by dynamically managing data between HBM and persistent storage.
  6. Spectrum X Ethernet Photonics (Co-packaged Optics - CPO): Integrates optical components directly into the Ethernet switch package. This addresses signal integrity challenges at high speeds by converting electrical signals to optical closer to the chip, ensuring clean and efficient communication within large-scale AI clusters.

The Rubin platform aims to consolidate these components into a unified, high-performance system for building "AI factories," promising enhanced speed, reliability, security, and deployment efficiency for massive AI supercomputers. The platform is pitched as a single "Rexscale product" or "DGX Spark"-like bundle.

Beyond the Rubin platform, the paper highlights advancements in specific domains:

Autonomous Driving:

  • AlphaMayo: An open model platform designed to accelerate autonomous driving development. It provides comprehensive data, simulations, tools, and datasets, covering the entire development loop from data collection and purification to testing, simulation, safety analysis, and compliance with regulations. It features a Vision Language Action (VLA) model composed of 10 trillion parameters, emphasizing explainable inference. The goal is to enable companies, especially those without extensive proprietary driving data like Tesla, to rapidly develop and deploy autonomous systems.
  • Drive Hyperion: A hardware platform aiming to accelerate autonomous driving from Level 2 to Level 4. It integrates computing capabilities with support for various sensors (cameras, radar, lidar) to process signals and apply them to vehicle platforms. Currently, it's being deployed in Mercedes-Benz vehicles, offering an open platform approach for OEMs, contrasting with Tesla's vertically integrated strategy.

DLSS 4.5:

  • Deep Learning Super Sampling is a gaming technology for Super Resolution Upsampling. DLSS 4.5 significantly improves image quality by leveraging advanced super-resolution models. A key feature is its ability to generate multiple frames (e.g., five additional frames from one rendered frame, resulting in a 6x increase in effective frame rate) using AI, thereby boosting frames per second for an enhanced gaming experience. This version is expected to be compatible with the upcoming RTX 50 series GPUs.

Physical AI (Cosmos):

  • This concept connects autonomous driving and robotics, emphasizing the development of models and tool systems (Cosmos) that learn and verify themselves in virtual environments and video simulations. It addresses the critical challenge in robotics and autonomous systems: the scarcity of real-world data for training. By leveraging video-based learning, robotics, and data simulation, Cosmos aims to provide a robust framework for pre-learning and verification, reducing the need for extensive real-world trials. This also ties back to the utilization of data centers and DGX Spark for such demanding AI workloads.

The overarching theme is NVIDIA's strategy to provide a comprehensive, vertically integrated ecosystem that accelerates AI development across various industries, from large-scale data centers to autonomous vehicles and robotics, by offering specialized hardware, software, and development platforms.