NVIDIA's Vera Rubin NVL72 coming to Lambda's Superintelligence Cloud
Blog post from Lambda
NVIDIA's Vera Rubin NVL72 platform is set to become a key component of Lambda's Superintelligence Cloud, marking a significant advancement in AI infrastructure. The Vera Rubin NVL72 rack functions as a singular, expansive GPU, enabling efficient scaling when integrated into Superclusters, which are designed to support AI teams in developing and deploying advanced models. This system offers significant improvements in performance and efficiency, crucial for engineers working on next-generation reasoning models. The architecture features a 72-GPU NVIDIA NVLink domain, which enhances model-parallel training and inference by minimizing communication overhead and increasing token throughput. It also supports massive-scale AI operations with up to 20.7 TB of HBM4 memory per rack and high bandwidth, while its integration into Lambda's infrastructure promises streamlined provisioning and management. The platform is anticipated to be available for production in the latter half of 2026, with facilities specifically designed for its high-power operation and cooling requirements, ensuring optimized performance and reliability.