AI Reasoning Models and Agents: Unlocking Next-Generation Computing with CoreWeave and NVIDIA
AI Reasoning Models and Agents are Set to Transform Industries
AI reasoning models and agents are set to transform industries, but delivering their full potential at scale requires massive compute and optimized software. The “reasoning” process involves multiple models, generating many additional tokens, and demands infrastructure with a combination of high-speed communication, memory, and compute to ensure real-time, high-quality results.
Introducing NVIDIA GB200 NVL72 on CoreWeave
To meet this demand, CoreWeave has launched NVIDIA GB200 NVL72-based instances, becoming the first cloud service provider to make the NVIDIA Blackwell platform generally available. With rack-scale NVIDIA NVLink across 72 NVIDIA Blackwell GPUs and 36 NVIDIA Grace CPUs, these instances provide the scale and performance needed to build and deploy the next generation of AI reasoning models and agents.
NVIDIA GB200 NVL72 on CoreWeave
NVIDIA GB200 NVL72 is a liquid-cooled, rack-scale solution with a 72-GPU NVLink domain, which enables the six dozen GPUs to act as a single massive GPU. NVIDIA Blackwell features many technological breakthroughs that accelerate inference token generation, boosting performance while reducing service costs.
CoreWeave’s Portfolio of Managed Cloud Services
CoreWeave’s portfolio of managed cloud services is purpose-built for Blackwell. CoreWeave Kubernetes Service optimizes workload orchestration by exposing NVLink domain IDs, ensuring efficient scheduling within the same rack. Slurm on Kubernetes (SUNK) supports the topology block plug-in, enabling intelligent workload distribution across GB200 NVL72 racks. In addition, CoreWeave’s Observability Platform provides real-time insights into NVLink performance, GPU utilization, and temperatures.
Full-Stack Accelerated Computing Platform for Enterprise AI
NVIDIA’s full-stack AI platform pairs cutting-edge software with Blackwell-powered infrastructure to help enterprises build fast, accurate, and scalable AI agents. NVIDIA Blueprints provides pre-defined, customizable, ready-to-deploy reference workflows to help developers create real-world applications. NVIDIA NIM is a set of easy-to-use microservices designed for secure, reliable deployment of high-performance AI models for inference. NVIDIA NeMo includes tools for training, customization, and continuous improvement of AI models for modern enterprise use cases. Enterprises can use NVIDIA Blueprints, NIM, and NeMo to build and fine-tune models for their specialized AI agents.
Bringing Next-Generation AI to the Cloud
The general availability of NVIDIA GB200 NVL72-based instances on CoreWeave underscores the latest in the companies’ collaboration, focused on delivering the latest accelerated computing solutions to the cloud. With the launch of these instances, enterprises now have access to the scale and performance needed to power the next wave of AI reasoning models and agents.
Conclusion
The availability of NVIDIA GB200 NVL72-based instances on CoreWeave marks a significant milestone in the development of next-generation AI. By providing the scale and performance needed to build and deploy AI reasoning models and agents, CoreWeave and NVIDIA are empowering enterprises to unlock the full potential of AI and transform industries.
FAQs
Q: What is NVIDIA GB200 NVL72?
A: NVIDIA GB200 NVL72 is a liquid-cooled, rack-scale solution with a 72-GPU NVLink domain, enabling six dozen GPUs to act as a single massive GPU.
Q: What is CoreWeave’s portfolio of managed cloud services?
A: CoreWeave’s portfolio of managed cloud services is purpose-built for Blackwell, providing optimized workload orchestration, intelligent workload distribution, and real-time insights into NVLink performance, GPU utilization, and temperatures.
Q: What is NVIDIA’s full-stack AI platform?
A: NVIDIA’s full-stack AI platform pairs cutting-edge software with Blackwell-powered infrastructure to help enterprises build fast, accurate, and scalable AI agents.
Q: How can enterprises access NVIDIA GB200 NVL72-based instances on CoreWeave?
A: Customers can start provisioning GB200 NVL72-based instances through CoreWeave Kubernetes Service in the US-WEST-01 region using the gb200-4x instance ID. To get started, contact CoreWeave.