Date:

NVIDIA Blackwell Now Generally Available in the Cloud

AI Reasoning Models and Agents: Unlocking Next-Generation Computing with CoreWeave and NVIDIA

AI Reasoning Models and Agents are Set to Transform Industries

AI reasoning models and agents are set to transform industries, but delivering their full potential at scale requires massive compute and optimized software. The “reasoning” process involves multiple models, generating many additional tokens, and demands infrastructure with a combination of high-speed communication, memory, and compute to ensure real-time, high-quality results.

Introducing NVIDIA GB200 NVL72 on CoreWeave

To meet this demand, CoreWeave has launched NVIDIA GB200 NVL72-based instances, becoming the first cloud service provider to make the NVIDIA Blackwell platform generally available. With rack-scale NVIDIA NVLink across 72 NVIDIA Blackwell GPUs and 36 NVIDIA Grace CPUs, these instances provide the scale and performance needed to build and deploy the next generation of AI reasoning models and agents.

NVIDIA GB200 NVL72 on CoreWeave

NVIDIA GB200 NVL72 is a liquid-cooled, rack-scale solution with a 72-GPU NVLink domain, which enables the six dozen GPUs to act as a single massive GPU. NVIDIA Blackwell features many technological breakthroughs that accelerate inference token generation, boosting performance while reducing service costs.

CoreWeave’s Portfolio of Managed Cloud Services

CoreWeave’s portfolio of managed cloud services is purpose-built for Blackwell. CoreWeave Kubernetes Service optimizes workload orchestration by exposing NVLink domain IDs, ensuring efficient scheduling within the same rack. Slurm on Kubernetes (SUNK) supports the topology block plug-in, enabling intelligent workload distribution across GB200 NVL72 racks. In addition, CoreWeave’s Observability Platform provides real-time insights into NVLink performance, GPU utilization, and temperatures.

Full-Stack Accelerated Computing Platform for Enterprise AI

NVIDIA’s full-stack AI platform pairs cutting-edge software with Blackwell-powered infrastructure to help enterprises build fast, accurate, and scalable AI agents. NVIDIA Blueprints provides pre-defined, customizable, ready-to-deploy reference workflows to help developers create real-world applications. NVIDIA NIM is a set of easy-to-use microservices designed for secure, reliable deployment of high-performance AI models for inference. NVIDIA NeMo includes tools for training, customization, and continuous improvement of AI models for modern enterprise use cases. Enterprises can use NVIDIA Blueprints, NIM, and NeMo to build and fine-tune models for their specialized AI agents.

Bringing Next-Generation AI to the Cloud

The general availability of NVIDIA GB200 NVL72-based instances on CoreWeave underscores the latest in the companies’ collaboration, focused on delivering the latest accelerated computing solutions to the cloud. With the launch of these instances, enterprises now have access to the scale and performance needed to power the next wave of AI reasoning models and agents.

Conclusion

The availability of NVIDIA GB200 NVL72-based instances on CoreWeave marks a significant milestone in the development of next-generation AI. By providing the scale and performance needed to build and deploy AI reasoning models and agents, CoreWeave and NVIDIA are empowering enterprises to unlock the full potential of AI and transform industries.

FAQs

Q: What is NVIDIA GB200 NVL72?
A: NVIDIA GB200 NVL72 is a liquid-cooled, rack-scale solution with a 72-GPU NVLink domain, enabling six dozen GPUs to act as a single massive GPU.

Q: What is CoreWeave’s portfolio of managed cloud services?
A: CoreWeave’s portfolio of managed cloud services is purpose-built for Blackwell, providing optimized workload orchestration, intelligent workload distribution, and real-time insights into NVLink performance, GPU utilization, and temperatures.

Q: What is NVIDIA’s full-stack AI platform?
A: NVIDIA’s full-stack AI platform pairs cutting-edge software with Blackwell-powered infrastructure to help enterprises build fast, accurate, and scalable AI agents.

Q: How can enterprises access NVIDIA GB200 NVL72-based instances on CoreWeave?
A: Customers can start provisioning GB200 NVL72-based instances through CoreWeave Kubernetes Service in the US-WEST-01 region using the gb200-4x instance ID. To get started, contact CoreWeave.

Latest stories

Read More

Titanfall Battle Royale

A Surprising Turn: How Titanfall 3 Became Apex Legends The...

AI-Powered Skin Cancer Prevention

AI-Assisted Cancer Diagnosis: The Future of Skin Cancer Detection Remarkable...

ASUS Zenbook A14: A Certain ‘Air’ About It

The Thin and Light Laptop Has Met Its Match The...

Try Apple’s Invites App for Event Planning

When you're inviting friends, family, or co-workers to an...

Declarative Pipelines for Data Teams

Declarative Data Pipelines: Moving from Code to Configuration Jonathan Bhaskar...

AI Agent Boosts Data Task Efficiency

New AI Agent Reduces Data Tasks by Up to...

Google Cloud Sales Disappoint

Alphabet's Earnings Fall Short of Expectations Revenue and Profit Alphabet, Google's...

Google’s Ambitious Plan for Search Evolution in 2025

Google Search Embarks on AI Journey Google Search is in...

LEAVE A REPLY

Please enter your comment!
Please enter your name here