Date:

NVIDIA Blackwell Now Generally Available in the Cloud

AI Reasoning Models and Agents: Unlocking Next-Generation Computing with CoreWeave and NVIDIA

AI Reasoning Models and Agents are Set to Transform Industries

AI reasoning models and agents are set to transform industries, but delivering their full potential at scale requires massive compute and optimized software. The “reasoning” process involves multiple models, generating many additional tokens, and demands infrastructure with a combination of high-speed communication, memory, and compute to ensure real-time, high-quality results.

Introducing NVIDIA GB200 NVL72 on CoreWeave

To meet this demand, CoreWeave has launched NVIDIA GB200 NVL72-based instances, becoming the first cloud service provider to make the NVIDIA Blackwell platform generally available. With rack-scale NVIDIA NVLink across 72 NVIDIA Blackwell GPUs and 36 NVIDIA Grace CPUs, these instances provide the scale and performance needed to build and deploy the next generation of AI reasoning models and agents.

NVIDIA GB200 NVL72 on CoreWeave

NVIDIA GB200 NVL72 is a liquid-cooled, rack-scale solution with a 72-GPU NVLink domain, which enables the six dozen GPUs to act as a single massive GPU. NVIDIA Blackwell features many technological breakthroughs that accelerate inference token generation, boosting performance while reducing service costs.

CoreWeave’s Portfolio of Managed Cloud Services

CoreWeave’s portfolio of managed cloud services is purpose-built for Blackwell. CoreWeave Kubernetes Service optimizes workload orchestration by exposing NVLink domain IDs, ensuring efficient scheduling within the same rack. Slurm on Kubernetes (SUNK) supports the topology block plug-in, enabling intelligent workload distribution across GB200 NVL72 racks. In addition, CoreWeave’s Observability Platform provides real-time insights into NVLink performance, GPU utilization, and temperatures.

Full-Stack Accelerated Computing Platform for Enterprise AI

NVIDIA’s full-stack AI platform pairs cutting-edge software with Blackwell-powered infrastructure to help enterprises build fast, accurate, and scalable AI agents. NVIDIA Blueprints provides pre-defined, customizable, ready-to-deploy reference workflows to help developers create real-world applications. NVIDIA NIM is a set of easy-to-use microservices designed for secure, reliable deployment of high-performance AI models for inference. NVIDIA NeMo includes tools for training, customization, and continuous improvement of AI models for modern enterprise use cases. Enterprises can use NVIDIA Blueprints, NIM, and NeMo to build and fine-tune models for their specialized AI agents.

Bringing Next-Generation AI to the Cloud

The general availability of NVIDIA GB200 NVL72-based instances on CoreWeave underscores the latest in the companies’ collaboration, focused on delivering the latest accelerated computing solutions to the cloud. With the launch of these instances, enterprises now have access to the scale and performance needed to power the next wave of AI reasoning models and agents.

Conclusion

The availability of NVIDIA GB200 NVL72-based instances on CoreWeave marks a significant milestone in the development of next-generation AI. By providing the scale and performance needed to build and deploy AI reasoning models and agents, CoreWeave and NVIDIA are empowering enterprises to unlock the full potential of AI and transform industries.

FAQs

Q: What is NVIDIA GB200 NVL72?
A: NVIDIA GB200 NVL72 is a liquid-cooled, rack-scale solution with a 72-GPU NVLink domain, enabling six dozen GPUs to act as a single massive GPU.

Q: What is CoreWeave’s portfolio of managed cloud services?
A: CoreWeave’s portfolio of managed cloud services is purpose-built for Blackwell, providing optimized workload orchestration, intelligent workload distribution, and real-time insights into NVLink performance, GPU utilization, and temperatures.

Q: What is NVIDIA’s full-stack AI platform?
A: NVIDIA’s full-stack AI platform pairs cutting-edge software with Blackwell-powered infrastructure to help enterprises build fast, accurate, and scalable AI agents.

Q: How can enterprises access NVIDIA GB200 NVL72-based instances on CoreWeave?
A: Customers can start provisioning GB200 NVL72-based instances through CoreWeave Kubernetes Service in the US-WEST-01 region using the gb200-4x instance ID. To get started, contact CoreWeave.

Latest stories

Read More

Hitachi Ventures Raises $400M Fund

Hitachi Ventures Secures $400 Million for Fourth Fund Hitachi Ventures...

Gesture Drawing Essentials: 2 & 5 Minute Practice Methods

Gesture Drawing Basics One of an artist's greatest tools to...

AI Startups Raised $8 Billion in 2024

Artificial Intelligence Summit: France's Thriving AI Ecosystem The Rise of...

We Need to Talk About Austen

Penguin's 'TikTok-ified' Covers for Jane Austen's Novels Spark Outrage Publishers...

Revamped ChatGPT: The Ultimate Messaging Revolution

New ChatGPT Experience in WhatsApp On Monday, OpenAI announced that...

Pixelated Perfection: ASCII Art Revival

Amid all the fuss over DeepSeek, OpenAI has pushed...

Titanfall Battle Royale

A Surprising Turn: How Titanfall 3 Became Apex Legends The...

AI-Powered Skin Cancer Prevention

AI-Assisted Cancer Diagnosis: The Future of Skin Cancer Detection Remarkable...

LEAVE A REPLY

Please enter your comment!
Please enter your name here