Generative AI Redefines Computing
Generative AI is redefining computing, unlocking new ways to build, train, and optimize AI models on PCs and workstations. From content creation and large and small language models to software development, AI-powered PCs and workstations are transforming workflows and enhancing productivity.
Develop and Deploy on RTX
RTX GPUs are built with specialized AI hardware called Tensor Cores that provide the compute performance needed to run the latest and most demanding AI models. These high-performance GPUs can help build digital humans, chatbots, AI-generated podcasts, and more.
Model Behavior
Large language models (LLMs) can be used for an abundance of use cases — and scale to tackle complex tasks like writing code or translating Japanese into Greek. But since they’re typically trained with a wide spectrum of knowledge for broad applications, they may not be the right fit for specific tasks, like nonplayer character dialog generation in a video game. In contrast, small language models balance need with reduced size, maintaining accuracy while running locally on more devices.
Maximizing AI Performance on Windows Workstations
Optimizing AI inference and model execution on Windows-based workstations requires strategic software and hardware tuning due to diverse hardware configurations and software environments. The session "Optimizing AI Workloads on Windows Workstations: Strategies and Best Practices" will explore best practices for AI optimization, including model quantization, inference pipeline enhancements, and hardware-aware tuning.
Advancing Local AI Development
Building, testing, and deploying AI models on local infrastructure ensures security and performance even without a connection to cloud-based services. Accelerated with NVIDIA RTX GPUs, Z by HP’s AI solutions provide the tools needed to develop AI on premises while maintaining control over data and IP.
Conclusion
GTC 2025 is set to put a spotlight on AI and all its benefits, from keynote sessions to technical hands-on training and unique networking events. Attend GTC 2025 to learn more about developing and deploying AI models on RTX AI PCs and workstations.
Frequently Asked Questions
Q: What is Generative AI?
A: Generative AI is a type of AI that can generate new and original content, such as text, images, or music.
Q: What is RTX?
A: RTX is a line of graphics processing units (GPUs) designed for AI and deep learning applications.
Q: What are Tensor Cores?
A: Tensor Cores are specialized AI hardware that provide the compute performance needed to run the latest and most demanding AI models.
Q: What are small language models?
A: Small language models are designed for specific tasks and are more accurate and efficient than large language models.
Q: What is NIM microservices?
A: NIM microservices are optimized, prepackaged models for generative AI, including the Llama 3.1 LLM, NVIDIA Riva Parakeet for automatic speech recognition (ASR), and YOLOX for computer vision.

