Unlocking Productivity and Creativity With AI-Powered Chatbots
AI Decoded earlier this year explored what Large Language Models (LLMs) are, why they matter, and how to use them.
For many, tools like ChatGPT were their first introduction to AI. LLM-powered chatbots have transformed computing from basic, rule-based interactions to dynamic conversations. They can suggest vacation ideas, write customer service emails, spin up original poetry, and even write code for users.
Introduced in March, ChatRTX is a demo app that lets users personalize a GPT LLM with their own content, such as documents, notes, and images. With features like retrieval-augmented generation (RAG), NVIDIA TensorRT-LLM, and RTX acceleration, ChatRTX enables users to quickly search and ask questions about their own data. And since the app runs locally on RTX PCs or workstations, results are both fast and private.
NVIDIA offers the broadest selection of foundation models for enthusiasts and developers, including Gemma 2, Mistral, and Llama-3. These models can run locally on NVIDIA GeForce and RTX GPUs for fast, secure performance without needing to rely on cloud services. Download ChatRTX today.
Introducing RTX-Accelerated Partner Applications
AI is being incorporated into more and more apps and use cases, including games, content creation apps, software development, and productivity tools.
This expansion is fueled by the wide selection of RTX-accelerated developer and community tools, software development kits, models, and frameworks that have made it easier than ever to run models locally in popular applications.
AI Decoded in October spotlighted how Brave Browser’s Leo AI, powered by NVIDIA RTX GPUs and the open-source Ollama platform, enables users to run local LLMs like Llama 3 directly on their RTX PCs or workstations. This local setup offers fast, responsive AI performance while keeping user data private – without relying on the cloud.
Agentic AI – Enabling Complex Problem-Solving
Agentic AI is the next frontier of AI, capable of using sophisticated reasoning and iterative planning to autonomously solve complex, multi-step problems.
Partner applications like AnythingLLM showcase how AI is going beyond simple question-answering to improving productivity and creativity. Users can harness the application to deploy built-in agents that can tackle tasks like searching the web or scheduling meetings.
AnythingLLM lets users interact with documents through intuitive interfaces, automate complex tasks with AI agents, and run advanced LLMs locally. Harnessing the power of RTX GPUs, it delivers faster, smarter, and more responsive AI workflows – all within a single local desktop application. The application also works offline and is fast and private, capable of using local data and tools typically inaccessible with cloud-based solutions.
AI Decoded Wrapped
Over 600 Windows apps and games today are already running AI locally on more than 100 million GeForce RTX AI PCs and workstations worldwide, delivering fast, reliable, and low-latency performance. Learn more about NVIDIA GeForce RTX AI PCs and NVIDIA RTX AI workstations.
Tune into the CES keynote delivered by NVIDIA founder and CEO Jensen Huang on Jan. 6 to discover how the latest in AI is supercharging gaming, content creation, and development.
Generative AI is transforming gaming, videoconferencing, and interactive experiences of all kinds. Make sense of what’s new and what’s next by subscribing to the AI Decoded newsletter.
FAQs
Q: What are LLMs, and how do they work?
A: Large Language Models (LLMs) are AI models that can process and generate human-like text. They are trained on vast amounts of text data and can be used for a variety of tasks, such as language translation, text summarization, and chatbots.
Q: What is ChatRTX, and how does it work?
A: ChatRTX is a demo app that lets users personalize a GPT LLM with their own content, such as documents, notes, and images. It uses features like retrieval-augmented generation (RAG), NVIDIA TensorRT-LLM, and RTX acceleration to enable users to quickly search and ask questions about their own data.
Q: What are RTX-accelerated partner applications, and how do they work?
A: RTX-accelerated partner applications are software applications that use NVIDIA RTX GPUs to accelerate AI workloads. These applications can be used for a variety of tasks, such as gaming, content creation, and software development.
Q: What is agentic AI, and how does it work?
A: Agentic AI is a type of AI that uses sophisticated reasoning and iterative planning to autonomously solve complex, multi-step problems. It can be used to automate tasks, improve productivity, and create new experiences.

