Create your own personalized AI chatbot.
Nvidia is going beyond just building chips designed for AI tasks. Today, they have released Chat With RTX is a demo app that lets you personalize a GPT large language model (LLM) connected to your own content — docs, notes, videos, or other data.
This means you can now run an AI chatbot without an internet connection and paying a dime for services like ChatGPT or Gemini. Chat with RTX can utilize open-source LLM models like Mistral or Llama.
Here are the software and hardware requirements needed to run it on your machine:
Download and install the Chat with RTX software from the Nvidia website.
The Chat with RTX tech demo is builtfrom a publicly available developer reference project found on GitHub, called TensorRT-LLM RAG. This offers some exciting possibilities for developers:
Note: Working with these tools likely requires a solid background in AI development. It’s not exactly a drag-and-drop system where everyone can make custom applications quickly.
It is important to note that not a lot of people can use it because of the high hardware requirements. You need a powerful computer with an NVIDIA RTX 30 or 40 Series GPU or NVIDIA RTX™ Ampere or Ada Generation GPU with at least 8GB of VRAM, as well as a latest-Gen Intel® Core™ i7 or AMD Ryzen™ 7 processor or better, 32GB of RAM or more, and 2TB of free storage space.
However, if you have the hardware, this is a game-changer. Nvidia has massive resources and a huge pool of funds to improve this technology, so the future of AI chatbots is bright. In the future, we can expect to see even more powerful and versatile AI chatbots that can be used for a wide variety of tasks.
‍
Software engineer, writer, solopreneur