2

With Nvidia's Chat with RTX, users can create personalized chatbots that run loc...

 7 months ago
source link: https://siliconangle.com/2024/02/13/nvidias-chat-rtx-users-can-create-personalized-chatbots-run-locally-pc/
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.
neoserver,ios ssh client

With Nvidia's Chat with RTX, users can create personalized chatbots that run locally on PCs

image_2024-02-14_082042388.png
AI

Nvidia Corp. is pioneering yet another innovation in artificial intelligence with the launch of a new feature called Chat with RTX, which gives users the ability to create their own personal AI assistant that resides on their laptop or personal computer, rather than in the cloud.

The company announced Chat with RTX as a free technology demonstration today, saying it allows users to tap into personalized AI capabilities hosted on their device. The offering also leverages retrieval-augmented generation or RAG techniques and Nvidia’s TensorRT-LLM software, and yet it’s said to go easy on computing resources, so users won’t notice any decrease in the performance of their machine.

Moreover, because Chat with RTX is hosted on the user’s machine, it means all chats are totally private – so no one will ever know what they discuss with their personal AI chatbot. Until now, generative AI chatbots such as ChatGPT have largely been restricted to the cloud, running on centralized servers powered by Nvidia’s graphics processing units.

That changes with Chat with RTX, which enables generative AI to run locally using the computing power of the GPU that sits inside the computer. To take advantage of it, users will need a laptop or PC that’s fitted with a GeForce RTX 30 Series GPU or a later model, such as the newly announced RTX 2000 Ada Generation GPU. They’ll also need to have at least 8 gigabytes of video random-access memory, or VRAM.

The main advantage of having a local chat assistant is that users can personalize it to their liking by deciding what sort of content it’s allowed to access to generate its responses. There are also the aforementioned privacy benefits, and it will generate responses faster too, as there’s none of the latency associated with the cloud.

Chat with RTX uses RAG techniques that enable it to augment its basic knowledge with additional data sources, including local files hosted on the computer, while the TensorRT-LLM and Nvidia RTX acceleration software provide a nice speed boost. In addition, Nvidia said users can choose from a range of underlying open-source LLMs, including Llama 2 and Mistral.

Nvidia said the personalized assistants will be able to handle the same kinds of queries that people normally use ChatGPT for, such as asking for restaurant recommendations and so on. It will also provide context to its responses when necessary, linking to the relevant file where it sourced the information.

Besides accessing local files, Chat with RTX users will also be able to specify which sources they want the chatbot to use on services such as YouTube. So they can ask their personal chat assistant to provide travel recommendations based on the content of their favorite YouTubers only, for example.

In addition to those specifications, users will need to be running Windows 10 or Windows 11, and have the latest Nvidia GPU drivers installed on their device.

Developers will also be able to experiment with Chat with RTX via the TensorRT-LLM RAG reference project on GitHub. The company is currently running a Generative AI on Nvidia RTX contest for developers, inviting them to submit applications that leverage the technology. Prizes include a GeForce RTX 4090 GPU and an invitation to the 2024 Nvidia GTC conference that’s slated to take place in March.

Image: Nvidia

A message from John Furrier, co-founder of SiliconANGLE:

Your vote of support is important to us and it helps us keep the content FREE.

One click below supports our mission to provide free, deep, and relevant content.  

Join our community on YouTube

Join the community that includes more than 15,000 #CubeAlumni experts, including Amazon.com CEO Andy Jassy, Dell Technologies founder and CEO Michael Dell, Intel CEO Pat Gelsinger, and many more luminaries and experts.

“TheCUBE is an important partner to the industry. You guys really are a part of our events and we really appreciate you coming and I know people appreciate the content you create as well” – Andy Jassy

THANK YOU


About Joyk


Aggregate valuable and interesting links.
Joyk means Joy of geeK