AMD Says It's Super Easy To Set Up Your Own AI Chatbot, Here's How

chatbot
Did you read our post last month about NVIDIA's Chat With RTX utility and shrug because you don't have a GeForce RTX graphics card? Well, don't sweat it, dear friend—AMD is here to offer you an open-source option that runs on just about anything. Known as LM Studio, it's a frontend that can download and run large language models right on your PC, no internet connection required (after you've finished the initial downloads, of course.)

lmstudio window

The setup process is really simple. In fact, it's so simple that we can lay it out for you in this paragraph. Just go download LM-Studio, and make sure to get the ROCm version if you have a Navi 21 or Navi 3x graphics card. (If you're not sure, that's the Radeon RX 6800 or better, or any Radeon RX 7000 card.) Run the installer, which will launch the application, then download a model. AMD recommends a couple that work well:

mistral conversation

Once you have a model downloaded, click on the chat bubble icon on the left to start chatting with it. You may have to click on the purple bar in the top to select the model you downloaded first. If you're using the ROCm version, don't forget to open the settings and move that "GPU Offload" slider all the way to the right so that LM Studio uploads the entire LLM to your GPU's local memory, which will drastically improve performance.

recommended models

If you're wondering why you'd bother doing this instead of simply talking to the more capable ChatGPT or Gemini services, it's primarily down to privacy and response time. Running an LLM on your local GPU guarantees a near-instant response time, and since everything happens on your own PC, there's no chance of anyone snooping on your conversation. You are also much less likely to run up against any guardrails placed on the model, meaning you could use the bot for whatever purpose your heart desires... within the limitations of its training data.

gpu offload
Don't forget to enable GPU Offload if your system has a fast GPU.

AMD also talks about using LM Studio on a "Ryzen AI PC". It's true; you can run LM Studio on a system with a Phoenix or Hawk Point processor. However, on those machines it will run on either the Zen 4 CPU cores or the integrated RDNA 3 GPU, as the llama.cpp project that LM Studio is based on doesn't support AMD's XDNA NPUs just yet. It also doesn't support Intel's NPUs either. It does look like progress is happening on both fronts though, so it may be the case that we have a real consumer application for NPUs soon.