-
|
Hi, I am a very inexperienced user if it comes to LLMs in general. I tried to follow the guide as close as I could, but I realized CPU is getting utilized instead GPU. Is it not possible to use an AMD GPU for Qwen2.5 or where can I find the config file for which hardware should be utilized? kind regards |
Beta Was this translation helpful? Give feedback.
Replies: 4 comments 1 reply
-
|
LLM inference is not a part of the Open-LLM-VTuber. If you are following the quick guide, I assume that you are using ollama for your LLM inference. I did some digging, and apparently ollama does not support your GPU on Windows as of now. Once these support got merged, you will probably be able to use GPU acceleration in ollama. For now, you can try:
|
Beta Was this translation helpful? Give feedback.
-
|
Thanks a lot for the answer. If I use LM Studio it utilizes my GPU, with the qwen2.5 7B i get up to 102 token/sec. Is there somewhere documented how to use LM Studio instead Ollama? I read it a few times but couldn't find an answer Edit: I meant how to use LM Studio as backend (if that's the right term) in LLM Vtuber |
Beta Was this translation helpful? Give feedback.
-
|
Yeah I just realized we don't have that in our doc. We should have one. It's pretty easy though, there is an option called llm_provider. Set that to lm Studio. There is also some options named lm Studio in conf yaml, you can change relevant settings there. In lm Studio, you might have to switch to dev mode and start the API server. I will add docs later today so if you don't know what I'm saying you can wait a bit |
Beta Was this translation helpful? Give feedback.
-
|
I am using LM Studio. To use it, it's just a simple edit of config.yaml
|
Beta Was this translation helpful? Give feedback.
Yeah I just realized we don't have that in our doc. We should have one. It's pretty easy though, there is an option called llm_provider. Set that to lm Studio. There is also some options named lm Studio in conf yaml, you can change relevant settings there.
In lm Studio, you might have to switch to dev mode and start the API server. I will add docs later today so if you don't know what I'm saying you can wait a bit