Replies: 2 comments
-
|
Did you see this #15396? |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
I am deeply sorry. As I stated in my post, this was the very first time that I posted anything on GitHub discussions, and obviously I did a mistake ! I did not realize that I was posting on llama.cpp's repo ... instead of ik_llama.cpp's ! I have no problem using GPT-OSS-120b on llama.cpp so I will close this discussion. Sorry ! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi everyone.
I tried many, many things but can't seem to find a way to use GPT-OSS-120b (model originally downloaded with LM Studio) so that it stops outputting the harmony tags like <|channel|>.
I use it with Open WebUI and Codex and the problem happens with the two.
I tried everything that I read on this pull request : https://github.com/ikawrakow/ik_llama.cpp/pull/723
Here is the command that I used :
Example of the issue :
<|channel|>analysis <|channel|>analysis<|message|>The user just says "Hey". Likely a greeting. Should respond accordingly. Keep tone friendly. Possibly ask how can help.<|end|><|start|>assistant<|channel|>finalHello! How can I assist you today?version :
build/bin/llama-server --version version: 3928 (16f30fcf3) built with cc (Ubuntu 11.4.0-1ubuntu1~22.04.2) 11.4.0 for x86_64-linux-gnuI opened this discussion and not an issue as I have seen that other issues are already opened but not primarily for this model. I hope that I don't make any mistake here, this is my first time posting on a GitHub repo.
Thank you for your help.
Beta Was this translation helpful? Give feedback.
All reactions