Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GGUF support needed #104

Open
clover1980 opened this issue Aug 30, 2024 · 0 comments
Open

GGUF support needed #104

clover1980 opened this issue Aug 30, 2024 · 0 comments

Comments

@clover1980
Copy link

clover1980 commented Aug 30, 2024

Hi,
This project is great, but i can't understand how it working with Ollama models but not with current widespread quantization standard as GGUF. Ollama itself don't support GGUF, it require to reformat models into it's own format, so it's very niche by itself and support the smallest number of models (mostly the few popular), all others models ever published are quantized in GGUF (also it's easy to do that from huggingface format with llama.cpp itself or change by AUTOGGUF). Also very limited variety of quality levels in ollama models, you won't find many models there in best no loss q8 quality. My model not presented in Ollama models list and i really don't want to reformat 30billions GGUF into ollama format, that's a huge waste of time and require huge storage space by my tests (not counting RAM).
Thanks.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant