Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Migrate from rustformes/llm to Candle #2735

Closed
karthik2804 opened this issue Aug 20, 2024 · 1 comment · Fixed by #2755
Closed

Migrate from rustformes/llm to Candle #2735

karthik2804 opened this issue Aug 20, 2024 · 1 comment · Fixed by #2755

Comments

@karthik2804
Copy link
Contributor

Spin relies on rustformers/llm. The library has now been archived and no longer actively maintained. The usage of that library has also constrained us to use only models in the ggml format which is not popular anymore and most models are not distributed in that format.

I suggest that we migrate to candle which will allow us to be able to use models in the much more popular safetensors format. This migration will also unlock the ability to use newer models like LLama3(.1).

The current implementation supports a few models as listed as features. I think we can begin with only supporting Llama models for inferencing first. We can then optionally add support for other models as desired and feature-gate it to keep the binary size in check.

I would love to hear your thoughts!

@technosophos
Copy link
Contributor

I like the idea of starting with Llama and feature-gating others as people express interest.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants