v2.7.0
This release adds support to the transformer backend for LLM as well!
For now instance you can run codellama-7b with transformers with:
docker run -ti -p 8080:8080 --gpus all localai/localai:v2.7.0-cublas-cuda12 codellama-7b
In the quickstart there are more examples available https://localai.io/basics/getting_started/#running-models.
Note: As llama.cpp is ongoing with changes that could possible cause breakage, this release does not includes changes from ggerganov/llama.cpp#5138 (the future versions will).
What's Changed
Bug fixes 🐛
Exciting New Features 🎉
- feat(transformers): support also text generation by @mudler in #1630
- transformers: correctly load automodels by @mudler in #1643
- feat(startup): fetch model definition remotely by @mudler in #1654
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1642
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1644
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1652
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1655
Other Changes
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1632
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1631
Full Changelog: v2.6.1...v2.6.2