Skip to content

Releases: mudler/LocalAI

v2.5.0

08 Jan 13:55
574fa67
Compare
Choose a tag to compare

What's Changed

This release adds more embedded models, and shrink image sizes.

You can run now phi-2 ( see here for the full list ) locally by starting localai with:

docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core phi-2

LocalAI accepts now as argument a list of short-hands models and/or URLs pointing to valid yaml file. A popular way to host those files are Github gists.

For instance, you can run llava, by starting local-ai with:

docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/llava.yaml

Exciting New Features 🎉

  • feat: more embedded models, coqui fixes, add model usage and description by @mudler in #1556

👒 Dependencies

  • deps(conda): use transformers-env with vllm,exllama(2) by @mudler in #1554
  • deps(conda): use transformers environment with autogptq by @mudler in #1555
  • ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1558

Other Changes

Full Changelog: v2.4.1...v2.5.0

v2.4.1

06 Jan 00:05
ce724a7
Compare
Choose a tag to compare

What's Changed

Exciting New Features 🎉

  • feat: embedded model configurations, add popular model examples, refactoring by @mudler in #1532

Other Changes

Full Changelog: v2.4.0...v2.4.1

v2.4.0

04 Jan 16:30
bcf0244
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

  • fix(entrypoint): cd to backend dir before start by @mudler in #1530
  • fix(diffusers): add omegaconf dependency by @mudler in #1540

Exciting New Features 🎉

👒 Dependencies

Other Changes

Full Changelog: v2.3.1...v2.4.0

v2.3.1

30 Dec 14:38
a95bb05
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

  • Fix: Set proper Homebrew install location for x86 Macs by @cnatale in #1510
  • fix(download): correctly check for not found error by @mudler in #1514

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v2.3.0...v2.3.1

v2.3.0

29 Dec 19:01
9723c3c
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

Exciting New Features 🎉

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v2.2.0...v2.3.0

v2.2.0

21 Dec 16:48
9ae47d3
Compare
Choose a tag to compare

What's Changed

This release brings updates to the backends and includes a fix for recompilation of LocalAI with go-rwkv (#1473). To note, it also tries to reduce the image size by allowing some backends (transformers-based) to share the same environment.

With this release inline templates and models as URLs are supported, for example:

name: mixtral

parameters:
  model: https://huggingface.co/TheBloke/Mixtral-8x7B-v0.1-GGUF/resolve/main/mixtral-8x7b-v0.1.Q2_K.gguf
# or huggingface://TheBloke/Mixtral-8x7B-v0.1-GGUF/mixtral-8x7b-v0.1.Q2_K.gguf@main
template:
  completion: |
            Complete the following: {{.Input}}

Exciting New Features 🎉

  • docs: automatically track latest versions by @mudler in #1451
  • feat: inline templates and accept URLs in models by @mudler in #1452
  • feat(conda): share envs with transformer-based backends by @mudler in #1465

👒 Dependencies

Other Changes

Full Changelog: v2.1.0...v2.2.0

v2.1.0

16 Dec 21:04
3d83128
Compare
Choose a tag to compare

What's Changed

Breaking Changes 🛠

  • feat(alias): alias llama to llama-cpp, update docs by @mudler in #1448

Bug fixes 🐛

  • fix(piper): pin petals, phonemize and espeak by @mudler in #1393
  • update(llama.cpp): update server, correctly propagate LLAMA_VERSION by @mudler in #1440

Exciting New Features 🎉

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v2.0.0...v2.1.0

v2.0.0

04 Dec 13:22
238fec2
Compare
Choose a tag to compare

What's Changed

Breaking Changes 🛠

  • 🔥 add LLaVA support and GPT vision API, Multiple requests for llama.cpp, return JSON types by @mudler in #1254
  • refactor: rename llama-stable to llama-ggml by @mudler in #1287

Bug fixes 🐛

  • fix: respect OpenAI spec for response format by @mudler in #1289
  • fix: handle grpc and llama-cpp with REBUILD=true by @mudler in #1328
  • fix: propagate CMAKE_ARGS when building grpc by @mudler in #1334
  • fix(vall-e-x): correctly install reqs in environment by @mudler in #1377

Exciting New Features 🎉

  • feat(certificates): add support for custom CA certificates by @vitorstone in #880
  • feat(conda): conda environments by @mudler in #1144
  • refactor: move backends into the backends directory by @mudler in #1279
  • feat: allow to run parallel requests by @mudler in #1290
  • feat(transformers): add embeddings with Automodel by @mudler in #1308
  • ci(core): add -core images without python deps by @mudler in #1309
  • feat: initial watchdog implementation by @mudler in #1341
  • feat: update whisper_cpp with CUBLAS, HIPBLAS, METAL, OPENBLAS, CLBLAST support by @wuxxin in #1302

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v1.40.0...v2.0.0

v2.0.0_beta

01 Dec 19:16
d6073ac
Compare
Choose a tag to compare
v2.0.0_beta Pre-release
Pre-release

What's Changed

Breaking Changes 🛠

  • 🔥 add LLaVA support and GPT vision API, Multiple requests for llama.cpp, return JSON types by @mudler in #1254
  • refactor: rename llama-stable to llama-ggml by @mudler in #1287

Bug fixes 🐛

  • fix: respect OpenAI spec for response format by @mudler in #1289
  • fix: handle grpc and llama-cpp with REBUILD=true by @mudler in #1328
  • fix: propagate CMAKE_ARGS when building grpc by @mudler in #1334

Exciting New Features 🎉

  • feat(certificates): add support for custom CA certificates by @vitorstone in #880
  • feat(conda): conda environments by @mudler in #1144
  • refactor: move backends into the backends directory by @mudler in #1279
  • feat: allow to run parallel requests by @mudler in #1290
  • feat(transformers): add embeddings with Automodel by @mudler in #1308
  • ci(core): add -core images without python deps by @mudler in #1309
  • feat: initial watchdog implementation by @mudler in #1341

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v1.40.0...v2.0.0_beta

v1.40.0

30 Oct 18:34
6ef7ea2
Compare
Choose a tag to compare

This release is a preparation before v2 - the efforts now will be to refactor, polish and add new backends. Follow up on: #1126

Hot topics

This release now brings the llama-cpp backend which is a c++ backend tied to llama.cpp. It follows more closely and tracks recent versions of llama.cpp. It is not feature compatible with the current llama backend but plans are to sunset the current llama backend in favor of this one. This one will be probably be the latest release containing the older llama backend written in go and c++. The major improvement with this change is that there are less layers that could be expose to potential bugs - and as well it ease out maintenance as well.

Support for ROCm/HIPBLAS

This release bring support for AMD thanks to @65a . See more details in #1100

More CLI commands

Thanks to @jespino now the local-ai binary has more subcommands allowing to manage the gallery or try out directly inferencing, check it out!

What's Changed

Bug fixes 🐛

  • fix(openai): Populate ID and Created fields in OpenAI compatible responses by @jespino in #1164
  • Fix backend/cpp/llama CMakeList.txt on OSX by @dave-gray101 in #1212

Exciting New Features 🎉

👒 Dependencies

  • fix(deps): update module github.com/onsi/gomega to v1.28.0 by @renovate in #1113
  • ⬆️ Update go-skynet/go-llama.cpp by @localai-bot in #1106
  • fix(deps): update github.com/tmc/langchaingo digest to e16b777 by @renovate in #1101
  • fix(deps): update github.com/go-skynet/go-llama.cpp digest to 79f9587 by @renovate in #1085
  • fix(deps): update module github.com/shirou/gopsutil/v3 to v3.23.9 by @renovate in #1120
  • fix(deps): update module github.com/sashabaranov/go-openai to v1.15.4 by @renovate in #1122
  • fix(deps): update module github.com/rs/zerolog to v1.31.0 by @renovate in #1102
  • ⬆️ Update go-skynet/go-llama.cpp by @localai-bot in #1130
  • fix(deps): update github.com/go-skynet/go-llama.cpp digest to 6018c9d by @renovate in #1129
  • ⬆️ Update go-skynet/go-llama.cpp by @localai-bot in #1136
  • fix(deps): update github.com/go-skynet/go-llama.cpp digest to 1676dcd by @renovate in #1135
  • fix(deps): update github.com/nomic-ai/gpt4all/gpt4all-bindings/golang digest to 56c0d28 by @renovate in #1140
  • fix(deps): update module github.com/onsi/ginkgo/v2 to v2.13.0 by @renovate in #1152
  • fix(deps): update module google.golang.org/grpc to v1.58.3 by @renovate in #1160
  • fix(deps): update github.com/go-skynet/go-llama.cpp digest to aeba71e by @renovate in #1155
  • fix(deps): update github.com/nomic-ai/gpt4all/gpt4all-bindings/golang digest to 10f9b49 by @renovate in #1158
  • fix(deps): update module github.com/sashabaranov/go-openai to v1.16.0 by @renovate in #1159
  • fix(deps): update github.com/nomic-ai/gpt4all/gpt4all-bindings/golang digest to 22de3c5 by @renovate in #1172
  • fix(deps): update github.com/tmc/langchaingo digest to a02d4fd by @renovate in #1175
  • fix(deps): update module github.com/gofiber/fiber/v2 to v2.50.0 by @renovate in #1177
  • fix(deps): update github.com/nomic-ai/gpt4all/gpt4all-bindings/golang digest to 9a19c74 by @renovate in #1179
  • fix(deps): update github.com/tmc/langchaingo digest to c636b3d by @renovate in #1188
  • fix(deps): update module google.golang.org/grpc to v1.59.0 by @renovate in #1189
  • chore(deps): update actions/checkout action to v4 by @renovate in #1006
  • feat(llama.cpp): update by @mudler in #1200
  • ⬆️ Update go-skynet/go-llama.cpp by @localai-bot in #1156
  • fix(deps): update github.com/nomic-ai/gpt4all/gpt4all-bindings/golang digest to c25dc51 by @renovate in #1191
  • ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1204
  • fix(deps): update module github.com/onsi/gomega to v1.28.1 by @renovate in #1205

Other Changes

  • fix(deps): update github.com/nomic-ai/gpt4all/gpt4all-bindings/golang digest to 6711bdd by @renovate in #1079
  • ci: cleanup worker by @mudler in #1166
  • docs(examples): Add mistral example by @mudler in #1214
  • feat(llama.cpp): Bump llama.cpp, adapt grpc server by @mudler in #1211
  • cleanup: drop bloomz and ggllm as now supported by llama.cpp by @mudler in #1217
  • ci: use self-hosted to build container images by @mudler in #1206
  • ci: run only cublas on selfhosted by @mudler in #1224
  • ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1207

New Contributors

Full Changelog: v1.30.0...v1.40.0