Skip to content

Commit 3fe85db

Browse files
authored
Merge pull request #98 from premAI-io/mudler-patch-1
mlops-engines: add LocalAI
2 parents 469e3b1 + 9295d86 commit 3fe85db

File tree

4 files changed

+18
-4
lines changed

4 files changed

+18
-4
lines changed

desktop-apps.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -208,7 +208,7 @@ koboldcpp Julius Model Configuration
208208

209209
[local.ai]: https://www.localai.app
210210

211-
The [local.ai] App from https://github.com/louisgv/local.ai ([not to be confused](https://github.com/louisgv/local.ai/discussions/71) with [LocalAI](https://localai.io) from https://github.com/mudler/LocalAI) is a simple application for loading LLMs after you manually download a `ggml` model from online.
211+
The [local.ai] App from https://github.com/louisgv/local.ai ([not to be confused](https://github.com/louisgv/local.ai/discussions/71) with [](mlops-engines.md#localai) from https://github.com/mudler/LocalAI) is a simple application for loading LLMs after you manually download a `ggml` model from online.
212212

213213
### UI and Chat
214214

mlops-engines.md

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -29,6 +29,7 @@ Inference Engine | Open-Source | GPU optimisations | Ease of use
2929
[](#vllm) | 🟢 Yes | Continuous Batching, Tensor Parallelism, Paged Attention | 🟢 Easy
3030
[](#bentoml) | 🟢 Yes | None | 🟢 Easy
3131
[](#modular) | 🔴 No | N/A | 🟡 Moderate
32+
[](#localai) | 🟢 Yes | 🟢 Yes | 🟢 Easy
3233
```
3334

3435
{{ table_feedback }}
@@ -127,6 +128,20 @@ Cons:
127128

128129
This is not an exhaustive list of MLOps engines by any means. There are many other tools and frameworks developer use to deploy their ML models. There is ongoing development in both the open-source and private sectors to improve the performance of LLMs. It's up to the community to test out different services to see which one works best for their use case.
129130

131+
## LocalAI
132+
133+
[LocalAI](https://localai.io) from https://github.com/mudler/LocalAI ([not to be confused](https://github.com/louisgv/local.ai/discussions/71) with [](desktop-apps.md#localai) from https://github.com/louisgv/local.ai) is the free, Open Source alternative to OpenAI. LocalAI act as a drop-in replacement REST API that's compatible with OpenAI API specifications for local inferencing. It can run LLMs (with various backend such as https://github.com/ggerganov/llama.cpp or [](#vllm)), generate images, generate audio, transcribe audio, and can be self-hosted (on-prem) with consumer-grade hardware.
134+
135+
Pros:
136+
137+
- [wide range of models supported](https://localai.io/model-compatibility)
138+
- support for [functions](https://localai.io/features/openai-functions) (self-hosted [OpenAI functions](https://platform.openai.com/docs/guides/gpt/function-calling))
139+
- [easy to integrate](https://localai.io/integrations)
140+
141+
Cons:
142+
143+
- binary version is harder to run and compile locally. https://github.com/mudler/LocalAI/issues/1196.
144+
- high learning curve due to high degree of customisation
130145

131146
## Challenges in Open Source
132147

model-formats.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -280,7 +280,7 @@ Some [clients & libraries supporting `GGUF`](https://huggingface.co/TheBloke/Lla
280280
- [LM Studio](https://lmstudio.ai) -- an easy-to-use and powerful local GUI with GPU acceleration on both Windows (NVidia and AMD), and macOS
281281

282282
```{seealso}
283-
For more info on `GGUF`, see https://github.com/ggerganov/llama.cpp/pull/2398 and its [spec](https://github.com/philpax/ggml/blob/gguf-spec/docs/gguf.md).
283+
For more info on `GGUF`, see https://github.com/ggerganov/llama.cpp/pull/2398 and its [spec](https://github.com/ggerganov/ggml/blob/master/docs/gguf.md).
284284
```
285285

286286
### Limitations

sdk.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -46,11 +46,10 @@ The list of vector stores that LangChain supports can be found [here](https://ap
4646

4747
### Models
4848

49-
This is the heart of most LLM models where the core functionality resides. There are broadly 3 different [models](https://docs.langchain.com/docs/components/models) that LLMs provide. They are Language, Chat, and Embedding model.
49+
This is the heart of most LLMs, where the core functionality resides. There are broadly [2 different types of models](https://python.langchain.com/docs/modules/model_io/models) which LangChain integrates with:
5050

5151
- **Language**: Inputs & outputs are `string`s
5252
- **Chat**: Run on top of a Language model. Inputs are a list of chat messages, and output is a chat message
53-
- **Embedding**: Inputs is a `string` and outputs are a list of `float`s (vector)
5453

5554
### Tools
5655

0 commit comments

Comments
 (0)