Releases: mudler/LocalAI
v2.7.0
This release adds support to the transformer backend for LLM as well!
For now instance you can run codellama-7b with transformers with:
docker run -ti -p 8080:8080 --gpus all localai/localai:v2.7.0-cublas-cuda12 codellama-7b
In the quickstart there are more examples available https://localai.io/basics/getting_started/#running-models.
Note: As llama.cpp is ongoing with changes that could possible cause breakage, this release does not includes changes from ggml-org/llama.cpp#5138 (the future versions will).
What's Changed
Bug fixes 🐛
Exciting New Features 🎉
- feat(transformers): support also text generation by @mudler in #1630
- transformers: correctly load automodels by @mudler in #1643
- feat(startup): fetch model definition remotely by @mudler in #1654
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1642
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1644
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1652
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1655
Other Changes
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1632
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1631
Full Changelog: v2.6.1...v2.6.2
v2.6.1
This is a patch release containing bug-fixes around parallel request support with llama.cpp models.
What's Changed
Bug fixes 🐛
- fix(llama.cpp): Enable parallel requests by @tauven in #1616
- fix(llama.cpp): enable cont batching when parallel is set by @mudler in #1622
Exciting New Features 🎉
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1623
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1619
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1620
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1626
New Contributors
Full Changelog: v2.6.0...v2.6.1
v2.6.0
What's Changed
Bug fixes 🐛
- move BUILD_GRPC_FOR_BACKEND_LLAMA logic to makefile: errors in this section now immediately fail the build by @dionysius in #1576
- prepend built binaries in PATH for BUILD_GRPC_FOR_BACKEND_LLAMA by @dionysius in #1593
Exciting New Features 🎉
- minor: replace shell pwd in Makefile with CURDIR for better windows compatibility by @dionysius in #1571
- Makefile: allow to build without GRPC_BACKENDS by @mudler in #1607
- feat: 🐍 add mamba support by @mudler in #1589
- feat(extra-backends): Improvements, adding mamba example by @mudler in #1618
👒 Dependencies
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1567
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1568
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1573
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1578
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1583
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1587
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1590
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1594
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1599
Other Changes
- Moving the how tos to self hosted by @lunamidori5 in #1574
- docs: missing golang requirement for local build for debian by @dionysius in #1596
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1597
- docs/examples: enhancements by @mudler in #1572
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1604
- Update README.md by @lunamidori5 in #1601
- docs: re-use original permalinks by @mudler in #1610
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1612
- Expanded and interlinked Docker documentation by @jamesbraza in #1614
- Modernized LlamaIndex integration by @jamesbraza in #1613
New Contributors
- @dionysius made their first contribution in #1571
Full Changelog: v2.5.1...v2.6.0
v2.5.1
Patch release to create /build/models
in the container images.
What's Changed
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1562
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1563
Full Changelog: v2.5.0...v2.5.1
v2.5.0
What's Changed
This release adds more embedded models, and shrink image sizes.
You can run now phi-2
( see here for the full list ) locally by starting localai with:
docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core phi-2
LocalAI accepts now as argument a list of short-hands models and/or URLs pointing to valid yaml file. A popular way to host those files are Github gists.
For instance, you can run llava
, by starting local-ai
with:
docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/llava.yaml
Exciting New Features 🎉
👒 Dependencies
- deps(conda): use transformers-env with vllm,exllama(2) by @mudler in #1554
- deps(conda): use transformers environment with autogptq by @mudler in #1555
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1558
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1557
Full Changelog: v2.4.1...v2.5.0
v2.4.1
What's Changed
Exciting New Features 🎉
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1546
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1547
- docs: improve getting started by @mudler in #1553
Full Changelog: v2.4.0...v2.4.1
v2.4.0
What's Changed
Bug fixes 🐛
- fix(entrypoint): cd to backend dir before start by @mudler in #1530
- fix(diffusers): add omegaconf dependency by @mudler in #1540
Exciting New Features 🎉
- feat(preload): prepare models in galleries by @mudler in #1515
- feat: share models by url by @mudler in #1522
- feat(prepare): allow to specify additional files to download by @mudler in #1526
- feat: add trimsuffix by @mudler in #1528
👒 Dependencies
- deps(llama.cpp): update and sync grpc server by @mudler in #1527
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1531
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1517
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1518
- docs(phi-2): add example by @mudler in #1525
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1533
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1544
- ci(dockerhub): push images also to dockerhub by @mudler in #1542
Full Changelog: v2.3.1...v2.4.0
v2.3.1
What's Changed
Bug fixes 🐛
- Fix: Set proper Homebrew install location for x86 Macs by @cnatale in #1510
- fix(download): correctly check for not found error by @mudler in #1514
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1509
Other Changes
- Update version.json (2.3.0) by @lunamidori5 in #1511
New Contributors
Full Changelog: v2.3.0...v2.3.1
v2.3.0
What's Changed
Bug fixes 🐛
- fix: exllama2 backend by @sozercan in #1484
- fix: guidance_scale not work in sd by @hibobmaster in #1488
- fix(image-gen): remove default to stablediffusion by @JZacharie in #1500
Exciting New Features 🎉
- env(conda): use transformers for vall-e-x by @mudler in #1481
- feat: add 🐸 coqui by @mudler in #1489
- feat: partial download by @muka in #1486
- feat: add tiny dream stable diffusion support by @M0Rf30 in #1283
👒 Dependencies
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1479
- ⬆️ Update donomii/go-rwkv.cpp by @localai-bot in #1478
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1469
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1482
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1483
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1492
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1495
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1496
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1501
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1503
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1477
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1480
- docs: add langchain4j integration by @mudler in #1476
- ci(apple): speedups by @mudler in #1471
New Contributors
- @hibobmaster made their first contribution in #1488
- @muka made their first contribution in #1486
- @JZacharie made their first contribution in #1500
Full Changelog: v2.2.0...v2.3.0
v2.2.0
What's Changed
This release brings updates to the backends and includes a fix for recompilation of LocalAI with go-rwkv (#1473). To note, it also tries to reduce the image size by allowing some backends (transformers-based) to share the same environment.
With this release inline templates and models as URLs are supported, for example:
name: mixtral
parameters:
model: https://huggingface.co/TheBloke/Mixtral-8x7B-v0.1-GGUF/resolve/main/mixtral-8x7b-v0.1.Q2_K.gguf
# or huggingface://TheBloke/Mixtral-8x7B-v0.1-GGUF/mixtral-8x7b-v0.1.Q2_K.gguf@main
template:
completion: |
Complete the following: {{.Input}}
Exciting New Features 🎉
- docs: automatically track latest versions by @mudler in #1451
- feat: inline templates and accept URLs in models by @mudler in #1452
- feat(conda): share envs with transformer-based backends by @mudler in #1465
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1450
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1455
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1457
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1462
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1461
- ⬆️ Update donomii/go-rwkv.cpp by @localai-bot in #1470
Other Changes
- How To (Updates and Fixes) by @lunamidori5 in #1456
- Revert ":arrow_up: Update donomii/go-rwkv.cpp" by @mudler in #1474
Full Changelog: v2.1.0...v2.2.0