Skip to content

Releases: mudler/LocalAI

v3.4.0

12 Aug 07:13
b2e8b6d
Compare
Choose a tag to compare




🚀 LocalAI 3.4.0

What’s New in LocalAI 3.4.0 🎉

  • WebUI improvements: now size can be set during image generation
  • New backends: KittenTTS, kokoro and dia now are available as backends and models can be installed directly from the gallery
    Note: these backends needs to be warmed up during the first call to download the model files.
  • Support for reasoning effort in the OpenAI chat completion
  • Diffusers backend now is available for l4t images and devices
  • During backend installation from the CLI can be supplied alias and name (--alias and --name`) to override configurations
  • Backends now can be sideloaded from the system: you can drag-and-drop the backends in the backends folder and they will just work!

The Complete Local Stack for Privacy-First AI

LocalAI Logo

LocalAI

The free, Open Source OpenAI alternative. Acts as a drop-in replacement REST API compatible with OpenAI specifications for local AI inferencing. No GPU required.

Link: https://github.com/mudler/LocalAI

LocalAGI Logo

LocalAGI

A powerful Local AI agent management platform. Serves as a drop-in replacement for OpenAI's Responses API, supercharged with advanced agentic capabilities and a no-code UI.

Link: https://github.com/mudler/LocalAGI

LocalRecall Logo

LocalRecall

A RESTful API and knowledge base management system providing persistent memory and storage capabilities for AI agents. Designed to work alongside LocalAI and LocalAGI.

Link: https://github.com/mudler/LocalRecall

Thank you! ❤️

A massive THANK YOU to our incredible community and our sponsors! LocalAI has over 34,500 stars, and LocalAGI has already rocketed past 1k+ stars!

As a reminder, LocalAI is real FOSS (Free and Open Source Software) and its sibling projects are community-driven and not backed by VCs or a company. We rely on contributors donating their spare time and our sponsors to provide us the hardware! If you love open-source, privacy-first AI, please consider starring the repos, contributing code, reporting bugs, or spreading the word!

👉 Check out the reborn LocalAGI v2 today: https://github.com/mudler/LocalAGI

Full changelog 👇

👉 Click to expand 👈

What's Changed

Bug fixes 🐛

  • fix(llama.cpp): do not default to linear rope by @mudler in #5982

Exciting New Features 🎉

  • feat(webui): allow to specify image size by @mudler in #5976
  • feat(backends): add KittenTTS by @mudler in #5977
  • feat(kokoro): complete kokoro integration by @mudler in #5978
  • feat: add reasoning effort and metadata to template by @mudler in #5981
  • feat(transformers): add support to Dia by @mudler in #5991
  • feat(diffusers): add builds for nvidia-l4t by @mudler in #6004
  • feat(backends install): allow to specify name and alias during manual installation by @mudler in #5971

🧠 Models

  • chore(models): add gpt-oss-20b by @mudler in #5973
  • chore(models): add gpt-oss-120b by @mudler in #5974
  • feat(models): add support to qwen-image by @mudler in #5975
  • chore(model gallery): add openai_gpt-oss-20b-neo by @mudler in #5986
  • fix(harmony): improve template by adding reasoning effort and system_prompt by @mudler in #5985
  • chore(model gallery): add qwen_qwen3-4b-instruct-2507 by @mudler in #5987
  • chore(model gallery): add qwen_qwen3-4b-thinking-2507 by @mudler in #5988
  • chore(model gallery): add huihui-ai_huihui-gpt-oss-20b-bf16-abliterated by @mudler in #5995
  • chore(model gallery): add openai-gpt-oss-20b-abliterated-uncensored-neo-imatrix by @mudler in #5996
  • chore(model gallery): add tarek07_nomad-llama-70b by @mudler in #5997
  • chore: add Dia to the model gallery, fix backend by @mudler in #5998
  • chore(model gallery): add chatterbox by @mudler in #5999
  • chore(model gallery): add outetts by @mudler in #6000
  • chore(model gallery): add impish_nemo_12b by @mudler in #6007
  • chore(model-gallery): ⬆️ update checksum by @localai-bot in #6010

👒 Dependencies

Other Changes

  • docs: ⬆️ update docs version mudler/LocalAI by @localai-bot in #5967
  • chore: ⬆️ Update ggml-org/llama.cpp to 41613437ffee0dbccad684fc744788bc504ec213 by @localai-bot in #5968
  • chore(deps): bump torch and diffusers by @mudler in #5970
  • chore(deps): bump torch and sentence-transformers by @mudler in #5969
  • chore: ⬆️ Update ggml-org/llama.cpp to fd1234cb468935ea087d6929b2487926c3afff4b by @localai-bot in #5972
  • chore: ⬆️ Update ggml-org/llama.cpp to e725a1a982ca870404a9c4935df52466327bbd02 by @localai-bot in #5984
  • feat(swagger): update swagger by @localai-bot in #5983
  • chore: ⬆️ Update ggml-org/llama.cpp to a0552c8beef74e843bb085c8ef0c63f9ed7a2b27 by @localai-bot in #5992
  • chore: ⬆️ Update ggml-org/whisper.cpp to 4245c77b654cd384ad9f53a4a302be716b3e5861 by @localai-bot in #5993
  • docs: update links in documentation by @lnnt in #5994
  • chore: ⬆️ Update ggml-org/llama.cpp to cd6983d56d2cce94ecb86bb114ae8379a609073c by @localai-bot in #6003
  • fix(l4t-diffusers): add sentencepiece by @mudler in #6005
  • chore: ⬆️ Update ggml-org/llama.cpp to 79c1160b073b8148a404f3dd2584be1606dccc66 by @localai-bot in #6006
  • chore: ⬆️ Update ggml-org/whisper.cpp to b02242d0adb5c6c4896d59ac86d9ec9fe0d0fe33 by @localai-bot in #6009
  • chore: ⬆️ Update ggml-org/llama.cpp to be48528b068111304e4a0bb82c028558b5705f05 by @localai-bot in #6012

New Contributors

Full Changelog: v3.3.2...v3.4.0

v3.3.2

04 Aug 14:52
d6274ea
Compare
Choose a tag to compare

What's Changed

Exciting New Features 🎉

  • feat(backends): install from local path by @mudler in #5962
  • feat(backends): allow backends to not have a metadata file by @mudler in #5963

📖 Documentation and examples

  • fix(docs): Improve responsiveness of tables by @dedyf5 in #5954

👒 Dependencies

Other Changes

  • docs: ⬆️ update docs version mudler/LocalAI by @localai-bot in #5956
  • chore: ⬆️ Update ggml-org/whisper.cpp to 0becabc8d68d9ffa6ddfba5240e38cd7a2642046 by @localai-bot in #5958
  • chore: ⬆️ Update ggml-org/llama.cpp to 5c0eb5ef544aeefd81c303e03208f768e158d93c by @localai-bot in #5959
  • chore: ⬆️ Update ggml-org/llama.cpp to d31192b4ee1441bbbecd3cbf9e02633368bdc4f5 by @localai-bot in #5965

Full Changelog: v3.3.1...v3.3.2

v3.3.1

01 Aug 13:02
0b08508
Compare
Choose a tag to compare

This is a minor release, however we have addressed some important bug regarding Intel-GPU Images, and we have changed naming of the container images.

This release also adds support for Flux Kontext and Flux krea!

⚠️ Breaking change

Intel GPU images has been renamed from latest-gpu-intel-f32 and latest-gpu-intel-f16 to a single one, latest-gpu-intel, for example:

docker run -ti --name local-ai -p 8080:8080 --device=/dev/dri/card1 --device=/dev/dri/renderD128 localai/localai:latest-gpu-intel

and for AIO (All-In-One) images:

docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-gpu-intel

🖼️ Flux kontext

From this release LocalAI supports Flux Kontext and can be used to edit images via the API:

Install with:

local-ai run flux.1-kontext-dev

To test:

curl http://localhost:8080/v1/images/generations -H "Content-Type: application/json" -d '{
  "model": "flux.1-kontext-dev",
  "prompt": "change 'flux.cpp' to 'LocalAI'",
  "size": "256x256",
  "ref_images": [
  	"https://raw.githubusercontent.com/leejet/stable-diffusion.cpp/master/assets/flux/flux1-dev-q8_0.png"
  ]
}'
b64567298114 (1) b641424088517

What's Changed

Breaking Changes 🛠

  • fix(intel): Set GPU vendor on Intel images and cleanup by @richiejp in #5945

Exciting New Features 🎉

  • feat(stablediffusion-ggml): add support to ref images (flux Kontext) by @mudler in #5935

🧠 Models

  • chore(model gallery): add qwen_qwen3-30b-a3b-instruct-2507 by @mudler in #5936
  • chore(model gallery): add arcee-ai_afm-4.5b by @mudler in #5938
  • chore(model gallery): add qwen_qwen3-30b-a3b-thinking-2507 by @mudler in #5939
  • chore(model gallery): add flux.1-dev-ggml-q8_0 by @mudler in #5947
  • chore(model gallery): add flux.1-dev-ggml-abliterated-v2-q8_0 by @mudler in #5948
  • chore(model gallery): add flux.1-krea-dev-ggml by @mudler in #5949

Other Changes

  • docs: ⬆️ update docs version mudler/LocalAI by @localai-bot in #5929
  • chore: ⬆️ Update ggml-org/llama.cpp to 8ad7b3e65b5834e5574c2f5640056c9047b5d93b by @localai-bot in #5931
  • chore: ⬆️ Update leejet/stable-diffusion.cpp to f6b9aa1a4373e322ff12c15b8a0749e6dd6f0253 by @localai-bot in #5930
  • chore: ⬆️ Update ggml-org/whisper.cpp to d0a9d8c7f8f7b91c51d77bbaa394b915f79cde6b by @localai-bot in #5932
  • chore: ⬆️ Update ggml-org/llama.cpp to aa79524c51fb014f8df17069d31d7c44b9ea6cb8 by @localai-bot in #5934
  • chore: ⬆️ Update ggml-org/llama.cpp to e9192bec564780bd4313ad6524d20a0ab92797db by @localai-bot in #5940
  • chore: ⬆️ Update ggml-org/whisper.cpp to f7502dca872866a310fe69d30b163fa87d256319 by @localai-bot in #5941
  • chore: update swagger by @mudler in #5946
  • feat(stablediffusion-ggml): allow to load loras by @mudler in #5943
  • chore(capability): improve messages by @mudler in #5944
  • feat(swagger): update swagger by @localai-bot in #5950
  • chore: ⬆️ Update ggml-org/llama.cpp to daf2dd788066b8b239cb7f68210e090c2124c199 by @localai-bot in #5951

Full Changelog: v3.3.0...v3.3.1

v3.3.0

28 Jul 15:03
36179ff
Compare
Choose a tag to compare




🚀 LocalAI 3.3.0

What’s New in LocalAI 3.3.0 🎉

  • Object detection! From 3.3.0, now LocalAI supports with a new API - also fast object detection! Just install the rfdetr-base model - See the documentation to learn more
  • Backends now have defined mirrors for download - this helps when primary registries fails during download
  • Bug fixes: worked hard into squashing bugfixes in this release! Ranging from container images to backends and installation scripts

The Complete Local Stack for Privacy-First AI

LocalAI Logo

LocalAI

The free, Open Source OpenAI alternative. Acts as a drop-in replacement REST API compatible with OpenAI specifications for local AI inferencing. No GPU required.

Link: https://github.com/mudler/LocalAI

LocalAGI Logo

LocalAGI

A powerful Local AI agent management platform. Serves as a drop-in replacement for OpenAI's Responses API, supercharged with advanced agentic capabilities and a no-code UI.

Link: https://github.com/mudler/LocalAGI

LocalRecall Logo

LocalRecall

A RESTful API and knowledge base management system providing persistent memory and storage capabilities for AI agents. Designed to work alongside LocalAI and LocalAGI.

Link: https://github.com/mudler/LocalRecall

Thank you! ❤️

A massive THANK YOU to our incredible community and our sponsors! LocalAI has over 34,100 stars, and LocalAGI has already rocketed past 900+ stars!

As a reminder, LocalAI is real FOSS (Free and Open Source Software) and its sibling projects are community-driven and not backed by VCs or a company. We rely on contributors donating their spare time and our sponsors to provide us the hardware! If you love open-source, privacy-first AI, please consider starring the repos, contributing code, reporting bugs, or spreading the word!

👉 Check out the reborn LocalAGI v2 today: https://github.com/mudler/LocalAGI

Full changelog 👇

👉 Click to expand 👈

What's Changed

Bug fixes 🐛

  • fix(backend gallery): intel images for python-based backends, re-add exllama2 by @mudler in #5928

Exciting New Features 🎉

Other Changes

  • docs: ⬆️ update docs version mudler/LocalAI by @localai-bot in #5920
  • chore: ⬆️ Update ggml-org/whisper.cpp to e7bf0294ec9099b5fc21f5ba969805dfb2108cea by @localai-bot in #5922
  • chore: ⬆️ Update ggml-org/llama.cpp to 11dd5a44eb180e1d69fac24d3852b5222d66fb7f by @localai-bot in #5921
  • chore: drop assistants endpoint by @mudler in #5926
  • chore: ⬆️ Update ggml-org/llama.cpp to bf78f5439ee8e82e367674043303ebf8e92b4805 by @localai-bot in #5927

Full Changelog: v3.2.3...v3.3.0

v3.2.3

26 Jul 06:31
a8057b9
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

  • fix(cuda): be consistent with image tag naming by @mudler in #5916

📖 Documentation and examples

  • chore(docs): add documentation on backend detection override by @mudler in #5915

Other Changes

  • chore: ⬆️ Update ggml-org/llama.cpp to c7f3169cd523140a288095f2d79befb20a0b73f4 by @localai-bot in #5913

Full Changelog: v3.2.2...v3.2.3

v3.2.2

25 Jul 21:05
47ccfcc
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

  • fix(backends gallery): trim string when reading cap from file by @mudler in #5909
  • fix(vulkan): use correct image suffix by @mudler in #5911
  • fix(ci): add nvidia-l4t capability to l4t images by @mudler in #5914

Exciting New Features 🎉

Other Changes

Full Changelog: v3.2.1...v3.2.2

v3.2.1

25 Jul 14:39
ee625fc
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

  • fix(install.sh): update to use the new binary naming by @mudler in #5903
  • fix(backends gallery): pass-by backend galleries to the model service by @mudler in #5906

Other Changes

  • chore: ⬆️ Update ggml-org/llama.cpp to 3f4fc97f1d745f1d5d3c853949503136d419e6de by @localai-bot in #5900
  • chore: ⬆️ Update leejet/stable-diffusion.cpp to eed97a5e1d054f9c1e7ac01982ae480411d4157e by @localai-bot in #5901
  • chore: ⬆️ Update ggml-org/whisper.cpp to 7de8dd783f7b2eab56bff6bbc5d3369e34f0e77f by @localai-bot in #5902

Full Changelog: v3.2.0...v3.2.1

v3.2.0

24 Jul 20:36
a28f276
Compare
Choose a tag to compare




🚀 LocalAI 3.2.0

Welcome to LocalAI 3.2.0! This is a release that refactors our architecture to be more flexible and lightweight.

The core is now separated from all the backends, making LocalAI faster to download, easier to manage, portable, and much more smaller.

TL;DR – What’s New in LocalAI 3.2.0 🎉

  • 🧩 Modular Backends: All backends now live outside the main binary in our new Backend Gallery. This means you can update, add, or manage backends independently of LocalAI releases.
  • 📉 Leaner Than Ever: The LocalAI binary and container images are drastically smaller, making for faster downloads and a reduced footprint.
  • 🤖 Smart Backend Installation: It just works! When you install a model, LocalAI automatically detects your hardware (CPU, NVIDIA, AMD, Intel) and downloads the necessary backend. No more manual configuration!
  • 🛠️ Simplified Build Process: The new modular architecture significantly simplifies the build process for contributors and power users.
  • ⚡️ Intel GPU Support for Whisper: Transcription with Whisper can now be accelerated on Intel GPUs using SYCL, bringing more hardware options to our users.
  • 🗣️ Enhanced Realtime Audio: We've added speech started and stopped events for more interactive applications and OpenAI-compatible support for the input_audio field in the chat API.
  • 🧠 Massive Model Expansion: The gallery has been updated with over 50 new models, including the latest from Qwen3, Gemma, Mistral, Nemotron, and more!

Note: CI is in the process of building all the backends for this release and will be available soon - if you hit any issue, please try in a few, thanks for understanding!
Note: Some parts of the documentation and the installation scripts (that download the release binaries) have to yet be adapted to the latest changes and/or might not reflect the current state

A New Modular Architecture 🧩

The biggest change in v3.2.0 is the complete separation of inference backends from the core LocalAI binary. Backends like llama.cpp, whisper.cpp, piper, and stablediffusion-ggml are no longer bundled in.

This fundamental shift makes LocalAI:

  • Lighter: Significantly smaller binary and container image sizes.
  • More Flexible: Update backends anytime from the gallery without waiting for a new LocalAI release.
  • Easier to Maintain: A cleaner, more streamlined codebase for faster development.
  • Easier to Customize: you can build your own backends and install them in your LocalAI instances.

Smart, Automatic Backend Installation 🤖

To make the new modular system seamless, LocalAI now features automatic backend installation.

When you install a model from the gallery (or a YAML file), LocalAI intelligently detects the required backend and your system's capabilities, then downloads the correct version for you. Whether you're running on a standard CPU, an NVIDIA GPU, an AMD GPU, or an Intel GPU, LocalAI handles it automatically.

For advanced use cases or to override auto-detection, you can use the LOCALAI_FORCE_META_BACKEND_CAPABILITY environment variable. Here are the available options:

  • default: Forces CPU-only backend. This is the fallback if no specific hardware is detected.
  • nvidia: Forces backends compiled with CUDA support for NVIDIA GPUs.
  • amd: Forces backends compiled with ROCm support for AMD GPUs.
  • intel: Forces backends compiled with SYCL/oneAPI support for Intel GPUs.

The Backend Gallery & CLI Control 🖼️

You are in full control. You can browse, install, and manage all available backends directly from the WebUI or using the new CLI commands:

# List all available backends in the gallery
local-ai backends list

# Install a specific backend (e.g., llama-cpp)
local-ai backends install llama-cpp

# Uninstall a backend
local-ai backends uninstall llama-cpp

For development, offline or air-gapped environments, you can now also install backends directly from a local OCI tar file:

local-ai backends install "ocifile://<PATH_TO_TAR_FILE>"

Other Key Improvements

  • 🗣️ Enhanced Realtime and Audio APIs: Building voice-activated applications is now easier.
    • The new speech started and stopped events give you precise control over realtime audio streams.
    • We now support the input_audio field in the /v1/chat/completions endpoint for multimodal audio inputs, improving OpenAI compatibility.
  • ⚡️ Intel GPU Acceleration for Whisper: Our Whisper backend now supports SYCL, enabling hardware-accelerated transcriptions on Intel GPUs.
  • ✅ UI and Bug Fixes: We've squashed several bugs for a smoother experience, including a fix that correctly shows the download status for backend images in the gallery, so you always know what's happening.
  • 🧠 Massive Model Gallery Expansion: Our model gallery has never been bigger! We've added over 50 new and updated models, with a focus on powerful new releases like qwen3, devstral-small, and nemotron.

🚨 Important Note for Upgrading

Due to the new modular architecture, if you have existing models installed with a version prior to 3.2.0, they might not have a specific backend assigned.

After upgrading, you may need to install the required backend manually for these models to work. You can do this easily from the WebUI or via the CLI: local-ai backends install <backend_name>.

The Complete Local Stack for Privacy-First AI

LocalAI Logo

LocalAI

The free, Open Source OpenAI alternative. Acts as a drop-in replacement REST API compatible with OpenAI specifications for local AI inferencing. No GPU required.

Link: https://github.com/mudler/LocalAI

LocalAGI Logo

LocalAGI

A powerful Local AI agent management platform. Serves as a drop-in replacement for OpenAI's Responses API, supercharged with advanced agentic capabilities and a no-code UI.

Link: https://github.com/mudler/LocalAGI

LocalRecall Logo

LocalRecall

A RESTful API and knowledge base management system providing persistent memory and storage capabilities for AI agents. Designed to work alongside LocalAI and LocalAGI.

Link: https://github.com/mudler/LocalRecall

Thank you! ❤️

A massive THANK YOU to our incredible community and our sponsors! LocalAI has over 34,100 stars, and LocalAGI has already rocketed past 900+ stars!

As a reminder, LocalAI is real FOSS (Free and Open Source Software) and its sibling projects are community-driven and not backed by VCs or a company. We rely on contributors donating their spare time and our sponsors to provide us the hardware! If you love open-source, privacy-first AI, please consider starring the repos, contributing code, reporting bugs, or spreading the word!

👉 Check out the reborn LocalAGI v2 today: https://github.com/mudler/LocalAGI

Full changelog 👇

👉 Click to expand 👈

What's Changed

Breaking Changes 🛠

  • feat: do not bundle llama-cpp anymore by @mudler in #5790
  • feat: refactor build process, drop embedded backends by @mudler in #5875

Bug fixes 🐛

  • fix(gallery): automatically install model from name by @mudler in #5757
  • fix: Diffusers and XPU fixes by @richiejp in #5737
  • fix(gallery): correctly show status for downloading OCI images by @mudler in #5774
  • fix: explorer page should not have login by @mudler in #5855
  • fix: dockerfile typo by @LeonSijiaLu in #5823
  • fix(docs): Resolve logo overlap on tablet view by @dedyf5 in #5853
  • fix: do not pass by environ to ffmpeg by @mudler in #5871
  • fix(p2p): adapt to backend changes, general improvements by @mudler in #5889

Exciting New Features 🎉

  • feat(llama.cpp): allow to set kv-overrides by @mudler in #5745
  • feat(backends): add metas in the gallery by @mudler in #5784
  • feat(system): detect and allow to override capabilities by @mudler in #5785
  • chore(cli): add backends CLI to manipulate and install backends by @mudler in https://gith...
Read more

v3.1.1

27 Jun 21:48
cd2b0c0
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

  • fix(backends gallery): correctly identify gpu vendor by @mudler in #5739
  • fix(backends gallery): meta packages do not have URIs by @mudler in #5740

Exciting New Features 🎉

  • feat(gallery): automatically install missing backends along models by @mudler in #5736

👒 Dependencies

  • chore: ⬆️ Update ggml-org/whisper.cpp to c88ffbf9baeaae8c2cc0a4f496618314bb2ee9e0 by @localai-bot in #5742
  • chore: ⬆️ Update ggml-org/llama.cpp to 72babea5dea56c8a8e8420ccf731b12a5cf37854 by @localai-bot in #5743

Other Changes

Full Changelog: v3.1.0...v3.1.1

v3.1.0

26 Jun 20:14
6a650e6
Compare
Choose a tag to compare




🚀 LocalAI 3.1

🚀 Highlights

Support for Gemma 3n!

Gemma 3n has been released and it's now available in LocalAI (currently only for text generation, install it with:

  • local-ai run gemma-3n-e2b-it
  • local-ai run gemma-3n-e4b-it

⚠️ Breaking Changes

Several important changes that reduce image size, simplify the ecosystem, and pave the way for a leaner LocalAI core:

🧰 Container Image Changes

  • Sources are no longer bundled in the container images. This significantly reduces image sizes.
    • Need to rebuild locally? Just follow the docs to build from scratch. We're working towards migrating all backends to the gallery, slimming down the default image further.

📁 Directory Structure Updated

New default model and backend paths for container images:

  • Models: /models/ (was /build/models)
  • Backends: /backends/ (was /build/backends)

🏷 Unified Image Tag Naming for master (development) builds

We've cleaned up and standardized container image tags for clarity and consistency:

  • gpu-nvidia-cuda11 and gpu-nvidia-cuda12 (previously cublas-cuda11, cublas-cuda12)
  • gpu-intel-f16 and gpu-intel-f32 (previously sycl-f16, sycl-f32)

Meta packages in backend galleries

We’ve introduced meta-packages to the backend gallery!
These packages automatically install the most suitable backend depending on the GPU detected in your system — saving time, reducing errors, and ensuring you get the right setup out of the box. These will be added as soon as the 3.1.0 images are going to be published, stay tuned!

For instance, you will be able to install vllm just by installing the vllm backend in the gallery ( no need to select anymore the correct GPU version)

The Complete Local Stack for Privacy-First AI

With LocalAGI rejoining LocalAI alongside LocalRecall, our ecosystem provides a complete, open-source stack for private, secure, and intelligent AI operations:

LocalAI Logo

LocalAI

The free, Open Source OpenAI alternative. Acts as a drop-in replacement REST API compatible with OpenAI specifications for local AI inferencing. No GPU required.

Link: https://github.com/mudler/LocalAI

LocalAGI Logo

LocalAGI

A powerful Local AI agent management platform. Serves as a drop-in replacement for OpenAI's Responses API, supercharged with advanced agentic capabilities and a no-code UI.

Link: https://github.com/mudler/LocalAGI

LocalRecall Logo

LocalRecall

A RESTful API and knowledge base management system providing persistent memory and storage capabilities for AI agents. Designed to work alongside LocalAI and LocalAGI.

Link: https://github.com/mudler/LocalRecall

Join the Movement! ❤️

A massive THANK YOU to our incredible community and our sponsors! LocalAI has over 33,500 stars, and LocalAGI has already rocketed past 800+ stars!

As a reminder, LocalAI is real FOSS (Free and Open Source Software) and its sibling projects are community-driven and not backed by VCs or a company. We rely on contributors donating their spare time and our sponsors to provide us the hardware! If you love open-source, privacy-first AI, please consider starring the repos, contributing code, reporting bugs, or spreading the word!

👉 Check out the reborn LocalAGI v2 today: https://github.com/mudler/LocalAGI

Full changelog 👇

👉 Click to expand 👈

What's Changed

Breaking Changes 🛠

  • chore(ci): ⚠️ fix latest tag by using docker meta action by @mudler in #5722
  • feat: ⚠️ reduce images size and stop bundling sources by @mudler in #5721

Bug fixes 🐛

  • fix(backends gallery): delete dangling dirs if installation failed by @mudler in #5729

Exciting New Features 🎉

  • feat(backend gallery): add meta packages by @mudler in #5696

🧠 Models

  • chore(model gallery): add qwen3-the-josiefied-omega-directive-22b-uncensored-abliterated-i1 by @mudler in #5704
  • chore(model gallery): add menlo_jan-nano by @mudler in #5705
  • chore(model gallery): add qwen3-the-xiaolong-omega-directive-22b-uncensored-abliterated-i1 by @mudler in #5706
  • chore(model gallery): add allura-org_q3-8b-kintsugi by @mudler in #5707
  • chore(model gallery): add ds-r1-qwen3-8b-arliai-rpr-v4-small-iq-imatrix by @mudler in #5708
  • chore(model gallery): add mistralai_mistral-small-3.2-24b-instruct-2506 by @mudler in #5714
  • chore(model gallery): add skywork_skywork-swe-32b by @mudler in #5715
  • chore(model gallery): add astrosage-70b by @mudler in #5716
  • chore(model gallery): add delta-vector_austral-24b-winton by @mudler in #5717
  • chore(model gallery): add menlo_jan-nano-128k by @mudler in #5723
  • chore(model gallery): add gemma-3n-e2b-it by @mudler in #5730
  • chore(model gallery): add gemma-3n-e4b-it by @mudler in #5731

👒 Dependencies

  • chore: ⬆️ Update ggml-org/whisper.cpp to 3e65f518ddf840b13b74794158aa95a2c8aa30cc by @localai-bot in #5691
  • chore: ⬆️ Update ggml-org/llama.cpp to 8f71d0f3e86ccbba059350058af8758cafed73e6 by @localai-bot in #5692
  • chore: ⬆️ Update ggml-org/llama.cpp to 06cbedfca1587473df9b537f1dd4d6bfa2e3de13 by @localai-bot in #5697
  • chore: ⬆️ Update ggml-org/whisper.cpp to e6c10cf3d5d60dc647eb6cd5e73d3c347149f746 by @localai-bot in #5702
  • chore: ⬆️ Update ggml-org/llama.cpp to aa0ef5c578eef4c2adc7be1282f21bab5f3e8d26 by @localai-bot in #5703
  • chore: ⬆️ Update ggml-org/llama.cpp to 238005c2dc67426cf678baa2d54c881701693288 by @localai-bot in #5710
  • chore: ⬆️ Update ggml-org/whisper.cpp to a422176937c5bb20eb58d969995765f90d3c1a9b by @localai-bot in #5713
  • chore: ⬆️ Update ggml-org/llama.cpp to ce82bd0117bd3598300b3a089d13d401b90279c7 by @localai-bot in #5712
  • chore: ⬆️ Update ggml-org/llama.cpp to 73e53dc834c0a2336cd104473af6897197b96277 by @localai-bot in #5719
  • chore: ⬆️ Update ggml-org/whisper.cpp to 0083335ba0e9d6becbe0958903b0a27fc2ebaeed by @localai-bot in #5718
  • chore: ⬆️ Update leejet/stable-diffusion.cpp to 10c6501bd05a697e014f1bee3a84e5664290c489 by @localai-bot in #4925
  • chore: ⬆️ Update ggml-org/llama.cpp to 2bf9d539dd158345e3a3b096e16474af535265b4 by @localai-bot in #5724
  • chore: ⬆️ Update ggml-org/whisper.cpp to 4daf7050ca2bf17f5166f45ac6da651c4e33f293 by @localai-bot in #5725
  • Revert "chore: ⬆️ Update leejet/stable-diffusion.cpp to 10c6501bd05a697e014f1bee3a84e5664290c489" by @mudler in #5727
  • chore: ⬆️ Update ggml-org/llama.cpp to 8846aace4934ad29651ea61b8c7e3f6b0556e3d2 by @localai-bot in #5734
  • chore: ⬆️ Update ggml-org/whisper.cpp to 32cf4e2aba799aff069011f37ca025401433cf9f by @localai-bot in #5733

Other Changes

**Full...

Read more