2023-04-19 17:03:12 +00:00
< h1 align = "center" >
< br >
2025-07-06 16:07:36 +00:00
< img width = "300" src = "./core/http/static/logo.png" > < br >
2023-04-19 17:03:12 +00:00
< br >
< / h1 >
2023-03-20 20:30:55 +00:00
2023-08-06 22:01:01 +00:00
< p align = "center" >
< a href = "https://github.com/go-skynet/LocalAI/fork" target = "blank" >
< img src = "https://img.shields.io/github/forks/go-skynet/LocalAI?style=for-the-badge" alt = "LocalAI forks" / >
< / a >
< a href = "https://github.com/go-skynet/LocalAI/stargazers" target = "blank" >
< img src = "https://img.shields.io/github/stars/go-skynet/LocalAI?style=for-the-badge" alt = "LocalAI stars" / >
< / a >
< a href = "https://github.com/go-skynet/LocalAI/pulls" target = "blank" >
< img src = "https://img.shields.io/github/issues-pr/go-skynet/LocalAI?style=for-the-badge" alt = "LocalAI pull-requests" / >
< / a >
< a href = 'https://github.com/go-skynet/LocalAI/releases' >
< img src = 'https://img.shields.io/github/release/go-skynet/LocalAI?&label=Latest&style=for-the-badge' >
< / a >
< / p >
2023-04-24 16:10:58 +00:00
2024-03-25 21:04:32 +00:00
< p align = "center" >
< a href = "https://hub.docker.com/r/localai/localai" target = "blank" >
< img src = "https://img.shields.io/badge/dockerhub-images-important.svg?logo=Docker" alt = "LocalAI Docker hub" / >
< / a >
< a href = "https://quay.io/repository/go-skynet/local-ai?tab=tags&tag=latest" target = "blank" >
< img src = "https://img.shields.io/badge/quay.io-images-important.svg?" alt = "LocalAI Quay.io" / >
< / a >
< / p >
2023-06-28 17:26:25 +00:00
2023-08-06 22:01:01 +00:00
< p align = "center" >
< a href = "https://twitter.com/LocalAI_API" target = "blank" >
2025-05-07 20:20:06 +00:00
< img src = "https://img.shields.io/badge/X-%23000000.svg?style=for-the-badge&logo=X&logoColor=white&label=LocalAI_API" alt = "Follow LocalAI_API" / >
2023-08-06 22:01:01 +00:00
< / a >
< a href = "https://discord.gg/uJAeKSAGDy" target = "blank" >
2025-12-12 11:50:55 +00:00
< img src = "https://img.shields.io/badge/dynamic/json?color=blue&label=Discord&style=for-the-badge&query=approximate_member_count&url=https%3A%2F%2Fdiscordapp.com%2Fapi%2Finvites%2FuJAeKSAGDy%3Fwith_counts%3Dtrue&logo=discord" alt = "Join LocalAI Discord Community" / >
2023-08-06 22:01:01 +00:00
< / a >
2024-03-25 21:04:32 +00:00
< / p >
2024-11-21 09:04:45 +00:00
< p align = "center" >
2025-01-22 17:07:30 +00:00
< a href = "https://trendshift.io/repositories/5539" target = "_blank" > < img src = "https://trendshift.io/api/badge/repositories/5539" alt = "mudler%2FLocalAI | Trendshift" style = "width: 250px; height: 55px;" width = "250" height = "55" / > < / a >
2024-11-21 09:04:45 +00:00
< / p >
2024-03-26 17:45:25 +00:00
> :bulb: Get help - [❓FAQ](https://localai.io/faq/) [💭Discussions](https://github.com/go-skynet/LocalAI/discussions) [:speech_balloon: Discord](https://discord.gg/uJAeKSAGDy) [:book: Documentation website](https://localai.io/)
>
2025-11-07 14:57:08 +00:00
> [💻 Quickstart](https://localai.io/basics/getting_started/) [🖼️ Models](https://models.localai.io/) [🚀 Roadmap](https://github.com/mudler/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3Aroadmap) [🛫 Examples](https://github.com/mudler/LocalAI-examples) Try on
2025-05-07 20:20:06 +00:00
[](https://t.me/localaiofficial_bot)
2024-03-26 17:45:25 +00:00
2024-03-25 21:04:32 +00:00
[](https://github.com/go-skynet/LocalAI/actions/workflows/test.yml)[](https://github.com/go-skynet/LocalAI/actions/workflows/release.yaml)[](https://github.com/go-skynet/LocalAI/actions/workflows/image.yml)[](https://github.com/go-skynet/LocalAI/actions/workflows/bump_deps.yaml)[](https://artifacthub.io/packages/search?repo=localai)
2023-08-06 22:01:01 +00:00
2025-04-15 15:51:24 +00:00
**LocalAI** is the free, Open Source OpenAI alternative. LocalAI act as a drop-in replacement REST API that's compatible with OpenAI (Elevenlabs, Anthropic... ) API specifications for local AI inferencing. It allows you to run LLMs, generate images, audio (and not only) locally or on-prem with consumer grade hardware, supporting multiple model families. Does not require GPU. It is created and maintained by [Ettore Di Giacinto ](https://github.com/mudler ).
2026-03-08 08:32:26 +00:00
< details >
< summary > < strong > Table of Contents< / strong > < / summary >
- [Local Stack Family ](#local-stack-family )
- [Screenshots / Video ](#screenshots--video )
- [Quickstart ](#-quickstart )
- [macOS Download ](#macos-download )
- [Containers (Docker, podman, ...) ](#containers-docker-podman- )
- [Latest project news ](#-latest-project-news )
- [Features ](#-features )
- [Supported Backends & Acceleration ](#-supported-backends--acceleration )
- [Text Generation & Language Models ](#text-generation--language-models )
- [Audio & Speech Processing ](#audio--speech-processing )
- [Image & Video Generation ](#image--video-generation )
- [Specialized AI Tasks ](#specialized-ai-tasks )
- [Hardware Acceleration Matrix ](#hardware-acceleration-matrix )
- [Community and integrations ](#-community-and-integrations )
- [Resources ](#-resources )
- [Media, Blogs, Social ](#book--media-blogs-social )
- [Autonomous Development Team ](#-autonomous-development-team )
- [Citation ](#citation )
- [Sponsors ](#️ -sponsors )
- [Individual sponsors ](#individual-sponsors )
- [Star history ](#-star-history )
- [License ](#-license )
- [Acknowledgements ](#-acknowledgements )
- [Contributors ](#-contributors )
< / details >
2025-04-15 15:51:24 +00:00
2026-01-24 21:46:40 +00:00
## Local Stack Family
2026-01-24 21:49:27 +00:00
Liking LocalAI? LocalAI is part of an integrated suite of AI infrastructure tools, you might also like:
2026-01-24 21:46:40 +00:00
- **[LocalAGI](https://github.com/mudler/LocalAGI)** - AI agent orchestration platform with OpenAI Responses API compatibility and advanced agentic capabilities
- **[LocalRecall](https://github.com/mudler/LocalRecall)** - MCP/REST API knowledge base system providing persistent memory and storage for AI agents
- 🆕 ** [Cogito ](https://github.com/mudler/cogito )** - Go library for building intelligent, co-operative agentic software and LLM-powered workflows, focusing on improving results for small, open source language models that scales to any LLM. Powers LocalAGI and LocalAI MCP/Agentic capabilities
- 🆕 ** [Wiz ](https://github.com/mudler/wiz )** - Terminal-based AI agent accessible via Ctrl+Space keybinding. Portable, local-LLM friendly shell assistant with TUI/CLI modes, tool execution with approval, MCP protocol support, and multi-shell compatibility (zsh, bash, fish)
- 🆕 ** [SkillServer ](https://github.com/mudler/skillserver )** - Simple, centralized skills database for AI agents via MCP. Manages skills as Markdown files with MCP server integration, web UI for editing, Git synchronization, and full-text search capabilities
2025-04-15 15:51:24 +00:00
2025-12-08 15:56:34 +00:00
## Screenshots / Video
2025-04-15 15:51:24 +00:00
2025-12-08 16:08:15 +00:00
### Youtube video
< h1 align = "center" >
< br >
< a href = "https://www.youtube.com/watch?v=PDqYhB9nNHA" target = "_blank" > < img width = "300" src = "https://img.youtube.com/vi/PDqYhB9nNHA/0.jpg" > < / a > < br >
< br >
< / h1 >
### Screenshots
2023-12-01 19:05:58 +00:00
2025-03-31 17:35:34 +00:00
| Talk Interface | Generate Audio |
| --- | --- |
2026-03-08 16:58:59 +00:00
|  |  |
2025-03-31 17:35:34 +00:00
| Models Overview | Generate Images |
| --- | --- |
2025-04-15 15:51:24 +00:00
|  |  |
2025-03-31 17:35:34 +00:00
2025-04-15 15:51:24 +00:00
| Chat Interface | Home |
2025-03-31 17:35:34 +00:00
| --- | --- |
2025-04-15 15:51:24 +00:00
|  |  |
2025-03-31 17:35:34 +00:00
| Login | Swarm |
| --- | --- |
2025-04-15 15:51:24 +00:00
| |  |
2025-03-31 17:35:34 +00:00
2025-04-15 15:51:24 +00:00
## 💻 Quickstart
2024-05-19 14:37:10 +00:00
2026-01-14 21:08:48 +00:00
2025-04-21 20:11:43 +00:00
2025-09-08 22:19:37 +00:00
### macOS Download:
< a href = "https://github.com/mudler/LocalAI/releases/latest/download/LocalAI.dmg" >
< img src = "https://img.shields.io/badge/Download-macOS-blue?style=for-the-badge&logo=apple&logoColor=white" alt = "Download LocalAI for macOS" / >
< / a >
2025-11-04 11:06:01 +00:00
> Note: the DMGs are not signed by Apple as quarantined. See https://github.com/mudler/LocalAI/issues/6268 for a workaround, fix is tracked here: https://github.com/mudler/LocalAI/issues/6244
2026-01-14 21:10:59 +00:00
### Containers (Docker, podman, ...)
2025-04-14 08:48:10 +00:00
2025-10-10 19:31:11 +00:00
> **💡 Docker Run vs Docker Start**
>
> - `docker run` creates and starts a new container. If a container with the same name already exists, this command will fail.
> - `docker start` starts an existing container that was previously created with `docker run`.
>
> If you've already run LocalAI before and want to start it again, use: `docker start -i local-ai`
2026-01-14 21:11:58 +00:00
#### CPU only image:
2025-05-14 17:28:30 +00:00
2024-05-19 14:37:10 +00:00
```bash
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest
2025-04-14 08:48:10 +00:00
```
2025-05-14 17:28:30 +00:00
2026-01-14 21:11:58 +00:00
#### NVIDIA GPU Images:
2025-05-14 17:28:30 +00:00
2025-04-14 08:48:10 +00:00
```bash
2025-12-25 09:00:07 +00:00
# CUDA 13.0
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-13
2025-06-19 19:46:09 +00:00
# CUDA 12.0
2024-11-21 08:56:05 +00:00
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-12
2025-05-14 17:28:30 +00:00
# NVIDIA Jetson (L4T) ARM64
2025-12-25 09:00:07 +00:00
# CUDA 12 (for Nvidia AGX Orin and similar platforms)
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-nvidia-l4t-arm64
2025-12-25 09:00:07 +00:00
# CUDA 13 (for Nvidia DGX Spark)
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-nvidia-l4t-arm64-cuda-13
2025-04-14 08:48:10 +00:00
```
2025-05-14 17:28:30 +00:00
2026-01-14 21:11:58 +00:00
#### AMD GPU Images (ROCm):
2025-05-14 17:28:30 +00:00
2025-04-14 08:48:10 +00:00
```bash
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-gpu-hipblas
```
2026-01-14 21:11:58 +00:00
#### Intel GPU Images (oneAPI):
2025-05-14 17:28:30 +00:00
```bash
2025-07-31 17:44:46 +00:00
docker run -ti --name local-ai -p 8080:8080 --device=/dev/dri/card1 --device=/dev/dri/renderD128 localai/localai:latest-gpu-intel
2025-04-14 08:48:10 +00:00
```
2025-05-14 17:28:30 +00:00
2026-01-14 21:11:58 +00:00
#### Vulkan GPU Images:
2025-05-14 17:28:30 +00:00
2025-04-14 08:48:10 +00:00
```bash
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-gpu-vulkan
```
2026-01-14 21:11:58 +00:00
#### AIO Images (pre-downloaded models):
2025-05-14 17:28:30 +00:00
```bash
# CPU version
2024-11-21 08:56:05 +00:00
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-cpu
2025-05-14 17:28:30 +00:00
2025-12-25 09:00:07 +00:00
# NVIDIA CUDA 13 version
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-aio-gpu-nvidia-cuda-13
2025-05-14 17:28:30 +00:00
# NVIDIA CUDA 12 version
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-aio-gpu-nvidia-cuda-12
# Intel GPU version
2025-07-31 17:44:46 +00:00
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-gpu-intel
2025-05-14 17:28:30 +00:00
# AMD GPU version
docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-aio-gpu-hipblas
2024-05-19 14:37:10 +00:00
```
2025-05-14 17:28:30 +00:00
For more information about the AIO images and pre-downloaded models, see [Container Documentation ](https://localai.io/basics/container/ ).
2024-10-17 07:46:26 +00:00
To load models:
```bash
2024-10-17 15:49:03 +00:00
# From the model gallery (see available models with `local-ai models list`, in the WebUI from the model tab, or visiting https://models.localai.io)
local-ai run llama-3.2-1b-instruct:q4_k_m
# Start LocalAI with the phi-2 model directly from huggingface
2024-10-17 07:46:26 +00:00
local-ai run huggingface://TheBloke/phi-2-GGUF/phi-2.Q8_0.gguf
# Install and run a model from the Ollama OCI registry
local-ai run ollama://gemma:2b
# Run a model from a configuration file
local-ai run https://gist.githubusercontent.com/.../phi-2.yaml
# Install and run a model from a standard OCI registry (e.g., Docker Hub)
local-ai run oci://localai/phi-2:latest
```
2025-07-26 06:18:31 +00:00
> ⚡ **Automatic Backend Detection**: When you install models from the gallery or YAML files, LocalAI automatically detects your system's GPU capabilities (NVIDIA, AMD, Intel) and downloads the appropriate backend. For advanced configuration options, see [GPU Acceleration](https://localai.io/features/gpu-acceleration/#automatic-backend-detection).
2025-11-07 15:50:47 +00:00
For more information, see [💻 Getting started ](https://localai.io/basics/getting_started/index.html ), if you are interested in our roadmap items and future enhancements, you can see the [Issues labeled as Roadmap here ](https://github.com/mudler/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3Aroadmap )
2024-05-19 14:37:10 +00:00
2024-10-04 17:52:43 +00:00
## 📰 Latest project news
2026-03-06 23:03:08 +00:00
- March 2026: [Agent management ](https://github.com/mudler/LocalAI/pull/8820 ), [New React UI ](https://github.com/mudler/LocalAI/pull/8772 ), [WebRTC ](https://github.com/mudler/LocalAI/pull/8790 ),[MLX-distributed via P2P and RDMA](https://github.com/mudler/LocalAI/pull/8801)
2026-02-05 11:04:53 +00:00
- February 2026: [Realtime API for audio-to-audio with tool calling ](https://github.com/mudler/LocalAI/pull/6245 ), [ACE-Step 1.5 support ](https://github.com/mudler/LocalAI/pull/8396 )
- January 2026: **LocalAI 3.10.0** - Major release with Anthropic API support, Open Responses API for stateful agents, video & image generation suite (LTX-2), unified GPU backends, tool streaming & XML parsing, system-aware backend gallery, crash fixes for AVX-only CPUs and AMD VRAM reporting, request tracing, and new backends: **Moonshine** (ultra-fast transcription), **Pocket-TTS** (lightweight TTS). Vulkan arm64 builds now available. [Release notes ](https://github.com/mudler/LocalAI/releases/tag/v3.10.0 ).
2025-12-16 08:16:42 +00:00
- December 2025: [Dynamic Memory Resource reclaimer ](https://github.com/mudler/LocalAI/pull/7583 ), [Automatic fitting of models to multiple GPUS(llama.cpp) ](https://github.com/mudler/LocalAI/pull/7584 ), [Added Vibevoice backend ](https://github.com/mudler/LocalAI/pull/7494 )
2025-11-25 18:31:05 +00:00
- November 2025: Major improvements to the UX. Among these: [Import models via URL ](https://github.com/mudler/LocalAI/pull/7245 ) and [Multiple chats and history ](https://github.com/mudler/LocalAI/pull/7325 )
2025-10-05 16:50:23 +00:00
- October 2025: 🔌 [Model Context Protocol (MCP) ](https://localai.io/docs/features/mcp/ ) support added for agentic capabilities with external tools
2025-10-05 15:51:41 +00:00
- September 2025: New Launcher application for MacOS and Linux, extended support to many backends for Mac and Nvidia L4T devices. Models: Added MLX-Audio, WAN 2.2. WebUI improvements and Python-based backends now ships portable python environments.
2025-08-24 09:50:20 +00:00
- August 2025: MLX, MLX-VLM, Diffusers and llama.cpp are now supported on Mac M1/M2/M3+ chips ( with `development` suffix in the gallery ): https://github.com/mudler/LocalAI/pull/6049 https://github.com/mudler/LocalAI/pull/6119 https://github.com/mudler/LocalAI/pull/6121 https://github.com/mudler/LocalAI/pull/6060
2025-07-27 20:02:51 +00:00
- July/August 2025: 🔍 [Object Detection ](https://localai.io/features/object-detection/ ) added to the API featuring [rf-detr ](https://github.com/roboflow/rf-detr )
2025-07-25 09:51:23 +00:00
- July 2025: All backends migrated outside of the main binary. LocalAI is now more lightweight, small, and automatically downloads the required backend to run the model. [Read the release notes ](https://github.com/mudler/LocalAI/releases/tag/v3.2.0 )
2025-06-15 14:06:43 +00:00
- June 2025: [Backend management ](https://github.com/mudler/LocalAI/pull/5607 ) has been added. Attention: extras images are going to be deprecated from the next release! Read [the backend management PR ](https://github.com/mudler/LocalAI/pull/5607 ).
2025-05-26 18:00:31 +00:00
- May 2025: [Audio input ](https://github.com/mudler/LocalAI/pull/5466 ) and [Reranking ](https://github.com/mudler/LocalAI/pull/5396 ) in llama.cpp backend, [Realtime API ](https://github.com/mudler/LocalAI/pull/5392 ), Support to Gemma, SmollVLM, and more multimodal models (available in the gallery).
- May 2025: Important: image name changes [See release ](https://github.com/mudler/LocalAI/releases/tag/v2.29.0 )
- Apr 2025: Rebrand, WebUI enhancements
2025-04-15 15:51:24 +00:00
- Apr 2025: [LocalAGI ](https://github.com/mudler/LocalAGI ) and [LocalRecall ](https://github.com/mudler/LocalRecall ) join the LocalAI family stack.
2025-03-31 19:51:09 +00:00
- Apr 2025: WebUI overhaul, AIO images updates
- Feb 2025: Backend cleanup, Breaking changes, new backends (kokoro, OutelTTS, faster-whisper), Nvidia L4T images
2025-01-17 14:11:10 +00:00
- Jan 2025: LocalAI model release: https://huggingface.co/mudler/LocalAI-functioncall-phi-4-v0.3, SANA support in diffusers: https://github.com/mudler/LocalAI/pull/4603
2024-12-04 10:31:08 +00:00
- Dec 2024: stablediffusion.cpp backend (ggml) added ( https://github.com/mudler/LocalAI/pull/4289 )
- Nov 2024: Bark.cpp backend added ( https://github.com/mudler/LocalAI/pull/4287 )
2024-11-21 10:05:00 +00:00
- Nov 2024: Voice activity detection models (**VAD**) added to the API: https://github.com/mudler/LocalAI/pull/4204
2024-10-30 08:10:33 +00:00
- Oct 2024: examples moved to [LocalAI-examples ](https://github.com/mudler/LocalAI-examples )
2024-08-28 12:48:16 +00:00
- Aug 2024: 🆕 FLUX-1, [P2P Explorer ](https://explorer.localai.io )
2025-01-15 14:46:27 +00:00
- July 2024: 🔥🔥 🆕 P2P Dashboard, LocalAI Federated mode and AI Swarms: https://github.com/mudler/LocalAI/pull/2723. P2P Global community pools: https://github.com/mudler/LocalAI/issues/3113
2024-07-17 13:14:22 +00:00
- May 2024: 🔥🔥 Decentralized P2P llama.cpp: https://github.com/mudler/LocalAI/pull/2343 (peer2peer llama.cpp!) 👉 Docs https://localai.io/features/distribute/
- May 2024: 🔥🔥 Distributed inferencing: https://github.com/mudler/LocalAI/pull/2324
- April 2024: Reranker API: https://github.com/mudler/LocalAI/pull/2121
2023-12-01 19:05:58 +00:00
2024-10-04 17:52:43 +00:00
Roadmap items: [List of issues ](https://github.com/mudler/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3Aroadmap )
2023-08-06 22:31:46 +00:00
## 🚀 [Features](https://localai.io/features/)
2023-05-19 17:33:53 +00:00
2025-06-28 07:16:25 +00:00
- 🧩 [Backend Gallery ](https://localai.io/backends/ ): Install/remove backends on the fly, powered by OCI images — fully customizable and API-driven.
2024-12-27 14:13:06 +00:00
- 📖 [Text generation with GPTs ](https://localai.io/features/text-generation/ ) (`llama.cpp`, `transformers` , `vllm` ... [:book: and more ](https://localai.io/model-compatibility/index.html#model-compatibility-table ))
2023-08-06 22:01:01 +00:00
- 🗣 [Text to Audio ](https://localai.io/features/text-to-audio/ )
2026-02-15 08:57:50 +00:00
- 🔈 [Audio to Text ](https://localai.io/features/audio-to-text/ )
2024-12-27 14:13:06 +00:00
- 🎨 [Image generation ](https://localai.io/features/image-generation )
2024-05-25 14:13:04 +00:00
- 🔥 [OpenAI-alike tools API ](https://localai.io/features/openai-functions/ )
2026-01-29 07:44:53 +00:00
- ⚡ [Realtime API ](https://localai.io/features/openai-realtime/ ) (Speech-to-speech)
2023-08-06 22:01:01 +00:00
- 🧠 [Embeddings generation for vector databases ](https://localai.io/features/embeddings/ )
- ✍️ [Constrained grammars ](https://localai.io/features/constrained_grammars/ )
- 🖼️ [Download Models directly from Huggingface ](https://localai.io/models/ )
2024-04-28 13:54:15 +00:00
- 🥽 [Vision API ](https://localai.io/features/gpt-vision/ )
2025-07-27 20:02:51 +00:00
- 🔍 [Object Detection ](https://localai.io/features/object-detection/ )
2024-05-25 14:11:59 +00:00
- 📈 [Reranker API ](https://localai.io/features/reranker/ )
- 🆕🖧 [P2P Inferencing ](https://localai.io/features/distribute/ )
2025-10-27 20:44:41 +00:00
- 🆕🔌 [Model Context Protocol (MCP) ](https://localai.io/docs/features/mcp/ ) - Agentic capabilities with external tools and [LocalAGI's Agentic capabilities ](https://github.com/mudler/LocalAGI )
2026-03-06 23:03:08 +00:00
- 🆕🤖 [Built-in Agents ](https://localai.io/features/agents/ ) - Autonomous AI agents with tool use, knowledge base (RAG), skills, SSE streaming, import/export, and [Agent Hub ](https://agenthub.localai.io ) — powered by [LocalAGI ](https://github.com/mudler/LocalAGI )
2024-12-27 14:17:02 +00:00
- 🔊 Voice activity detection (Silero-VAD support)
2024-07-07 21:01:01 +00:00
- 🌍 Integrated WebUI!
2023-07-02 22:52:26 +00:00
2025-08-24 18:09:19 +00:00
## 🧩 Supported Backends & Acceleration
LocalAI supports a comprehensive range of AI backends with multiple acceleration options:
### Text Generation & Language Models
| Backend | Description | Acceleration Support |
|---------|-------------|---------------------|
2026-01-06 14:26:42 +00:00
| **llama.cpp** | LLM inference in C/C++ | CUDA 12/13, ROCm, Intel SYCL, Vulkan, Metal, CPU |
2025-12-25 09:00:07 +00:00
| **vLLM** | Fast LLM inference with PagedAttention | CUDA 12/13, ROCm, Intel |
2026-01-06 14:26:42 +00:00
| **transformers** | HuggingFace transformers framework | CUDA 12/13, ROCm, Intel, CPU |
2025-08-24 18:09:19 +00:00
| **MLX** | Apple Silicon LLM inference | Metal (M1/M2/M3+) |
| **MLX-VLM** | Apple Silicon Vision-Language Models | Metal (M1/M2/M3+) |
### Audio & Speech Processing
| Backend | Description | Acceleration Support |
|---------|-------------|---------------------|
2025-12-25 09:00:07 +00:00
| **whisper.cpp** | OpenAI Whisper in C/C++ | CUDA 12/13, ROCm, Intel SYCL, Vulkan, CPU |
| **faster-whisper** | Fast Whisper with CTranslate2 | CUDA 12/13, ROCm, Intel, CPU |
2026-02-05 11:04:53 +00:00
| **moonshine** | Ultra-fast transcription engine for low-end devices | CUDA 12/13, Metal, CPU |
2025-12-25 09:00:07 +00:00
| **coqui** | Advanced TTS with 1100+ languages | CUDA 12/13, ROCm, Intel, CPU |
| **kokoro** | Lightweight TTS model | CUDA 12/13, ROCm, Intel, CPU |
2026-01-06 14:26:42 +00:00
| **chatterbox** | Production-grade TTS | CUDA 12/13, CPU |
2025-08-24 18:09:19 +00:00
| **piper** | Fast neural TTS system | CPU |
| **kitten-tts** | Kitten TTS models | CPU |
| **silero-vad** | Voice Activity Detection | CPU |
2025-12-25 09:00:07 +00:00
| **neutts** | Text-to-speech with voice cloning | CUDA 12/13, ROCm, CPU |
| **vibevoice** | Real-time TTS with voice cloning | CUDA 12/13, ROCm, Intel, CPU |
2026-01-13 22:35:19 +00:00
| **pocket-tts** | Lightweight CPU-based TTS | CUDA 12/13, ROCm, Intel, CPU |
2026-01-23 14:18:41 +00:00
| **qwen-tts** | High-quality TTS with custom voice, voice design, and voice cloning | CUDA 12/13, ROCm, Intel, CPU |
2026-02-05 11:04:53 +00:00
| **ace-step** | Music generation from text descriptions, lyrics, or audio samples | CUDA 12/13, ROCm, Intel, Metal, CPU |
2025-08-24 18:09:19 +00:00
### Image & Video Generation
| Backend | Description | Acceleration Support |
|---------|-------------|---------------------|
2025-12-25 09:00:07 +00:00
| **stablediffusion.cpp** | Stable Diffusion in C/C++ | CUDA 12/13, Intel SYCL, Vulkan, CPU |
2026-01-06 14:26:42 +00:00
| **diffusers** | HuggingFace diffusion models | CUDA 12/13, ROCm, Intel, Metal, CPU |
2025-08-24 18:09:19 +00:00
### Specialized AI Tasks
| Backend | Description | Acceleration Support |
|---------|-------------|---------------------|
2025-12-25 09:00:07 +00:00
| **rfdetr** | Real-time object detection | CUDA 12/13, Intel, CPU |
2026-01-06 14:26:42 +00:00
| **rerankers** | Document reranking API | CUDA 12/13, ROCm, Intel, CPU |
2025-08-24 18:09:19 +00:00
| **local-store** | Vector database | CPU |
| **huggingface** | HuggingFace API integration | API-based |
### Hardware Acceleration Matrix
| Acceleration Type | Supported Backends | Hardware Support |
|-------------------|-------------------|------------------|
| **NVIDIA CUDA 12** | All CUDA-compatible backends | Nvidia hardware |
2025-12-25 09:00:07 +00:00
| **NVIDIA CUDA 13** | All CUDA-compatible backends | Nvidia hardware |
2026-02-05 11:04:53 +00:00
| **AMD ROCm** | llama.cpp, whisper, vllm, transformers, diffusers, rerankers, coqui, kokoro, neutts, vibevoice, pocket-tts, qwen-tts, ace-step | AMD Graphics |
| **Intel oneAPI** | llama.cpp, whisper, stablediffusion, vllm, transformers, diffusers, rfdetr, rerankers, coqui, kokoro, vibevoice, pocket-tts, qwen-tts, ace-step | Intel Arc, Intel iGPUs |
| **Apple Metal** | llama.cpp, whisper, diffusers, MLX, MLX-VLM, moonshine, ace-step | Apple M1/M2/M3+ |
2025-08-24 18:09:19 +00:00
| **Vulkan** | llama.cpp, whisper, stablediffusion | Cross-platform GPUs |
2026-02-05 11:04:53 +00:00
| **NVIDIA Jetson (CUDA 12)** | llama.cpp, whisper, stablediffusion, diffusers, rfdetr, ace-step | ARM64 embedded AI (AGX Orin, etc.) |
2025-12-25 09:00:07 +00:00
| **NVIDIA Jetson (CUDA 13)** | llama.cpp, whisper, stablediffusion, diffusers, rfdetr | ARM64 embedded AI (DGX Spark) |
2025-08-24 18:09:19 +00:00
| **CPU Optimized** | All backends | AVX/AVX2/AVX512, quantization support |
2023-04-12 23:13:14 +00:00
2023-12-01 18:11:45 +00:00
### 🔗 Community and integrations
2023-04-24 21:42:03 +00:00
2023-12-12 17:58:57 +00:00
Build and deploy custom containers:
- https://github.com/sozercan/aikit
2023-12-01 18:11:45 +00:00
WebUIs:
2023-11-28 22:14:16 +00:00
- https://github.com/Jirubizu/localai-admin
- https://github.com/go-skynet/LocalAI-frontend
2024-06-10 17:18:47 +00:00
- QA-Pilot(An interactive chat project that leverages LocalAI LLMs for rapid understanding and navigation of GitHub code repository) https://github.com/reid41/QA-Pilot
2023-11-28 22:14:16 +00:00
2025-10-05 17:38:47 +00:00
Agentic Libraries:
- https://github.com/mudler/cogito
MCPs:
- https://github.com/mudler/MCPs
2026-01-24 21:49:27 +00:00
OS Assistant:
- https://github.com/mudler/Keygeist - Keygeist is an AI-powered keyboard operator that listens for key combinations and responds with AI-generated text typed directly into your Linux box.
2023-11-28 22:14:16 +00:00
Model galleries
- https://github.com/go-skynet/model-gallery
2025-08-27 14:24:23 +00:00
Voice:
- https://github.com/richiejp/VoxInput
2023-11-28 22:14:16 +00:00
Other:
- Helm chart https://github.com/go-skynet/helm-charts
2023-12-28 22:03:10 +00:00
- VSCode extension https://github.com/badgooooor/localai-vscode-plugin
2025-01-10 08:20:33 +00:00
- Langchain: https://python.langchain.com/docs/integrations/providers/localai/
2024-05-07 06:39:58 +00:00
- Terminal utility https://github.com/djcopley/ShellOracle
2023-12-25 08:04:35 +00:00
- Local Smart assistant https://github.com/mudler/LocalAGI
2026-03-01 07:28:58 +00:00
- Home Assistant https://github.com/drndos/hass-openai-custom-conversation / https://github.com/valentinfrlch/ha-llmvision / https://github.com/loryanstrant/HA-LocalAI-Monitor
2023-12-28 22:01:52 +00:00
- Discord bot https://github.com/mudler/LocalAGI/tree/main/examples/discord
- Slack bot https://github.com/mudler/LocalAGI/tree/main/examples/slack
2024-06-10 17:18:47 +00:00
- Shell-Pilot(Interact with LLM using LocalAI models via pure shell scripts on your Linux or MacOS system) https://github.com/reid41/shell-pilot
2023-12-28 22:01:52 +00:00
- Telegram bot https://github.com/mudler/LocalAI/tree/master/examples/telegram-bot
2024-11-22 10:32:50 +00:00
- Another Telegram Bot https://github.com/JackBekket/Hellper
- Auto-documentation https://github.com/JackBekket/Reflexia
- Github bot which answer on issues, with code and documentation as context https://github.com/JackBekket/GitHelper
2024-07-17 03:57:01 +00:00
- Github Actions: https://github.com/marketplace/actions/start-localai
2023-12-28 22:01:52 +00:00
- Examples: https://github.com/mudler/LocalAI/tree/master/examples/
2024-03-05 09:14:30 +00:00
2023-05-03 13:51:54 +00:00
2023-08-06 22:01:01 +00:00
### 🔗 Resources
2023-05-19 17:33:53 +00:00
2024-05-31 20:59:51 +00:00
- [LLM finetuning guide ](https://localai.io/docs/advanced/fine-tuning/ )
2023-08-06 22:01:01 +00:00
- [How to build locally ](https://localai.io/basics/build/index.html )
- [How to install in Kubernetes ](https://localai.io/basics/getting_started/index.html#run-localai-in-kubernetes )
2024-02-03 23:54:49 +00:00
- [Projects integrating LocalAI ](https://localai.io/docs/integrations/ )
2024-01-11 08:25:18 +00:00
- [How tos section ](https://io.midori-ai.xyz/howtos/ ) (curated by our community)
2023-12-12 17:58:57 +00:00
2023-12-01 18:11:45 +00:00
## :book: 🎥 [Media, Blogs, Social](https://localai.io/basics/news/#media-blogs-social)
2026-03-04 22:52:36 +00:00
- 🆕 [LocalAI Autonomous Dev Team Blog Post ](https://mudler.pm/posts/2026/02/28/a-call-to-open-source-maintainers-stop-babysitting-ai-how-i-built-a-100-local-autonomous-dev-team-to-maintain-localai-and-why-you-should-too/ )
2024-08-01 16:44:39 +00:00
- [Run Visual studio code with LocalAI (SUSE) ](https://www.suse.com/c/running-ai-locally/ )
2024-05-31 20:59:51 +00:00
- 🆕 [Run LocalAI on Jetson Nano Devkit ](https://mudler.pm/posts/local-ai-jetson-nano-devkit/ )
2024-05-05 13:45:55 +00:00
- [Run LocalAI on AWS EKS with Pulumi ](https://www.pulumi.com/blog/low-code-llm-apps-with-local-ai-flowise-and-pulumi/ )
2024-02-27 14:43:15 +00:00
- [Run LocalAI on AWS ](https://staleks.hashnode.dev/installing-localai-on-aws-ec2-instance )
2023-12-01 18:11:45 +00:00
- [Create a slackbot for teams and OSS projects that answer to documentation ](https://mudler.pm/posts/smart-slackbot-for-teams/ )
- [LocalAI meets k8sgpt ](https://www.youtube.com/watch?v=PKrDNuJ_dfE )
- [Question Answering on Documents locally with LangChain, LocalAI, Chroma, and GPT4All ](https://mudler.pm/posts/localai-question-answering/ )
- [Tutorial to use k8sgpt with LocalAI ](https://medium.com/@tyler_97636/k8sgpt-localai-unlock-kubernetes-superpowers-for-free-584790de9b65 )
2026-03-04 22:52:36 +00:00
## 🤖 Autonomous Development Team
LocalAI is now helped being maintained (for small tasks!) by a full team of autonomous AI agents led by an AI Scrum Master! This experiment demonstrates how open source projects can leverage AI agents for sustainable, long-term maintenance.
- **📊 Live Reports**: [Automatically generated reports ](http://reports.localai.io )
- **📋 Project Board**: [Agent task tracking ](https://github.com/users/mudler/projects/6 )
- **📝 Blog Post**: [Learn about the autonomous dev team experiment ](https://mudler.pm/posts/2026/02/28/a-call-to-open-source-maintainers-stop-babysitting-ai-how-i-built-a-100-local-autonomous-dev-team-to-maintain-localai-and-why-you-should-too/ )
2023-09-09 17:17:19 +00:00
## Citation
If you utilize this repository, data in a downstream project, please consider citing it with:
2023-09-10 07:21:47 +00:00
```
2023-09-09 17:17:19 +00:00
@misc {localai,
author = {Ettore Di Giacinto},
title = {LocalAI: The free, Open source OpenAI alternative},
year = {2023},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/go-skynet/LocalAI}},
2023-09-10 07:21:47 +00:00
```
2023-09-09 17:17:19 +00:00
2023-08-06 22:01:01 +00:00
## ❤️ Sponsors
2023-07-09 12:14:54 +00:00
> Do you find LocalAI useful?
Support the project by becoming [a backer or sponsor ](https://github.com/sponsors/mudler ). Your logo will show up here with a link to your website.
2024-06-04 13:23:00 +00:00
A huge thank you to our generous sponsors who support this project covering CI expenses, and our [Sponsor list ](https://github.com/sponsors/mudler ):
2023-07-09 12:14:54 +00:00
2024-06-04 13:23:00 +00:00
< p align = "center" >
< a href = "https://www.spectrocloud.com/" target = "blank" >
2025-02-20 14:31:41 +00:00
< img height = "200" src = "https://github.com/user-attachments/assets/72eab1dd-8b93-4fc0-9ade-84db49f24962" >
2024-06-04 13:23:00 +00:00
< / a >
< a href = "https://www.premai.io/" target = "blank" >
< img height = "200" src = "https://github.com/mudler/LocalAI/assets/2420543/42e4ca83-661e-4f79-8e46-ae43689683d6" > < br >
< / a >
< / p >
2023-09-16 21:00:42 +00:00
2025-12-30 22:01:22 +00:00
### Individual sponsors
A special thanks to individual sponsors that contributed to the project, a full list is in [Github ](https://github.com/sponsors/mudler ) and [buymeacoffee ](https://buymeacoffee.com/mudler ), a special shout out goes to [drikster80 ](https://github.com/drikster80 ) for being generous. Thank you everyone!
2023-08-06 22:01:01 +00:00
## 🌟 Star history
2023-04-27 08:39:01 +00:00
2023-04-27 04:18:18 +00:00
[](https://star-history.com/#go-skynet/LocalAI& Date)
2023-08-06 22:01:01 +00:00
## 📖 License
2023-03-30 16:46:11 +00:00
2023-05-29 21:09:19 +00:00
LocalAI is a community-driven project created by [Ettore Di Giacinto ](https://github.com/mudler/ ).
2023-04-27 08:39:01 +00:00
2024-06-03 17:55:01 +00:00
MIT - Author Ettore Di Giacinto < mudler @ localai . io >
2023-05-03 09:45:22 +00:00
2023-08-06 22:01:01 +00:00
## 🙇 Acknowledgements
2023-03-30 16:46:11 +00:00
2023-05-16 17:32:53 +00:00
LocalAI couldn't have been built without the help of great software already available from the community. Thank you!
2023-03-30 16:46:11 +00:00
- [llama.cpp ](https://github.com/ggerganov/llama.cpp )
- https://github.com/tatsu-lab/stanford_alpaca
- https://github.com/cornelk/llama-go for the initial ideas
2023-05-16 17:32:53 +00:00
- https://github.com/antimatter15/alpaca.cpp
- https://github.com/EdVince/Stable-Diffusion-NCNN
- https://github.com/ggerganov/whisper.cpp
2023-08-06 22:01:01 +00:00
- https://github.com/rhasspy/piper
2023-04-28 08:54:39 +00:00
2023-08-06 22:01:01 +00:00
## 🤗 Contributors
2023-04-28 08:54:39 +00:00
2023-08-06 22:01:01 +00:00
This is a community project, a special thanks to our contributors! 🤗
2023-04-28 08:54:39 +00:00
< a href = "https://github.com/go-skynet/LocalAI/graphs/contributors" >
< img src = "https://contrib.rocks/image?repo=go-skynet/LocalAI" / >
< / a >