2023-04-19 17:03:12 +00:00
< h1 align = "center" >
< br >
2025-07-06 16:07:36 +00:00
< img width = "300" src = "./core/http/static/logo.png" > < br >
2023-04-19 17:03:12 +00:00
< br >
< / h1 >
2023-03-20 20:30:55 +00:00
2023-08-06 22:01:01 +00:00
< p align = "center" >
< a href = "https://github.com/go-skynet/LocalAI/stargazers" target = "blank" >
< img src = "https://img.shields.io/github/stars/go-skynet/LocalAI?style=for-the-badge" alt = "LocalAI stars" / >
< / a >
< a href = 'https://github.com/go-skynet/LocalAI/releases' >
< img src = 'https://img.shields.io/github/release/go-skynet/LocalAI?&label=Latest&style=for-the-badge' >
< / a >
2026-03-08 20:17:36 +00:00
< a href = "LICENSE" target = "blank" >
< img src = "https://img.shields.io/badge/License-MIT-yellow.svg?style=for-the-badge" alt = "LocalAI License" / >
< / a >
< / p >
2023-08-06 22:01:01 +00:00
< p align = "center" >
< a href = "https://twitter.com/LocalAI_API" target = "blank" >
2025-05-07 20:20:06 +00:00
< img src = "https://img.shields.io/badge/X-%23000000.svg?style=for-the-badge&logo=X&logoColor=white&label=LocalAI_API" alt = "Follow LocalAI_API" / >
2023-08-06 22:01:01 +00:00
< / a >
< a href = "https://discord.gg/uJAeKSAGDy" target = "blank" >
2025-12-12 11:50:55 +00:00
< img src = "https://img.shields.io/badge/dynamic/json?color=blue&label=Discord&style=for-the-badge&query=approximate_member_count&url=https%3A%2F%2Fdiscordapp.com%2Fapi%2Finvites%2FuJAeKSAGDy%3Fwith_counts%3Dtrue&logo=discord" alt = "Join LocalAI Discord Community" / >
2023-08-06 22:01:01 +00:00
< / a >
2024-03-25 21:04:32 +00:00
< / p >
2024-11-21 09:04:45 +00:00
< p align = "center" >
2025-01-22 17:07:30 +00:00
< a href = "https://trendshift.io/repositories/5539" target = "_blank" > < img src = "https://trendshift.io/api/badge/repositories/5539" alt = "mudler%2FLocalAI | Trendshift" style = "width: 250px; height: 55px;" width = "250" height = "55" / > < / a >
2024-11-21 09:04:45 +00:00
< / p >
2026-03-22 20:28:07 +00:00
**LocalAI** is the open-source AI engine. Run any model - LLMs, vision, voice, image, video - on any hardware. No GPU required.
2024-03-26 17:45:25 +00:00
2026-03-22 20:24:44 +00:00
- **Drop-in API compatibility** — OpenAI, Anthropic, ElevenLabs APIs
- **35+ backends** — llama.cpp, vLLM, transformers, whisper, diffusers, MLX...
- **Any hardware** — NVIDIA, AMD, Intel, Apple Silicon, Vulkan, or CPU-only
- **Multi-user ready** — API key auth, user quotas, role-based access
- **Built-in AI agents** — autonomous agents with tool use, RAG, MCP, and skills
- **Privacy-first** — your data never leaves your infrastructure
2023-08-06 22:01:01 +00:00
2026-03-22 20:24:44 +00:00
Created and maintained by [Ettore Di Giacinto ](https://github.com/mudler ).
2026-03-09 08:26:16 +00:00
2026-03-22 20:24:44 +00:00
> [:book: Documentation](https://localai.io/) | [:speech_balloon: Discord](https://discord.gg/uJAeKSAGDy) | [💻 Quickstart](https://localai.io/basics/getting_started/) | [🖼️ Models](https://models.localai.io/) | [❓FAQ](https://localai.io/faq/)
2025-04-15 15:51:24 +00:00
2026-04-03 08:23:03 +00:00
## Guided tour
2026-01-24 21:46:40 +00:00
2026-04-03 08:24:08 +00:00
https://github.com/user-attachments/assets/08cbb692-57da-48f7-963d-2e7b43883c18
2026-01-24 21:46:40 +00:00
2026-04-03 08:23:03 +00:00
< details >
< summary >
Click to see more!
< / summary >
#### User and auth
2026-01-24 21:46:40 +00:00
2026-04-03 08:23:03 +00:00
https://github.com/user-attachments/assets/228fa9ad-81a3-4d43-bfb9-31557e14a36c
#### Agents
2025-04-15 15:51:24 +00:00
2026-03-15 20:47:08 +00:00
https://github.com/user-attachments/assets/6270b331-e21d-4087-a540-6290006b381a
2025-04-15 15:51:24 +00:00
2026-04-03 08:23:03 +00:00
#### Usage metrics per user
https://github.com/user-attachments/assets/cbb03379-23b4-4e3d-bd26-d152f057007f
#### Fine-tuning and Quantization
https://github.com/user-attachments/assets/5ba4ace9-d3df-4795-b7d4-b0b404ea71ee
#### WebRTC
2026-04-03 08:24:08 +00:00
https://github.com/user-attachments/assets/ed88e34c-fed3-4b83-8a67-4716a9feeb7b
2026-04-03 08:23:03 +00:00
< / details >
2026-03-22 20:24:44 +00:00
## Quickstart
2025-12-08 16:08:15 +00:00
2026-03-22 20:24:44 +00:00
### macOS
2025-09-08 22:19:37 +00:00
< a href = "https://github.com/mudler/LocalAI/releases/latest/download/LocalAI.dmg" >
< img src = "https://img.shields.io/badge/Download-macOS-blue?style=for-the-badge&logo=apple&logoColor=white" alt = "Download LocalAI for macOS" / >
< / a >
2026-03-22 20:24:44 +00:00
> **Note:** The DMG is not signed by Apple. After installing, run: `sudo xattr -d com.apple.quarantine /Applications/LocalAI.app`. See [#6268](https://github.com/mudler/LocalAI/issues/6268) for details.
2025-11-04 11:06:01 +00:00
2026-01-14 21:10:59 +00:00
### Containers (Docker, podman, ...)
2025-04-14 08:48:10 +00:00
2026-03-22 20:24:44 +00:00
> Already ran LocalAI before? Use `docker start -i local-ai` to restart an existing container.
2025-10-10 19:31:11 +00:00
2026-03-22 20:24:44 +00:00
#### CPU only:
2025-05-14 17:28:30 +00:00
2024-05-19 14:37:10 +00:00
```bash
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest
2025-04-14 08:48:10 +00:00
```
2025-05-14 17:28:30 +00:00
2026-03-22 20:24:44 +00:00
#### NVIDIA GPU:
2025-05-14 17:28:30 +00:00
2025-04-14 08:48:10 +00:00
```bash
2026-03-22 20:24:44 +00:00
# CUDA 13
2025-12-25 09:00:07 +00:00
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-13
2026-03-22 20:24:44 +00:00
# CUDA 12
2024-11-21 08:56:05 +00:00
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-12
2025-05-14 17:28:30 +00:00
2026-03-22 20:24:44 +00:00
# NVIDIA Jetson ARM64 (CUDA 12, for AGX Orin and similar)
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-nvidia-l4t-arm64
2025-12-25 09:00:07 +00:00
2026-03-22 20:24:44 +00:00
# NVIDIA Jetson ARM64 (CUDA 13, for DGX Spark)
2025-12-25 09:00:07 +00:00
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-nvidia-l4t-arm64-cuda-13
2025-04-14 08:48:10 +00:00
```
2025-05-14 17:28:30 +00:00
2026-03-22 20:24:44 +00:00
#### AMD GPU (ROCm):
2025-05-14 17:28:30 +00:00
2025-04-14 08:48:10 +00:00
```bash
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-gpu-hipblas
```
2026-03-22 20:24:44 +00:00
#### Intel GPU (oneAPI):
2025-05-14 17:28:30 +00:00
```bash
2025-07-31 17:44:46 +00:00
docker run -ti --name local-ai -p 8080:8080 --device=/dev/dri/card1 --device=/dev/dri/renderD128 localai/localai:latest-gpu-intel
2025-04-14 08:48:10 +00:00
```
2025-05-14 17:28:30 +00:00
2026-03-22 20:24:44 +00:00
#### Vulkan GPU:
2025-05-14 17:28:30 +00:00
2025-04-14 08:48:10 +00:00
```bash
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-gpu-vulkan
```
2026-03-22 20:24:44 +00:00
### Loading models
2024-10-17 07:46:26 +00:00
```bash
2026-03-22 20:24:44 +00:00
# From the model gallery (see available models with `local-ai models list` or at https://models.localai.io)
2024-10-17 15:49:03 +00:00
local-ai run llama-3.2-1b-instruct:q4_k_m
2026-03-22 20:24:44 +00:00
# From Huggingface
2024-10-17 07:46:26 +00:00
local-ai run huggingface://TheBloke/phi-2-GGUF/phi-2.Q8_0.gguf
2026-03-22 20:24:44 +00:00
# From the Ollama OCI registry
2024-10-17 07:46:26 +00:00
local-ai run ollama://gemma:2b
2026-03-22 20:24:44 +00:00
# From a YAML config
2024-10-17 07:46:26 +00:00
local-ai run https://gist.githubusercontent.com/.../phi-2.yaml
2026-03-22 20:24:44 +00:00
# From a standard OCI registry (e.g., Docker Hub)
2024-10-17 07:46:26 +00:00
local-ai run oci://localai/phi-2:latest
```
2026-03-22 20:24:44 +00:00
> **Automatic Backend Detection**: LocalAI automatically detects your GPU capabilities and downloads the appropriate backend. For advanced options, see [GPU Acceleration](https://localai.io/features/gpu-acceleration/).
For more details, see the [Getting Started guide ](https://localai.io/basics/getting_started/ ).
## Latest News
- **March 2026**: [Agent management ](https://github.com/mudler/LocalAI/pull/8820 ), [New React UI ](https://github.com/mudler/LocalAI/pull/8772 ), [WebRTC ](https://github.com/mudler/LocalAI/pull/8790 ), [MLX-distributed via P2P and RDMA ](https://github.com/mudler/LocalAI/pull/8801 ), [MCP Apps, MCP Client-side ](https://github.com/mudler/LocalAI/pull/8947 )
- **February 2026**: [Realtime API for audio-to-audio with tool calling ](https://github.com/mudler/LocalAI/pull/6245 ), [ACE-Step 1.5 support ](https://github.com/mudler/LocalAI/pull/8396 )
- **January 2026**: **LocalAI 3.10.0** — Anthropic API support, Open Responses API, video & image generation (LTX-2), unified GPU backends, tool streaming, Moonshine, Pocket-TTS. [Release notes ](https://github.com/mudler/LocalAI/releases/tag/v3.10.0 )
- **December 2025**: [Dynamic Memory Resource reclaimer ](https://github.com/mudler/LocalAI/pull/7583 ), [Automatic multi-GPU model fitting (llama.cpp) ](https://github.com/mudler/LocalAI/pull/7584 ), [Vibevoice backend ](https://github.com/mudler/LocalAI/pull/7494 )
- **November 2025**: [Import models via URL ](https://github.com/mudler/LocalAI/pull/7245 ), [Multiple chats and history ](https://github.com/mudler/LocalAI/pull/7325 )
- **October 2025**: [Model Context Protocol (MCP) ](https://localai.io/docs/features/mcp/ ) support for agentic capabilities
- **September 2025**: New Launcher for macOS and Linux, extended backend support for Mac and Nvidia L4T, MLX-Audio, WAN 2.2
- **August 2025**: MLX, MLX-VLM, Diffusers, llama.cpp now supported on Apple Silicon
- **July 2025**: All backends migrated outside the main binary — [lightweight, modular architecture ](https://github.com/mudler/LocalAI/releases/tag/v3.2.0 )
For older news and full release notes, see [GitHub Releases ](https://github.com/mudler/LocalAI/releases ) and the [News page ](https://localai.io/basics/news/ ).
## Features
- [Text generation ](https://localai.io/features/text-generation/ ) (`llama.cpp`, `transformers` , `vllm` ... [and more ](https://localai.io/model-compatibility/ ))
- [Text to Audio ](https://localai.io/features/text-to-audio/ )
- [Audio to Text ](https://localai.io/features/audio-to-text/ )
- [Image generation ](https://localai.io/features/image-generation )
- [OpenAI-compatible tools API ](https://localai.io/features/openai-functions/ )
- [Realtime API ](https://localai.io/features/openai-realtime/ ) (Speech-to-speech)
- [Embeddings generation ](https://localai.io/features/embeddings/ )
- [Constrained grammars ](https://localai.io/features/constrained_grammars/ )
- [Download models from Huggingface ](https://localai.io/models/ )
- [Vision API ](https://localai.io/features/gpt-vision/ )
- [Object Detection ](https://localai.io/features/object-detection/ )
- [Reranker API ](https://localai.io/features/reranker/ )
- [P2P Inferencing ](https://localai.io/features/distribute/ )
2026-03-29 22:47:27 +00:00
- [Distributed Mode ](https://localai.io/features/distributed-mode/ ) — Horizontal scaling with PostgreSQL + NATS
2026-03-22 20:24:44 +00:00
- [Model Context Protocol (MCP) ](https://localai.io/docs/features/mcp/ )
- [Built-in Agents ](https://localai.io/features/agents/ ) — Autonomous AI agents with tool use, RAG, skills, SSE streaming, and [Agent Hub ](https://agenthub.localai.io )
- [Backend Gallery ](https://localai.io/backends/ ) — Install/remove backends on the fly via OCI images
- Voice Activity Detection (Silero-VAD)
- Integrated WebUI
## Supported Backends & Acceleration
LocalAI supports **35+ backends** including llama.cpp, vLLM, transformers, whisper.cpp, diffusers, MLX, MLX-VLM, and many more. Hardware acceleration is available for **NVIDIA** (CUDA 12/13), **AMD** (ROCm), **Intel** (oneAPI/SYCL), **Apple Silicon** (Metal), **Vulkan** , and **NVIDIA Jetson** (L4T). All backends can be installed on-the-fly from the [Backend Gallery ](https://localai.io/backends/ ).
See the full [Backend & Model Compatibility Table ](https://localai.io/model-compatibility/ ) and [GPU Acceleration guide ](https://localai.io/features/gpu-acceleration/ ).
## Resources
- [Documentation ](https://localai.io/ )
- [LLM fine-tuning guide ](https://localai.io/docs/advanced/fine-tuning/ )
- [Build from source ](https://localai.io/basics/build/ )
- [Kubernetes installation ](https://localai.io/basics/getting_started/#run-localai-in-kubernetes )
- [Integrations & community projects ](https://localai.io/docs/integrations/ )
- [Media & blog posts ](https://localai.io/basics/news/#media-blogs-social )
- [Examples ](https://github.com/mudler/LocalAI-examples )
## Autonomous Development Team
LocalAI is helped being maintained by a team of autonomous AI agents led by an AI Scrum Master.
- **Live Reports**: [reports.localai.io ](http://reports.localai.io )
- **Project Board**: [Agent task tracking ](https://github.com/users/mudler/projects/6 )
- **Blog Post**: [Learn about the experiment ](https://mudler.pm/posts/2026/02/28/a-call-to-open-source-maintainers-stop-babysitting-ai-how-i-built-a-100-local-autonomous-dev-team-to-maintain-localai-and-why-you-should-too/ )
2026-03-04 22:52:36 +00:00
2023-09-09 17:17:19 +00:00
## Citation
If you utilize this repository, data in a downstream project, please consider citing it with:
2023-09-10 07:21:47 +00:00
```
2023-09-09 17:17:19 +00:00
@misc {localai,
author = {Ettore Di Giacinto},
title = {LocalAI: The free, Open source OpenAI alternative},
year = {2023},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/go-skynet/LocalAI}},
2023-09-10 07:21:47 +00:00
```
2023-09-09 17:17:19 +00:00
2026-03-22 20:24:44 +00:00
## Sponsors
2023-07-09 12:14:54 +00:00
> Do you find LocalAI useful?
Support the project by becoming [a backer or sponsor ](https://github.com/sponsors/mudler ). Your logo will show up here with a link to your website.
2024-06-04 13:23:00 +00:00
A huge thank you to our generous sponsors who support this project covering CI expenses, and our [Sponsor list ](https://github.com/sponsors/mudler ):
2023-07-09 12:14:54 +00:00
2024-06-04 13:23:00 +00:00
< p align = "center" >
< a href = "https://www.spectrocloud.com/" target = "blank" >
2025-02-20 14:31:41 +00:00
< img height = "200" src = "https://github.com/user-attachments/assets/72eab1dd-8b93-4fc0-9ade-84db49f24962" >
2024-06-04 13:23:00 +00:00
< / a >
< a href = "https://www.premai.io/" target = "blank" >
< img height = "200" src = "https://github.com/mudler/LocalAI/assets/2420543/42e4ca83-661e-4f79-8e46-ae43689683d6" > < br >
< / a >
< / p >
2023-09-16 21:00:42 +00:00
2025-12-30 22:01:22 +00:00
### Individual sponsors
2026-03-22 20:24:44 +00:00
A special thanks to individual sponsors, a full list is on [GitHub ](https://github.com/sponsors/mudler ) and [buymeacoffee ](https://buymeacoffee.com/mudler ). Special shout out to [drikster80 ](https://github.com/drikster80 ) for being generous. Thank you everyone!
2025-12-30 22:01:22 +00:00
2026-03-22 20:24:44 +00:00
## Star history
2023-04-27 08:39:01 +00:00
2023-04-27 04:18:18 +00:00
[](https://star-history.com/#go-skynet/LocalAI& Date)
2026-03-22 20:24:44 +00:00
## License
2023-03-30 16:46:11 +00:00
2023-05-29 21:09:19 +00:00
LocalAI is a community-driven project created by [Ettore Di Giacinto ](https://github.com/mudler/ ).
2023-04-27 08:39:01 +00:00
2024-06-03 17:55:01 +00:00
MIT - Author Ettore Di Giacinto < mudler @ localai . io >
2023-05-03 09:45:22 +00:00
2026-03-22 20:24:44 +00:00
## Acknowledgements
2023-03-30 16:46:11 +00:00
2023-05-16 17:32:53 +00:00
LocalAI couldn't have been built without the help of great software already available from the community. Thank you!
2023-03-30 16:46:11 +00:00
- [llama.cpp ](https://github.com/ggerganov/llama.cpp )
- https://github.com/tatsu-lab/stanford_alpaca
- https://github.com/cornelk/llama-go for the initial ideas
2023-05-16 17:32:53 +00:00
- https://github.com/antimatter15/alpaca.cpp
- https://github.com/EdVince/Stable-Diffusion-NCNN
- https://github.com/ggerganov/whisper.cpp
2023-08-06 22:01:01 +00:00
- https://github.com/rhasspy/piper
2026-03-09 16:29:32 +00:00
- [exo ](https://github.com/exo-explore/exo ) for the MLX distributed auto-parallel sharding implementation
2023-04-28 08:54:39 +00:00
2026-03-22 20:24:44 +00:00
## Contributors
2023-04-28 08:54:39 +00:00
2026-03-22 20:24:44 +00:00
This is a community project, a special thanks to our contributors!
2023-04-28 08:54:39 +00:00
< a href = "https://github.com/go-skynet/LocalAI/graphs/contributors" >
< img src = "https://contrib.rocks/image?repo=go-skynet/LocalAI" / >
< / a >