2023-04-19 17:03:12 +00:00
< h1 align = "center" >
< br >
2025-07-06 16:07:36 +00:00
< img width = "300" src = "./core/http/static/logo.png" > < br >
2023-04-19 17:03:12 +00:00
< br >
< / h1 >
2023-03-20 20:30:55 +00:00
2023-08-06 22:01:01 +00:00
< p align = "center" >
< a href = "https://github.com/go-skynet/LocalAI/fork" target = "blank" >
< img src = "https://img.shields.io/github/forks/go-skynet/LocalAI?style=for-the-badge" alt = "LocalAI forks" / >
< / a >
< a href = "https://github.com/go-skynet/LocalAI/stargazers" target = "blank" >
< img src = "https://img.shields.io/github/stars/go-skynet/LocalAI?style=for-the-badge" alt = "LocalAI stars" / >
< / a >
< a href = "https://github.com/go-skynet/LocalAI/pulls" target = "blank" >
< img src = "https://img.shields.io/github/issues-pr/go-skynet/LocalAI?style=for-the-badge" alt = "LocalAI pull-requests" / >
< / a >
< a href = 'https://github.com/go-skynet/LocalAI/releases' >
< img src = 'https://img.shields.io/github/release/go-skynet/LocalAI?&label=Latest&style=for-the-badge' >
< / a >
< / p >
2023-04-24 16:10:58 +00:00
2024-03-25 21:04:32 +00:00
< p align = "center" >
< a href = "https://hub.docker.com/r/localai/localai" target = "blank" >
< img src = "https://img.shields.io/badge/dockerhub-images-important.svg?logo=Docker" alt = "LocalAI Docker hub" / >
< / a >
< a href = "https://quay.io/repository/go-skynet/local-ai?tab=tags&tag=latest" target = "blank" >
< img src = "https://img.shields.io/badge/quay.io-images-important.svg?" alt = "LocalAI Quay.io" / >
< / a >
< / p >
2023-06-28 17:26:25 +00:00
2023-08-06 22:01:01 +00:00
< p align = "center" >
< a href = "https://twitter.com/LocalAI_API" target = "blank" >
2025-05-07 20:20:06 +00:00
< img src = "https://img.shields.io/badge/X-%23000000.svg?style=for-the-badge&logo=X&logoColor=white&label=LocalAI_API" alt = "Follow LocalAI_API" / >
2023-08-06 22:01:01 +00:00
< / a >
< a href = "https://discord.gg/uJAeKSAGDy" target = "blank" >
2025-12-12 11:50:55 +00:00
< img src = "https://img.shields.io/badge/dynamic/json?color=blue&label=Discord&style=for-the-badge&query=approximate_member_count&url=https%3A%2F%2Fdiscordapp.com%2Fapi%2Finvites%2FuJAeKSAGDy%3Fwith_counts%3Dtrue&logo=discord" alt = "Join LocalAI Discord Community" / >
2023-08-06 22:01:01 +00:00
< / a >
2024-03-25 21:04:32 +00:00
< / p >
2024-11-21 09:04:45 +00:00
< p align = "center" >
2025-01-22 17:07:30 +00:00
< a href = "https://trendshift.io/repositories/5539" target = "_blank" > < img src = "https://trendshift.io/api/badge/repositories/5539" alt = "mudler%2FLocalAI | Trendshift" style = "width: 250px; height: 55px;" width = "250" height = "55" / > < / a >
2024-11-21 09:04:45 +00:00
< / p >
2024-03-26 17:45:25 +00:00
> :bulb: Get help - [❓FAQ](https://localai.io/faq/) [💭Discussions](https://github.com/go-skynet/LocalAI/discussions) [:speech_balloon: Discord](https://discord.gg/uJAeKSAGDy) [:book: Documentation website](https://localai.io/)
>
2025-11-07 14:57:08 +00:00
> [💻 Quickstart](https://localai.io/basics/getting_started/) [🖼️ Models](https://models.localai.io/) [🚀 Roadmap](https://github.com/mudler/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3Aroadmap) [🛫 Examples](https://github.com/mudler/LocalAI-examples) Try on
2025-05-07 20:20:06 +00:00
[](https://t.me/localaiofficial_bot)
2024-03-26 17:45:25 +00:00
2024-03-25 21:04:32 +00:00
[](https://github.com/go-skynet/LocalAI/actions/workflows/test.yml)[](https://github.com/go-skynet/LocalAI/actions/workflows/release.yaml)[](https://github.com/go-skynet/LocalAI/actions/workflows/image.yml)[](https://github.com/go-skynet/LocalAI/actions/workflows/bump_deps.yaml)[](https://artifacthub.io/packages/search?repo=localai)
2023-08-06 22:01:01 +00:00
2025-04-15 15:51:24 +00:00
**LocalAI** is the free, Open Source OpenAI alternative. LocalAI act as a drop-in replacement REST API that's compatible with OpenAI (Elevenlabs, Anthropic... ) API specifications for local AI inferencing. It allows you to run LLMs, generate images, audio (and not only) locally or on-prem with consumer grade hardware, supporting multiple model families. Does not require GPU. It is created and maintained by [Ettore Di Giacinto ](https://github.com/mudler ).
## 📚🆕 Local Stack Family
🆕 LocalAI is now part of a comprehensive suite of AI tools designed to work together:
< table >
< tr >
< td width = "50%" valign = "top" >
< a href = "https://github.com/mudler/LocalAGI" >
< img src = "https://raw.githubusercontent.com/mudler/LocalAGI/refs/heads/main/webui/react-ui/public/logo_2.png" width = "300" alt = "LocalAGI Logo" >
< / a >
< / td >
< td width = "50%" valign = "top" >
< h3 > < a href = "https://github.com/mudler/LocalAGI" > LocalAGI< / a > < / h3 >
< p > A powerful Local AI agent management platform that serves as a drop-in replacement for OpenAI's Responses API, enhanced with advanced agentic capabilities.< / p >
< / td >
< / tr >
< tr >
< td width = "50%" valign = "top" >
< a href = "https://github.com/mudler/LocalRecall" >
< img src = "https://raw.githubusercontent.com/mudler/LocalRecall/refs/heads/main/static/localrecall_horizontal.png" width = "300" alt = "LocalRecall Logo" >
< / a >
< / td >
< td width = "50%" valign = "top" >
< h3 > < a href = "https://github.com/mudler/LocalRecall" > LocalRecall< / a > < / h3 >
< p > A REST-ful API and knowledge base management system that provides persistent memory and storage capabilities for AI agents.< / p >
< / td >
< / tr >
< / table >
2025-12-08 15:56:34 +00:00
## Screenshots / Video
2025-04-15 15:51:24 +00:00
2025-12-08 16:08:15 +00:00
### Youtube video
< h1 align = "center" >
< br >
< a href = "https://www.youtube.com/watch?v=PDqYhB9nNHA" target = "_blank" > < img width = "300" src = "https://img.youtube.com/vi/PDqYhB9nNHA/0.jpg" > < / a > < br >
< br >
< / h1 >
### Screenshots
2023-12-01 19:05:58 +00:00
2025-03-31 17:35:34 +00:00
| Talk Interface | Generate Audio |
| --- | --- |
2025-04-15 15:51:24 +00:00
|  |  |
2025-03-31 17:35:34 +00:00
| Models Overview | Generate Images |
| --- | --- |
2025-04-15 15:51:24 +00:00
|  |  |
2025-03-31 17:35:34 +00:00
2025-04-15 15:51:24 +00:00
| Chat Interface | Home |
2025-03-31 17:35:34 +00:00
| --- | --- |
2025-04-15 15:51:24 +00:00
|  |  |
2025-03-31 17:35:34 +00:00
| Login | Swarm |
| --- | --- |
2025-04-15 15:51:24 +00:00
| |  |
2025-03-31 17:35:34 +00:00
2025-04-15 15:51:24 +00:00
## 💻 Quickstart
2024-05-19 14:37:10 +00:00
2024-06-22 08:29:46 +00:00
Run the installer script:
```bash
2025-04-21 20:11:43 +00:00
# Basic installation
2024-06-22 08:29:46 +00:00
curl https://localai.io/install.sh | sh
```
2025-11-21 16:29:36 +00:00
For more installation options, see [Installer Options ](https://localai.io/installation/ ).
2025-04-21 20:11:43 +00:00
2025-09-08 22:19:37 +00:00
### macOS Download:
< a href = "https://github.com/mudler/LocalAI/releases/latest/download/LocalAI.dmg" >
< img src = "https://img.shields.io/badge/Download-macOS-blue?style=for-the-badge&logo=apple&logoColor=white" alt = "Download LocalAI for macOS" / >
< / a >
2025-11-04 11:06:01 +00:00
> Note: the DMGs are not signed by Apple as quarantined. See https://github.com/mudler/LocalAI/issues/6268 for a workaround, fix is tracked here: https://github.com/mudler/LocalAI/issues/6244
2024-06-22 08:29:46 +00:00
Or run with docker:
2025-04-14 08:48:10 +00:00
2025-10-10 19:31:11 +00:00
> **💡 Docker Run vs Docker Start**
>
> - `docker run` creates and starts a new container. If a container with the same name already exists, this command will fail.
> - `docker start` starts an existing container that was previously created with `docker run`.
>
> If you've already run LocalAI before and want to start it again, use: `docker start -i local-ai`
2025-04-14 08:48:10 +00:00
### CPU only image:
2025-05-14 17:28:30 +00:00
2024-05-19 14:37:10 +00:00
```bash
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest
2025-04-14 08:48:10 +00:00
```
2025-05-14 17:28:30 +00:00
### NVIDIA GPU Images:
2025-04-14 08:48:10 +00:00
```bash
2025-12-25 09:00:07 +00:00
# CUDA 13.0
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-13
2025-06-19 19:46:09 +00:00
# CUDA 12.0
2024-11-21 08:56:05 +00:00
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-12
2025-05-14 17:28:30 +00:00
2025-06-19 19:46:09 +00:00
# CUDA 11.7
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-11
# NVIDIA Jetson (L4T) ARM64
2025-12-25 09:00:07 +00:00
# CUDA 12 (for Nvidia AGX Orin and similar platforms)
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-nvidia-l4t-arm64
2025-12-25 09:00:07 +00:00
# CUDA 13 (for Nvidia DGX Spark)
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-nvidia-l4t-arm64-cuda-13
2025-04-14 08:48:10 +00:00
```
2025-05-14 17:28:30 +00:00
### AMD GPU Images (ROCm):
2025-04-14 08:48:10 +00:00
```bash
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-gpu-hipblas
```
### Intel GPU Images (oneAPI):
```bash
2025-07-31 17:44:46 +00:00
docker run -ti --name local-ai -p 8080:8080 --device=/dev/dri/card1 --device=/dev/dri/renderD128 localai/localai:latest-gpu-intel
2025-04-14 08:48:10 +00:00
```
2025-05-14 17:28:30 +00:00
### Vulkan GPU Images:
2025-04-14 08:48:10 +00:00
```bash
2025-05-14 17:28:30 +00:00
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-gpu-vulkan
```
### AIO Images (pre-downloaded models):
```bash
# CPU version
2024-11-21 08:56:05 +00:00
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-cpu
2025-05-14 17:28:30 +00:00
2025-12-25 09:00:07 +00:00
# NVIDIA CUDA 13 version
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-aio-gpu-nvidia-cuda-13
2025-05-14 17:28:30 +00:00
# NVIDIA CUDA 12 version
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-aio-gpu-nvidia-cuda-12
# NVIDIA CUDA 11 version
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-aio-gpu-nvidia-cuda-11
# Intel GPU version
2025-07-31 17:44:46 +00:00
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-gpu-intel
2025-05-14 17:28:30 +00:00
# AMD GPU version
docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-aio-gpu-hipblas
2024-05-19 14:37:10 +00:00
```
2025-05-14 17:28:30 +00:00
For more information about the AIO images and pre-downloaded models, see [Container Documentation ](https://localai.io/basics/container/ ).
2024-10-17 07:46:26 +00:00
To load models:
```bash
2024-10-17 15:49:03 +00:00
# From the model gallery (see available models with `local-ai models list`, in the WebUI from the model tab, or visiting https://models.localai.io)
local-ai run llama-3.2-1b-instruct:q4_k_m
# Start LocalAI with the phi-2 model directly from huggingface
2024-10-17 07:46:26 +00:00
local-ai run huggingface://TheBloke/phi-2-GGUF/phi-2.Q8_0.gguf
# Install and run a model from the Ollama OCI registry
local-ai run ollama://gemma:2b
# Run a model from a configuration file
local-ai run https://gist.githubusercontent.com/.../phi-2.yaml
# Install and run a model from a standard OCI registry (e.g., Docker Hub)
local-ai run oci://localai/phi-2:latest
```
2025-07-26 06:18:31 +00:00
> ⚡ **Automatic Backend Detection**: When you install models from the gallery or YAML files, LocalAI automatically detects your system's GPU capabilities (NVIDIA, AMD, Intel) and downloads the appropriate backend. For advanced configuration options, see [GPU Acceleration](https://localai.io/features/gpu-acceleration/#automatic-backend-detection).
2025-11-07 15:50:47 +00:00
For more information, see [💻 Getting started ](https://localai.io/basics/getting_started/index.html ), if you are interested in our roadmap items and future enhancements, you can see the [Issues labeled as Roadmap here ](https://github.com/mudler/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3Aroadmap )
2024-05-19 14:37:10 +00:00
2024-10-04 17:52:43 +00:00
## 📰 Latest project news
2023-12-01 19:05:58 +00:00
2025-12-16 08:16:42 +00:00
- December 2025: [Dynamic Memory Resource reclaimer ](https://github.com/mudler/LocalAI/pull/7583 ), [Automatic fitting of models to multiple GPUS(llama.cpp) ](https://github.com/mudler/LocalAI/pull/7584 ), [Added Vibevoice backend ](https://github.com/mudler/LocalAI/pull/7494 )
2025-11-25 18:31:05 +00:00
- November 2025: Major improvements to the UX. Among these: [Import models via URL ](https://github.com/mudler/LocalAI/pull/7245 ) and [Multiple chats and history ](https://github.com/mudler/LocalAI/pull/7325 )
2025-10-05 16:50:23 +00:00
- October 2025: 🔌 [Model Context Protocol (MCP) ](https://localai.io/docs/features/mcp/ ) support added for agentic capabilities with external tools
2025-10-05 15:51:41 +00:00
- September 2025: New Launcher application for MacOS and Linux, extended support to many backends for Mac and Nvidia L4T devices. Models: Added MLX-Audio, WAN 2.2. WebUI improvements and Python-based backends now ships portable python environments.
2025-08-24 09:50:20 +00:00
- August 2025: MLX, MLX-VLM, Diffusers and llama.cpp are now supported on Mac M1/M2/M3+ chips ( with `development` suffix in the gallery ): https://github.com/mudler/LocalAI/pull/6049 https://github.com/mudler/LocalAI/pull/6119 https://github.com/mudler/LocalAI/pull/6121 https://github.com/mudler/LocalAI/pull/6060
2025-07-27 20:02:51 +00:00
- July/August 2025: 🔍 [Object Detection ](https://localai.io/features/object-detection/ ) added to the API featuring [rf-detr ](https://github.com/roboflow/rf-detr )
2025-07-25 09:51:23 +00:00
- July 2025: All backends migrated outside of the main binary. LocalAI is now more lightweight, small, and automatically downloads the required backend to run the model. [Read the release notes ](https://github.com/mudler/LocalAI/releases/tag/v3.2.0 )
2025-06-15 14:06:43 +00:00
- June 2025: [Backend management ](https://github.com/mudler/LocalAI/pull/5607 ) has been added. Attention: extras images are going to be deprecated from the next release! Read [the backend management PR ](https://github.com/mudler/LocalAI/pull/5607 ).
2025-05-26 18:00:31 +00:00
- May 2025: [Audio input ](https://github.com/mudler/LocalAI/pull/5466 ) and [Reranking ](https://github.com/mudler/LocalAI/pull/5396 ) in llama.cpp backend, [Realtime API ](https://github.com/mudler/LocalAI/pull/5392 ), Support to Gemma, SmollVLM, and more multimodal models (available in the gallery).
- May 2025: Important: image name changes [See release ](https://github.com/mudler/LocalAI/releases/tag/v2.29.0 )
- Apr 2025: Rebrand, WebUI enhancements
2025-04-15 15:51:24 +00:00
- Apr 2025: [LocalAGI ](https://github.com/mudler/LocalAGI ) and [LocalRecall ](https://github.com/mudler/LocalRecall ) join the LocalAI family stack.
2025-03-31 19:51:09 +00:00
- Apr 2025: WebUI overhaul, AIO images updates
- Feb 2025: Backend cleanup, Breaking changes, new backends (kokoro, OutelTTS, faster-whisper), Nvidia L4T images
2025-01-17 14:11:10 +00:00
- Jan 2025: LocalAI model release: https://huggingface.co/mudler/LocalAI-functioncall-phi-4-v0.3, SANA support in diffusers: https://github.com/mudler/LocalAI/pull/4603
2024-12-04 10:31:08 +00:00
- Dec 2024: stablediffusion.cpp backend (ggml) added ( https://github.com/mudler/LocalAI/pull/4289 )
- Nov 2024: Bark.cpp backend added ( https://github.com/mudler/LocalAI/pull/4287 )
2024-11-21 10:05:00 +00:00
- Nov 2024: Voice activity detection models (**VAD**) added to the API: https://github.com/mudler/LocalAI/pull/4204
2024-10-30 08:10:33 +00:00
- Oct 2024: examples moved to [LocalAI-examples ](https://github.com/mudler/LocalAI-examples )
2024-08-28 12:48:16 +00:00
- Aug 2024: 🆕 FLUX-1, [P2P Explorer ](https://explorer.localai.io )
2025-01-15 14:46:27 +00:00
- July 2024: 🔥🔥 🆕 P2P Dashboard, LocalAI Federated mode and AI Swarms: https://github.com/mudler/LocalAI/pull/2723. P2P Global community pools: https://github.com/mudler/LocalAI/issues/3113
2024-07-17 13:14:22 +00:00
- May 2024: 🔥🔥 Decentralized P2P llama.cpp: https://github.com/mudler/LocalAI/pull/2343 (peer2peer llama.cpp!) 👉 Docs https://localai.io/features/distribute/
- May 2024: 🔥🔥 Distributed inferencing: https://github.com/mudler/LocalAI/pull/2324
- April 2024: Reranker API: https://github.com/mudler/LocalAI/pull/2121
2023-12-01 19:05:58 +00:00
2024-10-04 17:52:43 +00:00
Roadmap items: [List of issues ](https://github.com/mudler/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3Aroadmap )
2023-08-06 22:31:46 +00:00
## 🚀 [Features](https://localai.io/features/)
2023-05-19 17:33:53 +00:00
2025-06-28 07:16:25 +00:00
- 🧩 [Backend Gallery ](https://localai.io/backends/ ): Install/remove backends on the fly, powered by OCI images — fully customizable and API-driven.
2024-12-27 14:13:06 +00:00
- 📖 [Text generation with GPTs ](https://localai.io/features/text-generation/ ) (`llama.cpp`, `transformers` , `vllm` ... [:book: and more ](https://localai.io/model-compatibility/index.html#model-compatibility-table ))
2023-08-06 22:01:01 +00:00
- 🗣 [Text to Audio ](https://localai.io/features/text-to-audio/ )
- 🔈 [Audio to Text ](https://localai.io/features/audio-to-text/ ) (Audio transcription with `whisper.cpp` )
2024-12-27 14:13:06 +00:00
- 🎨 [Image generation ](https://localai.io/features/image-generation )
2024-05-25 14:13:04 +00:00
- 🔥 [OpenAI-alike tools API ](https://localai.io/features/openai-functions/ )
2023-08-06 22:01:01 +00:00
- 🧠 [Embeddings generation for vector databases ](https://localai.io/features/embeddings/ )
- ✍️ [Constrained grammars ](https://localai.io/features/constrained_grammars/ )
- 🖼️ [Download Models directly from Huggingface ](https://localai.io/models/ )
2024-04-28 13:54:15 +00:00
- 🥽 [Vision API ](https://localai.io/features/gpt-vision/ )
2025-07-27 20:02:51 +00:00
- 🔍 [Object Detection ](https://localai.io/features/object-detection/ )
2024-05-25 14:11:59 +00:00
- 📈 [Reranker API ](https://localai.io/features/reranker/ )
- 🆕🖧 [P2P Inferencing ](https://localai.io/features/distribute/ )
2025-10-27 20:44:41 +00:00
- 🆕🔌 [Model Context Protocol (MCP) ](https://localai.io/docs/features/mcp/ ) - Agentic capabilities with external tools and [LocalAGI's Agentic capabilities ](https://github.com/mudler/LocalAGI )
2024-12-27 14:17:02 +00:00
- 🔊 Voice activity detection (Silero-VAD support)
2024-07-07 21:01:01 +00:00
- 🌍 Integrated WebUI!
2023-07-02 22:52:26 +00:00
2025-08-24 18:09:19 +00:00
## 🧩 Supported Backends & Acceleration
LocalAI supports a comprehensive range of AI backends with multiple acceleration options:
### Text Generation & Language Models
| Backend | Description | Acceleration Support |
|---------|-------------|---------------------|
2025-12-25 09:00:07 +00:00
| **llama.cpp** | LLM inference in C/C++ | CUDA 11/12/13, ROCm, Intel SYCL, Vulkan, Metal, CPU |
| **vLLM** | Fast LLM inference with PagedAttention | CUDA 12/13, ROCm, Intel |
| **transformers** | HuggingFace transformers framework | CUDA 11/12/13, ROCm, Intel, CPU |
| **exllama2** | GPTQ inference library | CUDA 12/13 |
2025-08-24 18:09:19 +00:00
| **MLX** | Apple Silicon LLM inference | Metal (M1/M2/M3+) |
| **MLX-VLM** | Apple Silicon Vision-Language Models | Metal (M1/M2/M3+) |
### Audio & Speech Processing
| Backend | Description | Acceleration Support |
|---------|-------------|---------------------|
2025-12-25 09:00:07 +00:00
| **whisper.cpp** | OpenAI Whisper in C/C++ | CUDA 12/13, ROCm, Intel SYCL, Vulkan, CPU |
| **faster-whisper** | Fast Whisper with CTranslate2 | CUDA 12/13, ROCm, Intel, CPU |
| **bark** | Text-to-audio generation | CUDA 12/13, ROCm, Intel |
2025-08-24 18:09:19 +00:00
| **bark-cpp** | C++ implementation of Bark | CUDA, Metal, CPU |
2025-12-25 09:00:07 +00:00
| **coqui** | Advanced TTS with 1100+ languages | CUDA 12/13, ROCm, Intel, CPU |
| **kokoro** | Lightweight TTS model | CUDA 12/13, ROCm, Intel, CPU |
| **chatterbox** | Production-grade TTS | CUDA 11/12/13, CPU |
2025-08-24 18:09:19 +00:00
| **piper** | Fast neural TTS system | CPU |
| **kitten-tts** | Kitten TTS models | CPU |
| **silero-vad** | Voice Activity Detection | CPU |
2025-12-25 09:00:07 +00:00
| **neutts** | Text-to-speech with voice cloning | CUDA 12/13, ROCm, CPU |
| **vibevoice** | Real-time TTS with voice cloning | CUDA 12/13, ROCm, Intel, CPU |
2025-08-24 18:09:19 +00:00
### Image & Video Generation
| Backend | Description | Acceleration Support |
|---------|-------------|---------------------|
2025-12-25 09:00:07 +00:00
| **stablediffusion.cpp** | Stable Diffusion in C/C++ | CUDA 12/13, Intel SYCL, Vulkan, CPU |
| **diffusers** | HuggingFace diffusion models | CUDA 11/12/13, ROCm, Intel, Metal, CPU |
2025-08-24 18:09:19 +00:00
### Specialized AI Tasks
| Backend | Description | Acceleration Support |
|---------|-------------|---------------------|
2025-12-25 09:00:07 +00:00
| **rfdetr** | Real-time object detection | CUDA 12/13, Intel, CPU |
| **rerankers** | Document reranking API | CUDA 11/12/13, ROCm, Intel, CPU |
2025-08-24 18:09:19 +00:00
| **local-store** | Vector database | CPU |
| **huggingface** | HuggingFace API integration | API-based |
### Hardware Acceleration Matrix
| Acceleration Type | Supported Backends | Hardware Support |
|-------------------|-------------------|------------------|
| **NVIDIA CUDA 11** | llama.cpp, whisper, stablediffusion, diffusers, rerankers, bark, chatterbox | Nvidia hardware |
| **NVIDIA CUDA 12** | All CUDA-compatible backends | Nvidia hardware |
2025-12-25 09:00:07 +00:00
| **NVIDIA CUDA 13** | All CUDA-compatible backends | Nvidia hardware |
| **AMD ROCm** | llama.cpp, whisper, vllm, transformers, diffusers, rerankers, coqui, kokoro, bark, neutts, vibevoice | AMD Graphics |
| **Intel oneAPI** | llama.cpp, whisper, stablediffusion, vllm, transformers, diffusers, rfdetr, rerankers, exllama2, coqui, kokoro, bark, vibevoice | Intel Arc, Intel iGPUs |
2025-08-24 18:09:19 +00:00
| **Apple Metal** | llama.cpp, whisper, diffusers, MLX, MLX-VLM, bark-cpp | Apple M1/M2/M3+ |
| **Vulkan** | llama.cpp, whisper, stablediffusion | Cross-platform GPUs |
2025-12-25 09:00:07 +00:00
| **NVIDIA Jetson (CUDA 12)** | llama.cpp, whisper, stablediffusion, diffusers, rfdetr | ARM64 embedded AI (AGX Orin, etc.) |
| **NVIDIA Jetson (CUDA 13)** | llama.cpp, whisper, stablediffusion, diffusers, rfdetr | ARM64 embedded AI (DGX Spark) |
2025-08-24 18:09:19 +00:00
| **CPU Optimized** | All backends | AVX/AVX2/AVX512, quantization support |
2023-04-12 23:13:14 +00:00
2023-12-01 18:11:45 +00:00
### 🔗 Community and integrations
2023-04-24 21:42:03 +00:00
2023-12-12 17:58:57 +00:00
Build and deploy custom containers:
- https://github.com/sozercan/aikit
2023-12-01 18:11:45 +00:00
WebUIs:
2023-11-28 22:14:16 +00:00
- https://github.com/Jirubizu/localai-admin
- https://github.com/go-skynet/LocalAI-frontend
2024-06-10 17:18:47 +00:00
- QA-Pilot(An interactive chat project that leverages LocalAI LLMs for rapid understanding and navigation of GitHub code repository) https://github.com/reid41/QA-Pilot
2023-11-28 22:14:16 +00:00
2025-10-05 17:38:47 +00:00
Agentic Libraries:
- https://github.com/mudler/cogito
MCPs:
- https://github.com/mudler/MCPs
2023-11-28 22:14:16 +00:00
Model galleries
- https://github.com/go-skynet/model-gallery
2025-08-27 14:24:23 +00:00
Voice:
- https://github.com/richiejp/VoxInput
2023-11-28 22:14:16 +00:00
Other:
- Helm chart https://github.com/go-skynet/helm-charts
2023-12-28 22:03:10 +00:00
- VSCode extension https://github.com/badgooooor/localai-vscode-plugin
2025-01-10 08:20:33 +00:00
- Langchain: https://python.langchain.com/docs/integrations/providers/localai/
2024-05-07 06:39:58 +00:00
- Terminal utility https://github.com/djcopley/ShellOracle
2023-12-25 08:04:35 +00:00
- Local Smart assistant https://github.com/mudler/LocalAGI
2024-05-23 13:21:01 +00:00
- Home Assistant https://github.com/sammcj/homeassistant-localai / https://github.com/drndos/hass-openai-custom-conversation / https://github.com/valentinfrlch/ha-gpt4vision
2023-12-28 22:01:52 +00:00
- Discord bot https://github.com/mudler/LocalAGI/tree/main/examples/discord
- Slack bot https://github.com/mudler/LocalAGI/tree/main/examples/slack
2024-06-10 17:18:47 +00:00
- Shell-Pilot(Interact with LLM using LocalAI models via pure shell scripts on your Linux or MacOS system) https://github.com/reid41/shell-pilot
2023-12-28 22:01:52 +00:00
- Telegram bot https://github.com/mudler/LocalAI/tree/master/examples/telegram-bot
2024-11-22 10:32:50 +00:00
- Another Telegram Bot https://github.com/JackBekket/Hellper
- Auto-documentation https://github.com/JackBekket/Reflexia
- Github bot which answer on issues, with code and documentation as context https://github.com/JackBekket/GitHelper
2024-07-17 03:57:01 +00:00
- Github Actions: https://github.com/marketplace/actions/start-localai
2023-12-28 22:01:52 +00:00
- Examples: https://github.com/mudler/LocalAI/tree/master/examples/
2024-03-05 09:14:30 +00:00
2023-05-03 13:51:54 +00:00
2023-08-06 22:01:01 +00:00
### 🔗 Resources
2023-05-19 17:33:53 +00:00
2024-05-31 20:59:51 +00:00
- [LLM finetuning guide ](https://localai.io/docs/advanced/fine-tuning/ )
2023-08-06 22:01:01 +00:00
- [How to build locally ](https://localai.io/basics/build/index.html )
- [How to install in Kubernetes ](https://localai.io/basics/getting_started/index.html#run-localai-in-kubernetes )
2024-02-03 23:54:49 +00:00
- [Projects integrating LocalAI ](https://localai.io/docs/integrations/ )
2024-01-11 08:25:18 +00:00
- [How tos section ](https://io.midori-ai.xyz/howtos/ ) (curated by our community)
2023-12-12 17:58:57 +00:00
2023-12-01 18:11:45 +00:00
## :book: 🎥 [Media, Blogs, Social](https://localai.io/basics/news/#media-blogs-social)
2024-08-01 16:44:39 +00:00
- [Run Visual studio code with LocalAI (SUSE) ](https://www.suse.com/c/running-ai-locally/ )
2024-05-31 20:59:51 +00:00
- 🆕 [Run LocalAI on Jetson Nano Devkit ](https://mudler.pm/posts/local-ai-jetson-nano-devkit/ )
2024-05-05 13:45:55 +00:00
- [Run LocalAI on AWS EKS with Pulumi ](https://www.pulumi.com/blog/low-code-llm-apps-with-local-ai-flowise-and-pulumi/ )
2024-02-27 14:43:15 +00:00
- [Run LocalAI on AWS ](https://staleks.hashnode.dev/installing-localai-on-aws-ec2-instance )
2023-12-01 18:11:45 +00:00
- [Create a slackbot for teams and OSS projects that answer to documentation ](https://mudler.pm/posts/smart-slackbot-for-teams/ )
- [LocalAI meets k8sgpt ](https://www.youtube.com/watch?v=PKrDNuJ_dfE )
- [Question Answering on Documents locally with LangChain, LocalAI, Chroma, and GPT4All ](https://mudler.pm/posts/localai-question-answering/ )
- [Tutorial to use k8sgpt with LocalAI ](https://medium.com/@tyler_97636/k8sgpt-localai-unlock-kubernetes-superpowers-for-free-584790de9b65 )
2023-09-09 17:17:19 +00:00
## Citation
If you utilize this repository, data in a downstream project, please consider citing it with:
2023-09-10 07:21:47 +00:00
```
2023-09-09 17:17:19 +00:00
@misc {localai,
author = {Ettore Di Giacinto},
title = {LocalAI: The free, Open source OpenAI alternative},
year = {2023},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/go-skynet/LocalAI}},
2023-09-10 07:21:47 +00:00
```
2023-09-09 17:17:19 +00:00
2023-08-06 22:01:01 +00:00
## ❤️ Sponsors
2023-07-09 12:14:54 +00:00
> Do you find LocalAI useful?
Support the project by becoming [a backer or sponsor ](https://github.com/sponsors/mudler ). Your logo will show up here with a link to your website.
2024-06-04 13:23:00 +00:00
A huge thank you to our generous sponsors who support this project covering CI expenses, and our [Sponsor list ](https://github.com/sponsors/mudler ):
2023-07-09 12:14:54 +00:00
2024-06-04 13:23:00 +00:00
< p align = "center" >
< a href = "https://www.spectrocloud.com/" target = "blank" >
2025-02-20 14:31:41 +00:00
< img height = "200" src = "https://github.com/user-attachments/assets/72eab1dd-8b93-4fc0-9ade-84db49f24962" >
2024-06-04 13:23:00 +00:00
< / a >
< a href = "https://www.premai.io/" target = "blank" >
< img height = "200" src = "https://github.com/mudler/LocalAI/assets/2420543/42e4ca83-661e-4f79-8e46-ae43689683d6" > < br >
< / a >
< / p >
2023-09-16 21:00:42 +00:00
2023-08-06 22:01:01 +00:00
## 🌟 Star history
2023-04-27 08:39:01 +00:00
2023-04-27 04:18:18 +00:00
[](https://star-history.com/#go-skynet/LocalAI& Date)
2023-08-06 22:01:01 +00:00
## 📖 License
2023-03-30 16:46:11 +00:00
2023-05-29 21:09:19 +00:00
LocalAI is a community-driven project created by [Ettore Di Giacinto ](https://github.com/mudler/ ).
2023-04-27 08:39:01 +00:00
2024-06-03 17:55:01 +00:00
MIT - Author Ettore Di Giacinto < mudler @ localai . io >
2023-05-03 09:45:22 +00:00
2023-08-06 22:01:01 +00:00
## 🙇 Acknowledgements
2023-03-30 16:46:11 +00:00
2023-05-16 17:32:53 +00:00
LocalAI couldn't have been built without the help of great software already available from the community. Thank you!
2023-03-30 16:46:11 +00:00
- [llama.cpp ](https://github.com/ggerganov/llama.cpp )
- https://github.com/tatsu-lab/stanford_alpaca
- https://github.com/cornelk/llama-go for the initial ideas
2023-05-16 17:32:53 +00:00
- https://github.com/antimatter15/alpaca.cpp
- https://github.com/EdVince/Stable-Diffusion-NCNN
- https://github.com/ggerganov/whisper.cpp
2023-08-06 22:01:01 +00:00
- https://github.com/rhasspy/piper
2023-04-28 08:54:39 +00:00
2023-08-06 22:01:01 +00:00
## 🤗 Contributors
2023-04-28 08:54:39 +00:00
2023-08-06 22:01:01 +00:00
This is a community project, a special thanks to our contributors! 🤗
2023-04-28 08:54:39 +00:00
< a href = "https://github.com/go-skynet/LocalAI/graphs/contributors" >
< img src = "https://contrib.rocks/image?repo=go-skynet/LocalAI" / >
< / a >