mirror of
https://github.com/mudler/LocalAI
synced 2026-04-21 13:27:21 +00:00
* feat: split remaining backends and drop embedded backends - Drop silero-vad, huggingface, and stores backend from embedded binaries - Refactor Makefile and Dockerfile to avoid building grpc backends - Drop golang code that was used to embed backends - Simplify building by using goreleaser Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * chore(gallery): be specific with llama-cpp backend templates Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * chore(docs): update Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * chore(ci): minor fixes Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * chore: drop all ffmpeg references Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix: run protogen-go Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Always enable p2p mode Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Update gorelease file Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix(stores): do not always load Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Fix linting issues Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Simplify Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Mac OS fixup Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
258 lines
8 KiB
Go
258 lines
8 KiB
Go
package routes
|
|
|
|
import (
|
|
"github.com/mudler/LocalAI/core/config"
|
|
"github.com/mudler/LocalAI/core/gallery"
|
|
"github.com/mudler/LocalAI/core/http/elements"
|
|
"github.com/mudler/LocalAI/core/http/endpoints/localai"
|
|
"github.com/mudler/LocalAI/core/http/utils"
|
|
"github.com/mudler/LocalAI/core/p2p"
|
|
"github.com/mudler/LocalAI/core/services"
|
|
"github.com/mudler/LocalAI/internal"
|
|
"github.com/mudler/LocalAI/pkg/model"
|
|
|
|
"github.com/gofiber/fiber/v2"
|
|
)
|
|
|
|
func RegisterUIRoutes(app *fiber.App,
|
|
cl *config.BackendConfigLoader,
|
|
ml *model.ModelLoader,
|
|
appConfig *config.ApplicationConfig,
|
|
galleryService *services.GalleryService) {
|
|
|
|
// keeps the state of ops that are started from the UI
|
|
var processingOps = services.NewOpCache(galleryService)
|
|
|
|
app.Get("/", localai.WelcomeEndpoint(appConfig, cl, ml, processingOps))
|
|
|
|
// P2P
|
|
app.Get("/p2p", func(c *fiber.Ctx) error {
|
|
summary := fiber.Map{
|
|
"Title": "LocalAI - P2P dashboard",
|
|
"BaseURL": utils.BaseURL(c),
|
|
"Version": internal.PrintableVersion(),
|
|
//"Nodes": p2p.GetAvailableNodes(""),
|
|
//"FederatedNodes": p2p.GetAvailableNodes(p2p.FederatedID),
|
|
|
|
"P2PToken": appConfig.P2PToken,
|
|
"NetworkID": appConfig.P2PNetworkID,
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/p2p", summary)
|
|
})
|
|
|
|
/* show nodes live! */
|
|
app.Get("/p2p/ui/workers", func(c *fiber.Ctx) error {
|
|
return c.SendString(elements.P2PNodeBoxes(p2p.GetAvailableNodes(p2p.NetworkID(appConfig.P2PNetworkID, p2p.WorkerID))))
|
|
})
|
|
app.Get("/p2p/ui/workers-federation", func(c *fiber.Ctx) error {
|
|
return c.SendString(elements.P2PNodeBoxes(p2p.GetAvailableNodes(p2p.NetworkID(appConfig.P2PNetworkID, p2p.FederatedID))))
|
|
})
|
|
|
|
app.Get("/p2p/ui/workers-stats", func(c *fiber.Ctx) error {
|
|
return c.SendString(elements.P2PNodeStats(p2p.GetAvailableNodes(p2p.NetworkID(appConfig.P2PNetworkID, p2p.WorkerID))))
|
|
})
|
|
app.Get("/p2p/ui/workers-federation-stats", func(c *fiber.Ctx) error {
|
|
return c.SendString(elements.P2PNodeStats(p2p.GetAvailableNodes(p2p.NetworkID(appConfig.P2PNetworkID, p2p.FederatedID))))
|
|
})
|
|
|
|
// End P2P
|
|
|
|
if !appConfig.DisableGalleryEndpoint {
|
|
registerGalleryRoutes(app, cl, appConfig, galleryService, processingOps)
|
|
registerBackendGalleryRoutes(app, appConfig, galleryService, processingOps)
|
|
}
|
|
|
|
app.Get("/talk/", func(c *fiber.Ctx) error {
|
|
backendConfigs, _ := services.ListModels(cl, ml, config.NoFilterFn, services.SKIP_IF_CONFIGURED)
|
|
|
|
if len(backendConfigs) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(utils.BaseURL(c))
|
|
}
|
|
|
|
summary := fiber.Map{
|
|
"Title": "LocalAI - Talk",
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"Model": backendConfigs[0],
|
|
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/talk", summary)
|
|
})
|
|
|
|
app.Get("/chat/", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
if len(backendConfigs)+len(modelsWithoutConfig) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(utils.BaseURL(c))
|
|
}
|
|
modelThatCanBeUsed := ""
|
|
galleryConfigs := map[string]*gallery.ModelConfig{}
|
|
|
|
for _, m := range backendConfigs {
|
|
cfg, err := gallery.GetLocalModelConfiguration(ml.ModelPath, m.Name)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
galleryConfigs[m.Name] = cfg
|
|
}
|
|
|
|
title := "LocalAI - Chat"
|
|
|
|
for _, b := range backendConfigs {
|
|
if b.HasUsecases(config.FLAG_CHAT) {
|
|
modelThatCanBeUsed = b.Name
|
|
title = "LocalAI - Chat with " + modelThatCanBeUsed
|
|
break
|
|
}
|
|
}
|
|
|
|
summary := fiber.Map{
|
|
"Title": title,
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"GalleryConfig": galleryConfigs,
|
|
"ModelsConfig": backendConfigs,
|
|
"Model": modelThatCanBeUsed,
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/chat", summary)
|
|
})
|
|
|
|
// Show the Chat page
|
|
app.Get("/chat/:model", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
galleryConfigs := map[string]*gallery.ModelConfig{}
|
|
|
|
for _, m := range backendConfigs {
|
|
cfg, err := gallery.GetLocalModelConfiguration(ml.ModelPath, m.Name)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
galleryConfigs[m.Name] = cfg
|
|
}
|
|
|
|
summary := fiber.Map{
|
|
"Title": "LocalAI - Chat with " + c.Params("model"),
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"GalleryConfig": galleryConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": c.Params("model"),
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/chat", summary)
|
|
})
|
|
|
|
app.Get("/text2image/:model", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
summary := fiber.Map{
|
|
"Title": "LocalAI - Generate images with " + c.Params("model"),
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": c.Params("model"),
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/text2image", summary)
|
|
})
|
|
|
|
app.Get("/text2image/", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
if len(backendConfigs)+len(modelsWithoutConfig) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(utils.BaseURL(c))
|
|
}
|
|
|
|
modelThatCanBeUsed := ""
|
|
title := "LocalAI - Generate images"
|
|
|
|
for _, b := range backendConfigs {
|
|
if b.HasUsecases(config.FLAG_IMAGE) {
|
|
modelThatCanBeUsed = b.Name
|
|
title = "LocalAI - Generate images with " + modelThatCanBeUsed
|
|
break
|
|
}
|
|
}
|
|
|
|
summary := fiber.Map{
|
|
"Title": title,
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": modelThatCanBeUsed,
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/text2image", summary)
|
|
})
|
|
|
|
app.Get("/tts/:model", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
summary := fiber.Map{
|
|
"Title": "LocalAI - Generate images with " + c.Params("model"),
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": c.Params("model"),
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/tts", summary)
|
|
})
|
|
|
|
app.Get("/tts/", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
if len(backendConfigs)+len(modelsWithoutConfig) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(utils.BaseURL(c))
|
|
}
|
|
|
|
modelThatCanBeUsed := ""
|
|
title := "LocalAI - Generate audio"
|
|
|
|
for _, b := range backendConfigs {
|
|
if b.HasUsecases(config.FLAG_TTS) {
|
|
modelThatCanBeUsed = b.Name
|
|
title = "LocalAI - Generate audio with " + modelThatCanBeUsed
|
|
break
|
|
}
|
|
}
|
|
summary := fiber.Map{
|
|
"Title": title,
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": modelThatCanBeUsed,
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/tts", summary)
|
|
})
|
|
}
|