mirror of
https://github.com/mudler/LocalAI.git
synced 2025-06-17 00:05:00 +00:00

* feat: Add backend gallery This PR add support to manage backends as similar to models. There is now available a backend gallery which can be used to install and remove extra backends. The backend gallery can be configured similarly as a model gallery, and API calls allows to install and remove new backends in runtime, and as well during the startup phase of LocalAI. Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Add backends docs Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * wip: Backend Dockerfile for python backends Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * feat: drop extras images, build python backends separately Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fixup on all backends Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * test CI Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Tweaks Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Drop old backends leftovers Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Fixup CI Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Move dockerfile upper Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Fix proto Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Feature dropped for consistency - we prefer model galleries Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Add missing packages in the build image Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * exllama is ponly available on cublas Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * pin torch on chatterbox Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Fixups to index Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * CI Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Debug CI * Install accellerators deps Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Add target arch * Add cuda minor version Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Use self-hosted runners Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * ci: use quay for test images Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fixups for vllm and chatterbox Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Small fixups on CI Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * chatterbox is only available for nvidia Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Simplify CI builds Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Adapt test, use qwen3 Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * chore(model gallery): add jina-reranker-v1-tiny-en-gguf Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix(gguf-parser): recover from potential panics that can happen while reading ggufs with gguf-parser Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Use reranker from llama.cpp in AIO images Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Limit concurrent jobs Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io> Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
263 lines
8.4 KiB
Go
263 lines
8.4 KiB
Go
package routes
|
|
|
|
import (
|
|
"github.com/mudler/LocalAI/core/config"
|
|
"github.com/mudler/LocalAI/core/gallery"
|
|
"github.com/mudler/LocalAI/core/http/elements"
|
|
"github.com/mudler/LocalAI/core/http/endpoints/localai"
|
|
"github.com/mudler/LocalAI/core/http/utils"
|
|
"github.com/mudler/LocalAI/core/p2p"
|
|
"github.com/mudler/LocalAI/core/services"
|
|
"github.com/mudler/LocalAI/internal"
|
|
"github.com/mudler/LocalAI/pkg/model"
|
|
|
|
"github.com/gofiber/fiber/v2"
|
|
)
|
|
|
|
func RegisterUIRoutes(app *fiber.App,
|
|
cl *config.BackendConfigLoader,
|
|
ml *model.ModelLoader,
|
|
appConfig *config.ApplicationConfig,
|
|
galleryService *services.GalleryService) {
|
|
|
|
// keeps the state of ops that are started from the UI
|
|
var processingOps = services.NewOpCache(galleryService)
|
|
|
|
app.Get("/", localai.WelcomeEndpoint(appConfig, cl, ml, processingOps))
|
|
|
|
if p2p.IsP2PEnabled() {
|
|
app.Get("/p2p", func(c *fiber.Ctx) error {
|
|
summary := fiber.Map{
|
|
"Title": "LocalAI - P2P dashboard",
|
|
"BaseURL": utils.BaseURL(c),
|
|
"Version": internal.PrintableVersion(),
|
|
//"Nodes": p2p.GetAvailableNodes(""),
|
|
//"FederatedNodes": p2p.GetAvailableNodes(p2p.FederatedID),
|
|
"IsP2PEnabled": p2p.IsP2PEnabled(),
|
|
"P2PToken": appConfig.P2PToken,
|
|
"NetworkID": appConfig.P2PNetworkID,
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/p2p", summary)
|
|
})
|
|
|
|
/* show nodes live! */
|
|
app.Get("/p2p/ui/workers", func(c *fiber.Ctx) error {
|
|
return c.SendString(elements.P2PNodeBoxes(p2p.GetAvailableNodes(p2p.NetworkID(appConfig.P2PNetworkID, p2p.WorkerID))))
|
|
})
|
|
app.Get("/p2p/ui/workers-federation", func(c *fiber.Ctx) error {
|
|
return c.SendString(elements.P2PNodeBoxes(p2p.GetAvailableNodes(p2p.NetworkID(appConfig.P2PNetworkID, p2p.FederatedID))))
|
|
})
|
|
|
|
app.Get("/p2p/ui/workers-stats", func(c *fiber.Ctx) error {
|
|
return c.SendString(elements.P2PNodeStats(p2p.GetAvailableNodes(p2p.NetworkID(appConfig.P2PNetworkID, p2p.WorkerID))))
|
|
})
|
|
app.Get("/p2p/ui/workers-federation-stats", func(c *fiber.Ctx) error {
|
|
return c.SendString(elements.P2PNodeStats(p2p.GetAvailableNodes(p2p.NetworkID(appConfig.P2PNetworkID, p2p.FederatedID))))
|
|
})
|
|
}
|
|
|
|
if !appConfig.DisableGalleryEndpoint {
|
|
registerGalleryRoutes(app, cl, appConfig, galleryService, processingOps)
|
|
registerBackendGalleryRoutes(app, appConfig, galleryService, processingOps)
|
|
}
|
|
|
|
app.Get("/talk/", func(c *fiber.Ctx) error {
|
|
backendConfigs, _ := services.ListModels(cl, ml, config.NoFilterFn, services.SKIP_IF_CONFIGURED)
|
|
|
|
if len(backendConfigs) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(utils.BaseURL(c))
|
|
}
|
|
|
|
summary := fiber.Map{
|
|
"Title": "LocalAI - Talk",
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"Model": backendConfigs[0],
|
|
"IsP2PEnabled": p2p.IsP2PEnabled(),
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/talk", summary)
|
|
})
|
|
|
|
app.Get("/chat/", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
if len(backendConfigs)+len(modelsWithoutConfig) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(utils.BaseURL(c))
|
|
}
|
|
modelThatCanBeUsed := ""
|
|
galleryConfigs := map[string]*gallery.ModelConfig{}
|
|
|
|
for _, m := range backendConfigs {
|
|
cfg, err := gallery.GetLocalModelConfiguration(ml.ModelPath, m.Name)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
galleryConfigs[m.Name] = cfg
|
|
}
|
|
|
|
title := "LocalAI - Chat"
|
|
|
|
for _, b := range backendConfigs {
|
|
if b.HasUsecases(config.FLAG_CHAT) {
|
|
modelThatCanBeUsed = b.Name
|
|
title = "LocalAI - Chat with " + modelThatCanBeUsed
|
|
break
|
|
}
|
|
}
|
|
|
|
summary := fiber.Map{
|
|
"Title": title,
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"GalleryConfig": galleryConfigs,
|
|
"ModelsConfig": backendConfigs,
|
|
"Model": modelThatCanBeUsed,
|
|
"Version": internal.PrintableVersion(),
|
|
"IsP2PEnabled": p2p.IsP2PEnabled(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/chat", summary)
|
|
})
|
|
|
|
// Show the Chat page
|
|
app.Get("/chat/:model", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
galleryConfigs := map[string]*gallery.ModelConfig{}
|
|
|
|
for _, m := range backendConfigs {
|
|
cfg, err := gallery.GetLocalModelConfiguration(ml.ModelPath, m.Name)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
galleryConfigs[m.Name] = cfg
|
|
}
|
|
|
|
summary := fiber.Map{
|
|
"Title": "LocalAI - Chat with " + c.Params("model"),
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"GalleryConfig": galleryConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": c.Params("model"),
|
|
"Version": internal.PrintableVersion(),
|
|
"IsP2PEnabled": p2p.IsP2PEnabled(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/chat", summary)
|
|
})
|
|
|
|
app.Get("/text2image/:model", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
summary := fiber.Map{
|
|
"Title": "LocalAI - Generate images with " + c.Params("model"),
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": c.Params("model"),
|
|
"Version": internal.PrintableVersion(),
|
|
"IsP2PEnabled": p2p.IsP2PEnabled(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/text2image", summary)
|
|
})
|
|
|
|
app.Get("/text2image/", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
if len(backendConfigs)+len(modelsWithoutConfig) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(utils.BaseURL(c))
|
|
}
|
|
|
|
modelThatCanBeUsed := ""
|
|
title := "LocalAI - Generate images"
|
|
|
|
for _, b := range backendConfigs {
|
|
if b.HasUsecases(config.FLAG_IMAGE) {
|
|
modelThatCanBeUsed = b.Name
|
|
title = "LocalAI - Generate images with " + modelThatCanBeUsed
|
|
break
|
|
}
|
|
}
|
|
|
|
summary := fiber.Map{
|
|
"Title": title,
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": modelThatCanBeUsed,
|
|
"Version": internal.PrintableVersion(),
|
|
"IsP2PEnabled": p2p.IsP2PEnabled(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/text2image", summary)
|
|
})
|
|
|
|
app.Get("/tts/:model", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
summary := fiber.Map{
|
|
"Title": "LocalAI - Generate images with " + c.Params("model"),
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": c.Params("model"),
|
|
"Version": internal.PrintableVersion(),
|
|
"IsP2PEnabled": p2p.IsP2PEnabled(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/tts", summary)
|
|
})
|
|
|
|
app.Get("/tts/", func(c *fiber.Ctx) error {
|
|
backendConfigs := cl.GetAllBackendConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
if len(backendConfigs)+len(modelsWithoutConfig) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(utils.BaseURL(c))
|
|
}
|
|
|
|
modelThatCanBeUsed := ""
|
|
title := "LocalAI - Generate audio"
|
|
|
|
for _, b := range backendConfigs {
|
|
if b.HasUsecases(config.FLAG_TTS) {
|
|
modelThatCanBeUsed = b.Name
|
|
title = "LocalAI - Generate audio with " + modelThatCanBeUsed
|
|
break
|
|
}
|
|
}
|
|
summary := fiber.Map{
|
|
"Title": title,
|
|
"BaseURL": utils.BaseURL(c),
|
|
"ModelsConfig": backendConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": modelThatCanBeUsed,
|
|
"IsP2PEnabled": p2p.IsP2PEnabled(),
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render("views/tts", summary)
|
|
})
|
|
}
|