mirror of
https://github.com/mudler/LocalAI.git
synced 2025-05-20 10:35:01 +00:00

Some checks are pending
Explorer deployment / build-linux (push) Waiting to run
GPU tests / ubuntu-latest (1.21.x) (push) Waiting to run
generate and publish intel docker caches / generate_caches (intel/oneapi-basekit:2025.1.0-0-devel-ubuntu22.04, linux/amd64, ubuntu-latest) (push) Waiting to run
build container images / hipblas-jobs (-aio-gpu-hipblas, rocm/dev-ubuntu-22.04:6.1, hipblas, true, ubuntu:22.04, extras, latest-gpu-hipblas-extras, latest-aio-gpu-hipblas, --jobs=3 --output-sync=target, linux/amd64, arc-runner-set, auto, -hipblas-extras) (push) Waiting to run
build container images / hipblas-jobs (rocm/dev-ubuntu-22.04:6.1, hipblas, true, ubuntu:22.04, core, latest-gpu-hipblas, --jobs=3 --output-sync=target, linux/amd64, arc-runner-set, false, -hipblas) (push) Waiting to run
build container images / self-hosted-jobs (-aio-gpu-intel-f16, quay.io/go-skynet/intel-oneapi-base:latest, sycl_f16, true, ubuntu:22.04, extras, latest-gpu-intel-f16-extras, latest-aio-gpu-intel-f16, --jobs=3 --output-sync=target, linux/amd64, arc-runner-set, false, -sycl-f16-… (push) Waiting to run
build container images / self-hosted-jobs (-aio-gpu-intel-f32, quay.io/go-skynet/intel-oneapi-base:latest, sycl_f32, true, ubuntu:22.04, extras, latest-gpu-intel-f32-extras, latest-aio-gpu-intel-f32, --jobs=3 --output-sync=target, linux/amd64, arc-runner-set, false, -sycl-f32-… (push) Waiting to run
build container images / self-hosted-jobs (-aio-gpu-nvidia-cuda-11, ubuntu:22.04, cublas, 11, 7, true, extras, latest-gpu-nvidia-cuda-11-extras, latest-aio-gpu-nvidia-cuda-11, --jobs=3 --output-sync=target, linux/amd64, arc-runner-set, false, -cublas-cuda11-extras) (push) Waiting to run
build container images / self-hosted-jobs (-aio-gpu-nvidia-cuda-12, ubuntu:22.04, cublas, 12, 0, true, extras, latest-gpu-nvidia-cuda-12-extras, latest-aio-gpu-nvidia-cuda-12, --jobs=3 --output-sync=target, linux/amd64, arc-runner-set, false, -cublas-cuda12-extras) (push) Waiting to run
build container images / self-hosted-jobs (quay.io/go-skynet/intel-oneapi-base:latest, sycl_f16, true, ubuntu:22.04, core, latest-gpu-intel-f16, --jobs=3 --output-sync=target, linux/amd64, arc-runner-set, false, -sycl-f16) (push) Waiting to run
build container images / self-hosted-jobs (quay.io/go-skynet/intel-oneapi-base:latest, sycl_f32, true, ubuntu:22.04, core, latest-gpu-intel-f32, --jobs=3 --output-sync=target, linux/amd64, arc-runner-set, false, -sycl-f32) (push) Waiting to run
build container images / core-image-build (-aio-cpu, ubuntu:22.04, , true, core, latest-cpu, latest-aio-cpu, --jobs=4 --output-sync=target, linux/amd64,linux/arm64, arc-runner-set, false, auto, ) (push) Waiting to run
build container images / core-image-build (ubuntu:22.04, cublas, 11, 7, true, core, latest-gpu-nvidia-cuda-12, --jobs=4 --output-sync=target, linux/amd64, arc-runner-set, false, false, -cublas-cuda11) (push) Waiting to run
build container images / core-image-build (ubuntu:22.04, cublas, 12, 0, true, core, latest-gpu-nvidia-cuda-12, --jobs=4 --output-sync=target, linux/amd64, arc-runner-set, false, false, -cublas-cuda12) (push) Waiting to run
build container images / core-image-build (ubuntu:22.04, vulkan, true, core, latest-gpu-vulkan, --jobs=4 --output-sync=target, linux/amd64, arc-runner-set, false, false, -vulkan) (push) Waiting to run
build container images / gh-runner (nvcr.io/nvidia/l4t-jetpack:r36.4.0, cublas, 12, 0, true, core, latest-nvidia-l4t-arm64, --jobs=4 --output-sync=target, linux/arm64, ubuntu-24.04-arm, true, false, -nvidia-l4t-arm64) (push) Waiting to run
Security Scan / tests (push) Waiting to run
Tests extras backends / tests-transformers (push) Waiting to run
Tests extras backends / tests-rerankers (push) Waiting to run
Tests extras backends / tests-diffusers (push) Waiting to run
Tests extras backends / tests-coqui (push) Waiting to run
tests / tests-linux (1.21.x) (push) Waiting to run
tests / tests-aio-container (push) Waiting to run
tests / tests-apple (1.21.x) (push) Waiting to run
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
214 lines
6.9 KiB
Go
214 lines
6.9 KiB
Go
package localai
|
|
|
|
import (
|
|
"encoding/json"
|
|
"fmt"
|
|
"slices"
|
|
|
|
"github.com/gofiber/fiber/v2"
|
|
"github.com/google/uuid"
|
|
"github.com/mudler/LocalAI/core/config"
|
|
"github.com/mudler/LocalAI/core/gallery"
|
|
"github.com/mudler/LocalAI/core/http/utils"
|
|
"github.com/mudler/LocalAI/core/schema"
|
|
"github.com/mudler/LocalAI/core/services"
|
|
"github.com/rs/zerolog/log"
|
|
)
|
|
|
|
type ModelGalleryEndpointService struct {
|
|
galleries []config.Gallery
|
|
modelPath string
|
|
galleryApplier *services.GalleryService
|
|
}
|
|
|
|
type GalleryModel struct {
|
|
ID string `json:"id"`
|
|
ConfigURL string `json:"config_url"`
|
|
gallery.GalleryModel
|
|
}
|
|
|
|
func CreateModelGalleryEndpointService(galleries []config.Gallery, modelPath string, galleryApplier *services.GalleryService) ModelGalleryEndpointService {
|
|
return ModelGalleryEndpointService{
|
|
galleries: galleries,
|
|
modelPath: modelPath,
|
|
galleryApplier: galleryApplier,
|
|
}
|
|
}
|
|
|
|
// GetOpStatusEndpoint returns the job status
|
|
// @Summary Returns the job status
|
|
// @Success 200 {object} gallery.GalleryOpStatus "Response"
|
|
// @Router /models/jobs/{uuid} [get]
|
|
func (mgs *ModelGalleryEndpointService) GetOpStatusEndpoint() func(c *fiber.Ctx) error {
|
|
return func(c *fiber.Ctx) error {
|
|
status := mgs.galleryApplier.GetStatus(c.Params("uuid"))
|
|
if status == nil {
|
|
return fmt.Errorf("could not find any status for ID")
|
|
}
|
|
return c.JSON(status)
|
|
}
|
|
}
|
|
|
|
// GetAllStatusEndpoint returns all the jobs status progress
|
|
// @Summary Returns all the jobs status progress
|
|
// @Success 200 {object} map[string]gallery.GalleryOpStatus "Response"
|
|
// @Router /models/jobs [get]
|
|
func (mgs *ModelGalleryEndpointService) GetAllStatusEndpoint() func(c *fiber.Ctx) error {
|
|
return func(c *fiber.Ctx) error {
|
|
return c.JSON(mgs.galleryApplier.GetAllStatus())
|
|
}
|
|
}
|
|
|
|
// ApplyModelGalleryEndpoint installs a new model to a LocalAI instance from the model gallery
|
|
// @Summary Install models to LocalAI.
|
|
// @Param request body GalleryModel true "query params"
|
|
// @Success 200 {object} schema.GalleryResponse "Response"
|
|
// @Router /models/apply [post]
|
|
func (mgs *ModelGalleryEndpointService) ApplyModelGalleryEndpoint() func(c *fiber.Ctx) error {
|
|
return func(c *fiber.Ctx) error {
|
|
input := new(GalleryModel)
|
|
// Get input data from the request body
|
|
if err := c.BodyParser(input); err != nil {
|
|
return err
|
|
}
|
|
|
|
uuid, err := uuid.NewUUID()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
mgs.galleryApplier.C <- gallery.GalleryOp{
|
|
Req: input.GalleryModel,
|
|
Id: uuid.String(),
|
|
GalleryModelName: input.ID,
|
|
Galleries: mgs.galleries,
|
|
ConfigURL: input.ConfigURL,
|
|
}
|
|
|
|
return c.JSON(schema.GalleryResponse{ID: uuid.String(), StatusURL: fmt.Sprintf("%smodels/jobs/%s", utils.BaseURL(c), uuid.String())})
|
|
}
|
|
}
|
|
|
|
// DeleteModelGalleryEndpoint lets delete models from a LocalAI instance
|
|
// @Summary delete models to LocalAI.
|
|
// @Param name path string true "Model name"
|
|
// @Success 200 {object} schema.GalleryResponse "Response"
|
|
// @Router /models/delete/{name} [post]
|
|
func (mgs *ModelGalleryEndpointService) DeleteModelGalleryEndpoint() func(c *fiber.Ctx) error {
|
|
return func(c *fiber.Ctx) error {
|
|
modelName := c.Params("name")
|
|
|
|
mgs.galleryApplier.C <- gallery.GalleryOp{
|
|
Delete: true,
|
|
GalleryModelName: modelName,
|
|
}
|
|
|
|
uuid, err := uuid.NewUUID()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return c.JSON(schema.GalleryResponse{ID: uuid.String(), StatusURL: fmt.Sprintf("%smodels/jobs/%s", utils.BaseURL(c), uuid.String())})
|
|
}
|
|
}
|
|
|
|
// ListModelFromGalleryEndpoint list the available models for installation from the active galleries
|
|
// @Summary List installable models.
|
|
// @Success 200 {object} []gallery.GalleryModel "Response"
|
|
// @Router /models/available [get]
|
|
func (mgs *ModelGalleryEndpointService) ListModelFromGalleryEndpoint() func(c *fiber.Ctx) error {
|
|
return func(c *fiber.Ctx) error {
|
|
|
|
models, err := gallery.AvailableGalleryModels(mgs.galleries, mgs.modelPath)
|
|
if err != nil {
|
|
log.Error().Err(err).Msg("could not list models from galleries")
|
|
return err
|
|
}
|
|
|
|
log.Debug().Msgf("Available %d models from %d galleries\n", len(models), len(mgs.galleries))
|
|
|
|
m := []gallery.Metadata{}
|
|
|
|
for _, mm := range models {
|
|
m = append(m, mm.Metadata)
|
|
}
|
|
|
|
log.Debug().Msgf("Models %#v", m)
|
|
|
|
dat, err := json.Marshal(m)
|
|
if err != nil {
|
|
return fmt.Errorf("could not marshal models: %w", err)
|
|
}
|
|
return c.Send(dat)
|
|
}
|
|
}
|
|
|
|
// ListModelGalleriesEndpoint list the available galleries configured in LocalAI
|
|
// @Summary List all Galleries
|
|
// @Success 200 {object} []config.Gallery "Response"
|
|
// @Router /models/galleries [get]
|
|
// NOTE: This is different (and much simpler!) than above! This JUST lists the model galleries that have been loaded, not their contents!
|
|
func (mgs *ModelGalleryEndpointService) ListModelGalleriesEndpoint() func(c *fiber.Ctx) error {
|
|
return func(c *fiber.Ctx) error {
|
|
log.Debug().Msgf("Listing model galleries %+v", mgs.galleries)
|
|
dat, err := json.Marshal(mgs.galleries)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return c.Send(dat)
|
|
}
|
|
}
|
|
|
|
// AddModelGalleryEndpoint adds a gallery in LocalAI
|
|
// @Summary Adds a gallery in LocalAI
|
|
// @Param request body config.Gallery true "Gallery details"
|
|
// @Success 200 {object} []config.Gallery "Response"
|
|
// @Router /models/galleries [post]
|
|
func (mgs *ModelGalleryEndpointService) AddModelGalleryEndpoint() func(c *fiber.Ctx) error {
|
|
return func(c *fiber.Ctx) error {
|
|
input := new(config.Gallery)
|
|
// Get input data from the request body
|
|
if err := c.BodyParser(input); err != nil {
|
|
return err
|
|
}
|
|
if slices.ContainsFunc(mgs.galleries, func(gallery config.Gallery) bool {
|
|
return gallery.Name == input.Name
|
|
}) {
|
|
return fmt.Errorf("%s already exists", input.Name)
|
|
}
|
|
dat, err := json.Marshal(mgs.galleries)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
log.Debug().Msgf("Adding %+v to gallery list", *input)
|
|
mgs.galleries = append(mgs.galleries, *input)
|
|
return c.Send(dat)
|
|
}
|
|
}
|
|
|
|
// RemoveModelGalleryEndpoint remove a gallery in LocalAI
|
|
// @Summary removes a gallery from LocalAI
|
|
// @Param request body config.Gallery true "Gallery details"
|
|
// @Success 200 {object} []config.Gallery "Response"
|
|
// @Router /models/galleries [delete]
|
|
func (mgs *ModelGalleryEndpointService) RemoveModelGalleryEndpoint() func(c *fiber.Ctx) error {
|
|
return func(c *fiber.Ctx) error {
|
|
input := new(config.Gallery)
|
|
// Get input data from the request body
|
|
if err := c.BodyParser(input); err != nil {
|
|
return err
|
|
}
|
|
if !slices.ContainsFunc(mgs.galleries, func(gallery config.Gallery) bool {
|
|
return gallery.Name == input.Name
|
|
}) {
|
|
return fmt.Errorf("%s is not currently registered", input.Name)
|
|
}
|
|
mgs.galleries = slices.DeleteFunc(mgs.galleries, func(gallery config.Gallery) bool {
|
|
return gallery.Name == input.Name
|
|
})
|
|
dat, err := json.Marshal(mgs.galleries)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return c.Send(dat)
|
|
}
|
|
}
|