package startup import ( "github.com/go-skynet/LocalAI/core/services" "github.com/go-skynet/LocalAI/internal" "github.com/go-skynet/LocalAI/pkg/assets" "github.com/go-skynet/LocalAI/pkg/model" "github.com/go-skynet/LocalAI/pkg/schema" "github.com/rs/zerolog" "github.com/rs/zerolog/log" ) func Startup(opts ...schema.AppOption) (*services.ConfigLoader, *model.ModelLoader, *schema.StartupOptions, error) { options := schema.NewStartupOptions(opts...) ml := model.NewModelLoader(options.ModelPath) zerolog.SetGlobalLevel(zerolog.InfoLevel) if options.Debug { zerolog.SetGlobalLevel(zerolog.DebugLevel) } log.Info().Msgf("Starting LocalAI using %d threads, with models path: %s", options.Threads, options.ModelPath) log.Info().Msgf("LocalAI version: %s", internal.PrintableVersion()) cl := services.NewConfigLoader() if err := cl.LoadConfigs(options.ModelPath); err != nil { log.Error().Msgf("error loading config files: %s", err.Error()) } if options.ConfigFile != "" { if err := cl.LoadConfigFile(options.ConfigFile); err != nil { log.Error().Msgf("error loading config file: %s", err.Error()) } } if err := cl.Preload(options.ModelPath); err != nil { log.Error().Msgf("error downloading models: %s", err.Error()) } if options.PreloadJSONModels != "" { if err := services.ApplyGalleryFromString(options.ModelPath, options.PreloadJSONModels, cl, options.Galleries); err != nil { return nil, nil, nil, err } } if options.PreloadModelsFromPath != "" { if err := services.ApplyGalleryFromFile(options.ModelPath, options.PreloadModelsFromPath, cl, options.Galleries); err != nil { return nil, nil, nil, err } } if options.Debug { for _, v := range cl.ListConfigs() { cfg, _ := cl.GetConfig(v) log.Debug().Msgf("Model: %s (config: %+v)", v, cfg) } } if options.AssetsDestination != "" { // Extract files from the embedded FS err := assets.ExtractFiles(options.BackendAssets, options.AssetsDestination) log.Debug().Msgf("Extracting backend assets files to %s", options.AssetsDestination) if err != nil { log.Warn().Msgf("Failed extracting backend assets files: %s (might be required for some backends to work properly, like gpt4all)", err) } } // turn off any process that was started by GRPC if the context is canceled go func() { <-options.Context.Done() log.Debug().Msgf("Context canceled, shutting down") ml.StopAllGRPC() }() if options.WatchDog { wd := model.NewWatchDog( ml, options.WatchDogBusyTimeout, options.WatchDogIdleTimeout, options.WatchDogBusy, options.WatchDogIdle) ml.SetWatchDog(wd) go wd.Run() go func() { <-options.Context.Done() log.Debug().Msgf("Context canceled, shutting down") wd.Shutdown() }() } return cl, ml, options, nil }