diff --git a/api/api_test.go b/api/api_test.go index ca840b53..06e978ba 100644 --- a/api/api_test.go +++ b/api/api_test.go @@ -555,9 +555,10 @@ var _ = Describe("API test", func() { }) It("returns errors", func() { + backends := len(model.AutoLoadBackends) _, err := client.CreateCompletion(context.TODO(), openai.CompletionRequest{Model: "foomodel", Prompt: "abcdedfghikl"}) Expect(err).To(HaveOccurred()) - Expect(err.Error()).To(ContainSubstring("error, status code: 500, message: could not load model - all backends returned error: 12 errors occurred:")) + Expect(err.Error()).To(ContainSubstring(fmt.Sprintf("error, status code: 500, message: could not load model - all backends returned error: %d errors occurred:", backends))) }) It("transcribes audio", func() { if runtime.GOOS != "linux" { diff --git a/pkg/model/initializers.go b/pkg/model/initializers.go index 8ca3769d..a4d21fa5 100644 --- a/pkg/model/initializers.go +++ b/pkg/model/initializers.go @@ -48,7 +48,7 @@ const ( //GGLLMFalconBackend = "falcon" ) -var autoLoadBackends []string = []string{ +var AutoLoadBackends []string = []string{ LlamaBackend, Gpt4All, RwkvBackend, @@ -56,6 +56,7 @@ var autoLoadBackends []string = []string{ WhisperBackend, GPTNeoXBackend, BertEmbeddingsBackend, + LlamaMasterBackend, FalconGGMLBackend, GPTJBackend, Gpt2Backend, @@ -218,7 +219,7 @@ func (ml *ModelLoader) GreedyLoader(opts ...Option) (*grpc.Client, error) { ml.mu.Unlock() var err error - for _, b := range autoLoadBackends { + for _, b := range AutoLoadBackends { if b == BloomzBackend || b == WhisperBackend || b == RwkvBackend { // do not autoload bloomz/whisper/rwkv continue }