mirror of
https://github.com/mudler/LocalAI.git
synced 2025-06-30 06:30:43 +00:00
Not anywhere close to done, but getting back to this and staging some progress. Preliminary pass on loading a config and overriding params with request json.
This commit is contained in:
parent
fc3c105d42
commit
20a0cd2f66
13 changed files with 812 additions and 388 deletions
174
apiv2/localai.go
174
apiv2/localai.go
|
@ -3,20 +3,36 @@ package apiv2
|
|||
import (
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
"io"
|
||||
"net/http"
|
||||
"runtime"
|
||||
"strings"
|
||||
|
||||
"github.com/mitchellh/mapstructure"
|
||||
)
|
||||
|
||||
type LocalAIServer struct {
|
||||
configMerger *ConfigMerger
|
||||
configManager *ConfigManager
|
||||
}
|
||||
|
||||
var _ ServerInterface = (*LocalAIServer)(nil)
|
||||
|
||||
type Error struct {
|
||||
Code int `json:"code"`
|
||||
Message string `json:"message"`
|
||||
}
|
||||
|
||||
type ModelOnlyRequest struct {
|
||||
Model string `json:"model" yaml:"model"`
|
||||
}
|
||||
|
||||
// This function grabs the name of the function that calls it, skipping up the callstack `skip` levels.
|
||||
// This is probably a go war crime, but NJ method and all. It's an awesome way to index EndpointConfigMap
|
||||
func printCurrentFunctionName(skip int) string {
|
||||
pc, _, _, _ := runtime.Caller(skip)
|
||||
funcName := runtime.FuncForPC(pc).Name()
|
||||
fmt.Println("Current function:", funcName)
|
||||
return funcName
|
||||
}
|
||||
|
||||
func sendError(w http.ResponseWriter, code int, message string) {
|
||||
localAiError := Error{
|
||||
Code: code,
|
||||
|
@ -26,39 +42,91 @@ func sendError(w http.ResponseWriter, code int, message string) {
|
|||
json.NewEncoder(w).Encode(localAiError)
|
||||
}
|
||||
|
||||
// It won't work, but it's worth a try.
|
||||
const nyiErrorMessageFormatString = "%s is not yet implemented by LocalAI\nThere is no need to contact support about this error and retrying will not help.\nExpect an update at https://github.com/go-skynet/LocalAI if this changes!"
|
||||
// TODO: Is it a good idea to return "" in cases where the model isn't provided?
|
||||
// Or is that actually an error condition?
|
||||
// NO is a decent guess as any to start with?
|
||||
// r *http.Request
|
||||
func (server *LocalAIServer) getRequestModelName(body []byte) string {
|
||||
var modelOnlyRequest = ModelOnlyRequest{}
|
||||
if err := json.Unmarshal(body, &modelOnlyRequest); err != nil {
|
||||
fmt.Printf("ERR in getRequestModelName, %+v", err)
|
||||
return ""
|
||||
}
|
||||
return modelOnlyRequest.Model
|
||||
}
|
||||
|
||||
// Do we want or need an additional "wontfix" template that is even stronger than this?
|
||||
const nyiDepreciatedErrorMessageFormatString = "%s is a depreciated portion of the OpenAI API, and is not yet implemented by LocalAI\nThere is no need to contact support about this error and retrying will not help."
|
||||
func (server *LocalAIServer) combineRequestAndConfig(endpointName string, body []byte) (interface{}, error) {
|
||||
model := server.getRequestModelName(body)
|
||||
|
||||
lookup := ConfigRegistration{Model: model, Endpoint: endpointName}
|
||||
|
||||
config, exists := server.configManager.GetConfig(lookup)
|
||||
|
||||
if !exists {
|
||||
return nil, fmt.Errorf("Config not found for %+v", lookup)
|
||||
}
|
||||
|
||||
// fmt.Printf("Model: %s\nConfig: %+v\n", model, config)
|
||||
|
||||
request := config.GetRequestDefaults()
|
||||
// fmt.Printf("BEFORE rD: %T\n%+v\n\n", request, request)
|
||||
tmpUnmarshal := map[string]interface{}{}
|
||||
if err := json.Unmarshal(body, &tmpUnmarshal); err != nil {
|
||||
return nil, fmt.Errorf("error unmarshalling json to temp map\n%w", err)
|
||||
}
|
||||
// fmt.Printf("$$$ tmpUnmarshal: %+v\n", tmpUnmarshal)
|
||||
mapstructure.Decode(tmpUnmarshal, &request)
|
||||
fmt.Printf("AFTER rD: %T\n%+v\n\n", request, request)
|
||||
return request, nil
|
||||
}
|
||||
|
||||
func (server *LocalAIServer) getRequest(w http.ResponseWriter, r *http.Request) (interface{}, error) {
|
||||
body, err := io.ReadAll(r.Body)
|
||||
if err != nil {
|
||||
sendError(w, http.StatusBadRequest, "Failed to read body")
|
||||
}
|
||||
|
||||
splitFnName := strings.Split(printCurrentFunctionName(2), ".")
|
||||
|
||||
endpointName := splitFnName[len(splitFnName)-1]
|
||||
|
||||
return server.combineRequestAndConfig(endpointName, body)
|
||||
}
|
||||
|
||||
// CancelFineTune implements ServerInterface
|
||||
func (*LocalAIServer) CancelFineTune(w http.ResponseWriter, r *http.Request, fineTuneId string) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "Fine Tune"))
|
||||
return
|
||||
}
|
||||
|
||||
// CreateAnswer implements ServerInterface
|
||||
func (*LocalAIServer) CreateAnswer(w http.ResponseWriter, r *http.Request) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiDepreciatedErrorMessageFormatString, "CreateAnswer"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// CreateChatCompletion implements ServerInterface
|
||||
func (*LocalAIServer) CreateChatCompletion(w http.ResponseWriter, r *http.Request) {
|
||||
var chatRequest CreateChatCompletionRequest
|
||||
if err := json.NewDecoder(r.Body).Decode(&chatRequest); err != nil {
|
||||
sendError(w, http.StatusBadRequest, "Invalid CreateChatCompletionRequest")
|
||||
func (server *LocalAIServer) CreateChatCompletion(w http.ResponseWriter, r *http.Request) {
|
||||
fmt.Println("HIT APIv2 CreateChatCompletion!")
|
||||
|
||||
request, err := server.getRequest(w, r)
|
||||
|
||||
if err != nil {
|
||||
sendError(w, http.StatusBadRequest, err.Error())
|
||||
}
|
||||
|
||||
// fmt.Printf("\n!!! Survived to attempt cast. BEFORE:\n\tType: %T\n\t%+v", request, request)
|
||||
|
||||
chatRequest, castSuccess := request.(CreateChatCompletionRequest)
|
||||
|
||||
if !castSuccess {
|
||||
sendError(w, http.StatusInternalServerError, "Cast Fail???")
|
||||
return
|
||||
}
|
||||
configMerger.GetConfig(chatRequest.Model)
|
||||
|
||||
fmt.Printf("\n\n!! AFTER !!\ntemperature %f\n top_p %f \n %d\n", *chatRequest.Temperature, *chatRequest.TopP, *chatRequest.XLocalaiExtensions.TopK)
|
||||
|
||||
fmt.Printf("chatRequest: %+v\nlen(messages): %d", chatRequest, len(chatRequest.Messages))
|
||||
for i, m := range chatRequest.Messages {
|
||||
fmt.Printf("message #%d: %+v", i, m)
|
||||
}
|
||||
}
|
||||
|
||||
// CreateClassification implements ServerInterface
|
||||
func (*LocalAIServer) CreateClassification(w http.ResponseWriter, r *http.Request) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiDepreciatedErrorMessageFormatString, "CreateClassification"))
|
||||
return
|
||||
}
|
||||
// switch chatRequest := requestDefault.(type) {
|
||||
// case CreateChatCompletionRequest:
|
||||
|
||||
// CreateCompletion implements ServerInterface
|
||||
func (*LocalAIServer) CreateCompletion(w http.ResponseWriter, r *http.Request) {
|
||||
|
@ -67,8 +135,7 @@ func (*LocalAIServer) CreateCompletion(w http.ResponseWriter, r *http.Request) {
|
|||
|
||||
// CreateEdit implements ServerInterface
|
||||
func (*LocalAIServer) CreateEdit(w http.ResponseWriter, r *http.Request) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "CreateEdit"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// CreateEmbedding implements ServerInterface
|
||||
|
@ -78,14 +145,12 @@ func (*LocalAIServer) CreateEmbedding(w http.ResponseWriter, r *http.Request) {
|
|||
|
||||
// CreateFile implements ServerInterface
|
||||
func (*LocalAIServer) CreateFile(w http.ResponseWriter, r *http.Request) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "Create File"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// CreateFineTune implements ServerInterface
|
||||
func (*LocalAIServer) CreateFineTune(w http.ResponseWriter, r *http.Request) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "Create Fine Tune"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// CreateImage implements ServerInterface
|
||||
|
@ -105,14 +170,7 @@ func (*LocalAIServer) CreateImageVariation(w http.ResponseWriter, r *http.Reques
|
|||
|
||||
// CreateModeration implements ServerInterface
|
||||
func (*LocalAIServer) CreateModeration(w http.ResponseWriter, r *http.Request) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "CreateModeration"))
|
||||
return
|
||||
}
|
||||
|
||||
// CreateSearch implements ServerInterface
|
||||
func (*LocalAIServer) CreateSearch(w http.ResponseWriter, r *http.Request, engineId string) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiDepreciatedErrorMessageFormatString, "CreateSearch"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// CreateTranscription implements ServerInterface
|
||||
|
@ -127,44 +185,32 @@ func (*LocalAIServer) CreateTranslation(w http.ResponseWriter, r *http.Request)
|
|||
|
||||
// DeleteFile implements ServerInterface
|
||||
func (*LocalAIServer) DeleteFile(w http.ResponseWriter, r *http.Request, fileId string) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "DeleteFile"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// DeleteModel implements ServerInterface
|
||||
func (*LocalAIServer) DeleteModel(w http.ResponseWriter, r *http.Request, model string) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "DeleteModel"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// DownloadFile implements ServerInterface
|
||||
func (*LocalAIServer) DownloadFile(w http.ResponseWriter, r *http.Request, fileId string) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "DownloadFile"))
|
||||
return
|
||||
}
|
||||
|
||||
// ListEngines implements ServerInterface
|
||||
func (*LocalAIServer) ListEngines(w http.ResponseWriter, r *http.Request) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiDepreciatedErrorMessageFormatString, "List Engines"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// ListFiles implements ServerInterface
|
||||
func (*LocalAIServer) ListFiles(w http.ResponseWriter, r *http.Request) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "ListFiles"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// ListFineTuneEvents implements ServerInterface
|
||||
func (*LocalAIServer) ListFineTuneEvents(w http.ResponseWriter, r *http.Request, fineTuneId string, params ListFineTuneEventsParams) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "List Fine Tune Events"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// ListFineTunes implements ServerInterface
|
||||
func (*LocalAIServer) ListFineTunes(w http.ResponseWriter, r *http.Request) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "List Fine Tunes"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// ListModels implements ServerInterface
|
||||
|
@ -172,25 +218,19 @@ func (*LocalAIServer) ListModels(w http.ResponseWriter, r *http.Request) {
|
|||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// RetrieveEngine implements ServerInterface
|
||||
func (*LocalAIServer) RetrieveEngine(w http.ResponseWriter, r *http.Request, engineId string) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiDepreciatedErrorMessageFormatString, "RetrieveEngine"))
|
||||
return
|
||||
}
|
||||
|
||||
// RetrieveFile implements ServerInterface
|
||||
func (*LocalAIServer) RetrieveFile(w http.ResponseWriter, r *http.Request, fileId string) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "RetrieveFile"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// RetrieveFineTune implements ServerInterface
|
||||
func (*LocalAIServer) RetrieveFineTune(w http.ResponseWriter, r *http.Request, fineTuneId string) {
|
||||
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "Retrieve Fine Tune"))
|
||||
return
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
// RetrieveModel implements ServerInterface
|
||||
func (*LocalAIServer) RetrieveModel(w http.ResponseWriter, r *http.Request, model string) {
|
||||
panic("unimplemented")
|
||||
}
|
||||
|
||||
var _ ServerInterface = (*LocalAIServer)(nil)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue