Not anywhere close to done, but getting back to this and staging some progress. Preliminary pass on loading a config and overriding params with request json.

This commit is contained in:
Dave Lee 2023-06-01 03:31:55 -04:00
parent fc3c105d42
commit 20a0cd2f66
No known key found for this signature in database
13 changed files with 812 additions and 388 deletions

View file

@ -3,20 +3,36 @@ package apiv2
import (
"encoding/json"
"fmt"
"io"
"net/http"
"runtime"
"strings"
"github.com/mitchellh/mapstructure"
)
type LocalAIServer struct {
configMerger *ConfigMerger
configManager *ConfigManager
}
var _ ServerInterface = (*LocalAIServer)(nil)
type Error struct {
Code int `json:"code"`
Message string `json:"message"`
}
type ModelOnlyRequest struct {
Model string `json:"model" yaml:"model"`
}
// This function grabs the name of the function that calls it, skipping up the callstack `skip` levels.
// This is probably a go war crime, but NJ method and all. It's an awesome way to index EndpointConfigMap
func printCurrentFunctionName(skip int) string {
pc, _, _, _ := runtime.Caller(skip)
funcName := runtime.FuncForPC(pc).Name()
fmt.Println("Current function:", funcName)
return funcName
}
func sendError(w http.ResponseWriter, code int, message string) {
localAiError := Error{
Code: code,
@ -26,39 +42,91 @@ func sendError(w http.ResponseWriter, code int, message string) {
json.NewEncoder(w).Encode(localAiError)
}
// It won't work, but it's worth a try.
const nyiErrorMessageFormatString = "%s is not yet implemented by LocalAI\nThere is no need to contact support about this error and retrying will not help.\nExpect an update at https://github.com/go-skynet/LocalAI if this changes!"
// TODO: Is it a good idea to return "" in cases where the model isn't provided?
// Or is that actually an error condition?
// NO is a decent guess as any to start with?
// r *http.Request
func (server *LocalAIServer) getRequestModelName(body []byte) string {
var modelOnlyRequest = ModelOnlyRequest{}
if err := json.Unmarshal(body, &modelOnlyRequest); err != nil {
fmt.Printf("ERR in getRequestModelName, %+v", err)
return ""
}
return modelOnlyRequest.Model
}
// Do we want or need an additional "wontfix" template that is even stronger than this?
const nyiDepreciatedErrorMessageFormatString = "%s is a depreciated portion of the OpenAI API, and is not yet implemented by LocalAI\nThere is no need to contact support about this error and retrying will not help."
func (server *LocalAIServer) combineRequestAndConfig(endpointName string, body []byte) (interface{}, error) {
model := server.getRequestModelName(body)
lookup := ConfigRegistration{Model: model, Endpoint: endpointName}
config, exists := server.configManager.GetConfig(lookup)
if !exists {
return nil, fmt.Errorf("Config not found for %+v", lookup)
}
// fmt.Printf("Model: %s\nConfig: %+v\n", model, config)
request := config.GetRequestDefaults()
// fmt.Printf("BEFORE rD: %T\n%+v\n\n", request, request)
tmpUnmarshal := map[string]interface{}{}
if err := json.Unmarshal(body, &tmpUnmarshal); err != nil {
return nil, fmt.Errorf("error unmarshalling json to temp map\n%w", err)
}
// fmt.Printf("$$$ tmpUnmarshal: %+v\n", tmpUnmarshal)
mapstructure.Decode(tmpUnmarshal, &request)
fmt.Printf("AFTER rD: %T\n%+v\n\n", request, request)
return request, nil
}
func (server *LocalAIServer) getRequest(w http.ResponseWriter, r *http.Request) (interface{}, error) {
body, err := io.ReadAll(r.Body)
if err != nil {
sendError(w, http.StatusBadRequest, "Failed to read body")
}
splitFnName := strings.Split(printCurrentFunctionName(2), ".")
endpointName := splitFnName[len(splitFnName)-1]
return server.combineRequestAndConfig(endpointName, body)
}
// CancelFineTune implements ServerInterface
func (*LocalAIServer) CancelFineTune(w http.ResponseWriter, r *http.Request, fineTuneId string) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "Fine Tune"))
return
}
// CreateAnswer implements ServerInterface
func (*LocalAIServer) CreateAnswer(w http.ResponseWriter, r *http.Request) {
sendError(w, 501, fmt.Sprintf(nyiDepreciatedErrorMessageFormatString, "CreateAnswer"))
return
panic("unimplemented")
}
// CreateChatCompletion implements ServerInterface
func (*LocalAIServer) CreateChatCompletion(w http.ResponseWriter, r *http.Request) {
var chatRequest CreateChatCompletionRequest
if err := json.NewDecoder(r.Body).Decode(&chatRequest); err != nil {
sendError(w, http.StatusBadRequest, "Invalid CreateChatCompletionRequest")
func (server *LocalAIServer) CreateChatCompletion(w http.ResponseWriter, r *http.Request) {
fmt.Println("HIT APIv2 CreateChatCompletion!")
request, err := server.getRequest(w, r)
if err != nil {
sendError(w, http.StatusBadRequest, err.Error())
}
// fmt.Printf("\n!!! Survived to attempt cast. BEFORE:\n\tType: %T\n\t%+v", request, request)
chatRequest, castSuccess := request.(CreateChatCompletionRequest)
if !castSuccess {
sendError(w, http.StatusInternalServerError, "Cast Fail???")
return
}
configMerger.GetConfig(chatRequest.Model)
fmt.Printf("\n\n!! AFTER !!\ntemperature %f\n top_p %f \n %d\n", *chatRequest.Temperature, *chatRequest.TopP, *chatRequest.XLocalaiExtensions.TopK)
fmt.Printf("chatRequest: %+v\nlen(messages): %d", chatRequest, len(chatRequest.Messages))
for i, m := range chatRequest.Messages {
fmt.Printf("message #%d: %+v", i, m)
}
}
// CreateClassification implements ServerInterface
func (*LocalAIServer) CreateClassification(w http.ResponseWriter, r *http.Request) {
sendError(w, 501, fmt.Sprintf(nyiDepreciatedErrorMessageFormatString, "CreateClassification"))
return
}
// switch chatRequest := requestDefault.(type) {
// case CreateChatCompletionRequest:
// CreateCompletion implements ServerInterface
func (*LocalAIServer) CreateCompletion(w http.ResponseWriter, r *http.Request) {
@ -67,8 +135,7 @@ func (*LocalAIServer) CreateCompletion(w http.ResponseWriter, r *http.Request) {
// CreateEdit implements ServerInterface
func (*LocalAIServer) CreateEdit(w http.ResponseWriter, r *http.Request) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "CreateEdit"))
return
panic("unimplemented")
}
// CreateEmbedding implements ServerInterface
@ -78,14 +145,12 @@ func (*LocalAIServer) CreateEmbedding(w http.ResponseWriter, r *http.Request) {
// CreateFile implements ServerInterface
func (*LocalAIServer) CreateFile(w http.ResponseWriter, r *http.Request) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "Create File"))
return
panic("unimplemented")
}
// CreateFineTune implements ServerInterface
func (*LocalAIServer) CreateFineTune(w http.ResponseWriter, r *http.Request) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "Create Fine Tune"))
return
panic("unimplemented")
}
// CreateImage implements ServerInterface
@ -105,14 +170,7 @@ func (*LocalAIServer) CreateImageVariation(w http.ResponseWriter, r *http.Reques
// CreateModeration implements ServerInterface
func (*LocalAIServer) CreateModeration(w http.ResponseWriter, r *http.Request) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "CreateModeration"))
return
}
// CreateSearch implements ServerInterface
func (*LocalAIServer) CreateSearch(w http.ResponseWriter, r *http.Request, engineId string) {
sendError(w, 501, fmt.Sprintf(nyiDepreciatedErrorMessageFormatString, "CreateSearch"))
return
panic("unimplemented")
}
// CreateTranscription implements ServerInterface
@ -127,44 +185,32 @@ func (*LocalAIServer) CreateTranslation(w http.ResponseWriter, r *http.Request)
// DeleteFile implements ServerInterface
func (*LocalAIServer) DeleteFile(w http.ResponseWriter, r *http.Request, fileId string) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "DeleteFile"))
return
panic("unimplemented")
}
// DeleteModel implements ServerInterface
func (*LocalAIServer) DeleteModel(w http.ResponseWriter, r *http.Request, model string) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "DeleteModel"))
return
panic("unimplemented")
}
// DownloadFile implements ServerInterface
func (*LocalAIServer) DownloadFile(w http.ResponseWriter, r *http.Request, fileId string) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "DownloadFile"))
return
}
// ListEngines implements ServerInterface
func (*LocalAIServer) ListEngines(w http.ResponseWriter, r *http.Request) {
sendError(w, 501, fmt.Sprintf(nyiDepreciatedErrorMessageFormatString, "List Engines"))
return
panic("unimplemented")
}
// ListFiles implements ServerInterface
func (*LocalAIServer) ListFiles(w http.ResponseWriter, r *http.Request) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "ListFiles"))
return
panic("unimplemented")
}
// ListFineTuneEvents implements ServerInterface
func (*LocalAIServer) ListFineTuneEvents(w http.ResponseWriter, r *http.Request, fineTuneId string, params ListFineTuneEventsParams) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "List Fine Tune Events"))
return
panic("unimplemented")
}
// ListFineTunes implements ServerInterface
func (*LocalAIServer) ListFineTunes(w http.ResponseWriter, r *http.Request) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "List Fine Tunes"))
return
panic("unimplemented")
}
// ListModels implements ServerInterface
@ -172,25 +218,19 @@ func (*LocalAIServer) ListModels(w http.ResponseWriter, r *http.Request) {
panic("unimplemented")
}
// RetrieveEngine implements ServerInterface
func (*LocalAIServer) RetrieveEngine(w http.ResponseWriter, r *http.Request, engineId string) {
sendError(w, 501, fmt.Sprintf(nyiDepreciatedErrorMessageFormatString, "RetrieveEngine"))
return
}
// RetrieveFile implements ServerInterface
func (*LocalAIServer) RetrieveFile(w http.ResponseWriter, r *http.Request, fileId string) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "RetrieveFile"))
return
panic("unimplemented")
}
// RetrieveFineTune implements ServerInterface
func (*LocalAIServer) RetrieveFineTune(w http.ResponseWriter, r *http.Request, fineTuneId string) {
sendError(w, 501, fmt.Sprintf(nyiErrorMessageFormatString, "Retrieve Fine Tune"))
return
panic("unimplemented")
}
// RetrieveModel implements ServerInterface
func (*LocalAIServer) RetrieveModel(w http.ResponseWriter, r *http.Request, model string) {
panic("unimplemented")
}
var _ ServerInterface = (*LocalAIServer)(nil)