mirror of
https://github.com/mudler/LocalAI.git
synced 2025-05-20 10:35:01 +00:00
fix: use bytes in gRPC proto instead of strings (#813)
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
This commit is contained in:
parent
0af0df7423
commit
b96e30e66c
8 changed files with 20 additions and 16 deletions
|
@ -67,9 +67,9 @@ func ModelInference(ctx context.Context, s string, loader *model.ModelLoader, c
|
|||
opts.Prompt = s
|
||||
if tokenCallback != nil {
|
||||
ss := ""
|
||||
err := inferenceModel.PredictStream(ctx, opts, func(s string) {
|
||||
tokenCallback(s)
|
||||
ss += s
|
||||
err := inferenceModel.PredictStream(ctx, opts, func(s []byte) {
|
||||
tokenCallback(string(s))
|
||||
ss += string(s)
|
||||
})
|
||||
return ss, err
|
||||
} else {
|
||||
|
@ -77,7 +77,7 @@ func ModelInference(ctx context.Context, s string, loader *model.ModelLoader, c
|
|||
if err != nil {
|
||||
return "", err
|
||||
}
|
||||
return reply.Message, err
|
||||
return string(reply.Message), err
|
||||
}
|
||||
}
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue