LocalAI/core/http
Ettore Di Giacinto c89271b2e4
feat(llama.cpp): add distributed llama.cpp inferencing (#2324)
* feat(llama.cpp): support distributed llama.cpp

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* feat: let tweak how chat messages are merged together

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* refactor

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* Makefile: register to ALL_GRPC_BACKENDS

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* refactoring, allow disable auto-detection of backends

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* minor fixups

Signed-off-by: mudler <mudler@localai.io>

* feat: add cmd to start rpc-server from llama.cpp

Signed-off-by: mudler <mudler@localai.io>

* ci: add ccache

Signed-off-by: mudler <mudler@localai.io>

---------

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Signed-off-by: mudler <mudler@localai.io>
2024-05-15 01:17:02 +02:00
..
ctx Revert #1963 (#2056) 2024-04-17 23:33:49 +02:00
elements fix(ux): fix small glitches (#2265) 2024-05-08 19:34:33 +02:00
endpoints feat(llama.cpp): add distributed llama.cpp inferencing (#2324) 2024-05-15 01:17:02 +02:00
routes feat(ui): display number of available models for installation (#2298) 2024-05-12 14:24:36 +02:00
static feat(ui): prompt for chat, support vision, enhancements (#2259) 2024-05-08 00:42:34 +02:00
views feat(ui): display number of available models for installation (#2298) 2024-05-12 14:24:36 +02:00
app.go UI: flag trust_remote_code to users // favicon support (#2253) 2024-05-07 08:39:23 +02:00
app_test.go test: check the response URL during image gen in app_test.go (#2248) 2024-05-05 18:46:33 +00:00
http_suite_test.go fix: rename fiber entrypoint from http/api to http/app (#2096) 2024-04-21 22:39:28 +02:00
render.go fix: use bluemonday as recommended by blackfriday (#2142) 2024-04-26 10:34:50 +02:00