LocalAI/backend/cpp/llama
Ettore Di Giacinto 3826edb9da chore(deps): bump llama.cpp to '10f2e81809bbb69ecfe64fc8b4686285f84b0c07'
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2025-03-12 09:12:59 +01:00
..
patches chore(deps): bump llama.cpp to '10f2e81809bbb69ecfe64fc8b4686285f84b0c07' 2025-03-12 09:12:59 +01:00
CMakeLists.txt deps(llama.cpp): update, support Gemma models (#1734) 2024-02-21 17:23:38 +01:00
grpc-server.cpp fix(llama.cpp): correctly handle embeddings in batches (#4957) 2025-03-07 19:29:52 +01:00
json.hpp 🔥 add LaVA support and GPT vision API, Multiple requests for llama.cpp, return JSON types (#1254) 2023-11-11 13:14:59 +01:00
Makefile fix(llama.cpp): embed metal file into result binary for darwin (#4279) 2024-11-28 04:17:00 +00:00
prepare.sh chore(deps): bump llama.cpp to '10f2e81809bbb69ecfe64fc8b4686285f84b0c07' 2025-03-12 09:12:59 +01:00
utils.hpp chore(deps): update llama.cpp (#3497) 2024-09-12 20:55:27 +02:00