chore(deps): bump llama.cpp to 'f01bd02376f919b05ee635f438311be8dfc91d7c (#5110)

chore(deps): bump llama.cpp to 'f01bd02376f919b05ee635f438311be8dfc91d7c'

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
This commit is contained in:
Ettore Di Giacinto 2025-04-03 10:23:14 +02:00 committed by GitHub
parent 89e151f035
commit 18b320d577
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
2 changed files with 9 additions and 9 deletions

View file

@ -6,7 +6,7 @@ BINARY_NAME=local-ai
DETECT_LIBS?=true DETECT_LIBS?=true
# llama.cpp versions # llama.cpp versions
CPPLLAMA_VERSION?=f423981ac806bf031d83784bcb47d2721bc70f97 CPPLLAMA_VERSION?=f01bd02376f919b05ee635f438311be8dfc91d7c
# whisper.cpp version # whisper.cpp version
WHISPER_REPO?=https://github.com/ggerganov/whisper.cpp WHISPER_REPO?=https://github.com/ggerganov/whisper.cpp

View file

@ -509,15 +509,15 @@ struct llama_server_context
bool load_model(const common_params &params_) bool load_model(const common_params &params_)
{ {
params = params_; params = params_;
if (!params.mmproj.empty()) { if (!params.mmproj.path.empty()) {
multimodal = true; multimodal = true;
LOG_INFO("Multi Modal Mode Enabled", {}); LOG_INFO("Multi Modal Mode Enabled", {});
clp_ctx = clip_init(params.mmproj.c_str(), clip_context_params { clp_ctx = clip_init(params.mmproj.path.c_str(), clip_context_params {
/* use_gpu */ has_gpu, /* use_gpu */ has_gpu,
/*verbosity=*/ 1, /*verbosity=*/ 1,
}); });
if(clp_ctx == nullptr) { if(clp_ctx == nullptr) {
LOG_ERR("unable to load clip model: %s", params.mmproj.c_str()); LOG_ERR("unable to load clip model: %s", params.mmproj.path.c_str());
return false; return false;
} }
@ -531,7 +531,7 @@ struct llama_server_context
ctx = common_init.context.release(); ctx = common_init.context.release();
if (model == nullptr) if (model == nullptr)
{ {
LOG_ERR("unable to load model: %s", params.model.c_str()); LOG_ERR("unable to load model: %s", params.model.path.c_str());
return false; return false;
} }
@ -2326,11 +2326,11 @@ static void params_parse(const backend::ModelOptions* request,
// this is comparable to: https://github.com/ggerganov/llama.cpp/blob/d9b33fe95bd257b36c84ee5769cc048230067d6f/examples/server/server.cpp#L1809 // this is comparable to: https://github.com/ggerganov/llama.cpp/blob/d9b33fe95bd257b36c84ee5769cc048230067d6f/examples/server/server.cpp#L1809
params.model = request->modelfile(); params.model.path = request->modelfile();
if (!request->mmproj().empty()) { if (!request->mmproj().empty()) {
// get the directory of modelfile // get the directory of modelfile
std::string model_dir = params.model.substr(0, params.model.find_last_of("/\\")); std::string model_dir = params.model.path.substr(0, params.model.path.find_last_of("/\\"));
params.mmproj = model_dir + "/"+ request->mmproj(); params.mmproj.path = model_dir + "/"+ request->mmproj();
} }
// params.model_alias ?? // params.model_alias ??
params.model_alias = request->modelfile(); params.model_alias = request->modelfile();
@ -2405,7 +2405,7 @@ static void params_parse(const backend::ModelOptions* request,
scale_factor = request->lorascale(); scale_factor = request->lorascale();
} }
// get the directory of modelfile // get the directory of modelfile
std::string model_dir = params.model.substr(0, params.model.find_last_of("/\\")); std::string model_dir = params.model.path.substr(0, params.model.path.find_last_of("/\\"));
params.lora_adapters.push_back({ model_dir + "/"+request->loraadapter(), scale_factor }); params.lora_adapters.push_back({ model_dir + "/"+request->loraadapter(), scale_factor });
} }
params.use_mlock = request->mlock(); params.use_mlock = request->mlock();