mirror of
https://github.com/mudler/LocalAI.git
synced 2025-05-20 10:35:01 +00:00
chore(deps): bump llama.cpp to 'f01bd02376f919b05ee635f438311be8dfc91d7c (#5110)
chore(deps): bump llama.cpp to 'f01bd02376f919b05ee635f438311be8dfc91d7c' Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
This commit is contained in:
parent
89e151f035
commit
18b320d577
2 changed files with 9 additions and 9 deletions
2
Makefile
2
Makefile
|
@ -6,7 +6,7 @@ BINARY_NAME=local-ai
|
||||||
DETECT_LIBS?=true
|
DETECT_LIBS?=true
|
||||||
|
|
||||||
# llama.cpp versions
|
# llama.cpp versions
|
||||||
CPPLLAMA_VERSION?=f423981ac806bf031d83784bcb47d2721bc70f97
|
CPPLLAMA_VERSION?=f01bd02376f919b05ee635f438311be8dfc91d7c
|
||||||
|
|
||||||
# whisper.cpp version
|
# whisper.cpp version
|
||||||
WHISPER_REPO?=https://github.com/ggerganov/whisper.cpp
|
WHISPER_REPO?=https://github.com/ggerganov/whisper.cpp
|
||||||
|
|
|
@ -509,15 +509,15 @@ struct llama_server_context
|
||||||
bool load_model(const common_params ¶ms_)
|
bool load_model(const common_params ¶ms_)
|
||||||
{
|
{
|
||||||
params = params_;
|
params = params_;
|
||||||
if (!params.mmproj.empty()) {
|
if (!params.mmproj.path.empty()) {
|
||||||
multimodal = true;
|
multimodal = true;
|
||||||
LOG_INFO("Multi Modal Mode Enabled", {});
|
LOG_INFO("Multi Modal Mode Enabled", {});
|
||||||
clp_ctx = clip_init(params.mmproj.c_str(), clip_context_params {
|
clp_ctx = clip_init(params.mmproj.path.c_str(), clip_context_params {
|
||||||
/* use_gpu */ has_gpu,
|
/* use_gpu */ has_gpu,
|
||||||
/*verbosity=*/ 1,
|
/*verbosity=*/ 1,
|
||||||
});
|
});
|
||||||
if(clp_ctx == nullptr) {
|
if(clp_ctx == nullptr) {
|
||||||
LOG_ERR("unable to load clip model: %s", params.mmproj.c_str());
|
LOG_ERR("unable to load clip model: %s", params.mmproj.path.c_str());
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -531,7 +531,7 @@ struct llama_server_context
|
||||||
ctx = common_init.context.release();
|
ctx = common_init.context.release();
|
||||||
if (model == nullptr)
|
if (model == nullptr)
|
||||||
{
|
{
|
||||||
LOG_ERR("unable to load model: %s", params.model.c_str());
|
LOG_ERR("unable to load model: %s", params.model.path.c_str());
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2326,11 +2326,11 @@ static void params_parse(const backend::ModelOptions* request,
|
||||||
|
|
||||||
// this is comparable to: https://github.com/ggerganov/llama.cpp/blob/d9b33fe95bd257b36c84ee5769cc048230067d6f/examples/server/server.cpp#L1809
|
// this is comparable to: https://github.com/ggerganov/llama.cpp/blob/d9b33fe95bd257b36c84ee5769cc048230067d6f/examples/server/server.cpp#L1809
|
||||||
|
|
||||||
params.model = request->modelfile();
|
params.model.path = request->modelfile();
|
||||||
if (!request->mmproj().empty()) {
|
if (!request->mmproj().empty()) {
|
||||||
// get the directory of modelfile
|
// get the directory of modelfile
|
||||||
std::string model_dir = params.model.substr(0, params.model.find_last_of("/\\"));
|
std::string model_dir = params.model.path.substr(0, params.model.path.find_last_of("/\\"));
|
||||||
params.mmproj = model_dir + "/"+ request->mmproj();
|
params.mmproj.path = model_dir + "/"+ request->mmproj();
|
||||||
}
|
}
|
||||||
// params.model_alias ??
|
// params.model_alias ??
|
||||||
params.model_alias = request->modelfile();
|
params.model_alias = request->modelfile();
|
||||||
|
@ -2405,7 +2405,7 @@ static void params_parse(const backend::ModelOptions* request,
|
||||||
scale_factor = request->lorascale();
|
scale_factor = request->lorascale();
|
||||||
}
|
}
|
||||||
// get the directory of modelfile
|
// get the directory of modelfile
|
||||||
std::string model_dir = params.model.substr(0, params.model.find_last_of("/\\"));
|
std::string model_dir = params.model.path.substr(0, params.model.path.find_last_of("/\\"));
|
||||||
params.lora_adapters.push_back({ model_dir + "/"+request->loraadapter(), scale_factor });
|
params.lora_adapters.push_back({ model_dir + "/"+request->loraadapter(), scale_factor });
|
||||||
}
|
}
|
||||||
params.use_mlock = request->mlock();
|
params.use_mlock = request->mlock();
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue