From 83e2dd5dff7b36d8cc9528d63ed0468145ef79df Mon Sep 17 00:00:00 2001 From: "LocalAI [bot]" <139863280+localai-bot@users.noreply.github.com> Date: Sun, 19 Jan 2025 23:34:32 +0100 Subject: [PATCH 1/3] chore: :arrow_up: Update ggerganov/llama.cpp to `92bc493917d43b83e592349e138b54c90b1c3ea7` (#4640) :arrow_up: Update ggerganov/llama.cpp Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: mudler <2420543+mudler@users.noreply.github.com> --- Makefile | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Makefile b/Makefile index dfa91a15..7aaad492 100644 --- a/Makefile +++ b/Makefile @@ -8,7 +8,7 @@ DETECT_LIBS?=true # llama.cpp versions GOLLAMA_REPO?=https://github.com/go-skynet/go-llama.cpp GOLLAMA_VERSION?=2b57a8ae43e4699d3dc5d1496a1ccd42922993be -CPPLLAMA_VERSION?=a1649cc13f89946322358f92ea268ae1b7b5096c +CPPLLAMA_VERSION?=92bc493917d43b83e592349e138b54c90b1c3ea7 # whisper.cpp version WHISPER_REPO?=https://github.com/ggerganov/whisper.cpp From 30739d94a41139fe5c8cf68239cc7353d102c4fe Mon Sep 17 00:00:00 2001 From: Gianluca Boiano <491117+M0Rf30@users.noreply.github.com> Date: Mon, 20 Jan 2025 10:34:19 +0100 Subject: [PATCH 2/3] chore(model gallery): add InternLM3-8b-Q4_K_M (#4637) chore(model gallery): add InternLM3-8b-Q4_K_M Signed-off-by: Gianluca Boiano --- gallery/index.yaml | 27 ++++++++++++++++++++++++++- 1 file changed, 26 insertions(+), 1 deletion(-) diff --git a/gallery/index.yaml b/gallery/index.yaml index edd52725..61ecd107 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -10100,7 +10100,7 @@ urls: - https://huggingface.co/internlm/internlm2_5-7b-chat-1m - https://huggingface.co/bartowski/internlm2_5-7b-chat-1m-GGUF - icon: https://github.com/InternLM/InternLM/assets/22529082/b9788105-8892-4398-8b47-b513a292378e + icon: https://avatars.githubusercontent.com/u/135356492 tags: - internlm2 - gguf @@ -10121,6 +10121,31 @@ - filename: internlm2_5-7b-chat-1m-Q4_K_M.gguf uri: huggingface://bartowski/internlm2_5-7b-chat-1m-GGUF/internlm2_5-7b-chat-1m-Q4_K_M.gguf sha256: 10d5e18a4125f9d4d74a9284a21e0c820b150af06dee48665e54ff6e1be3a564 +### Internlm3 +- name: "internlm3-8b-instruct" + url: "github:mudler/LocalAI/gallery/chatml.yaml@master" + urls: + - https://huggingface.co/internlm/internlm3-8b-instruct + - https://huggingface.co/bartowski/internlm3-8b-instruct-GGUF + icon: https://avatars.githubusercontent.com/u/135356492 + tags: + - internlm3 + - gguf + - cpu + - gpu + description: | + InternLM3 has open-sourced an 8-billion parameter instruction model, InternLM3-8B-Instruct, designed for general-purpose usage and advanced reasoning. The model has the following characteristics: + + Enhanced performance at reduced cost: State-of-the-art performance on reasoning and knowledge-intensive tasks surpass models like Llama3.1-8B and Qwen2.5-7B. + + Deep thinking capability: InternLM3 supports both the deep thinking mode for solving complicated reasoning tasks via the long chain-of-thought and the normal response mode for fluent user interactions. + overrides: + parameters: + model: internlm3-8b-instruct-Q4_K_M.gguf + files: + - filename: internlm3-8b-instruct-Q4_K_M.gguf + uri: huggingface://bartowski/internlm3-8b-instruct-GGUF/internlm3-8b-instruct-Q4_K_M.gguf + sha256: 2a9644687318e8659c9cf9b40730d5cc2f5af06f786a50439c7c51359b23896e - &phi-3 ### START Phi-3 url: "github:mudler/LocalAI/gallery/phi-3-chat.yaml@master" From 390bb3f58bb5d878c852c71e473ae0754a8d817d Mon Sep 17 00:00:00 2001 From: Gianluca Boiano <491117+M0Rf30@users.noreply.github.com> Date: Mon, 20 Jan 2025 10:35:05 +0100 Subject: [PATCH 3/3] fix(model gallery): minicpm-v-2.6 is based on qwen2 (#4638) Signed-off-by: Gianluca Boiano --- gallery/index.yaml | 54 +++++++++++++++++++++++----------------------- 1 file changed, 27 insertions(+), 27 deletions(-) diff --git a/gallery/index.yaml b/gallery/index.yaml index 61ecd107..1c170f99 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -5583,6 +5583,33 @@ - filename: marco-o1-uncensored.Q4_K_M.gguf sha256: ad0440270a7254098f90779744d3e5b34fe49b7baf97c819909ba9c5648cc0d9 uri: huggingface://QuantFactory/marco-o1-uncensored-GGUF/marco-o1-uncensored.Q4_K_M.gguf +- !!merge <<: *qwen2 + name: "minicpm-v-2_6" + license: apache-2.0 + icon: https://raw.githubusercontent.com/OpenBMB/MiniCPM/main/assets/minicpm_logo.png + urls: + - https://huggingface.co/openbmb/MiniCPM-V-2_6-gguf + - https://huggingface.co/openbmb/MiniCPM-V-2_6 + description: | + MiniCPM-V 2.6 is the latest and most capable model in the MiniCPM-V series. The model is built on SigLip-400M and Qwen2-7B with a total of 8B parameters + tags: + - llm + - multimodal + - gguf + - gpu + - qwen2 + - cpu + overrides: + mmproj: minicpm-v-2_6-mmproj-f16.gguf + parameters: + model: minicpm-v-2_6-Q4_K_M.gguf + files: + - filename: minicpm-v-2_6-Q4_K_M.gguf + sha256: 3a4078d53b46f22989adbf998ce5a3fd090b6541f112d7e936eb4204a04100b1 + uri: huggingface://openbmb/MiniCPM-V-2_6-gguf/ggml-model-Q4_K_M.gguf + - filename: minicpm-v-2_6-mmproj-f16.gguf + sha256: f8a805e9e62085805c69c427287acefc284932eb4abfe6e1b1ce431d27e2f4e0 + uri: huggingface://openbmb/MiniCPM-V-2_6-gguf/mmproj-model-f16.gguf - &mistral03 ## START Mistral url: "github:mudler/LocalAI/gallery/mistral-0.3.yaml@master" @@ -9211,33 +9238,6 @@ - filename: minicpm-llama3-mmproj-f16.gguf sha256: 391d11736c3cd24a90417c47b0c88975e86918fcddb1b00494c4d715b08af13e uri: huggingface://openbmb/MiniCPM-Llama3-V-2_5-gguf/mmproj-model-f16.gguf -- !!merge <<: *llama3 - name: "minicpm-v-2_6" - license: apache-2.0 - icon: https://raw.githubusercontent.com/OpenBMB/MiniCPM/main/assets/minicpm_logo.png - urls: - - https://huggingface.co/openbmb/MiniCPM-V-2_6-gguf - - https://huggingface.co/openbmb/MiniCPM-V-2_6 - description: | - MiniCPM-V 2.6 is the latest and most capable model in the MiniCPM-V series. The model is built on SigLip-400M and Qwen2-7B with a total of 8B parameters - tags: - - llm - - multimodal - - gguf - - gpu - - llama3 - - cpu - overrides: - mmproj: minicpm-v-2_6-mmproj-f16.gguf - parameters: - model: minicpm-v-2_6-Q4_K_M.gguf - files: - - filename: minicpm-v-2_6-Q4_K_M.gguf - sha256: 3a4078d53b46f22989adbf998ce5a3fd090b6541f112d7e936eb4204a04100b1 - uri: huggingface://openbmb/MiniCPM-V-2_6-gguf/ggml-model-Q4_K_M.gguf - - filename: minicpm-v-2_6-mmproj-f16.gguf - sha256: f8a805e9e62085805c69c427287acefc284932eb4abfe6e1b1ce431d27e2f4e0 - uri: huggingface://openbmb/MiniCPM-V-2_6-gguf/mmproj-model-f16.gguf - !!merge <<: *llama3 name: "llama-3-cursedstock-v1.8-8b-iq-imatrix" urls: