From 6c6d840e6b273e43eacd0ce05dbffa9932b6b3f5 Mon Sep 17 00:00:00 2001 From: "LocalAI [bot]" <139863280+localai-bot@users.noreply.github.com> Date: Sat, 4 Jan 2025 22:43:08 +0100 Subject: [PATCH 1/5] chore: :arrow_up: Update ggerganov/llama.cpp to `9394bbd484f802ce80d2858033583af3ef700d25` (#4536) :arrow_up: Update ggerganov/llama.cpp Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: mudler <2420543+mudler@users.noreply.github.com> --- Makefile | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Makefile b/Makefile index fd9c7627..80458a29 100644 --- a/Makefile +++ b/Makefile @@ -8,7 +8,7 @@ DETECT_LIBS?=true # llama.cpp versions GOLLAMA_REPO?=https://github.com/go-skynet/go-llama.cpp GOLLAMA_VERSION?=2b57a8ae43e4699d3dc5d1496a1ccd42922993be -CPPLLAMA_VERSION?=4b0c638b9a68f577cb2066b638c9f622d91ee661 +CPPLLAMA_VERSION?=9394bbd484f802ce80d2858033583af3ef700d25 # whisper.cpp version WHISPER_REPO?=https://github.com/ggerganov/whisper.cpp From a10033e8a4ca429a30d51004c229bb9a5f11e892 Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Sun, 5 Jan 2025 09:12:12 +0100 Subject: [PATCH 2/5] chore(model gallery): add experimental-lwd-mirau-rp-14b-iq-imatrix (#4539) Signed-off-by: Ettore Di Giacinto --- gallery/index.yaml | 27 +++++++++++++++++++++++++++ 1 file changed, 27 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index 0242b5ff..3126e8fb 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -2596,6 +2596,33 @@ - filename: DRT-o1-7B.Q4_K_M.gguf sha256: f592a2523f92ae29630b45fbb501bba7f2fbd99355975cd05fa989faf8d3597d uri: huggingface://QuantFactory/DRT-o1-7B-GGUF/DRT-o1-7B.Q4_K_M.gguf +- !!merge <<: *qwen25 + name: "experimental-lwd-mirau-rp-14b-iq-imatrix" + icon: https://cdn-uploads.huggingface.co/production/uploads/65d4cf2693a0a3744a27536c/99YhsFSeaGDYCq7XVcTcq.png + urls: + - https://huggingface.co/AetherArchitectural/lwd-Mirau-RP-14B + - https://huggingface.co/Lewdiculous/experimental-lwd-Mirau-RP-14B-GGUF-IQ-Imatrix + description: | + This model is designed to improve the controllability and consistency of current roleplaying models. We developed a story flow thought chain approach that makes the system prompts combined with the entire user-BOT dialogue read like a first-person narrative told by the BOT. We found this design greatly enhances the model's consistency and expressiveness. + + Additionally, we allow users to play two roles simultaneously: one as the director of the entire plot (see Special Designs), and another as an actor dialoguing with the BOT. Users can be viewed as writers who need to draft outlines and plot summaries, while the BOT helps complete story details, requiring users to have powerful control over the BOT. + + The model's output is divided into two parts: the model's inner monologue (which it believes is invisible to users) and the final response. + + Overall, mirau features: + + Superior character consistency + + Powerful long-context memory capability + + Transparent thinking with hidden thought chains + overrides: + parameters: + model: lwd-Mirau-RP-Q4_K_M-imat.gguf + files: + - filename: lwd-Mirau-RP-Q4_K_M-imat.gguf + sha256: 22ff461e9034b9ebded07b2a9d3d88c2f75359d5c069ebb3ee4e9c6ec5c45cf8 + uri: huggingface://Lewdiculous/experimental-lwd-Mirau-RP-14B-GGUF-IQ-Imatrix/lwd-Mirau-RP-Q4_K_M-imat.gguf - &smollm ## SmolLM url: "github:mudler/LocalAI/gallery/chatml.yaml@master" From e845cc040181983fcf79e4c17b786c137bf53eee Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Sun, 5 Jan 2025 09:19:05 +0100 Subject: [PATCH 3/5] chore(model gallery): add llama-deepsync-3b (#4540) Signed-off-by: Ettore Di Giacinto --- gallery/index.yaml | 14 ++++++++++++++ 1 file changed, 14 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index 3126e8fb..89a4118f 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -1055,6 +1055,20 @@ - filename: Codepy-Deepthink-3B.Q4_K_M.gguf sha256: 6202976de1a1b23bb09448dd6f188b849e10f3f99366f829415533ea4445e853 uri: huggingface://QuantFactory/Codepy-Deepthink-3B-GGUF/Codepy-Deepthink-3B.Q4_K_M.gguf +- !!merge <<: *llama32 + name: "llama-deepsync-3b" + urls: + - https://huggingface.co/prithivMLmods/Llama-Deepsync-3B + - https://huggingface.co/prithivMLmods/Llama-Deepsync-3B-GGUF + description: | + The Llama-Deepsync-3B-GGUF is a fine-tuned version of the Llama-3.2-3B-Instruct base model, designed for text generation tasks that require deep reasoning, logical structuring, and problem-solving. This model leverages its optimized architecture to provide accurate and contextually relevant outputs for complex queries, making it ideal for applications in education, programming, and creative writing. + overrides: + parameters: + model: Llama-Deepsync-3B.Q4_K_M.gguf + files: + - filename: Llama-Deepsync-3B.Q4_K_M.gguf + sha256: f11c4d9b10a732845d8e64dc9badfcbb7d94053bc5fe11f89bb8e99ed557f711 + uri: huggingface://prithivMLmods/Llama-Deepsync-3B-GGUF/Llama-Deepsync-3B.Q4_K_M.gguf - &qwen25 ## Qwen2.5 name: "qwen2.5-14b-instruct" From b5992255ac62927960072096820131222c56e13a Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Sun, 5 Jan 2025 09:22:00 +0100 Subject: [PATCH 4/5] chore(model gallery): add qwentile2.5-32b-instruct (#4541) Signed-off-by: Ettore Di Giacinto --- gallery/index.yaml | 25 +++++++++++++++++++++++++ 1 file changed, 25 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index 89a4118f..1a194c9d 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -2387,6 +2387,31 @@ - filename: miscii-14b-1225.Q4_K_M.gguf sha256: f21fe73450be394055aeb87b7619e98a09e5c190b48f145bdebef4e12df871fe uri: huggingface://mradermacher/miscii-14b-1225-GGUF/miscii-14b-1225.Q4_K_M.gguf +- !!merge <<: *qwen25 + name: "qwentile2.5-32b-instruct" + icon: https://cdn-uploads.huggingface.co/production/uploads/65b19c1b098c85365af5a83e/sF7RDZA7lFYOmGy4bGy1s.png + urls: + - https://huggingface.co/maldv/Qwentile2.5-32B-Instruct + - https://huggingface.co/bartowski/Qwentile2.5-32B-Instruct-GGUF + description: | + Qwentile 2.5 32B Instruct is a normalized denoised fourier interpolation of the following models: + - { "model": "AiCloser/Qwen2.5-32B-AGI", "base": "Qwen/Qwen2.5-32B", "alpha": 0.3 } + - { "model": "EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2", "base": "Qwen/Qwen2.5-32B", "alpha": 0.7 } + - { "model": "fblgit/TheBeagle-v2beta-32B-MGS", "base": "Qwen/Qwen2.5-32B", "alpha": 0.6 } + - { "model": "huihui-ai/Qwen2.5-32B-Instruct-abliterated", "base": "Qwen/Qwen2.5-32B-Instruct", "alpha": 1.0 } + - { "model": "huihui-ai/QwQ-32B-Preview-abliterated", "base": "Qwen/Qwen2.5-32B", "alpha": 1.0 } + - { "model": "Qwen/QwQ-32B-Preview", "base": "Qwen/Qwen2.5-32B", "alpha": 0.8, "is_input": true } + - { "model": "rombodawg/Rombos-LLM-V2.5-Qwen-32b", "base": "Qwen/Qwen2.5-32B", "alpha": 1.0, "is_output": true } + - { "model": "nbeerbower/Qwen2.5-Gutenberg-Doppel-32B", "base": "Qwen/Qwen2.5-32B-Instruct", "alpha": 0.4 } + I started my experiment because of QwQ is a really nifty model, but it was giving me problems with xml output - which is what I use for my thought tokens. So, I thought... lets just merge it in! + The first model worked pretty well, but I got a sense that the balances could be tweaked. Why not throw in some other models as well for fun and see if I can't run out of disk space in the process? + overrides: + parameters: + model: Qwentile2.5-32B-Instruct-Q4_K_M.gguf + files: + - filename: Qwentile2.5-32B-Instruct-Q4_K_M.gguf + sha256: e476d6e3c15c78fc3f986d7ae8fa35c16116843827f2e6243c05767cef2f3615 + uri: huggingface://bartowski/Qwentile2.5-32B-Instruct-GGUF/Qwentile2.5-32B-Instruct-Q4_K_M.gguf - &archfunct license: apache-2.0 tags: From 0d7550ad5451c4ca1ccc2794d1d5e598a830208e Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Sun, 5 Jan 2025 15:01:49 +0100 Subject: [PATCH 5/5] chore(deps): bump grpcio to 1.69.0 (#4543) Signed-off-by: Ettore Di Giacinto --- backend/python/autogptq/requirements.txt | 2 +- backend/python/bark/requirements.txt | 2 +- backend/python/common/template/requirements.txt | 2 +- backend/python/coqui/requirements.txt | 2 +- backend/python/diffusers/requirements.txt | 2 +- backend/python/exllama2/requirements.txt | 2 +- backend/python/mamba/requirements.txt | 2 +- backend/python/openvoice/requirements-intel.txt | 2 +- backend/python/openvoice/requirements.txt | 2 +- backend/python/parler-tts/requirements.txt | 2 +- backend/python/rerankers/requirements.txt | 2 +- backend/python/sentencetransformers/requirements.txt | 2 +- backend/python/transformers-musicgen/requirements.txt | 2 +- backend/python/transformers/requirements.txt | 2 +- backend/python/vall-e-x/requirements.txt | 2 +- backend/python/vllm/requirements.txt | 2 +- 16 files changed, 16 insertions(+), 16 deletions(-) diff --git a/backend/python/autogptq/requirements.txt b/backend/python/autogptq/requirements.txt index 22408f47..c857a867 100644 --- a/backend/python/autogptq/requirements.txt +++ b/backend/python/autogptq/requirements.txt @@ -1,6 +1,6 @@ accelerate auto-gptq==0.7.1 -grpcio==1.68.1 +grpcio==1.69.0 protobuf certifi transformers \ No newline at end of file diff --git a/backend/python/bark/requirements.txt b/backend/python/bark/requirements.txt index 3fca1de5..81c1273d 100644 --- a/backend/python/bark/requirements.txt +++ b/backend/python/bark/requirements.txt @@ -1,4 +1,4 @@ bark==0.1.5 -grpcio==1.68.1 +grpcio==1.69.0 protobuf certifi \ No newline at end of file diff --git a/backend/python/common/template/requirements.txt b/backend/python/common/template/requirements.txt index 893dc812..0f43df10 100644 --- a/backend/python/common/template/requirements.txt +++ b/backend/python/common/template/requirements.txt @@ -1,3 +1,3 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf grpcio-tools \ No newline at end of file diff --git a/backend/python/coqui/requirements.txt b/backend/python/coqui/requirements.txt index 57638588..76c9ba4b 100644 --- a/backend/python/coqui/requirements.txt +++ b/backend/python/coqui/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf certifi packaging==24.1 \ No newline at end of file diff --git a/backend/python/diffusers/requirements.txt b/backend/python/diffusers/requirements.txt index 71832ead..d49155ed 100644 --- a/backend/python/diffusers/requirements.txt +++ b/backend/python/diffusers/requirements.txt @@ -1,5 +1,5 @@ setuptools -grpcio==1.68.1 +grpcio==1.69.0 pillow protobuf certifi diff --git a/backend/python/exllama2/requirements.txt b/backend/python/exllama2/requirements.txt index 408eb318..77464406 100644 --- a/backend/python/exllama2/requirements.txt +++ b/backend/python/exllama2/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf certifi wheel diff --git a/backend/python/mamba/requirements.txt b/backend/python/mamba/requirements.txt index 8e4eabf1..afc8b2a9 100644 --- a/backend/python/mamba/requirements.txt +++ b/backend/python/mamba/requirements.txt @@ -1,3 +1,3 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf certifi \ No newline at end of file diff --git a/backend/python/openvoice/requirements-intel.txt b/backend/python/openvoice/requirements-intel.txt index 43fad1ad..39b2b8b0 100644 --- a/backend/python/openvoice/requirements-intel.txt +++ b/backend/python/openvoice/requirements-intel.txt @@ -4,7 +4,7 @@ torch==2.3.1+cxx11.abi torchaudio==2.3.1+cxx11.abi oneccl_bind_pt==2.3.100+xpu optimum[openvino] -grpcio==1.68.1 +grpcio==1.69.0 protobuf librosa==0.9.1 faster-whisper==0.9.0 diff --git a/backend/python/openvoice/requirements.txt b/backend/python/openvoice/requirements.txt index a105da94..62b886bb 100644 --- a/backend/python/openvoice/requirements.txt +++ b/backend/python/openvoice/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf librosa faster-whisper diff --git a/backend/python/parler-tts/requirements.txt b/backend/python/parler-tts/requirements.txt index faf4ea3d..e6ba016b 100644 --- a/backend/python/parler-tts/requirements.txt +++ b/backend/python/parler-tts/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.68.1 +grpcio==1.69.0 certifi llvmlite==0.43.0 setuptools \ No newline at end of file diff --git a/backend/python/rerankers/requirements.txt b/backend/python/rerankers/requirements.txt index 8e4eabf1..afc8b2a9 100644 --- a/backend/python/rerankers/requirements.txt +++ b/backend/python/rerankers/requirements.txt @@ -1,3 +1,3 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf certifi \ No newline at end of file diff --git a/backend/python/sentencetransformers/requirements.txt b/backend/python/sentencetransformers/requirements.txt index b9dacf9b..6e03c63f 100644 --- a/backend/python/sentencetransformers/requirements.txt +++ b/backend/python/sentencetransformers/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf certifi datasets diff --git a/backend/python/transformers-musicgen/requirements.txt b/backend/python/transformers-musicgen/requirements.txt index 2e46b08f..f58e1e80 100644 --- a/backend/python/transformers-musicgen/requirements.txt +++ b/backend/python/transformers-musicgen/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf scipy==1.14.0 certifi \ No newline at end of file diff --git a/backend/python/transformers/requirements.txt b/backend/python/transformers/requirements.txt index d981fd99..a1eea776 100644 --- a/backend/python/transformers/requirements.txt +++ b/backend/python/transformers/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf certifi setuptools \ No newline at end of file diff --git a/backend/python/vall-e-x/requirements.txt b/backend/python/vall-e-x/requirements.txt index d981fd99..a1eea776 100644 --- a/backend/python/vall-e-x/requirements.txt +++ b/backend/python/vall-e-x/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf certifi setuptools \ No newline at end of file diff --git a/backend/python/vllm/requirements.txt b/backend/python/vllm/requirements.txt index d981fd99..a1eea776 100644 --- a/backend/python/vllm/requirements.txt +++ b/backend/python/vllm/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.68.1 +grpcio==1.69.0 protobuf certifi setuptools \ No newline at end of file