From 585e0745dae17670d0067038739ce7a5347976de Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Wed, 16 Oct 2024 18:28:51 +0200 Subject: [PATCH 1/5] models(gallery): add astral-fusion-neural-happy-l3.1-8b (#3848) Signed-off-by: Ettore Di Giacinto --- gallery/index.yaml | 16 ++++++++++++++++ 1 file changed, 16 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index 198d8b2b90f7..80fec85b51a8 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -1620,6 +1620,22 @@ - filename: Doctoraifinetune-3.1-8B.i1-Q4_K_M.gguf sha256: 282456efcb6c7e54d34ac25ae7fc022a94152ed77281ae4625b9628091e0a3d6 uri: huggingface://mradermacher/Doctoraifinetune-3.1-8B-i1-GGUF/Doctoraifinetune-3.1-8B.i1-Q4_K_M.gguf +- !!merge <<: *llama31 + name: "astral-fusion-neural-happy-l3.1-8b" + urls: + - https://huggingface.co/ZeroXClem/Astral-Fusion-Neural-Happy-L3.1-8B + - https://huggingface.co/mradermacher/Astral-Fusion-Neural-Happy-L3.1-8B-GGUF + description: | + Astral-Fusion-Neural-Happy-L3.1-8B is a celestial blend of magic, creativity, and dynamic storytelling. Designed to excel in instruction-following, immersive roleplaying, and magical narrative generation, this model is a fusion of the finest qualities from Astral-Fusion, NIHAPPY, and NeuralMahou. ✨🚀 + + This model is perfect for anyone seeking a cosmic narrative experience, with the ability to generate both precise instructional content and fantastical stories in one cohesive framework. Whether you're crafting immersive stories, creating AI roleplaying characters, or working on interactive storytelling, this model brings out the magic. 🌟 + overrides: + parameters: + model: Astral-Fusion-Neural-Happy-L3.1-8B.Q4_K_M.gguf + files: + - filename: Astral-Fusion-Neural-Happy-L3.1-8B.Q4_K_M.gguf + sha256: 14a3b07c1723ef1ca24f99382254b1227d95974541e23792a4e7ff621896055d + uri: huggingface://mradermacher/Astral-Fusion-Neural-Happy-L3.1-8B-GGUF/Astral-Fusion-Neural-Happy-L3.1-8B.Q4_K_M.gguf - &deepseek ## Deepseek url: "github:mudler/LocalAI/gallery/deepseek.yaml@master" From 773cec77a2aa63094c8c6b0b5e287ee4cd635a99 Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Wed, 16 Oct 2024 18:31:50 +0200 Subject: [PATCH 2/5] models(gallery): add tsunami-0.5x-7b-instruct-i1 (#3849) Signed-off-by: Ettore Di Giacinto --- gallery/index.yaml | 20 ++++++++++++++++++++ 1 file changed, 20 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index 80fec85b51a8..c2e105c43b13 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -572,6 +572,26 @@ - filename: EdgeRunner-Command-Nested.i1-Q4_K_M.gguf sha256: a1cc4d2b601dc20e58cbb549bd3e9bc460995840c0aaf1cd3c1cb5414c900ac7 uri: huggingface://mradermacher/EdgeRunner-Command-Nested-i1-GGUF/EdgeRunner-Command-Nested.i1-Q4_K_M.gguf +- !!merge <<: *qwen25 + name: "tsunami-0.5x-7b-instruct-i1" + icon: https://huggingface.co/Tsunami-th/Tsunami-0.5x-7B-Instruct/resolve/main/Tsunami.webp + urls: + - https://huggingface.co/Tsunami-th/Tsunami-0.5x-7B-Instruct + - https://huggingface.co/mradermacher/Tsunami-0.5x-7B-Instruct-i1-GGUF + description: | + TSUNAMI: Transformative Semantic Understanding and Natural Augmentation Model for Intelligence. + + TSUNAMI full name was created by ChatGPT. + infomation + + Tsunami-0.5x-7B-Instruct is Thai Large Language Model that fine-tuned from Qwen2.5-7B around 100,000 rows in Thai dataset. + overrides: + parameters: + model: Tsunami-0.5x-7B-Instruct.i1-Q4_K_M.gguf + files: + - filename: Tsunami-0.5x-7B-Instruct.i1-Q4_K_M.gguf + sha256: 22e2003ecec7f1e91f2e9aaec334613c0f37fb3000d0e628b5a9980e53322fa7 + uri: huggingface://mradermacher/Tsunami-0.5x-7B-Instruct-i1-GGUF/Tsunami-0.5x-7B-Instruct.i1-Q4_K_M.gguf - &archfunct license: apache-2.0 tags: From fdf1452c6b5076e351838c79478b9ba8f409b39c Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Wed, 16 Oct 2024 18:37:01 +0200 Subject: [PATCH 3/5] models(gallery): add mahou-1.5-llama3.1-70b-i1 (#3850) Signed-off-by: Ettore Di Giacinto --- gallery/index.yaml | 16 ++++++++++++++++ 1 file changed, 16 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index c2e105c43b13..b1dbf20b3175 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -1656,6 +1656,22 @@ - filename: Astral-Fusion-Neural-Happy-L3.1-8B.Q4_K_M.gguf sha256: 14a3b07c1723ef1ca24f99382254b1227d95974541e23792a4e7ff621896055d uri: huggingface://mradermacher/Astral-Fusion-Neural-Happy-L3.1-8B-GGUF/Astral-Fusion-Neural-Happy-L3.1-8B.Q4_K_M.gguf +- !!merge <<: *llama31 + name: "mahou-1.5-llama3.1-70b-i1" + url: "github:mudler/LocalAI/gallery/chatml.yaml@master" + icon: https://huggingface.co/flammenai/Mahou-1.0-mistral-7B/resolve/main/mahou1.png + urls: + - https://huggingface.co/flammenai/Mahou-1.5-llama3.1-70B + - https://huggingface.co/mradermacher/Mahou-1.5-llama3.1-70B-i1-GGUF + description: | + Mahou is designed to provide short messages in a conversational context. It is capable of casual conversation and character roleplay. + overrides: + parameters: + model: Mahou-1.5-llama3.1-70B.i1-Q4_K_M.gguf + files: + - filename: Mahou-1.5-llama3.1-70B.i1-Q4_K_M.gguf + sha256: c2711c4c9c8d011edbeaa391b4418d433e273a318d1de3dbdda9b85baf4996f2 + uri: huggingface://mradermacher/Mahou-1.5-llama3.1-70B-i1-GGUF/Mahou-1.5-llama3.1-70B.i1-Q4_K_M.gguf - &deepseek ## Deepseek url: "github:mudler/LocalAI/gallery/deepseek.yaml@master" From 1b44a5a3b7a94f102ec0bfe11d2cfdfc5afac788 Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Wed, 16 Oct 2024 18:39:28 +0200 Subject: [PATCH 4/5] chore(deps): bump grpcio to 1.67.0 (#3851) Signed-off-by: Ettore Di Giacinto --- backend/python/autogptq/requirements.txt | 2 +- backend/python/bark/requirements.txt | 2 +- backend/python/common/template/requirements.txt | 2 +- backend/python/coqui/requirements.txt | 2 +- backend/python/diffusers/requirements.txt | 2 +- backend/python/exllama2/requirements.txt | 2 +- backend/python/mamba/requirements.txt | 2 +- backend/python/openvoice/requirements-intel.txt | 2 +- backend/python/openvoice/requirements.txt | 2 +- backend/python/parler-tts/requirements.txt | 2 +- backend/python/rerankers/requirements.txt | 2 +- backend/python/sentencetransformers/requirements.txt | 2 +- backend/python/transformers-musicgen/requirements.txt | 2 +- backend/python/transformers/requirements.txt | 2 +- backend/python/vall-e-x/requirements.txt | 2 +- backend/python/vllm/install.sh | 2 +- backend/python/vllm/requirements.txt | 2 +- 17 files changed, 17 insertions(+), 17 deletions(-) diff --git a/backend/python/autogptq/requirements.txt b/backend/python/autogptq/requirements.txt index 9cb6ce94f589..7e66f084958e 100644 --- a/backend/python/autogptq/requirements.txt +++ b/backend/python/autogptq/requirements.txt @@ -1,6 +1,6 @@ accelerate auto-gptq==0.7.1 -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi transformers \ No newline at end of file diff --git a/backend/python/bark/requirements.txt b/backend/python/bark/requirements.txt index 6e46924ae77b..d1a90719f9f4 100644 --- a/backend/python/bark/requirements.txt +++ b/backend/python/bark/requirements.txt @@ -1,4 +1,4 @@ bark==0.1.5 -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi \ No newline at end of file diff --git a/backend/python/common/template/requirements.txt b/backend/python/common/template/requirements.txt index 540c0eb56cf6..16716764c30a 100644 --- a/backend/python/common/template/requirements.txt +++ b/backend/python/common/template/requirements.txt @@ -1,2 +1,2 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf \ No newline at end of file diff --git a/backend/python/coqui/requirements.txt b/backend/python/coqui/requirements.txt index 29484f7d26f6..0ad62f7027e9 100644 --- a/backend/python/coqui/requirements.txt +++ b/backend/python/coqui/requirements.txt @@ -1,4 +1,4 @@ coqui-tts -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi \ No newline at end of file diff --git a/backend/python/diffusers/requirements.txt b/backend/python/diffusers/requirements.txt index 730e316f68dd..624b048e7848 100644 --- a/backend/python/diffusers/requirements.txt +++ b/backend/python/diffusers/requirements.txt @@ -1,5 +1,5 @@ setuptools -grpcio==1.66.2 +grpcio==1.67.0 pillow protobuf certifi diff --git a/backend/python/exllama2/requirements.txt b/backend/python/exllama2/requirements.txt index e3db2b2f2016..8a0d9a17d563 100644 --- a/backend/python/exllama2/requirements.txt +++ b/backend/python/exllama2/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi wheel diff --git a/backend/python/mamba/requirements.txt b/backend/python/mamba/requirements.txt index 83ae4279db55..6be5d8ac6b89 100644 --- a/backend/python/mamba/requirements.txt +++ b/backend/python/mamba/requirements.txt @@ -1,3 +1,3 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi \ No newline at end of file diff --git a/backend/python/openvoice/requirements-intel.txt b/backend/python/openvoice/requirements-intel.txt index 687efe78b52b..b446386f9761 100644 --- a/backend/python/openvoice/requirements-intel.txt +++ b/backend/python/openvoice/requirements-intel.txt @@ -2,7 +2,7 @@ intel-extension-for-pytorch torch optimum[openvino] -grpcio==1.66.2 +grpcio==1.67.0 protobuf librosa==0.9.1 faster-whisper==1.0.3 diff --git a/backend/python/openvoice/requirements.txt b/backend/python/openvoice/requirements.txt index 6ee29ce4d185..fd1268d0c215 100644 --- a/backend/python/openvoice/requirements.txt +++ b/backend/python/openvoice/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf librosa faster-whisper diff --git a/backend/python/parler-tts/requirements.txt b/backend/python/parler-tts/requirements.txt index d7f36feb3837..ff9adca994b2 100644 --- a/backend/python/parler-tts/requirements.txt +++ b/backend/python/parler-tts/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi llvmlite==0.43.0 \ No newline at end of file diff --git a/backend/python/rerankers/requirements.txt b/backend/python/rerankers/requirements.txt index 83ae4279db55..6be5d8ac6b89 100644 --- a/backend/python/rerankers/requirements.txt +++ b/backend/python/rerankers/requirements.txt @@ -1,3 +1,3 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi \ No newline at end of file diff --git a/backend/python/sentencetransformers/requirements.txt b/backend/python/sentencetransformers/requirements.txt index 40a387f12c20..36ce8a6f30e5 100644 --- a/backend/python/sentencetransformers/requirements.txt +++ b/backend/python/sentencetransformers/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi datasets diff --git a/backend/python/transformers-musicgen/requirements.txt b/backend/python/transformers-musicgen/requirements.txt index a3f666512907..ea0e3fa9632e 100644 --- a/backend/python/transformers-musicgen/requirements.txt +++ b/backend/python/transformers-musicgen/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf scipy==1.14.0 certifi \ No newline at end of file diff --git a/backend/python/transformers/requirements.txt b/backend/python/transformers/requirements.txt index 084cc034e7b8..d006cf0e1e44 100644 --- a/backend/python/transformers/requirements.txt +++ b/backend/python/transformers/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi setuptools==69.5.1 # https://github.com/mudler/LocalAI/issues/2406 \ No newline at end of file diff --git a/backend/python/vall-e-x/requirements.txt b/backend/python/vall-e-x/requirements.txt index 83ae4279db55..6be5d8ac6b89 100644 --- a/backend/python/vall-e-x/requirements.txt +++ b/backend/python/vall-e-x/requirements.txt @@ -1,3 +1,3 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi \ No newline at end of file diff --git a/backend/python/vllm/install.sh b/backend/python/vllm/install.sh index 9078b81b84a1..69e74a0624be 100755 --- a/backend/python/vllm/install.sh +++ b/backend/python/vllm/install.sh @@ -22,7 +22,7 @@ if [ "x${BUILD_TYPE}" == "x" ] && [ "x${FROM_SOURCE}" == "xtrue" ]; then git clone https://github.com/vllm-project/vllm fi pushd vllm - uv pip install wheel packaging ninja "setuptools>=49.4.0" numpy typing-extensions pillow setuptools-scm grpcio==1.66.2 protobuf bitsandbytes + uv pip install wheel packaging ninja "setuptools>=49.4.0" numpy typing-extensions pillow setuptools-scm grpcio==1.67.0 protobuf bitsandbytes uv pip install -v -r requirements-cpu.txt --extra-index-url https://download.pytorch.org/whl/cpu VLLM_TARGET_DEVICE=cpu python setup.py install popd diff --git a/backend/python/vllm/requirements.txt b/backend/python/vllm/requirements.txt index 8fb8a4185fe7..95447f746da3 100644 --- a/backend/python/vllm/requirements.txt +++ b/backend/python/vllm/requirements.txt @@ -1,4 +1,4 @@ -grpcio==1.66.2 +grpcio==1.67.0 protobuf certifi setuptools \ No newline at end of file From a60b9b7a381d5ed7cd657f4d0ba042b931215e50 Mon Sep 17 00:00:00 2001 From: "LocalAI [bot]" <139863280+localai-bot@users.noreply.github.com> Date: Wed, 16 Oct 2024 23:41:30 +0200 Subject: [PATCH 5/5] chore: :arrow_up: Update ggerganov/llama.cpp to `9e041024481f6b249ab8918e18b9477f873b5a5e` (#3853) :arrow_up: Update ggerganov/llama.cpp Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: mudler <2420543+mudler@users.noreply.github.com> --- Makefile | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Makefile b/Makefile index 34da8b57002e..aea5e1572542 100644 --- a/Makefile +++ b/Makefile @@ -8,7 +8,7 @@ DETECT_LIBS?=true # llama.cpp versions GOLLAMA_REPO?=https://github.com/go-skynet/go-llama.cpp GOLLAMA_VERSION?=2b57a8ae43e4699d3dc5d1496a1ccd42922993be -CPPLLAMA_VERSION?=755a9b2bf00fbae988e03a47e852b66eaddd113a +CPPLLAMA_VERSION?=9e041024481f6b249ab8918e18b9477f873b5a5e # go-rwkv version RWKV_REPO?=https://github.com/donomii/go-rwkv.cpp