From 698f1f5bfdac0412116983c3d8927638b2ecec6d Mon Sep 17 00:00:00 2001 From: hahuyhoang411 Date: Mon, 29 Jan 2024 10:33:23 +0700 Subject: [PATCH] add(stablezephyr): add new stable 3b model --- models/dolphin-phi-2/model.json | 12 +++++------ models/stable-zephyr-3b/model.json | 34 ++++++++++++++++++++++++++++++ 2 files changed, 40 insertions(+), 6 deletions(-) create mode 100644 models/stable-zephyr-3b/model.json diff --git a/models/dolphin-phi-2/model.json b/models/dolphin-phi-2/model.json index 3842e0a56..c7348c359 100644 --- a/models/dolphin-phi-2/model.json +++ b/models/dolphin-phi-2/model.json @@ -1,20 +1,20 @@ { "source": [ { - "url": "https://huggingface.co/TheBloke/dolphin-2_6-phi-2-GGUF/resolve/main/dolphin-2_6-phi-2.Q4_K_M.gguf", - "filename": "dolphin-2_6-phi-2.Q4_K_M.gguf" + "url": "https://huggingface.co/TheBloke/dolphin-2_6-phi-2-GGUF/resolve/main/dolphin-2_6-phi-2.Q8_0.gguf", + "filename": "dolphin-2_6-phi-2.Q8_0.gguf" } ], "id": "dolphin-phi-2", "object": "model", - "name": "Dolphin Phi-2 2.7B Q4", + "name": "Dolphin Phi-2 2.7B Q8", "version": "1.0", "description": "Dolphin Phi-2 is a 2.7B model, fine-tuned for chat, excelling in common sense and logical reasoning benchmarks.", "format": "gguf", "settings": { "ctx_len": 4096, "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant\n", - "llama_model_path": "dolphin-2_6-phi-2.Q4_K_M.gguf" + "llama_model_path": "dolphin-2_6-phi-2.Q8_0.gguf" }, "parameters": { "max_tokens": 4096, @@ -24,9 +24,9 @@ "author": "Cognitive Computations, Microsoft", "tags": [ "3B", - "Chat Model" + "Finetuned" ], - "size": 1790000000 + "size": 2960000000 }, "engine": "nitro" } \ No newline at end of file diff --git a/models/stable-zephyr-3b/model.json b/models/stable-zephyr-3b/model.json new file mode 100644 index 000000000..f5df3dd76 --- /dev/null +++ b/models/stable-zephyr-3b/model.json @@ -0,0 +1,34 @@ +{ + "source": [ + { + "url": "https://huggingface.co/TheBloke/stablelm-zephyr-3b-GGUF/resolve/main/stablelm-zephyr-3b.Q8_0.gguf", + "filename": "stablelm-zephyr-3b.Q8_0.gguf" + } + ], + "id": "stable-zephyr-3b", + "object": "model", + "name": "Stable Zephyr 3B Q8", + "version": "1.0", + "description": "StableLM Zephyr 3B is trained on a mix of publicly and synthetic datasets. It's focus on safety and reliability.", + "format": "gguf", + "settings": { + "ctx_len": 4096, + "prompt_template": "<|user|>\n{prompt}<|endoftext|>\n<|assistant|>", + "llama_model_path": "stablelm-zephyr-3b.Q8_0.gguf" + }, + "parameters": { + "temperature": 0.7, + "top_p": 0.95, + "stream": true, + "max_tokens": 4096, + "stop": ["<|endoftext|>"], + "frequency_penalty": 0, + "presence_penalty": 0 + }, + "metadata": { + "author": "StabilityAI", + "tags": ["3B", "Finetuned"], + "size": 2970000000 + }, + "engine": "nitro" + } \ No newline at end of file