From dbc12d27c89f31e1939150344cac57885c8025ee Mon Sep 17 00:00:00 2001 From: hahuyhoang411 Date: Mon, 4 Dec 2023 12:32:56 +0700 Subject: [PATCH] update mistral 7b chat --- models/lzlv-70b/model.json | 4 ++-- models/mistral-ins-7b-q4/model.json | 24 ++++++++++++++++++++++++ models/mistral-ins-7b-q5/model.json | 24 ++++++++++++++++++++++++ 3 files changed, 50 insertions(+), 2 deletions(-) create mode 100644 models/mistral-ins-7b-q4/model.json create mode 100644 models/mistral-ins-7b-q5/model.json diff --git a/models/lzlv-70b/model.json b/models/lzlv-70b/model.json index 166e3d8ed..98d581f24 100644 --- a/models/lzlv-70b/model.json +++ b/models/lzlv-70b/model.json @@ -7,13 +7,13 @@ "description": "lzlv_70B is a sophisticated AI model designed for roleplaying and creative tasks. This merge aims to combine intelligence with creativity, seemingly outperforming its individual components in complex scenarios and creative outputs.", "format": "gguf", "settings": { - "ctx_len": 4096, + "ctx_len": 2048, "system_prompt": "", "user_prompt": "USER: ", "ai_prompt": "ASSISTANT: " }, "parameters": { - "max_tokens": 4096 + "max_tokens": 2048 }, "metadata": { "author": "lizpreciatior, The Bloke", diff --git a/models/mistral-ins-7b-q4/model.json b/models/mistral-ins-7b-q4/model.json new file mode 100644 index 000000000..51b5c231e --- /dev/null +++ b/models/mistral-ins-7b-q4/model.json @@ -0,0 +1,24 @@ +{ + "source_url": "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF/resolve/main/mistral-7b-instruct-v0.1.Q4_K_M.gguf", + "id": "mistral-ins-7b-q4", + "object": "model", + "name": "Mistral Instruct 7B Q4", + "version": "1.0", + "description": "This is a 4-bit quantized version of MistralAI's Mistral Instruct 7B model.", + "format": "gguf", + "settings": { + "ctx_len": 2048, + "system_prompt": "", + "user_prompt": "[INST]", + "ai_prompt": "[/INST]" + }, + "parameters": { + "max_tokens": 2048 + }, + "metadata": { + "author": "MistralAI, The Bloke", + "tags": ["Foundational Model", "General", "Code"], + "size": 4370000000 + } + } + \ No newline at end of file diff --git a/models/mistral-ins-7b-q5/model.json b/models/mistral-ins-7b-q5/model.json new file mode 100644 index 000000000..46c5f5378 --- /dev/null +++ b/models/mistral-ins-7b-q5/model.json @@ -0,0 +1,24 @@ +{ + "source_url": "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF/resolve/main/mistral-7b-instruct-v0.1.Q5_K_M.gguf", + "id": "mistral-ins-7b-q5", + "object": "model", + "name": "Mistral Instruct 7B Q5", + "version": "1.0", + "description": "This is a 5-bit quantized version of MistralAI's Mistral Instruct 7B model.", + "format": "gguf", + "settings": { + "ctx_len": 2048, + "system_prompt": "", + "user_prompt": "[INST]", + "ai_prompt": "[/INST]" + }, + "parameters": { + "max_tokens": 2048 + }, + "metadata": { + "author": "MistralAI, The Bloke", + "tags": ["Foundational Model", "General", "Code"], + "size": 5130000000 + } + } + \ No newline at end of file