diff --git a/extensions/model-extension/package.json b/extensions/model-extension/package.json index e7f113e49..733eee45f 100644 --- a/extensions/model-extension/package.json +++ b/extensions/model-extension/package.json @@ -1,6 +1,6 @@ { "name": "@janhq/model-extension", - "version": "1.0.17", + "version": "1.0.18", "description": "Model Management Extension provides model exploration and seamless downloads", "main": "dist/index.js", "module": "dist/module.js", diff --git a/models/dolphin-2.7-mixtral-8x7b/model.json b/models/dolphin-2.7-mixtral-8x7b/model.json new file mode 100644 index 000000000..01f42d9b3 --- /dev/null +++ b/models/dolphin-2.7-mixtral-8x7b/model.json @@ -0,0 +1,22 @@ +{ + "source_url": "https://huggingface.co/TheBloke/dolphin-2.7-mixtral-8x7b-GGUF/resolve/main/dolphin-2.7-mixtral-8x7b.Q4_K_M.gguf", + "id": "dolphin-2.7-mixtral-8x7b", + "object": "model", + "name": "Dolphin 8x7B Q4", + "version": "1.0", + "description": "This model is an uncensored model based on Mixtral-8x7b. Dolphin is really good at coding", + "format": "gguf", + "settings": { + "ctx_len": 4096, + "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant" + }, + "parameters": { + "max_tokens": 4096 + }, + "metadata": { + "author": "Cognitive Computations, TheBloke", + "tags": ["70B", "Fintuned"], + "size": 26440000000 + }, + "engine": "nitro" + } diff --git a/models/lzlv-70b/model.json b/models/lzlv-70b/model.json deleted file mode 100644 index bae737328..000000000 --- a/models/lzlv-70b/model.json +++ /dev/null @@ -1,23 +0,0 @@ -{ - "source_url": "https://huggingface.co/TheBloke/lzlv_70B-GGUF/resolve/main/lzlv_70b_fp16_hf.Q5_K_M.gguf", - "id": "lzlv-70b", - "object": "model", - "name": "Lzlv 70B Q4", - "version": "1.0", - "description": "lzlv_70B is a sophisticated AI model designed for roleplaying and creative tasks. This merge aims to combine intelligence with creativity, seemingly outperforming its individual components in complex scenarios and creative outputs.", - "format": "gguf", - "settings": { - "ctx_len": 4096, - "prompt_template": "USER:\n{prompt}\nASSISTANT:" - }, - "parameters": { - "max_tokens": 4096 - }, - "metadata": { - "author": "Lizpreciatior, The Bloke", - "tags": ["70B", "Finetuned"], - "size": 48750000000 - }, - "engine": "nitro" - } - \ No newline at end of file diff --git a/models/mistral-ins-7b-q4/model.json b/models/mistral-ins-7b-q4/model.json index 39427152b..5345b8232 100644 --- a/models/mistral-ins-7b-q4/model.json +++ b/models/mistral-ins-7b-q4/model.json @@ -1,5 +1,5 @@ { - "source_url": "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF/resolve/main/mistral-7b-instruct-v0.1.Q4_K_M.gguf", + "source_url": "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GGUF/resolve/main/mistral-7b-instruct-v0.2.Q4_K_M.gguf", "id": "mistral-ins-7b-q4", "object": "model", "name": "Mistral Instruct 7B Q4", @@ -8,9 +8,6 @@ "format": "gguf", "settings": { "ctx_len": 4096, - "system_prompt": "", - "user_prompt": "[INST]", - "ai_prompt": "[/INST]", "prompt_template": "[INST]{prompt}\n[/INST]" }, "parameters": { diff --git a/models/openchat-3.5-7b/model.json b/models/openchat-3.5-7b/model.json new file mode 100644 index 000000000..a9459006b --- /dev/null +++ b/models/openchat-3.5-7b/model.json @@ -0,0 +1,22 @@ +{ + "source_url": "https://huggingface.co/TheBloke/openchat-3.5-1210-GGUF/resolve/main/openchat-3.5-1210.Q4_K_M.gguf", + "id": "openchat-3.5-7b", + "object": "model", + "name": "Openchat-3.5 7B Q4", + "version": "1.0", + "description": "The performance of this open-source model surpasses that of ChatGPT-3.5 and Grok-1 across various benchmarks.", + "format": "gguf", + "settings": { + "ctx_len": 4096, + "prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:" + }, + "parameters": { + "max_tokens": 4096 + }, + "metadata": { + "author": "Openchat", + "tags": ["7B", "Fintuned", "Featured"], + "size": 4370000000 + }, + "engine": "nitro" + } diff --git a/models/stealth-v1.2-7b/model.json b/models/stealth-v1.2-7b/model.json index 76264b9c8..0fde33cb5 100644 --- a/models/stealth-v1.2-7b/model.json +++ b/models/stealth-v1.2-7b/model.json @@ -1,5 +1,5 @@ { - "source_url": "https://huggingface.co/janhq/stealth-v1.2-GGUF/resolve/main/stealth-v1.2.Q4_K_M.gguf", + "source_url": "https://huggingface.co/janhq/stealth-v1.3-GGUF/resolve/main/stealth-v1.3.Q4_K_M.gguf", "id": "stealth-v1.2-7b", "object": "model", "name": "Stealth-v1.2 7B Q4", diff --git a/models/tinyllama-1.1b/model.json b/models/tinyllama-1.1b/model.json index d924be046..bd59369f8 100644 --- a/models/tinyllama-1.1b/model.json +++ b/models/tinyllama-1.1b/model.json @@ -1,5 +1,5 @@ { - "source_url": "https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v0.6/resolve/main/ggml-model-q4_0.gguf", + "source_url": "https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf", "id": "tinyllama-1.1b", "object": "model", "name": "TinyLlama Chat 1.1B Q4", @@ -16,7 +16,7 @@ "metadata": { "author": "TinyLlama", "tags": ["Tiny", "Foundation Model"], - "size": 637000000 + "size": 669000000 }, "engine": "nitro" } \ No newline at end of file diff --git a/models/trinity-v1-7b/cover.png b/models/trinity-v1-7b/cover.png deleted file mode 100644 index a548e3c17..000000000 Binary files a/models/trinity-v1-7b/cover.png and /dev/null differ diff --git a/models/trinity-v1-7b/model.json b/models/trinity-v1-7b/model.json deleted file mode 100644 index 400ab1028..000000000 --- a/models/trinity-v1-7b/model.json +++ /dev/null @@ -1,22 +0,0 @@ -{ - "source_url": "https://huggingface.co/janhq/trinity-v1-GGUF/resolve/main/trinity-v1.Q4_K_M.gguf", - "id": "trinity-v1-7b", - "object": "model", - "name": "Trinity-v1 7B Q4", - "version": "1.0", - "description": "Please use the latest version Trinity v1.2 for the best experience. Trinity is an experimental model merge of GreenNodeLM & LeoScorpius using the Slerp method. Recommended for daily assistance purposes.", - "format": "gguf", - "settings": { - "ctx_len": 4096, - "prompt_template": "{system_message}\n### Instruction:\n{prompt}\n### Response:" - }, - "parameters": { - "max_tokens": 4096 - }, - "metadata": { - "author": "Jan", - "tags": ["7B", "Merged"], - "size": 4370000000 - }, - "engine": "nitro" - } \ No newline at end of file diff --git a/models/tulu-2-70b/model.json b/models/tulu-2-70b/model.json new file mode 100644 index 000000000..c85da8223 --- /dev/null +++ b/models/tulu-2-70b/model.json @@ -0,0 +1,22 @@ +{ + "source_url": "https://huggingface.co/TheBloke/tulu-2-dpo-70B-GGUF/resolve/main/tulu-2-dpo-70b.Q4_K_M.gguf", + "id": "tulu-2-70b", + "object": "model", + "name": "Tulu 2 70B Q4", + "version": "1.0", + "description": "Tulu V2 DPO 70B is a fine-tuned version of Llama 2 using (DPO). This model is a strong alternative to Llama 2 70b Chat to act as helpful assistants.", + "format": "gguf", + "settings": { + "ctx_len": 4096, + "prompt_template": "<|user|>\n{prompt}\n<|assistant|>" + }, + "parameters": { + "max_tokens": 4096 + }, + "metadata": { + "author": "Lizpreciatior, The Bloke", + "tags": ["70B", "Finetuned"], + "size": 41400000000 + }, + "engine": "nitro" + } \ No newline at end of file diff --git a/models/yarn-mistral-7b/model.json b/models/yarn-mistral-7b/model.json new file mode 100644 index 000000000..0d1af9c50 --- /dev/null +++ b/models/yarn-mistral-7b/model.json @@ -0,0 +1,23 @@ +{ + "source_url": "https://huggingface.co/TheBloke/Yarn-Mistral-7B-128k-GGUF/resolve/main/yarn-mistral-7b-128k.Q4_K_M.gguf", + "id": "yarn-mistral-7b", + "object": "model", + "name": "Yarn Mistral 7B Q4", + "version": "1.0", + "description": "Yarn Mistral 7B is a language model for long context and supports a 128k token context window.", + "format": "gguf", + "settings": { + "ctx_len": 4096, + "prompt_template": "{prompt}" + }, + "parameters": { + "max_tokens": 4096 + }, + "metadata": { + "author": "NousResearch, The Bloke", + "tags": ["7B","Finetuned"], + "size": 4370000000 + }, + "engine": "nitro" + } + \ No newline at end of file