Feat: Adjust model hub v0.4.13 (#2879)
* fix: correct phi3 * redundant phi2 dolphin * add: hermes llama3 * add: ngl settings * correct ctx len * correct ngl * correct maxlen + ngl * disable phi3 * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * remove redundant hermes pro * add ngl * add ngl * add ngl * remove miqu * add ngl * add ngl * add ngl * add ngl * remove redundant * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * version package bump * feat: resolve issue of cannot found model in the extensions due to the removal * feat: completely remove hermes-pro-7b * feat: completely remove openhermes-neural-7b and miqu-70b, and add llama3-hermes-8b via renaming from Rex * fix: correct description --------- Co-authored-by: Van-QA <van@jan.ai>
This commit is contained in:
parent
6af4a2d484
commit
1e0d4f3753
@ -1,7 +1,7 @@
|
|||||||
{
|
{
|
||||||
"name": "@janhq/inference-nitro-extension",
|
"name": "@janhq/inference-nitro-extension",
|
||||||
"productName": "Nitro Inference Engine",
|
"productName": "Nitro Inference Engine",
|
||||||
"version": "1.0.5",
|
"version": "1.0.6",
|
||||||
"description": "This extension embeds Nitro, a lightweight (3mb) inference engine written in C++. See https://nitro.jan.ai.\nAdditional dependencies could be installed to run without Cuda Toolkit installation.",
|
"description": "This extension embeds Nitro, a lightweight (3mb) inference engine written in C++. See https://nitro.jan.ai.\nAdditional dependencies could be installed to run without Cuda Toolkit installation.",
|
||||||
"main": "dist/index.js",
|
"main": "dist/index.js",
|
||||||
"node": "dist/node/index.cjs.js",
|
"node": "dist/node/index.cjs.js",
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 8192,
|
"ctx_len": 8192,
|
||||||
"prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:",
|
"prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:",
|
||||||
"llama_model_path": "codeninja-1.0-openchat-7b.Q4_K_M.gguf"
|
"llama_model_path": "codeninja-1.0-openchat-7b.Q4_K_M.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 131072,
|
"ctx_len": 131072,
|
||||||
"prompt_template": "<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{prompt}<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>",
|
"prompt_template": "<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{prompt}<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>",
|
||||||
"llama_model_path": "c4ai-command-r-v01-Q4_K_M.gguf"
|
"llama_model_path": "c4ai-command-r-v01-Q4_K_M.gguf",
|
||||||
|
"ngl": 40
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 16384,
|
"ctx_len": 16384,
|
||||||
"prompt_template": "### Instruction:\n{prompt}\n### Response:",
|
"prompt_template": "### Instruction:\n{prompt}\n### Response:",
|
||||||
"llama_model_path": "deepseek-coder-1.3b-instruct.Q8_0.gguf"
|
"llama_model_path": "deepseek-coder-1.3b-instruct.Q8_0.gguf",
|
||||||
|
"ngl": 24
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 16384,
|
"ctx_len": 16384,
|
||||||
"prompt_template": "### Instruction:\n{prompt}\n### Response:",
|
"prompt_template": "### Instruction:\n{prompt}\n### Response:",
|
||||||
"llama_model_path": "deepseek-coder-33b-instruct.Q4_K_M.gguf"
|
"llama_model_path": "deepseek-coder-33b-instruct.Q4_K_M.gguf",
|
||||||
|
"ngl": 62
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -1,32 +0,0 @@
|
|||||||
{
|
|
||||||
"sources": [
|
|
||||||
{
|
|
||||||
"url": "https://huggingface.co/TheBloke/dolphin-2_6-phi-2-GGUF/resolve/main/dolphin-2_6-phi-2.Q8_0.gguf",
|
|
||||||
"filename": "dolphin-2_6-phi-2.Q8_0.gguf"
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"id": "dolphin-phi-2",
|
|
||||||
"object": "model",
|
|
||||||
"name": "Dolphin Phi-2 2.7B Q8",
|
|
||||||
"version": "1.0",
|
|
||||||
"description": "Dolphin Phi-2 is a good alternative for Phi-2 in chatting",
|
|
||||||
"format": "gguf",
|
|
||||||
"settings": {
|
|
||||||
"ctx_len": 4096,
|
|
||||||
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
|
||||||
"llama_model_path": "dolphin-2_6-phi-2.Q8_0.gguf"
|
|
||||||
},
|
|
||||||
"parameters": {
|
|
||||||
"max_tokens": 4096,
|
|
||||||
"stop": ["<|im_end|>"]
|
|
||||||
},
|
|
||||||
"metadata": {
|
|
||||||
"author": "Cognitive Computations, Microsoft",
|
|
||||||
"tags": [
|
|
||||||
"3B",
|
|
||||||
"Finetuned"
|
|
||||||
],
|
|
||||||
"size": 2960000000
|
|
||||||
},
|
|
||||||
"engine": "nitro"
|
|
||||||
}
|
|
||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 8192,
|
"ctx_len": 8192,
|
||||||
"prompt_template": "<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model",
|
"prompt_template": "<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model",
|
||||||
"llama_model_path": "gemma-2b-it-q4_k_m.gguf"
|
"llama_model_path": "gemma-2b-it-q4_k_m.gguf",
|
||||||
|
"ngl": 18
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 8192,
|
"ctx_len": 8192,
|
||||||
"prompt_template": "<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model",
|
"prompt_template": "<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model",
|
||||||
"llama_model_path": "gemma-7b-it-q4_K_M.gguf"
|
"llama_model_path": "gemma-7b-it-q4_K_M.gguf",
|
||||||
|
"ngl": 28
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 4096,
|
"ctx_len": 4096,
|
||||||
"prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]",
|
"prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]",
|
||||||
"llama_model_path": "llama-2-70b-chat.Q4_K_M.gguf"
|
"llama_model_path": "llama-2-70b-chat.Q4_K_M.gguf",
|
||||||
|
"ngl": 80
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 4096,
|
"ctx_len": 4096,
|
||||||
"prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]",
|
"prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]",
|
||||||
"llama_model_path": "llama-2-7b-chat.Q4_K_M.gguf"
|
"llama_model_path": "llama-2-7b-chat.Q4_K_M.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -8,19 +8,20 @@
|
|||||||
"id": "llama3-8b-instruct",
|
"id": "llama3-8b-instruct",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
"name": "Llama 3 8B Q4",
|
"name": "Llama 3 8B Q4",
|
||||||
"version": "1.0",
|
"version": "1.1",
|
||||||
"description": "Meta's Llama 3 excels at general usage situations, including chat, general world knowledge, and coding.",
|
"description": "Meta's Llama 3 excels at general usage situations, including chat, general world knowledge, and coding.",
|
||||||
"format": "gguf",
|
"format": "gguf",
|
||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 8192,
|
"ctx_len": 8192,
|
||||||
"prompt_template": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
|
"prompt_template": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
|
||||||
"llama_model_path": "Meta-Llama-3-8B-Instruct-Q4_K_M.gguf"
|
"llama_model_path": "Meta-Llama-3-8B-Instruct-Q4_K_M.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
"top_p": 0.95,
|
"top_p": 0.95,
|
||||||
"stream": true,
|
"stream": true,
|
||||||
"max_tokens": 4096,
|
"max_tokens": 8192,
|
||||||
"stop": ["<|end_of_text|>","<|eot_id|>"],
|
"stop": ["<|end_of_text|>","<|eot_id|>"],
|
||||||
"frequency_penalty": 0,
|
"frequency_penalty": 0,
|
||||||
"presence_penalty": 0
|
"presence_penalty": 0
|
||||||
|
|||||||
@ -1,35 +1,38 @@
|
|||||||
{
|
{
|
||||||
"sources": [
|
"sources": [
|
||||||
{
|
{
|
||||||
"filename": "Hermes-2-Pro-Mistral-7B.Q4_K_M.gguf",
|
"filename": "Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf",
|
||||||
"url": "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B-GGUF/resolve/main/Hermes-2-Pro-Mistral-7B.Q4_K_M.gguf"
|
"url": "https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF/resolve/main/Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf"
|
||||||
}
|
}
|
||||||
],
|
],
|
||||||
"id": "hermes-pro-7b",
|
"id": "llama3-hermes-8b",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
"name": "Hermes Pro 7B Q4",
|
"name": "Hermes Pro Llama 3 8B Q4",
|
||||||
"version": "1.1",
|
"version": "1.1",
|
||||||
"description": "Hermes Pro is superior in Roleplaying, Reasoning and Explaining problem.",
|
"description": "Hermes Pro is well-designed for General chat and JSON output.",
|
||||||
"format": "gguf",
|
"format": "gguf",
|
||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 4096,
|
"ctx_len": 8192,
|
||||||
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
||||||
"llama_model_path": "Hermes-2-Pro-Mistral-7B.Q4_K_M.gguf"
|
"llama_model_path": "Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
"top_p": 0.95,
|
"top_p": 0.95,
|
||||||
"stream": true,
|
"stream": true,
|
||||||
"max_tokens": 4096,
|
"max_tokens": 8192,
|
||||||
"stop": [],
|
"stop": [],
|
||||||
"frequency_penalty": 0,
|
"frequency_penalty": 0,
|
||||||
"presence_penalty": 0
|
"presence_penalty": 0
|
||||||
},
|
},
|
||||||
"metadata": {
|
"metadata": {
|
||||||
"author": "NousResearch",
|
"author": "NousResearch",
|
||||||
"tags": ["7B", "Finetuned"],
|
"tags": [
|
||||||
"size": 4370000000
|
"7B",
|
||||||
|
"Finetuned"
|
||||||
|
],
|
||||||
|
"size": 4920000000
|
||||||
},
|
},
|
||||||
"engine": "nitro"
|
"engine": "nitro"
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 2048,
|
"ctx_len": 2048,
|
||||||
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
||||||
"llama_model_path": "llamacorn-1.1b-chat.Q8_0.gguf"
|
"llama_model_path": "llamacorn-1.1b-chat.Q8_0.gguf",
|
||||||
|
"ngl": 22
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -1,34 +0,0 @@
|
|||||||
{
|
|
||||||
"sources": [
|
|
||||||
{
|
|
||||||
"filename": "miqu-1-70b.q4_k_m.gguf",
|
|
||||||
"url": "https://huggingface.co/miqudev/miqu-1-70b/resolve/main/miqu-1-70b.q4_k_m.gguf"
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"id": "miqu-70b",
|
|
||||||
"object": "model",
|
|
||||||
"name": "Mistral 70B Q4",
|
|
||||||
"version": "1.0",
|
|
||||||
"description": "A leak weight of Mistral 70B model.",
|
|
||||||
"format": "gguf",
|
|
||||||
"settings": {
|
|
||||||
"ctx_len": 4096,
|
|
||||||
"prompt_template": "[INST] {prompt} [/INST]",
|
|
||||||
"llama_model_path": "miqu-1-70b.q4_k_m.gguf"
|
|
||||||
},
|
|
||||||
"parameters": {
|
|
||||||
"temperature": 0.7,
|
|
||||||
"top_p": 0.95,
|
|
||||||
"stream": true,
|
|
||||||
"max_tokens": 4096,
|
|
||||||
"frequency_penalty": 0,
|
|
||||||
"presence_penalty": 0
|
|
||||||
},
|
|
||||||
"metadata": {
|
|
||||||
"author": "miqudev",
|
|
||||||
"tags": ["70B", "Foundational Model"],
|
|
||||||
"size": 26440000000
|
|
||||||
},
|
|
||||||
"engine": "nitro"
|
|
||||||
}
|
|
||||||
|
|
||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 32768,
|
"ctx_len": 32768,
|
||||||
"prompt_template": "[INST] {prompt} [/INST]",
|
"prompt_template": "[INST] {prompt} [/INST]",
|
||||||
"llama_model_path": "mistral-7b-instruct-v0.2.Q4_K_M.gguf"
|
"llama_model_path": "mistral-7b-instruct-v0.2.Q4_K_M.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 32768,
|
"ctx_len": 32768,
|
||||||
"prompt_template": "[INST] {prompt} [/INST]",
|
"prompt_template": "[INST] {prompt} [/INST]",
|
||||||
"llama_model_path": "mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf"
|
"llama_model_path": "mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf",
|
||||||
|
"ngl": 100
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 32768,
|
"ctx_len": 32768,
|
||||||
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
||||||
"llama_model_path": "Noromaid-7B-0.4-DPO.q4_k_m.gguf"
|
"llama_model_path": "Noromaid-7B-0.4-DPO.q4_k_m.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 8192,
|
"ctx_len": 8192,
|
||||||
"prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:",
|
"prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:",
|
||||||
"llama_model_path": "openchat-3.5-0106.Q4_K_M.gguf"
|
"llama_model_path": "openchat-3.5-0106.Q4_K_M.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -1,34 +0,0 @@
|
|||||||
{
|
|
||||||
"sources": [
|
|
||||||
{
|
|
||||||
"filename": "openhermes-2.5-neural-chat-v3-3-slerp.Q4_K_M.gguf",
|
|
||||||
"url": "https://huggingface.co/janhq/openhermes-2.5-neural-chat-v3-3-slerp-GGUF/resolve/main/openhermes-2.5-neural-chat-v3-3-slerp.Q4_K_M.gguf"
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"id": "openhermes-neural-7b",
|
|
||||||
"object": "model",
|
|
||||||
"name": "OpenHermes Neural 7B Q4",
|
|
||||||
"version": "1.1",
|
|
||||||
"description": "OpenHermes Neural is a merged model using the TIES method. It performs well in various benchmarks.",
|
|
||||||
"format": "gguf",
|
|
||||||
"settings": {
|
|
||||||
"ctx_len": 4096,
|
|
||||||
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
|
||||||
"llama_model_path": "openhermes-2.5-neural-chat-v3-3-slerp.Q4_K_M.gguf"
|
|
||||||
},
|
|
||||||
"parameters": {
|
|
||||||
"temperature": 0.7,
|
|
||||||
"top_p": 0.95,
|
|
||||||
"stream": true,
|
|
||||||
"max_tokens": 4096,
|
|
||||||
"frequency_penalty": 0,
|
|
||||||
"presence_penalty": 0
|
|
||||||
},
|
|
||||||
"metadata": {
|
|
||||||
"author": "Intel, Jan",
|
|
||||||
"tags": ["7B", "Merged"],
|
|
||||||
"size": 4370000000,
|
|
||||||
"cover": "https://raw.githubusercontent.com/janhq/jan/dev/models/openhermes-neural-7b/cover.png"
|
|
||||||
},
|
|
||||||
"engine": "nitro"
|
|
||||||
}
|
|
||||||
@ -1,32 +0,0 @@
|
|||||||
{
|
|
||||||
"sources": [
|
|
||||||
{
|
|
||||||
"url": "https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf/resolve/main/Phi-3-mini-4k-instruct-q4.gguf",
|
|
||||||
"filename": "Phi-3-mini-4k-instruct-q4.gguf"
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"id": "phi3-3.8b",
|
|
||||||
"object": "model",
|
|
||||||
"name": "Phi-3 Mini",
|
|
||||||
"version": "1.0",
|
|
||||||
"description": "Phi-3 Mini is Microsoft's newest, compact model designed for mobile use.",
|
|
||||||
"format": "gguf",
|
|
||||||
"settings": {
|
|
||||||
"ctx_len": 4096,
|
|
||||||
"prompt_template": "<|system|>\n{system_message}<|end|>\n<|user|>\n{prompt}<|end|>\n<|assistant|>\n",
|
|
||||||
"llama_model_path": "Phi-3-mini-4k-instruct-q4.gguf"
|
|
||||||
},
|
|
||||||
"parameters": {
|
|
||||||
"max_tokens": 4096,
|
|
||||||
"stop": ["<|end|>"]
|
|
||||||
},
|
|
||||||
"metadata": {
|
|
||||||
"author": "Microsoft",
|
|
||||||
"tags": [
|
|
||||||
"3B",
|
|
||||||
"Finetuned"
|
|
||||||
],
|
|
||||||
"size": 2320000000
|
|
||||||
},
|
|
||||||
"engine": "nitro"
|
|
||||||
}
|
|
||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 16384,
|
"ctx_len": 16384,
|
||||||
"prompt_template": "### System Prompt\n{system_message}\n### User Message\n{prompt}\n### Assistant",
|
"prompt_template": "### System Prompt\n{system_message}\n### User Message\n{prompt}\n### Assistant",
|
||||||
"llama_model_path": "phind-codellama-34b-v2.Q4_K_M.gguf"
|
"llama_model_path": "phind-codellama-34b-v2.Q4_K_M.gguf",
|
||||||
|
"ngl": 48
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 32768,
|
"ctx_len": 32768,
|
||||||
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
||||||
"llama_model_path": "qwen1_5-7b-chat-q4_k_m.gguf"
|
"llama_model_path": "qwen1_5-7b-chat-q4_k_m.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 4096,
|
"ctx_len": 4096,
|
||||||
"prompt_template": "<|user|>\n{prompt}<|endoftext|>\n<|assistant|>",
|
"prompt_template": "<|user|>\n{prompt}<|endoftext|>\n<|assistant|>",
|
||||||
"llama_model_path": "stablelm-zephyr-3b.Q8_0.gguf"
|
"llama_model_path": "stablelm-zephyr-3b.Q8_0.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 32768,
|
"ctx_len": 32768,
|
||||||
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
||||||
"llama_model_path": "stealth-v1.3.Q4_K_M.gguf"
|
"llama_model_path": "stealth-v1.3.Q4_K_M.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 4096,
|
"ctx_len": 4096,
|
||||||
"prompt_template": "<|system|>\n{system_message}<|user|>\n{prompt}<|assistant|>",
|
"prompt_template": "<|system|>\n{system_message}<|user|>\n{prompt}<|assistant|>",
|
||||||
"llama_model_path": "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf"
|
"llama_model_path": "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf",
|
||||||
|
"ngl": 22
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 32768,
|
"ctx_len": 32768,
|
||||||
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
||||||
"llama_model_path": "trinity-v1.2.Q4_K_M.gguf"
|
"llama_model_path": "trinity-v1.2.Q4_K_M.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 32768,
|
"ctx_len": 32768,
|
||||||
"prompt_template": "[INST] <<SYS>>\n{system_message}\n<</SYS>>\n{prompt} [/INST]",
|
"prompt_template": "[INST] <<SYS>>\n{system_message}\n<</SYS>>\n{prompt} [/INST]",
|
||||||
"llama_model_path": "vistral-7b-chat-dpo.Q4_K_M.gguf"
|
"llama_model_path": "vistral-7b-chat-dpo.Q4_K_M.gguf",
|
||||||
|
"ngl": 32
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 16384,
|
"ctx_len": 16384,
|
||||||
"prompt_template": "### Instruction:\n{prompt}\n### Response:",
|
"prompt_template": "### Instruction:\n{prompt}\n### Response:",
|
||||||
"llama_model_path": "wizardcoder-python-13b-v1.0.Q4_K_M.gguf"
|
"llama_model_path": "wizardcoder-python-13b-v1.0.Q4_K_M.gguf",
|
||||||
|
"ngl": 40
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -14,7 +14,8 @@
|
|||||||
"settings": {
|
"settings": {
|
||||||
"ctx_len": 4096,
|
"ctx_len": 4096,
|
||||||
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
||||||
"llama_model_path": "yi-34b-chat.Q4_K_M.gguf"
|
"llama_model_path": "yi-34b-chat.Q4_K_M.gguf",
|
||||||
|
"ngl": 60
|
||||||
},
|
},
|
||||||
"parameters": {
|
"parameters": {
|
||||||
"temperature": 0.7,
|
"temperature": 0.7,
|
||||||
|
|||||||
@ -12,21 +12,17 @@ const codeninja7bJson = require('./resources/models/codeninja-1.0-7b/model.json'
|
|||||||
const commandr34bJson = require('./resources/models/command-r-34b/model.json')
|
const commandr34bJson = require('./resources/models/command-r-34b/model.json')
|
||||||
const deepseekCoder13bJson = require('./resources/models/deepseek-coder-1.3b/model.json')
|
const deepseekCoder13bJson = require('./resources/models/deepseek-coder-1.3b/model.json')
|
||||||
const deepseekCoder34bJson = require('./resources/models/deepseek-coder-34b/model.json')
|
const deepseekCoder34bJson = require('./resources/models/deepseek-coder-34b/model.json')
|
||||||
const dolphinPhi2Json = require('./resources/models/dolphin-phi-2/model.json')
|
|
||||||
const gemma2bJson = require('./resources/models/gemma-2b/model.json')
|
const gemma2bJson = require('./resources/models/gemma-2b/model.json')
|
||||||
const gemma7bJson = require('./resources/models/gemma-7b/model.json')
|
const gemma7bJson = require('./resources/models/gemma-7b/model.json')
|
||||||
const hermesPro7bJson = require('./resources/models/hermes-pro-7b/model.json')
|
|
||||||
const llama2Chat70bJson = require('./resources/models/llama2-chat-70b/model.json')
|
const llama2Chat70bJson = require('./resources/models/llama2-chat-70b/model.json')
|
||||||
const llama2Chat7bJson = require('./resources/models/llama2-chat-7b/model.json')
|
const llama2Chat7bJson = require('./resources/models/llama2-chat-7b/model.json')
|
||||||
const llamacorn1bJson = require('./resources/models/llamacorn-1.1b/model.json')
|
const llamacorn1bJson = require('./resources/models/llamacorn-1.1b/model.json')
|
||||||
const llava13bJson = require('./resources/models/llava-13b/model.json')
|
const llava13bJson = require('./resources/models/llava-13b/model.json')
|
||||||
const llava7bJson = require('./resources/models/llava-7b/model.json')
|
const llava7bJson = require('./resources/models/llava-7b/model.json')
|
||||||
const miqu70bJson = require('./resources/models/miqu-70b/model.json')
|
|
||||||
const mistralIns7bq4Json = require('./resources/models/mistral-ins-7b-q4/model.json')
|
const mistralIns7bq4Json = require('./resources/models/mistral-ins-7b-q4/model.json')
|
||||||
const mixtral8x7bInstructJson = require('./resources/models/mixtral-8x7b-instruct/model.json')
|
const mixtral8x7bInstructJson = require('./resources/models/mixtral-8x7b-instruct/model.json')
|
||||||
const noromaid7bJson = require('./resources/models/noromaid-7b/model.json')
|
const noromaid7bJson = require('./resources/models/noromaid-7b/model.json')
|
||||||
const openchat357bJson = require('./resources/models/openchat-3.5-7b/model.json')
|
const openchat357bJson = require('./resources/models/openchat-3.5-7b/model.json')
|
||||||
const openhermesNeural7bJson = require('./resources/models/openhermes-neural-7b/model.json')
|
|
||||||
const phind34bJson = require('./resources/models/phind-34b/model.json')
|
const phind34bJson = require('./resources/models/phind-34b/model.json')
|
||||||
const qwen7bJson = require('./resources/models/qwen-7b/model.json')
|
const qwen7bJson = require('./resources/models/qwen-7b/model.json')
|
||||||
const stableZephyr3bJson = require('./resources/models/stable-zephyr-3b/model.json')
|
const stableZephyr3bJson = require('./resources/models/stable-zephyr-3b/model.json')
|
||||||
@ -37,6 +33,7 @@ const vistral7bJson = require('./resources/models/vistral-7b/model.json')
|
|||||||
const wizardcoder13bJson = require('./resources/models/wizardcoder-13b/model.json')
|
const wizardcoder13bJson = require('./resources/models/wizardcoder-13b/model.json')
|
||||||
const yi34bJson = require('./resources/models/yi-34b/model.json')
|
const yi34bJson = require('./resources/models/yi-34b/model.json')
|
||||||
const llama3Json = require('./resources/models/llama3-8b-instruct/model.json')
|
const llama3Json = require('./resources/models/llama3-8b-instruct/model.json')
|
||||||
|
const llama3Hermes8bJson = require('./resources/models/llama3-hermes-8b/model.json')
|
||||||
|
|
||||||
export default [
|
export default [
|
||||||
{
|
{
|
||||||
@ -56,21 +53,17 @@ export default [
|
|||||||
commandr34bJson,
|
commandr34bJson,
|
||||||
deepseekCoder13bJson,
|
deepseekCoder13bJson,
|
||||||
deepseekCoder34bJson,
|
deepseekCoder34bJson,
|
||||||
dolphinPhi2Json,
|
|
||||||
gemma2bJson,
|
gemma2bJson,
|
||||||
gemma7bJson,
|
gemma7bJson,
|
||||||
hermesPro7bJson,
|
|
||||||
llama2Chat70bJson,
|
llama2Chat70bJson,
|
||||||
llama2Chat7bJson,
|
llama2Chat7bJson,
|
||||||
llamacorn1bJson,
|
llamacorn1bJson,
|
||||||
llava13bJson,
|
llava13bJson,
|
||||||
llava7bJson,
|
llava7bJson,
|
||||||
miqu70bJson,
|
|
||||||
mistralIns7bq4Json,
|
mistralIns7bq4Json,
|
||||||
mixtral8x7bInstructJson,
|
mixtral8x7bInstructJson,
|
||||||
noromaid7bJson,
|
noromaid7bJson,
|
||||||
openchat357bJson,
|
openchat357bJson,
|
||||||
openhermesNeural7bJson,
|
|
||||||
phind34bJson,
|
phind34bJson,
|
||||||
qwen7bJson,
|
qwen7bJson,
|
||||||
stableZephyr3bJson,
|
stableZephyr3bJson,
|
||||||
@ -80,7 +73,8 @@ export default [
|
|||||||
vistral7bJson,
|
vistral7bJson,
|
||||||
wizardcoder13bJson,
|
wizardcoder13bJson,
|
||||||
yi34bJson,
|
yi34bJson,
|
||||||
llama3Json
|
llama3Json,
|
||||||
|
llama3Hermes8bJson
|
||||||
]),
|
]),
|
||||||
NODE: JSON.stringify(`${packageJson.name}/${packageJson.node}`),
|
NODE: JSON.stringify(`${packageJson.name}/${packageJson.node}`),
|
||||||
DEFAULT_SETTINGS: JSON.stringify(defaultSettingJson),
|
DEFAULT_SETTINGS: JSON.stringify(defaultSettingJson),
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user