Feat: Adjust model hub v0.4.13 (#2879)

* fix: correct phi3

* redundant phi2 dolphin

* add: hermes llama3

* add: ngl settings

* correct ctx len

* correct ngl

* correct maxlen + ngl

* disable phi3

* add ngl

* add ngl

* add ngl

* add ngl

* add ngl

* add ngl

* add ngl

* remove redundant  hermes pro

* add ngl

* add ngl

* add ngl

* remove miqu

* add ngl

* add ngl

* add ngl

* add ngl

* remove redundant

* add ngl

* add ngl

* add ngl

* add ngl

* add ngl

* add ngl

* add ngl

* add ngl

* add ngl

* version package bump

* feat: resolve issue of cannot found model in the extensions due to the removal

* feat: completely remove hermes-pro-7b

* feat: completely remove openhermes-neural-7b and miqu-70b, and add llama3-hermes-8b via renaming from Rex

* fix: correct description

---------

Co-authored-by: Van-QA <van@jan.ai>
This commit is contained in:
Hoang Ha 2024-05-13 11:48:03 +07:00 committed by GitHub
parent 6af4a2d484
commit 1e0d4f3753
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
30 changed files with 66 additions and 178 deletions

View File

@ -1,7 +1,7 @@
{ {
"name": "@janhq/inference-nitro-extension", "name": "@janhq/inference-nitro-extension",
"productName": "Nitro Inference Engine", "productName": "Nitro Inference Engine",
"version": "1.0.5", "version": "1.0.6",
"description": "This extension embeds Nitro, a lightweight (3mb) inference engine written in C++. See https://nitro.jan.ai.\nAdditional dependencies could be installed to run without Cuda Toolkit installation.", "description": "This extension embeds Nitro, a lightweight (3mb) inference engine written in C++. See https://nitro.jan.ai.\nAdditional dependencies could be installed to run without Cuda Toolkit installation.",
"main": "dist/index.js", "main": "dist/index.js",
"node": "dist/node/index.cjs.js", "node": "dist/node/index.cjs.js",

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 8192, "ctx_len": 8192,
"prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:", "prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:",
"llama_model_path": "codeninja-1.0-openchat-7b.Q4_K_M.gguf" "llama_model_path": "codeninja-1.0-openchat-7b.Q4_K_M.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 131072, "ctx_len": 131072,
"prompt_template": "<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{prompt}<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>", "prompt_template": "<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{prompt}<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>",
"llama_model_path": "c4ai-command-r-v01-Q4_K_M.gguf" "llama_model_path": "c4ai-command-r-v01-Q4_K_M.gguf",
"ngl": 40
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 16384, "ctx_len": 16384,
"prompt_template": "### Instruction:\n{prompt}\n### Response:", "prompt_template": "### Instruction:\n{prompt}\n### Response:",
"llama_model_path": "deepseek-coder-1.3b-instruct.Q8_0.gguf" "llama_model_path": "deepseek-coder-1.3b-instruct.Q8_0.gguf",
"ngl": 24
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 16384, "ctx_len": 16384,
"prompt_template": "### Instruction:\n{prompt}\n### Response:", "prompt_template": "### Instruction:\n{prompt}\n### Response:",
"llama_model_path": "deepseek-coder-33b-instruct.Q4_K_M.gguf" "llama_model_path": "deepseek-coder-33b-instruct.Q4_K_M.gguf",
"ngl": 62
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -1,32 +0,0 @@
{
"sources": [
{
"url": "https://huggingface.co/TheBloke/dolphin-2_6-phi-2-GGUF/resolve/main/dolphin-2_6-phi-2.Q8_0.gguf",
"filename": "dolphin-2_6-phi-2.Q8_0.gguf"
}
],
"id": "dolphin-phi-2",
"object": "model",
"name": "Dolphin Phi-2 2.7B Q8",
"version": "1.0",
"description": "Dolphin Phi-2 is a good alternative for Phi-2 in chatting",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "dolphin-2_6-phi-2.Q8_0.gguf"
},
"parameters": {
"max_tokens": 4096,
"stop": ["<|im_end|>"]
},
"metadata": {
"author": "Cognitive Computations, Microsoft",
"tags": [
"3B",
"Finetuned"
],
"size": 2960000000
},
"engine": "nitro"
}

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 8192, "ctx_len": 8192,
"prompt_template": "<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model", "prompt_template": "<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model",
"llama_model_path": "gemma-2b-it-q4_k_m.gguf" "llama_model_path": "gemma-2b-it-q4_k_m.gguf",
"ngl": 18
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 8192, "ctx_len": 8192,
"prompt_template": "<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model", "prompt_template": "<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model",
"llama_model_path": "gemma-7b-it-q4_K_M.gguf" "llama_model_path": "gemma-7b-it-q4_K_M.gguf",
"ngl": 28
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 4096, "ctx_len": 4096,
"prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]", "prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]",
"llama_model_path": "llama-2-70b-chat.Q4_K_M.gguf" "llama_model_path": "llama-2-70b-chat.Q4_K_M.gguf",
"ngl": 80
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 4096, "ctx_len": 4096,
"prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]", "prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]",
"llama_model_path": "llama-2-7b-chat.Q4_K_M.gguf" "llama_model_path": "llama-2-7b-chat.Q4_K_M.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -8,19 +8,20 @@
"id": "llama3-8b-instruct", "id": "llama3-8b-instruct",
"object": "model", "object": "model",
"name": "Llama 3 8B Q4", "name": "Llama 3 8B Q4",
"version": "1.0", "version": "1.1",
"description": "Meta's Llama 3 excels at general usage situations, including chat, general world knowledge, and coding.", "description": "Meta's Llama 3 excels at general usage situations, including chat, general world knowledge, and coding.",
"format": "gguf", "format": "gguf",
"settings": { "settings": {
"ctx_len": 8192, "ctx_len": 8192,
"prompt_template": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n", "prompt_template": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
"llama_model_path": "Meta-Llama-3-8B-Instruct-Q4_K_M.gguf" "llama_model_path": "Meta-Llama-3-8B-Instruct-Q4_K_M.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,
"top_p": 0.95, "top_p": 0.95,
"stream": true, "stream": true,
"max_tokens": 4096, "max_tokens": 8192,
"stop": ["<|end_of_text|>","<|eot_id|>"], "stop": ["<|end_of_text|>","<|eot_id|>"],
"frequency_penalty": 0, "frequency_penalty": 0,
"presence_penalty": 0 "presence_penalty": 0

View File

@ -1,35 +1,38 @@
{ {
"sources": [ "sources": [
{ {
"filename": "Hermes-2-Pro-Mistral-7B.Q4_K_M.gguf", "filename": "Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf",
"url": "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B-GGUF/resolve/main/Hermes-2-Pro-Mistral-7B.Q4_K_M.gguf" "url": "https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF/resolve/main/Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf"
} }
], ],
"id": "hermes-pro-7b", "id": "llama3-hermes-8b",
"object": "model", "object": "model",
"name": "Hermes Pro 7B Q4", "name": "Hermes Pro Llama 3 8B Q4",
"version": "1.1", "version": "1.1",
"description": "Hermes Pro is superior in Roleplaying, Reasoning and Explaining problem.", "description": "Hermes Pro is well-designed for General chat and JSON output.",
"format": "gguf", "format": "gguf",
"settings": { "settings": {
"ctx_len": 4096, "ctx_len": 8192,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant", "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "Hermes-2-Pro-Mistral-7B.Q4_K_M.gguf" "llama_model_path": "Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,
"top_p": 0.95, "top_p": 0.95,
"stream": true, "stream": true,
"max_tokens": 4096, "max_tokens": 8192,
"stop": [], "stop": [],
"frequency_penalty": 0, "frequency_penalty": 0,
"presence_penalty": 0 "presence_penalty": 0
}, },
"metadata": { "metadata": {
"author": "NousResearch", "author": "NousResearch",
"tags": ["7B", "Finetuned"], "tags": [
"size": 4370000000 "7B",
"Finetuned"
],
"size": 4920000000
}, },
"engine": "nitro" "engine": "nitro"
} }

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 2048, "ctx_len": 2048,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant", "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "llamacorn-1.1b-chat.Q8_0.gguf" "llama_model_path": "llamacorn-1.1b-chat.Q8_0.gguf",
"ngl": 22
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -1,34 +0,0 @@
{
"sources": [
{
"filename": "miqu-1-70b.q4_k_m.gguf",
"url": "https://huggingface.co/miqudev/miqu-1-70b/resolve/main/miqu-1-70b.q4_k_m.gguf"
}
],
"id": "miqu-70b",
"object": "model",
"name": "Mistral 70B Q4",
"version": "1.0",
"description": "A leak weight of Mistral 70B model.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "[INST] {prompt} [/INST]",
"llama_model_path": "miqu-1-70b.q4_k_m.gguf"
},
"parameters": {
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "miqudev",
"tags": ["70B", "Foundational Model"],
"size": 26440000000
},
"engine": "nitro"
}

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 32768, "ctx_len": 32768,
"prompt_template": "[INST] {prompt} [/INST]", "prompt_template": "[INST] {prompt} [/INST]",
"llama_model_path": "mistral-7b-instruct-v0.2.Q4_K_M.gguf" "llama_model_path": "mistral-7b-instruct-v0.2.Q4_K_M.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 32768, "ctx_len": 32768,
"prompt_template": "[INST] {prompt} [/INST]", "prompt_template": "[INST] {prompt} [/INST]",
"llama_model_path": "mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf" "llama_model_path": "mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf",
"ngl": 100
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 32768, "ctx_len": 32768,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant", "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "Noromaid-7B-0.4-DPO.q4_k_m.gguf" "llama_model_path": "Noromaid-7B-0.4-DPO.q4_k_m.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 8192, "ctx_len": 8192,
"prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:", "prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:",
"llama_model_path": "openchat-3.5-0106.Q4_K_M.gguf" "llama_model_path": "openchat-3.5-0106.Q4_K_M.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -1,34 +0,0 @@
{
"sources": [
{
"filename": "openhermes-2.5-neural-chat-v3-3-slerp.Q4_K_M.gguf",
"url": "https://huggingface.co/janhq/openhermes-2.5-neural-chat-v3-3-slerp-GGUF/resolve/main/openhermes-2.5-neural-chat-v3-3-slerp.Q4_K_M.gguf"
}
],
"id": "openhermes-neural-7b",
"object": "model",
"name": "OpenHermes Neural 7B Q4",
"version": "1.1",
"description": "OpenHermes Neural is a merged model using the TIES method. It performs well in various benchmarks.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "openhermes-2.5-neural-chat-v3-3-slerp.Q4_K_M.gguf"
},
"parameters": {
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Intel, Jan",
"tags": ["7B", "Merged"],
"size": 4370000000,
"cover": "https://raw.githubusercontent.com/janhq/jan/dev/models/openhermes-neural-7b/cover.png"
},
"engine": "nitro"
}

View File

@ -1,32 +0,0 @@
{
"sources": [
{
"url": "https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf/resolve/main/Phi-3-mini-4k-instruct-q4.gguf",
"filename": "Phi-3-mini-4k-instruct-q4.gguf"
}
],
"id": "phi3-3.8b",
"object": "model",
"name": "Phi-3 Mini",
"version": "1.0",
"description": "Phi-3 Mini is Microsoft's newest, compact model designed for mobile use.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "<|system|>\n{system_message}<|end|>\n<|user|>\n{prompt}<|end|>\n<|assistant|>\n",
"llama_model_path": "Phi-3-mini-4k-instruct-q4.gguf"
},
"parameters": {
"max_tokens": 4096,
"stop": ["<|end|>"]
},
"metadata": {
"author": "Microsoft",
"tags": [
"3B",
"Finetuned"
],
"size": 2320000000
},
"engine": "nitro"
}

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 16384, "ctx_len": 16384,
"prompt_template": "### System Prompt\n{system_message}\n### User Message\n{prompt}\n### Assistant", "prompt_template": "### System Prompt\n{system_message}\n### User Message\n{prompt}\n### Assistant",
"llama_model_path": "phind-codellama-34b-v2.Q4_K_M.gguf" "llama_model_path": "phind-codellama-34b-v2.Q4_K_M.gguf",
"ngl": 48
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 32768, "ctx_len": 32768,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant", "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "qwen1_5-7b-chat-q4_k_m.gguf" "llama_model_path": "qwen1_5-7b-chat-q4_k_m.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 4096, "ctx_len": 4096,
"prompt_template": "<|user|>\n{prompt}<|endoftext|>\n<|assistant|>", "prompt_template": "<|user|>\n{prompt}<|endoftext|>\n<|assistant|>",
"llama_model_path": "stablelm-zephyr-3b.Q8_0.gguf" "llama_model_path": "stablelm-zephyr-3b.Q8_0.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 32768, "ctx_len": 32768,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant", "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "stealth-v1.3.Q4_K_M.gguf" "llama_model_path": "stealth-v1.3.Q4_K_M.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 4096, "ctx_len": 4096,
"prompt_template": "<|system|>\n{system_message}<|user|>\n{prompt}<|assistant|>", "prompt_template": "<|system|>\n{system_message}<|user|>\n{prompt}<|assistant|>",
"llama_model_path": "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf" "llama_model_path": "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf",
"ngl": 22
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 32768, "ctx_len": 32768,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant", "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "trinity-v1.2.Q4_K_M.gguf" "llama_model_path": "trinity-v1.2.Q4_K_M.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 32768, "ctx_len": 32768,
"prompt_template": "[INST] <<SYS>>\n{system_message}\n<</SYS>>\n{prompt} [/INST]", "prompt_template": "[INST] <<SYS>>\n{system_message}\n<</SYS>>\n{prompt} [/INST]",
"llama_model_path": "vistral-7b-chat-dpo.Q4_K_M.gguf" "llama_model_path": "vistral-7b-chat-dpo.Q4_K_M.gguf",
"ngl": 32
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 16384, "ctx_len": 16384,
"prompt_template": "### Instruction:\n{prompt}\n### Response:", "prompt_template": "### Instruction:\n{prompt}\n### Response:",
"llama_model_path": "wizardcoder-python-13b-v1.0.Q4_K_M.gguf" "llama_model_path": "wizardcoder-python-13b-v1.0.Q4_K_M.gguf",
"ngl": 40
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -14,7 +14,8 @@
"settings": { "settings": {
"ctx_len": 4096, "ctx_len": 4096,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant", "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "yi-34b-chat.Q4_K_M.gguf" "llama_model_path": "yi-34b-chat.Q4_K_M.gguf",
"ngl": 60
}, },
"parameters": { "parameters": {
"temperature": 0.7, "temperature": 0.7,

View File

@ -12,21 +12,17 @@ const codeninja7bJson = require('./resources/models/codeninja-1.0-7b/model.json'
const commandr34bJson = require('./resources/models/command-r-34b/model.json') const commandr34bJson = require('./resources/models/command-r-34b/model.json')
const deepseekCoder13bJson = require('./resources/models/deepseek-coder-1.3b/model.json') const deepseekCoder13bJson = require('./resources/models/deepseek-coder-1.3b/model.json')
const deepseekCoder34bJson = require('./resources/models/deepseek-coder-34b/model.json') const deepseekCoder34bJson = require('./resources/models/deepseek-coder-34b/model.json')
const dolphinPhi2Json = require('./resources/models/dolphin-phi-2/model.json')
const gemma2bJson = require('./resources/models/gemma-2b/model.json') const gemma2bJson = require('./resources/models/gemma-2b/model.json')
const gemma7bJson = require('./resources/models/gemma-7b/model.json') const gemma7bJson = require('./resources/models/gemma-7b/model.json')
const hermesPro7bJson = require('./resources/models/hermes-pro-7b/model.json')
const llama2Chat70bJson = require('./resources/models/llama2-chat-70b/model.json') const llama2Chat70bJson = require('./resources/models/llama2-chat-70b/model.json')
const llama2Chat7bJson = require('./resources/models/llama2-chat-7b/model.json') const llama2Chat7bJson = require('./resources/models/llama2-chat-7b/model.json')
const llamacorn1bJson = require('./resources/models/llamacorn-1.1b/model.json') const llamacorn1bJson = require('./resources/models/llamacorn-1.1b/model.json')
const llava13bJson = require('./resources/models/llava-13b/model.json') const llava13bJson = require('./resources/models/llava-13b/model.json')
const llava7bJson = require('./resources/models/llava-7b/model.json') const llava7bJson = require('./resources/models/llava-7b/model.json')
const miqu70bJson = require('./resources/models/miqu-70b/model.json')
const mistralIns7bq4Json = require('./resources/models/mistral-ins-7b-q4/model.json') const mistralIns7bq4Json = require('./resources/models/mistral-ins-7b-q4/model.json')
const mixtral8x7bInstructJson = require('./resources/models/mixtral-8x7b-instruct/model.json') const mixtral8x7bInstructJson = require('./resources/models/mixtral-8x7b-instruct/model.json')
const noromaid7bJson = require('./resources/models/noromaid-7b/model.json') const noromaid7bJson = require('./resources/models/noromaid-7b/model.json')
const openchat357bJson = require('./resources/models/openchat-3.5-7b/model.json') const openchat357bJson = require('./resources/models/openchat-3.5-7b/model.json')
const openhermesNeural7bJson = require('./resources/models/openhermes-neural-7b/model.json')
const phind34bJson = require('./resources/models/phind-34b/model.json') const phind34bJson = require('./resources/models/phind-34b/model.json')
const qwen7bJson = require('./resources/models/qwen-7b/model.json') const qwen7bJson = require('./resources/models/qwen-7b/model.json')
const stableZephyr3bJson = require('./resources/models/stable-zephyr-3b/model.json') const stableZephyr3bJson = require('./resources/models/stable-zephyr-3b/model.json')
@ -37,6 +33,7 @@ const vistral7bJson = require('./resources/models/vistral-7b/model.json')
const wizardcoder13bJson = require('./resources/models/wizardcoder-13b/model.json') const wizardcoder13bJson = require('./resources/models/wizardcoder-13b/model.json')
const yi34bJson = require('./resources/models/yi-34b/model.json') const yi34bJson = require('./resources/models/yi-34b/model.json')
const llama3Json = require('./resources/models/llama3-8b-instruct/model.json') const llama3Json = require('./resources/models/llama3-8b-instruct/model.json')
const llama3Hermes8bJson = require('./resources/models/llama3-hermes-8b/model.json')
export default [ export default [
{ {
@ -56,21 +53,17 @@ export default [
commandr34bJson, commandr34bJson,
deepseekCoder13bJson, deepseekCoder13bJson,
deepseekCoder34bJson, deepseekCoder34bJson,
dolphinPhi2Json,
gemma2bJson, gemma2bJson,
gemma7bJson, gemma7bJson,
hermesPro7bJson,
llama2Chat70bJson, llama2Chat70bJson,
llama2Chat7bJson, llama2Chat7bJson,
llamacorn1bJson, llamacorn1bJson,
llava13bJson, llava13bJson,
llava7bJson, llava7bJson,
miqu70bJson,
mistralIns7bq4Json, mistralIns7bq4Json,
mixtral8x7bInstructJson, mixtral8x7bInstructJson,
noromaid7bJson, noromaid7bJson,
openchat357bJson, openchat357bJson,
openhermesNeural7bJson,
phind34bJson, phind34bJson,
qwen7bJson, qwen7bJson,
stableZephyr3bJson, stableZephyr3bJson,
@ -80,7 +73,8 @@ export default [
vistral7bJson, vistral7bJson,
wizardcoder13bJson, wizardcoder13bJson,
yi34bJson, yi34bJson,
llama3Json llama3Json,
llama3Hermes8bJson
]), ]),
NODE: JSON.stringify(`${packageJson.name}/${packageJson.node}`), NODE: JSON.stringify(`${packageJson.name}/${packageJson.node}`),
DEFAULT_SETTINGS: JSON.stringify(defaultSettingJson), DEFAULT_SETTINGS: JSON.stringify(defaultSettingJson),