* fix: correct phi3 * redundant phi2 dolphin * add: hermes llama3 * add: ngl settings * correct ctx len * correct ngl * correct maxlen + ngl * disable phi3 * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * remove redundant hermes pro * add ngl * add ngl * add ngl * remove miqu * add ngl * add ngl * add ngl * add ngl * remove redundant * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * add ngl * version package bump * feat: resolve issue of cannot found model in the extensions due to the removal * feat: completely remove hermes-pro-7b * feat: completely remove openhermes-neural-7b and miqu-70b, and add llama3-hermes-8b via renaming from Rex * fix: correct description --------- Co-authored-by: Van-QA <van@jan.ai>
36 lines
1.0 KiB
JSON
36 lines
1.0 KiB
JSON
{
|
|
"sources": [
|
|
{
|
|
"filename": "trinity-v1.2.Q4_K_M.gguf",
|
|
"url": "https://huggingface.co/janhq/trinity-v1.2-GGUF/resolve/main/trinity-v1.2.Q4_K_M.gguf"
|
|
}
|
|
],
|
|
"id": "trinity-v1.2-7b",
|
|
"object": "model",
|
|
"name": "Trinity-v1.2 7B Q4",
|
|
"version": "1.1",
|
|
"description": "Trinity is an experimental model merge using the Slerp method. Recommended for daily assistance purposes.",
|
|
"format": "gguf",
|
|
"settings": {
|
|
"ctx_len": 32768,
|
|
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
|
|
"llama_model_path": "trinity-v1.2.Q4_K_M.gguf",
|
|
"ngl": 32
|
|
},
|
|
"parameters": {
|
|
"temperature": 0.7,
|
|
"top_p": 0.95,
|
|
"stream": true,
|
|
"max_tokens": 32768,
|
|
"frequency_penalty": 0,
|
|
"presence_penalty": 0
|
|
},
|
|
"metadata": {
|
|
"author": "Jan",
|
|
"tags": ["7B", "Merged"],
|
|
"size": 4370000000,
|
|
"cover": "https://raw.githubusercontent.com/janhq/jan/dev/models/trinity-v1.2-7b/cover.png"
|
|
},
|
|
"engine": "nitro"
|
|
}
|