decapriate some 70b models

This commit is contained in:
hahuyhoang411 2023-12-04 11:41:30 +07:00
parent e5a440fc8f
commit fc6f8d2c0a
3 changed files with 0 additions and 69 deletions

View File

@ -1,24 +0,0 @@
{
"source_url": "https://huggingface.co/TheBloke/GOAT-70B-Storytelling-GGUF/resolve/main/goat-70b-storytelling.Q5_K_M.gguf",
"id": "storytelling-70b",
"object": "model",
"name": "Storytelling 70B",
"version": "1.0",
"description": "The GOAT-70B-Storytelling model is designed for autonomous story-writing, including crafting books and movie scripts. Based on the LLaMA 2 70B architecture, this model excels in generating cohesive and engaging narratives using inputs like plot outlines and character profiles.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"system_prompt": "",
"user_prompt": "### USER: ",
"ai_prompt": "\n### ASSISTANT: "
},
"parameters": {
"max_tokens": 4096
},
"metadata": {
"author": "GOAT-AI, The Bloke",
"tags": ["General Use", "Writing"],
"size": 48750000000
}
}

View File

@ -1,24 +0,0 @@
{
"source_url": "https://huggingface.co/TheBloke/Xwin-LM-70B-V0.1-GGUF/resolve/main/xwin-lm-70b-v0.1.Q5_K_M.gguf",
"id": "xwin-70b",
"object": "model",
"name": "Xwin LM 70B",
"version": "1.0",
"description": "Xwin-LM, based on Llama2 models, emphasizes alignment and exhibits advanced language understanding, text generation, and role-playing abilities.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"system_prompt": "",
"user_prompt": "USER: ",
"ai_prompt": "ASSISTANT: "
},
"parameters": {
"max_tokens": 4096
},
"metadata": {
"author": "Xwin-LM, The Bloke",
"tags": ["General Use", "Role-playing"],
"size": 48750000000
}
}

View File

@ -1,21 +0,0 @@
{
"source_url": "https://huggingface.co/TheBloke/Yarn-Llama-2-70B-32k-GGUF/resolve/main/yarn-llama-2-70b-32k.Q5_K_M.gguf",
"id": "yarn-70b",
"object": "model",
"name": "Yarn 32k 70B",
"version": "1,0",
"description": "Yarn-Llama-2-70b-32k is designed specifically for handling long contexts. It represents an extension of the Llama-2-70b-hf model, now supporting a 32k token context window.",
"format": "gguf",
"settings": {
"ctx_len": 4096
},
"parameters": {
"max_tokens": 4096
},
"metadata": {
"author": "NousResearch, The Bloke",
"tags": ["General Use", "Big Context Length"],
"size": 48750000000
}
}