update name for models

This commit is contained in:
hahuyhoang411 2023-12-04 16:22:42 +07:00
parent 84c940be47
commit 7017e6426d
21 changed files with 23 additions and 23 deletions

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/Nous-Capybara-34B-GGUF/resolve/main/nous-capybara-34b.Q5_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/Nous-Capybara-34B-GGUF/resolve/main/nous-capybara-34b.Q5_K_M.gguf",
"id": "capybara-34b", "id": "capybara-34b",
"object": "model", "object": "model",
"name": "Capybara 200k 34B", "name": "Capybara 200k 34B Q5",
"version": "1.0", "version": "1.0",
"description": "Nous Capybara 34B, a variant of the Yi-34B model, is the first Nous model with a 200K context length, trained for three epochs on the innovative Capybara dataset.", "description": "Nous Capybara 34B, a variant of the Yi-34B model, is the first Nous model with a 200K context length, trained for three epochs on the innovative Capybara dataset.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/deepseek-coder-1.3b-base-GGUF/resolve/main/deepseek-coder-1.3b-base.Q4_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/deepseek-coder-1.3b-base-GGUF/resolve/main/deepseek-coder-1.3b-base.Q4_K_M.gguf",
"id": "deepseek-coder-1.3b", "id": "deepseek-coder-1.3b",
"object": "model", "object": "model",
"name": "Deepseek Coder 1.3B", "name": "Deepseek Coder 1.3B Q4",
"version": "1.0", "version": "1.0",
"description": "", "description": "",
"format": "gguf", "format": "gguf",
@ -16,7 +16,7 @@
"max_tokens": 2048 "max_tokens": 2048
}, },
"metadata": { "metadata": {
"author": "deepseek, The Bloke", "author": "Deepseek, The Bloke",
"tags": ["Community Recommended", "Code", "Small size"], "tags": ["Community Recommended", "Code", "Small size"],
"size": 870000000 "size": 870000000
} }

View File

@ -18,7 +18,7 @@
"metadata": { "metadata": {
"author": "MetaAI, The Bloke", "author": "MetaAI, The Bloke",
"tags": ["Foundational Model", "General", "Code"], "tags": ["Foundational Model", "General", "Code"],
"size": 4080000000 "size": 43920000000
} }
} }

View File

@ -4,7 +4,7 @@
"object": "model", "object": "model",
"name": "Llama 2 Chat 7B Q4", "name": "Llama 2 Chat 7B Q4",
"version": "1.0", "version": "1.0",
"description": "This is a 4-bit quantized version of Meta AI's Llama 2 Chat 7b model.", "description": "This is a 4-bit quantized iteration of Meta AI's Llama 2 Chat 7b model, specifically designed for a comprehensive understanding through training on extensive internet data.",
"format": "gguf", "format": "gguf",
"settings": { "settings": {
"ctx_len": 2048, "ctx_len": 2048,

View File

@ -4,7 +4,7 @@
"object": "model", "object": "model",
"name": "Llama 2 Chat 7B Q5", "name": "Llama 2 Chat 7B Q5",
"version": "1.0", "version": "1.0",
"description": "This is a 5-bit quantized version of Meta AI's Llama 2 Chat 7b model.", "description": "This is a 5-bit quantized iteration of Meta AI's Llama 2 Chat 7b model, specifically designed for a comprehensive understanding through training on extensive internet data.",
"format": "gguf", "format": "gguf",
"settings": { "settings": {
"ctx_len": 2048, "ctx_len": 2048,

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/lzlv_70B-GGUF/resolve/main/lzlv_70b_fp16_hf.Q5_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/lzlv_70B-GGUF/resolve/main/lzlv_70b_fp16_hf.Q5_K_M.gguf",
"id": "lzlv-70b", "id": "lzlv-70b",
"object": "model", "object": "model",
"name": "Lzlv 70B", "name": "Lzlv 70B Q4",
"version": "1.0", "version": "1.0",
"description": "lzlv_70B is a sophisticated AI model designed for roleplaying and creative tasks. This merge aims to combine intelligence with creativity, seemingly outperforming its individual components in complex scenarios and creative outputs.", "description": "lzlv_70B is a sophisticated AI model designed for roleplaying and creative tasks. This merge aims to combine intelligence with creativity, seemingly outperforming its individual components in complex scenarios and creative outputs.",
"format": "gguf", "format": "gguf",

View File

@ -4,7 +4,7 @@
"object": "model", "object": "model",
"name": "Mistral Instruct 7B Q4", "name": "Mistral Instruct 7B Q4",
"version": "1.0", "version": "1.0",
"description": "This is a 4-bit quantized version of MistralAI's Mistral Instruct 7B model.", "description": "This is a 4-bit quantized iteration of MistralAI's Mistral Instruct 7b model, specifically designed for a comprehensive understanding through training on extensive internet data.",
"format": "gguf", "format": "gguf",
"settings": { "settings": {
"ctx_len": 2048, "ctx_len": 2048,

View File

@ -4,7 +4,7 @@
"object": "model", "object": "model",
"name": "Mistral Instruct 7B Q5", "name": "Mistral Instruct 7B Q5",
"version": "1.0", "version": "1.0",
"description": "This is a 5-bit quantized version of MistralAI's Mistral Instruct 7B model.", "description": "This is a 5-bit quantized iteration of MistralAI's Mistral Instruct 7b model, specifically designed for a comprehensive understanding through training on extensive internet data.",
"format": "gguf", "format": "gguf",
"settings": { "settings": {
"ctx_len": 2048, "ctx_len": 2048,

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/neural-chat-7B-v3-1-GGUF/resolve/main/neural-chat-7b-v3-1.Q4_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/neural-chat-7B-v3-1-GGUF/resolve/main/neural-chat-7b-v3-1.Q4_K_M.gguf",
"id": "neural-chat-7b", "id": "neural-chat-7b",
"object": "model", "object": "model",
"name": "Neural Chat 7B", "name": "Neural Chat 7B Q4",
"version": "1.0", "version": "1.0",
"description": "The Neural Chat 7B model, developed on the foundation of mistralai/Mistral-7B-v0.1, has been fine-tuned using the Open-Orca/SlimOrca dataset and aligned with the Direct Preference Optimization (DPO) algorithm. It has demonstrated substantial improvements in various AI tasks and performance well on the open_llm_leaderboard.", "description": "The Neural Chat 7B model, developed on the foundation of mistralai/Mistral-7B-v0.1, has been fine-tuned using the Open-Orca/SlimOrca dataset and aligned with the Direct Preference Optimization (DPO) algorithm. It has demonstrated substantial improvements in various AI tasks and performance well on the open_llm_leaderboard.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/Noromaid-20B-v0.1.1-GGUF/resolve/main/noromaid-20b-v0.1.1.Q4_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/Noromaid-20B-v0.1.1-GGUF/resolve/main/noromaid-20b-v0.1.1.Q4_K_M.gguf",
"id": "noromaid-20b", "id": "noromaid-20b",
"object": "model", "object": "model",
"name": "Noromaid 20B", "name": "Noromaid 20B Q4",
"version": "1.0", "version": "1.0",
"description": "The Noromaid 20b model is designed for role-playing and general use, featuring a unique touch with the no_robots dataset that enhances human-like behavior.", "description": "The Noromaid 20b model is designed for role-playing and general use, featuring a unique touch with the no_robots dataset that enhances human-like behavior.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/OpenHermes-2.5-neural-chat-7B-v3-2-7B-GGUF/resolve/main/openhermes-2.5-neural-chat-7b-v3-2-7b.Q4_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/OpenHermes-2.5-neural-chat-7B-v3-2-7B-GGUF/resolve/main/openhermes-2.5-neural-chat-7b-v3-2-7b.Q4_K_M.gguf",
"id": "openhermes-neural-7b", "id": "openhermes-neural-7b",
"object": "model", "object": "model",
"name": "OpenHermes Neural 7B", "name": "OpenHermes Neural 7B Q4",
"version": "1.0", "version": "1.0",
"description": "OpenHermes Neural is a merged model from OpenHermes-2.5-Mistral-7B and neural-chat-7b-v3-2 with the TIES method.", "description": "OpenHermes Neural is a merged model from OpenHermes-2.5-Mistral-7B and neural-chat-7b-v3-2 with the TIES method.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/Orca-2-13B-GGUF/resolve/main/orca-2-13b.Q5_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/Orca-2-13B-GGUF/resolve/main/orca-2-13b.Q5_K_M.gguf",
"id": "openorca-13b", "id": "openorca-13b",
"object": "model", "object": "model",
"name": "Orca 2 13B", "name": "Orca 2 13B Q5",
"version": "1.0", "version": "1.0",
"description": "Orca 2 is a finetuned version of LLAMA-2, designed primarily for single-turn responses in reasoning, reading comprehension, math problem solving, and text summarization.", "description": "Orca 2 is a finetuned version of LLAMA-2, designed primarily for single-turn responses in reasoning, reading comprehension, math problem solving, and text summarization.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/Phind-CodeLlama-34B-v2-GGUF/resolve/main/phind-codellama-34b-v2.Q5_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/Phind-CodeLlama-34B-v2-GGUF/resolve/main/phind-codellama-34b-v2.Q5_K_M.gguf",
"id": "phind-34b", "id": "phind-34b",
"object": "model", "object": "model",
"name": "Phind 34B", "name": "Phind 34B Q5",
"version": "1.0", "version": "1.0",
"description": "Phind-CodeLlama-34B-v2 is an AI model fine-tuned on 1.5B tokens of high-quality programming data. It's a SOTA open-source model in coding. This multi-lingual model excels in various programming languages, including Python, C/C++, TypeScript, Java, and is designed to be steerable and user-friendly.", "description": "Phind-CodeLlama-34B-v2 is an AI model fine-tuned on 1.5B tokens of high-quality programming data. It's a SOTA open-source model in coding. This multi-lingual model excels in various programming languages, including Python, C/C++, TypeScript, Java, and is designed to be steerable and user-friendly.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/rocket-3B-GGUF/resolve/main/rocket-3b.Q4_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/rocket-3B-GGUF/resolve/main/rocket-3b.Q4_K_M.gguf",
"id": "rocket-3b", "id": "rocket-3b",
"object": "model", "object": "model",
"name": "Rocket 3B", "name": "Rocket 3B Q4",
"version": "1.0", "version": "1.0",
"description": "Rocket-3B is a GPT-like model, primarily English, fine-tuned on diverse public datasets. It outperforms larger models in benchmarks, showcasing superior understanding and text generation, making it an effective chat model for its size.", "description": "Rocket-3B is a GPT-like model, primarily English, fine-tuned on diverse public datasets. It outperforms larger models in benchmarks, showcasing superior understanding and text generation, making it an effective chat model for its size.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/Starling-LM-7B-alpha-GGUF/resolve/main/starling-lm-7b-alpha.Q4_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/Starling-LM-7B-alpha-GGUF/resolve/main/starling-lm-7b-alpha.Q4_K_M.gguf",
"id": "starling-7b", "id": "starling-7b",
"object": "model", "object": "model",
"name": "Strarling alpha 7B", "name": "Strarling alpha 7B Q4",
"version": "1.0", "version": "1.0",
"description": "Starling-RM-7B-alpha is a language model finetuned with Reinforcement Learning from AI Feedback from Openchat 3.5. It stands out for its impressive performance using GPT-4 as a judge, making it one of the top-performing models in its category.", "description": "Starling-RM-7B-alpha is a language model finetuned with Reinforcement Learning from AI Feedback from Openchat 3.5. It stands out for its impressive performance using GPT-4 as a judge, making it one of the top-performing models in its category.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/LLaMA2-13B-Tiefighter-GGUF/resolve/main/llama2-13b-tiefighter.Q5_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/LLaMA2-13B-Tiefighter-GGUF/resolve/main/llama2-13b-tiefighter.Q5_K_M.gguf",
"id": "tiefighter-13b", "id": "tiefighter-13b",
"object": "model", "object": "model",
"name": "Tiefighter 13B", "name": "Tiefighter 13B Q5",
"version": "1.0", "version": "1.0",
"description": "Tiefighter-13B is a highly creative, merged AI model achieved by combining various 'LORAs' on top of an existing merge, particularly focusing on storytelling and improvisation. This model excels in story writing, chatbots, and adventuring, and is designed to perform better with less detailed inputs, leveraging its inherent creativity.", "description": "Tiefighter-13B is a highly creative, merged AI model achieved by combining various 'LORAs' on top of an existing merge, particularly focusing on storytelling and improvisation. This model excels in story writing, chatbots, and adventuring, and is designed to perform better with less detailed inputs, leveraging its inherent creativity.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v0.6/resolve/main/ggml-model-q4_0.gguf", "source_url": "https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v0.6/resolve/main/ggml-model-q4_0.gguf",
"id": "tinyllama-1.1b", "id": "tinyllama-1.1b",
"object": "model", "object": "model",
"name": "TinyLlama Chat 1.1B", "name": "TinyLlama Chat 1.1B Q4",
"version": "1.0", "version": "1.0",
"description": "The TinyLlama project, featuring a 1.1B parameter Llama model, is pretrained on an expansive 3 trillion token dataset. Its design ensures easy integration with various Llama-based open-source projects. Despite its smaller size, it efficiently utilizes lower computational and memory resources, drawing on GPT-4's analytical prowess to enhance its conversational abilities and versatility.", "description": "The TinyLlama project, featuring a 1.1B parameter Llama model, is pretrained on an expansive 3 trillion token dataset. Its design ensures easy integration with various Llama-based open-source projects. Despite its smaller size, it efficiently utilizes lower computational and memory resources, drawing on GPT-4's analytical prowess to enhance its conversational abilities and versatility.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/WizardCoder-Python-13B-V1.0-GGUF/resolve/main/wizardcoder-python-13b-v1.0.Q5_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/WizardCoder-Python-13B-V1.0-GGUF/resolve/main/wizardcoder-python-13b-v1.0.Q5_K_M.gguf",
"id": "wizardcoder-13b", "id": "wizardcoder-13b",
"object": "model", "object": "model",
"name": "Wizard Coder Python 13B", "name": "Wizard Coder Python 13B Q5",
"version": "1.0", "version": "1.0",
"description": "WizardCoder-Python-13B is a Python coding model major models like ChatGPT-3.5. This model based on the Llama2 architecture, demonstrate high proficiency in specific domains like coding and mathematics.", "description": "WizardCoder-Python-13B is a Python coding model major models like ChatGPT-3.5. This model based on the Llama2 architecture, demonstrate high proficiency in specific domains like coding and mathematics.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/WizardCoder-Python-34B-V1.0-GGUF/resolve/main/wizardcoder-python-34b-v1.0.Q5_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/WizardCoder-Python-34B-V1.0-GGUF/resolve/main/wizardcoder-python-34b-v1.0.Q5_K_M.gguf",
"id": "wizardcoder-34b", "id": "wizardcoder-34b",
"object": "model", "object": "model",
"name": "Wizard Coder Python 34B", "name": "Wizard Coder Python 34B Q5",
"version": "1.0", "version": "1.0",
"description": "WizardCoder-Python-34B is a Python coding model major models like ChatGPT-3.5. This model based on the Llama2 architecture, demonstrate high proficiency in specific domains like coding and mathematics.", "description": "WizardCoder-Python-34B is a Python coding model major models like ChatGPT-3.5. This model based on the Llama2 architecture, demonstrate high proficiency in specific domains like coding and mathematics.",
"format": "gguf", "format": "gguf",

View File

@ -2,7 +2,7 @@
"source_url": "https://huggingface.co/TheBloke/Yi-34B-Chat-GGUF/resolve/main/yi-34b-chat.Q5_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/Yi-34B-Chat-GGUF/resolve/main/yi-34b-chat.Q5_K_M.gguf",
"id": "yi-34b", "id": "yi-34b",
"object": "model", "object": "model",
"name": "Yi 34B", "name": "Yi 34B Q5",
"version": "1.0", "version": "1.0",
"description": "Yi-34B, a specialized chat model, is known for its diverse and creative responses and excels across various NLP tasks and benchmarks.", "description": "Yi-34B, a specialized chat model, is known for its diverse and creative responses and excels across various NLP tasks and benchmarks.",
"format": "gguf", "format": "gguf",

View File

@ -2,9 +2,9 @@
"source_url": "https://huggingface.co/TheBloke/zephyr-7B-beta-GGUF/resolve/main/zephyr-7b-beta.Q4_K_M.gguf", "source_url": "https://huggingface.co/TheBloke/zephyr-7B-beta-GGUF/resolve/main/zephyr-7b-beta.Q4_K_M.gguf",
"id": "zephyr-beta-7b", "id": "zephyr-beta-7b",
"object": "model", "object": "model",
"name": "Zephyr Beta 7B", "name": "Zephyr Beta 7B Q4",
"version": "1.0", "version": "1.0",
"description": "The Zephyr-7B-β model marks the second iteration in the Zephyr series, designed to function as an effective assistant. It has been fine-tuned from the mistralai/Mistral-7B-v0.1 base model, utilizing a combination of public and synthetic datasets with the application of Direct Preference Optimization.", "description": "The Zephyr-7B-β model is trained by HuggingFace, designed to function as a practical assistant. It has been fine-tuned from the mistralai/Mistral-7B-v0.1 base model, utilizing a combination of public and synthetic datasets with the application of Direct Preference Optimization.",
"format": "gguf", "format": "gguf",
"settings": { "settings": {
"ctx_len": 2048, "ctx_len": 2048,