36 lines
1.1 KiB
JSON
36 lines
1.1 KiB
JSON
{
|
||
"sources": [
|
||
{
|
||
"filename": "DeepSeek-R1-Distill-Llama-70B-Q4_K_M.gguf",
|
||
"url": "https://huggingface.co/unsloth/DeepSeek-R1-Distill-Llama-70B-GGUF/resolve/main/DeepSeek-R1-Distill-Llama-70B-Q4_K_M.gguf"
|
||
}
|
||
],
|
||
"id": "deepseek-r1-distill-llama-70b",
|
||
"object": "model",
|
||
"name": "DeepSeek R1 Distill Llama 70B Q4",
|
||
"version": "1.0",
|
||
"description": "DeepSeek-R1 is a cheaper and open-source model that excels at agentic reasoning, superior multilingual capabilities, large context windows, and generalization across domains.",
|
||
"format": "gguf",
|
||
"settings": {
|
||
"ctx_len": 131072,
|
||
"prompt_template": "<|User|> {prompt} <|Assistant|>",
|
||
"llama_model_path": "DeepSeek-R1-Distill-Llama-70B-Q4_K_M.gguf",
|
||
"ngl": 81
|
||
},
|
||
"parameters": {
|
||
"temperature": 0.6,
|
||
"top_p": 0.95,
|
||
"stream": true,
|
||
"max_tokens": 131072,
|
||
"stop": [],
|
||
"frequency_penalty": 0,
|
||
"presence_penalty": 0
|
||
},
|
||
"metadata": {
|
||
"author": "DeepSeek",
|
||
"tags": ["70B", "Featured"],
|
||
"size": 42500000000
|
||
},
|
||
"engine": "llama-cpp"
|
||
}
|