Hoang Ha ecc866427b
Update Model.json (#1005)
* add(mixtral): add model.json for mixtral

* archived some models + update the model.json

* add(model): add pandora 10.7b

* fix(model): update description

* fix(model): pump vers and change the featured model to trinity

* fix(model): archive neuralchat

* fix(model): decapriated all old models

* fix(trinity): add cover image and change description

* fix(trinity): update cover png

* add(pandora): cover image

* fix(pandora): cover image

* add(mixtral): add model.json for mixtral

* archived some models + update the model.json

* add(model): add pandora 10.7b

* fix(model): update description

* fix(model): pump vers and change the featured model to trinity

* fix(model): archive neuralchat

* fix(model): decapriated all old models

* fix(trinity): add cover image and change description

* fix(trinity): update cover png

* add(pandora): cover image

* fix(pandora): cover image

* chore: model desc nits

* fix(models): adjust the size for solars and pandoras

* add(mixtral): description

---------

Co-authored-by: 0xSage <n@pragmatic.vc>
2023-12-15 14:19:49 +07:00

23 lines
860 B
JSON

{
"source_url": "https://huggingface.co/janhq/Solar-10.7B-Instruct-v1.0/resolve/main/solar-10.7b-instruct-v1.0.Q4_K_M.gguf",
"id": "solar-10.7b-instruct",
"object": "model",
"name": "Solar Instruct 10.7B Q4",
"version": "1.0",
"description": "SOLAR-10.7B model built on the Llama2 architecture with Depth Up-Scaling and integrated Mistral 7B weights. Its robustness and adaptability make it ideal for fine-tuning applications, significantly enhancing performance with simple instruction-based techniques.",
"format": "gguf",
"settings": {
"ctx_len": 2048,
"prompt_template": "### User: {prompt}\n### Assistant:"
},
"parameters": {
"max_tokens": 2048
},
"metadata": {
"author": "Upstage, Jan",
"tags": ["11B","Finetuned"],
"size": 6360000000
},
"engine": "nitro"
}