Merge pull request #4214 from janhq/chore/add-llava-ngl-in-model-json-files
chore: add NGL settings for vision models
This commit is contained in:
commit
bf5d121a9e
@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "@janhq/inference-cortex-extension",
|
||||
"productName": "Cortex Inference Engine",
|
||||
"version": "1.0.23",
|
||||
"version": "1.0.24",
|
||||
"description": "This extension embeds cortex.cpp, a lightweight inference engine written in C++. See https://jan.ai.\nAdditional dependencies could be installed to run without Cuda Toolkit installation.",
|
||||
"main": "dist/index.js",
|
||||
"node": "dist/node/index.cjs.js",
|
||||
|
||||
@ -21,7 +21,8 @@
|
||||
"ctx_len": 4096,
|
||||
"prompt_template": "\n### Instruction:\n{prompt}\n### Response:\n",
|
||||
"llama_model_path": "ggml-model-q5_k.gguf",
|
||||
"mmproj": "mmproj-model-f16.gguf"
|
||||
"mmproj": "mmproj-model-f16.gguf",
|
||||
"ngl": 33
|
||||
},
|
||||
"parameters": {
|
||||
"max_tokens": 4096
|
||||
|
||||
@ -21,7 +21,8 @@
|
||||
"ctx_len": 4096,
|
||||
"prompt_template": "\n### Instruction:\n{prompt}\n### Response:\n",
|
||||
"llama_model_path": "llava-v1.6-vicuna-13b.Q4_K_M.gguf",
|
||||
"mmproj": "mmproj-model-f16.gguf"
|
||||
"mmproj": "mmproj-model-f16.gguf",
|
||||
"ngl": 33
|
||||
},
|
||||
"parameters": {
|
||||
"max_tokens": 4096,
|
||||
|
||||
@ -21,7 +21,8 @@
|
||||
"ctx_len": 4096,
|
||||
"prompt_template": "\n### Instruction:\n{prompt}\n### Response:\n",
|
||||
"llama_model_path": "llava-v1.6-mistral-7b.Q4_K_M.gguf",
|
||||
"mmproj": "mmproj-model-f16.gguf"
|
||||
"mmproj": "mmproj-model-f16.gguf",
|
||||
"ngl": 33
|
||||
},
|
||||
"parameters": {
|
||||
"max_tokens": 4096,
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user