fix: change Phi-3 Mini Instruct 4K to cortexso repository (#3655)
* fix: change Phi-3 Mini Instruct 4K to cortexso repository * chore: correct model names * fix: Phi3-Medium prompt correction * chore: bump versions * chore: take out some feature tags
This commit is contained in:
parent
592b3e69a5
commit
ba1ba89fa3
@ -69,9 +69,9 @@
|
||||
],
|
||||
"id": "gemma-7b-it",
|
||||
"object": "model",
|
||||
"name": "Groq Gemma 7b Instruct",
|
||||
"version": "1.1",
|
||||
"description": "Groq Gemma 7b Instruct with supercharged speed!",
|
||||
"name": "Groq Gemma 7B Instruct",
|
||||
"version": "1.2",
|
||||
"description": "Groq Gemma 7B Instruct with supercharged speed!",
|
||||
"format": "api",
|
||||
"settings": {},
|
||||
"parameters": {
|
||||
@ -99,9 +99,9 @@
|
||||
],
|
||||
"id": "mixtral-8x7b-32768",
|
||||
"object": "model",
|
||||
"name": "Groq Mixtral 8x7b Instruct",
|
||||
"version": "1.1",
|
||||
"description": "Groq Mixtral 8x7b Instruct is Mixtral with supercharged speed!",
|
||||
"name": "Groq Mixtral 8x7B Instruct",
|
||||
"version": "1.2",
|
||||
"description": "Groq Mixtral 8x7B Instruct is Mixtral with supercharged speed!",
|
||||
"format": "api",
|
||||
"settings": {},
|
||||
"parameters": {
|
||||
|
||||
@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "@janhq/inference-cortex-extension",
|
||||
"productName": "Cortex Inference Engine",
|
||||
"version": "1.0.15",
|
||||
"version": "1.0.16",
|
||||
"description": "This extension embeds cortex.cpp, a lightweight inference engine written in C++. See https://jan.ai.\nAdditional dependencies could be installed to run without Cuda Toolkit installation.",
|
||||
"main": "dist/index.js",
|
||||
"node": "dist/node/index.cjs.js",
|
||||
|
||||
@ -8,7 +8,7 @@
|
||||
"id": "aya-23-8b",
|
||||
"object": "model",
|
||||
"name": "Aya 23 8B Q4",
|
||||
"version": "1.1",
|
||||
"version": "1.2",
|
||||
"description": "Aya 23 can talk upto 23 languages fluently.",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
@ -28,7 +28,7 @@
|
||||
},
|
||||
"metadata": {
|
||||
"author": "CohereForAI",
|
||||
"tags": ["7B", "Finetuned","Featured"],
|
||||
"tags": ["7B", "Finetuned"],
|
||||
"size": 5056982144
|
||||
},
|
||||
"engine": "nitro"
|
||||
|
||||
@ -7,8 +7,8 @@
|
||||
],
|
||||
"id": "deepseek-coder-1.3b",
|
||||
"object": "model",
|
||||
"name": "Deepseek Coder 1.3B Q8",
|
||||
"version": "1.2",
|
||||
"name": "Deepseek Coder 1.3B Instruct Q8",
|
||||
"version": "1.3",
|
||||
"description": "Deepseek Coder excelled in project-level code completion with advanced capabilities across multiple programming languages.",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
|
||||
@ -7,8 +7,8 @@
|
||||
],
|
||||
"id": "deepseek-coder-34b",
|
||||
"object": "model",
|
||||
"name": "Deepseek Coder 33B Q4",
|
||||
"version": "1.2",
|
||||
"name": "Deepseek Coder 33B Instruct Q4",
|
||||
"version": "1.3",
|
||||
"description": "Deepseek Coder excelled in project-level code completion with advanced capabilities across multiple programming languages.",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
|
||||
@ -8,7 +8,7 @@
|
||||
"id": "gemma-1.1-7b-it",
|
||||
"object": "model",
|
||||
"name": "Gemma 1.1 7B Q4",
|
||||
"version": "1.2",
|
||||
"version": "1.3",
|
||||
"description": "Google's Gemma is built for multilingual purpose",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
@ -28,7 +28,7 @@
|
||||
},
|
||||
"metadata": {
|
||||
"author": "Google",
|
||||
"tags": ["7B", "Finetuned", "Featured"],
|
||||
"tags": ["7B", "Finetuned"],
|
||||
"size": 5330000000
|
||||
},
|
||||
"engine": "nitro"
|
||||
|
||||
@ -7,8 +7,8 @@
|
||||
],
|
||||
"id": "llama3-8b-instruct",
|
||||
"object": "model",
|
||||
"name": "Llama 3 8B Q4",
|
||||
"version": "1.2",
|
||||
"name": "Llama 3 8B Instruct Q4",
|
||||
"version": "1.4",
|
||||
"description": "Meta's Llama 3 excels at general usage situations, including chat, general world knowledge, and coding.",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
@ -28,7 +28,7 @@
|
||||
},
|
||||
"metadata": {
|
||||
"author": "MetaAI",
|
||||
"tags": ["8B", "Featured"],
|
||||
"tags": ["8B"],
|
||||
"size": 4920000000
|
||||
},
|
||||
"engine": "nitro"
|
||||
|
||||
@ -7,8 +7,8 @@
|
||||
],
|
||||
"id": "llama3.1-70b-instruct",
|
||||
"object": "model",
|
||||
"name": "Llama 3.1 70B Q4 Instruct",
|
||||
"version": "1.0",
|
||||
"name": "Llama 3.1 70B Instruct Q4",
|
||||
"version": "1.1",
|
||||
"description": "Meta's Llama 3.1 excels at general usage situations, including chat, general world knowledge, and coding.",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
|
||||
@ -7,8 +7,8 @@
|
||||
],
|
||||
"id": "llama3.1-8b-instruct",
|
||||
"object": "model",
|
||||
"name": "Llama 3.1 8B Q4 Instruct",
|
||||
"version": "1.0",
|
||||
"name": "Llama 3.1 8B Instruct Q4",
|
||||
"version": "1.1",
|
||||
"description": "Meta's Llama 3.1 excels at general usage situations, including chat, general world knowledge, and coding.",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
|
||||
@ -7,9 +7,9 @@
|
||||
],
|
||||
"id": "mistral-ins-7b-q4",
|
||||
"object": "model",
|
||||
"name": "Mistral Instruct 7B Q4",
|
||||
"version": "1.3",
|
||||
"description": "Mistral Instruct 7b model, specifically designed for a comprehensive understanding of the world.",
|
||||
"name": "Mistral 7B Instruct Q4",
|
||||
"version": "1.5",
|
||||
"description": "Mistral 7B Instruct model, specifically designed for a comprehensive understanding of the world.",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
"ctx_len": 32768,
|
||||
@ -28,7 +28,7 @@
|
||||
},
|
||||
"metadata": {
|
||||
"author": "MistralAI",
|
||||
"tags": ["Featured", "7B", "Foundational Model"],
|
||||
"tags": ["7B", "Foundational Model"],
|
||||
"size": 4370000000,
|
||||
"cover": "https://raw.githubusercontent.com/janhq/jan/dev/models/mistral-ins-7b-q4/cover.png"
|
||||
},
|
||||
|
||||
@ -1,20 +1,20 @@
|
||||
{
|
||||
"sources": [
|
||||
{
|
||||
"url": "https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf/resolve/main/Phi-3-mini-4k-instruct-q4.gguf",
|
||||
"filename": "Phi-3-mini-4k-instruct-q4.gguf"
|
||||
"url": "https://huggingface.co/cortexso/phi3/resolve/main/model.gguf",
|
||||
"filename": "model.gguf"
|
||||
}
|
||||
],
|
||||
"id": "phi3-3.8b",
|
||||
"object": "model",
|
||||
"name": "Phi-3 Mini",
|
||||
"version": "1.2",
|
||||
"name": "Phi-3 Mini Instruct Q4",
|
||||
"version": "1.3",
|
||||
"description": "Phi-3 Mini is Microsoft's newest, compact model designed for mobile use.",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
"ctx_len": 4096,
|
||||
"prompt_template": "<|user|>\n{prompt}<|end|>\n<|assistant|>\n",
|
||||
"llama_model_path": "Phi-3-mini-4k-instruct-q4.gguf",
|
||||
"llama_model_path": "model.gguf",
|
||||
"ngl": 33
|
||||
},
|
||||
"parameters": {
|
||||
@ -35,4 +35,4 @@
|
||||
"size": 2320000000
|
||||
},
|
||||
"engine": "nitro"
|
||||
}
|
||||
}
|
||||
@ -7,13 +7,13 @@
|
||||
],
|
||||
"id": "phi3-medium",
|
||||
"object": "model",
|
||||
"name": "Phi-3 Medium",
|
||||
"version": "1.2",
|
||||
"name": "Phi-3 Medium Instruct Q4",
|
||||
"version": "1.3",
|
||||
"description": "Phi-3 Medium is Microsoft's latest SOTA model.",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
"ctx_len": 128000,
|
||||
"prompt_template": "<|user|>\n{prompt}<|end|>\n<|assistant|>\n",
|
||||
"prompt_template": "<|user|> {prompt}<|end|><|assistant|><|end|>",
|
||||
"llama_model_path": "Phi-3-medium-128k-instruct-Q4_K_M.gguf",
|
||||
"ngl": 33
|
||||
},
|
||||
|
||||
@ -7,8 +7,8 @@
|
||||
],
|
||||
"id": "qwen2-7b",
|
||||
"object": "model",
|
||||
"name": "Qwen 2 Instruct 7B Q4",
|
||||
"version": "1.1",
|
||||
"name": "Qwen 2 7B Instruct Q4",
|
||||
"version": "1.2",
|
||||
"description": "Qwen is optimized at Chinese, ideal for everyday tasks.",
|
||||
"format": "gguf",
|
||||
"settings": {
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user