diff --git a/models/capybara-34b/model.json b/models/capybara-34b/model.json
index e4263f957..e337191a9 100644
--- a/models/capybara-34b/model.json
+++ b/models/capybara-34b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "",
- "user_prompt": "USER:\n",
- "ai_prompt": "ASSISTANT:\n"
+ "prompt_template": "USER:\n{prompt}\nASSISTANT:"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/deepseek-coder-1.3b/model.json b/models/deepseek-coder-1.3b/model.json
index 70f86ec89..77fa2d1f7 100644
--- a/models/deepseek-coder-1.3b/model.json
+++ b/models/deepseek-coder-1.3b/model.json
@@ -9,9 +9,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "",
- "user_prompt": "### Instruction:\n",
- "ai_prompt": "### Response:\n"
+ "prompt_template": "### Instruction:\n{prompt}\n### Response:"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/deepseek-coder-34b/model.json b/models/deepseek-coder-34b/model.json
index 402419357..80b224691 100644
--- a/models/deepseek-coder-34b/model.json
+++ b/models/deepseek-coder-34b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "",
- "user_prompt": "### Instruction:\n",
- "ai_prompt": "### Response:\n"
+ "prompt_template": "### Instruction:\n{prompt}\n### Response:"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/llama2-chat-70b-q4/model.json b/models/llama2-chat-70b-q4/model.json
index 07886aed5..abe44ddad 100644
--- a/models/llama2-chat-70b-q4/model.json
+++ b/models/llama2-chat-70b-q4/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "[INST] <>\n",
- "user_prompt": "<>\n",
- "ai_prompt": "[/INST]"
+ "prompt_template": "[INST] <>\n{system_message}<>\n{prompt}[/INST]"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/llama2-chat-7b-q4/model.json b/models/llama2-chat-7b-q4/model.json
index 2d1a67236..9ec601429 100644
--- a/models/llama2-chat-7b-q4/model.json
+++ b/models/llama2-chat-7b-q4/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "[INST] <>\n",
- "user_prompt": "<>\n",
- "ai_prompt": "[/INST]"
+ "prompt_template": "[INST] <>\n{system_message}<>\n{prompt}[/INST]"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/llama2-chat-7b-q5/model.json b/models/llama2-chat-7b-q5/model.json
index 96c652ab2..1e999003f 100644
--- a/models/llama2-chat-7b-q5/model.json
+++ b/models/llama2-chat-7b-q5/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "[INST] <>\n",
- "user_prompt": "<>\n",
- "ai_prompt": "[/INST]"
+ "prompt_template": "[INST] <>\n{system_message}<>\n{prompt}[/INST]"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/lzlv-70b/model.json b/models/lzlv-70b/model.json
index ca6af617e..cb2387432 100644
--- a/models/lzlv-70b/model.json
+++ b/models/lzlv-70b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "",
- "user_prompt": "USER:\n",
- "ai_prompt": "ASSISTANT:\n"
+ "prompt_template": "USER:\n{prompt}\nASSISTANT:"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/mistral-ins-7b-q4/model.json b/models/mistral-ins-7b-q4/model.json
index 9fc86f1be..9f2c22a03 100644
--- a/models/mistral-ins-7b-q4/model.json
+++ b/models/mistral-ins-7b-q4/model.json
@@ -10,7 +10,8 @@
"ctx_len": 2048,
"system_prompt": "",
"user_prompt": "[INST]",
- "ai_prompt": "[/INST]"
+ "ai_prompt": "[/INST]",
+ "prompt_template": "[INST]{prompt}\n[/INST]"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/mistral-ins-7b-q5/model.json b/models/mistral-ins-7b-q5/model.json
index 291f2ad0e..328ba2926 100644
--- a/models/mistral-ins-7b-q5/model.json
+++ b/models/mistral-ins-7b-q5/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "",
- "user_prompt": "[INST]",
- "ai_prompt": "[/INST]"
+ "prompt_template": "[INST]{prompt}\n[/INST]"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/neural-chat-7b/model.json b/models/neural-chat-7b/model.json
index 1d6271469..20c14c507 100644
--- a/models/neural-chat-7b/model.json
+++ b/models/neural-chat-7b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "### System:\n",
- "user_prompt": "### User:\n",
- "ai_prompt": "### Assistant:\n"
+ "prompt_template": "### System:\n{system_message}### User:\n{prompt}### Assistant:"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/noromaid-20b/model.json b/models/noromaid-20b/model.json
index 5c937a831..961c4bd25 100644
--- a/models/noromaid-20b/model.json
+++ b/models/noromaid-20b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "",
- "user_prompt": "### Instruction:\n",
- "ai_prompt": "### Response:\n"
+ "prompt_template": "### Instruction:{prompt}\n### Response:"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/openhermes-neural-7b/model.json b/models/openhermes-neural-7b/model.json
index 8dcb51ad7..bc4a4e315 100644
--- a/models/openhermes-neural-7b/model.json
+++ b/models/openhermes-neural-7b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "<|im_start|>system\n",
- "user_prompt": "<|im_end|>\n<|im_start|>user\n",
- "ai_prompt": "<|im_end|>\n<|im_start|>assistant\n"
+ "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/openorca-13b/model.json b/models/openorca-13b/model.json
index 42c8bd96e..3478ad528 100644
--- a/models/openorca-13b/model.json
+++ b/models/openorca-13b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "<|im_start|>system\n",
- "user_prompt": "<|im_end|>\n<|im_start|>user\n",
- "ai_prompt": "<|im_end|>\n<|im_start|>assistant\n"
+ "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/phind-34b/model.json b/models/phind-34b/model.json
index eb3c1a18a..d3fc6664a 100644
--- a/models/phind-34b/model.json
+++ b/models/phind-34b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "### System Prompt\n",
- "user_prompt": "### User Message\n",
- "ai_prompt": "### Assistant\n"
+ "prompt_template": "### System Prompt\n{system_message}\n### User Message\n{prompt}\n### Assistant"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/rocket-3b/model.json b/models/rocket-3b/model.json
index 8d4fc76bf..02335a0ed 100644
--- a/models/rocket-3b/model.json
+++ b/models/rocket-3b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 4096,
- "system_prompt": "<|im_start|>system\n",
- "user_prompt": "<|im_end|>\n<|im_start|>user\n",
- "ai_prompt": "<|im_end|>\n<|im_start|>assistant\n"
+ "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/starling-7b/model.json b/models/starling-7b/model.json
index 52b03f8b1..cdb06c2c7 100644
--- a/models/starling-7b/model.json
+++ b/models/starling-7b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "",
- "user_prompt": "GPT4 User: ",
- "ai_prompt": "<|end_of_turn|>\nGPT4 Assistant: "
+ "prompt_template": "GPT4 User: {prompt}<|end_of_turn|>GPT4 Assistant:"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/tiefighter-13b/model.json b/models/tiefighter-13b/model.json
index 20075777c..3f5abe3e3 100644
--- a/models/tiefighter-13b/model.json
+++ b/models/tiefighter-13b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "",
- "user_prompt": "### Instruction: ",
- "ai_prompt": "\n### Response: "
+ "prompt_template": "### Instruction:\n{prompt}\n### Response:"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/tinyllama-1.1b/model.json b/models/tinyllama-1.1b/model.json
index bb6aeaf5c..d924be046 100644
--- a/models/tinyllama-1.1b/model.json
+++ b/models/tinyllama-1.1b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "<|system|>\n",
- "user_prompt": "<|user|>\n",
- "ai_prompt": "<|assistant|>\n"
+ "prompt_template": "<|system|>\n{system_message}<|user|>\n{prompt}<|assistant|>"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/wizardcoder-13b/model.json b/models/wizardcoder-13b/model.json
index 40f275037..774c1a424 100644
--- a/models/wizardcoder-13b/model.json
+++ b/models/wizardcoder-13b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "",
- "user_prompt": "### Instruction:\n",
- "ai_prompt": "### Response:\n"
+ "prompt_template": "### Instruction:\n{prompt}\n### Response:"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/yi-34b/model.json b/models/yi-34b/model.json
index ab111591c..6a570b4de 100644
--- a/models/yi-34b/model.json
+++ b/models/yi-34b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "<|im_start|>system\n",
- "user_prompt": "<|im_end|>\n<|im_start|>user\n",
- "ai_prompt": "<|im_end|>\n<|im_start|>assistant\n"
+ "prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant"
},
"parameters": {
"max_tokens": 2048
diff --git a/models/zephyr-beta-7b/model.json b/models/zephyr-beta-7b/model.json
index 4993366fd..b12f03c67 100644
--- a/models/zephyr-beta-7b/model.json
+++ b/models/zephyr-beta-7b/model.json
@@ -8,9 +8,7 @@
"format": "gguf",
"settings": {
"ctx_len": 2048,
- "system_prompt": "<|system|>\n",
- "user_prompt": "\n<|user|>\n",
- "ai_prompt": "\n<|assistant|>\n"
+ "prompt_template": "<|system|>\n{system_message}\n<|user|>\n{prompt}\n<|assistant|>"
},
"parameters": {
"max_tokens": 2048