diff --git a/models/capybara-34b/model.json b/models/capybara-34b/model.json index 7c669bbb7..74eb27eb9 100644 --- a/models/capybara-34b/model.json +++ b/models/capybara-34b/model.json @@ -9,8 +9,8 @@ "settings": { "ctx_len": 2048, "system_prompt": "", - "user_prompt": "USER: ", - "ai_prompt": "ASSISTANT: " + "user_prompt": "USER:\n", + "ai_prompt": "ASSISTANT:\n" }, "parameters": { "max_tokens": 2048 diff --git a/models/lzlv-70b/model.json b/models/lzlv-70b/model.json index 98d581f24..844bdadf4 100644 --- a/models/lzlv-70b/model.json +++ b/models/lzlv-70b/model.json @@ -9,8 +9,8 @@ "settings": { "ctx_len": 2048, "system_prompt": "", - "user_prompt": "USER: ", - "ai_prompt": "ASSISTANT: " + "user_prompt": "USER:\n", + "ai_prompt": "ASSISTANT:\n" }, "parameters": { "max_tokens": 2048 diff --git a/models/neural-chat-7b/model.json b/models/neural-chat-7b/model.json index 84c8d029f..cac717b8a 100644 --- a/models/neural-chat-7b/model.json +++ b/models/neural-chat-7b/model.json @@ -8,9 +8,9 @@ "format": "gguf", "settings": { "ctx_len": 2048, - "system_prompt": "### System: ", - "user_prompt": "### User: ", - "ai_prompt": "### Assistant: " + "system_prompt": "### System:\n", + "user_prompt": "### User:\n", + "ai_prompt": "### Assistant:\n" }, "parameters": { "max_tokens": 2048 diff --git a/models/neuralhermes-7b/model.json b/models/openhermes-neural-7b/model.json similarity index 52% rename from models/neuralhermes-7b/model.json rename to models/openhermes-neural-7b/model.json index 0cc83d7fb..cd4394dc5 100644 --- a/models/neuralhermes-7b/model.json +++ b/models/openhermes-neural-7b/model.json @@ -1,10 +1,10 @@ { - "source_url": "https://huggingface.co/TheBloke/NeuralHermes-2.5-Mistral-7B-GGUF/resolve/main/neuralhermes-2.5-mistral-7b.Q4_K_M.gguf", - "id": "neuralhermes-7b", + "source_url": "https://huggingface.co/TheBloke/OpenHermes-2.5-neural-chat-7B-v3-2-7B-GGUF/resolve/main/openhermes-2.5-neural-chat-7b-v3-2-7b.Q4_K_M.gguf", + "id": "openhermes-neural-7b", "object": "model", - "name": "NeuralHermes 7B", + "name": "OpenHermes Neural 7B", "version": "1.0", - "description": "NeuralHermes 2.5 has been enhanced using Direct Preference Optimization. This fine-tuning, inspired by the RLHF process of Neural-chat-7b and OpenHermes-2.5-Mistral-7B, has led to improved performance across several benchmarks.", + "description": "OpenHermes Neural is a merged model from OpenHermes-2.5-Mistral-7B and neural-chat-7b-v3-2 with the TIES method.", "format": "gguf", "settings": { "ctx_len": 2048,