From 72d178f3c389845092da81d4de768ff7307ae517 Mon Sep 17 00:00:00 2001 From: Sharun Date: Sat, 19 Oct 2024 19:43:33 -0400 Subject: [PATCH] update max_tokens for llama-3.1-8b-instant --- extensions/inference-groq-extension/resources/models.json | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/extensions/inference-groq-extension/resources/models.json b/extensions/inference-groq-extension/resources/models.json index d1c8c4289..04b60bfdd 100644 --- a/extensions/inference-groq-extension/resources/models.json +++ b/extensions/inference-groq-extension/resources/models.json @@ -106,7 +106,7 @@ "format": "api", "settings": {}, "parameters": { - "max_tokens": 8192, + "max_tokens": 8000, "temperature": 0.7, "top_p": 0.95, "stream": true,