add : img of using nitro in jan
This commit is contained in:
parent
ef36b6b5dc
commit
e9362892e0
BIN
docs/blog/img/nitro-on-jan.png
Normal file
BIN
docs/blog/img/nitro-on-jan.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 226 KiB |
@ -35,7 +35,7 @@ Mistral 7B continues to outshine [Meta's Llama-2 7B](https://huggingface.co/meta
|
||||
|
||||
Having a robust base model is critical. In our experiments, using Mistral as a starting point ensured the highest accuracy for subsequent specialized adaptations.
|
||||
|
||||

|
||||

|
||||
|
||||
*Figure 1. Mistral 7B excels in benchmarks, ranking among the top foundational models.*
|
||||
|
||||
@ -93,4 +93,8 @@ Training was done with supervised finetuning (SFT) from the [Hugging Face's alig
|
||||
|
||||
We used consumer-grade, dual Nvidia RTX 4090s for the training. The end-to-end training took 18 minutes. We found optimal hyperparameters in LoRA for this specific task to be `r = 256` and `alpha = 512`.
|
||||
|
||||
This final model can be found [here on Huggingface](https://huggingface.co/jan-hq/nitro-v1.2-e3).
|
||||
This final model can be found [here on Huggingface](https://huggingface.co/jan-hq/nitro-v1.2-e3).
|
||||
|
||||

|
||||
|
||||
*Figure 3. Using the new finetuned model in [Jan](https://jan.ai/)*
|
||||
Loading…
x
Reference in New Issue
Block a user