From f6d8929e9b643815424dd09e1d99a1297736548c Mon Sep 17 00:00:00 2001 From: vuonghoainam Date: Tue, 12 Sep 2023 18:09:24 +0700 Subject: [PATCH] fix: Fix README with minor bug in step to download model --- README.md | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index d032a6917..d9282bb67 100644 --- a/README.md +++ b/README.md @@ -136,14 +136,13 @@ pip install 'llama-cpp-python[server]' We recommend that Llama2-7B (4-bit quantized) as a basic model to get started. -You will need to download the models to the `models` folder. +You will need to download the models to the `models` folder at root level. ```shell -mkdir -p models # Downloads model (~4gb) # Download time depends on your internet connection and HuggingFace's bandwidth # In this part, please head over to any source contains `.gguf` format model - https://huggingface.co/models?search=gguf -wget LLM_MODEL_URL=https://huggingface.co/TheBloke/CodeLlama-13B-GGUF/resolve/main/codellama-13b.Q3_K_L.gguf -P models +wget https://huggingface.co/TheBloke/CodeLlama-13B-GGUF/resolve/main/codellama-13b.Q3_K_L.gguf -P models ``` - Run the model in host machine