From 6b9478ccd5812c3f6b83fe7736bbb3bc56851f8b Mon Sep 17 00:00:00 2001 From: akawrykow Date: Tue, 22 Aug 2023 17:45:58 -0700 Subject: [PATCH 1/2] Remove reference to tokenizer_checklist.chk --- README.md | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index 82e070ac3ef0d..92420bbd40ff9 100644 --- a/README.md +++ b/README.md @@ -504,7 +504,8 @@ Building the program with BLAS support may lead to some performance improvements ```bash # obtain the original LLaMA model weights and place them in ./models ls ./models -65B 30B 13B 7B tokenizer_checklist.chk tokenizer.model +65B 30B 13B 7B tokenizer.model + # [Optional] for models using BPE tokenizers ls ./models 65B 30B 13B 7B vocab.json From 4f8d62e4442392b38a56f75b8980e2b3558cafc4 Mon Sep 17 00:00:00 2001 From: akawrykow Date: Tue, 22 Aug 2023 17:48:48 -0700 Subject: [PATCH 2/2] Add --ctx param to Prepare Data & Run section --- README.md | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index 92420bbd40ff9..c16eced419e74 100644 --- a/README.md +++ b/README.md @@ -514,7 +514,8 @@ ls ./models python3 -m pip install -r requirements.txt # convert the 7B model to ggml FP16 format -python3 convert.py models/7B/ +# Use --ctx 4096 for LLaMA v2 +python3 convert.py models/7B/ --ctx 2048 # [Optional] for models using BPE tokenizers python convert.py models/7B/ --vocabtype bpe