summaryrefslogtreecommitdiff
path: root/llama.cpp/examples/convert-llama2c-to-ggml/README.md
diff options
context:
space:
mode:
Diffstat (limited to 'llama.cpp/examples/convert-llama2c-to-ggml/README.md')
-rw-r--r--llama.cpp/examples/convert-llama2c-to-ggml/README.md25
1 files changed, 25 insertions, 0 deletions
diff --git a/llama.cpp/examples/convert-llama2c-to-ggml/README.md b/llama.cpp/examples/convert-llama2c-to-ggml/README.md
new file mode 100644
index 0000000..46a42da
--- /dev/null
+++ b/llama.cpp/examples/convert-llama2c-to-ggml/README.md
@@ -0,0 +1,25 @@
+## Convert llama2.c model to ggml
+
+This example reads weights from project [llama2.c](https://github.com/karpathy/llama2.c) and saves them in ggml compatible format. The vocab that is available in `models/ggml-vocab.bin` is used by default.
+
+To convert the model first download the models from the [llama2.c](https://github.com/karpathy/llama2.c) repository.
+
+```
+usage: ./llama-convert-llama2c-to-ggml [options]
+
+options:
+ -h, --help show this help message and exit
+ --copy-vocab-from-model FNAME path of gguf llama model or llama2.c vocabulary from which to copy vocab (default 'models/7B/ggml-model-f16.gguf')
+ --llama2c-model FNAME [REQUIRED] model path from which to load Karpathy's llama2.c model
+ --llama2c-output-model FNAME model path to save the converted llama2.c model (default ak_llama_model.bin')
+```
+
+An example command using a model from [karpathy/tinyllamas](https://huggingface.co/karpathy/tinyllamas) is as follows:
+
+`$ ./llama-convert-llama2c-to-ggml --copy-vocab-from-model llama-2-7b-chat.gguf.q2_K.bin --llama2c-model stories42M.bin --llama2c-output-model stories42M.gguf.bin`
+
+Note: The vocabulary for `stories260K.bin` should be its own tokenizer `tok512.bin` found in [karpathy/tinyllamas/stories260K](https://huggingface.co/karpathy/tinyllamas/tree/main/stories260K).
+
+Now you can use the model with a command like:
+
+`$ ./llama-cli -m stories42M.gguf.bin -p "One day, Lily met a Shoggoth" -n 500 -c 256`