diff options
Diffstat (limited to 'llama.cpp/tools/export-lora/README.md')
| -rw-r--r-- | llama.cpp/tools/export-lora/README.md | 33 |
1 files changed, 33 insertions, 0 deletions
diff --git a/llama.cpp/tools/export-lora/README.md b/llama.cpp/tools/export-lora/README.md new file mode 100644 index 0000000..7dce99c --- /dev/null +++ b/llama.cpp/tools/export-lora/README.md @@ -0,0 +1,33 @@ +# export-lora + +Apply LORA adapters to base model and export the resulting model. + +``` +usage: llama-export-lora [options] + +options: + -m, --model model path from which to load base model (default '') + --lora FNAME path to LoRA adapter (can be repeated to use multiple adapters) + --lora-scaled FNAME S path to LoRA adapter with user defined scaling S (can be repeated to use multiple adapters) + -t, --threads N number of threads to use during computation (default: 4) + -o, --output FNAME output file (default: 'ggml-lora-merged-f16.gguf') +``` + +For example: + +```bash +./bin/llama-export-lora \ + -m open-llama-3b-v2.gguf \ + -o open-llama-3b-v2-english2tokipona-chat.gguf \ + --lora lora-open-llama-3b-v2-english2tokipona-chat-LATEST.gguf +``` + +Multiple LORA adapters can be applied by passing multiple `--lora FNAME` or `--lora-scaled FNAME S` command line parameters: + +```bash +./bin/llama-export-lora \ + -m your_base_model.gguf \ + -o your_merged_model.gguf \ + --lora-scaled lora_task_A.gguf 0.5 \ + --lora-scaled lora_task_B.gguf 0.5 +``` |
