summaryrefslogtreecommitdiff
path: root/llama.cpp/tools/export-lora/README.md
diff options
context:
space:
mode:
Diffstat (limited to 'llama.cpp/tools/export-lora/README.md')
-rw-r--r--llama.cpp/tools/export-lora/README.md33
1 files changed, 33 insertions, 0 deletions
diff --git a/llama.cpp/tools/export-lora/README.md b/llama.cpp/tools/export-lora/README.md
new file mode 100644
index 0000000..7dce99c
--- /dev/null
+++ b/llama.cpp/tools/export-lora/README.md
@@ -0,0 +1,33 @@
+# export-lora
+
+Apply LORA adapters to base model and export the resulting model.
+
+```
+usage: llama-export-lora [options]
+
+options:
+ -m, --model model path from which to load base model (default '')
+ --lora FNAME path to LoRA adapter (can be repeated to use multiple adapters)
+ --lora-scaled FNAME S path to LoRA adapter with user defined scaling S (can be repeated to use multiple adapters)
+ -t, --threads N number of threads to use during computation (default: 4)
+ -o, --output FNAME output file (default: 'ggml-lora-merged-f16.gguf')
+```
+
+For example:
+
+```bash
+./bin/llama-export-lora \
+ -m open-llama-3b-v2.gguf \
+ -o open-llama-3b-v2-english2tokipona-chat.gguf \
+ --lora lora-open-llama-3b-v2-english2tokipona-chat-LATEST.gguf
+```
+
+Multiple LORA adapters can be applied by passing multiple `--lora FNAME` or `--lora-scaled FNAME S` command line parameters:
+
+```bash
+./bin/llama-export-lora \
+ -m your_base_model.gguf \
+ -o your_merged_model.gguf \
+ --lora-scaled lora_task_A.gguf 0.5 \
+ --lora-scaled lora_task_B.gguf 0.5
+```