summaryrefslogtreecommitdiff
path: root/llama.cpp/tools/export-lora/README.md
diff options
context:
space:
mode:
authorMitja Felicijan <mitja.felicijan@gmail.com>2026-02-12 20:57:17 +0100
committerMitja Felicijan <mitja.felicijan@gmail.com>2026-02-12 20:57:17 +0100
commitb333b06772c89d96aacb5490d6a219fba7c09cc6 (patch)
tree211df60083a5946baa2ed61d33d8121b7e251b06 /llama.cpp/tools/export-lora/README.md
downloadllmnpc-b333b06772c89d96aacb5490d6a219fba7c09cc6.tar.gz
Engage!
Diffstat (limited to 'llama.cpp/tools/export-lora/README.md')
-rw-r--r--llama.cpp/tools/export-lora/README.md33
1 files changed, 33 insertions, 0 deletions
diff --git a/llama.cpp/tools/export-lora/README.md b/llama.cpp/tools/export-lora/README.md
new file mode 100644
index 0000000..7dce99c
--- /dev/null
+++ b/llama.cpp/tools/export-lora/README.md
@@ -0,0 +1,33 @@
+# export-lora
+
+Apply LORA adapters to base model and export the resulting model.
+
+```
+usage: llama-export-lora [options]
+
+options:
+ -m, --model model path from which to load base model (default '')
+ --lora FNAME path to LoRA adapter (can be repeated to use multiple adapters)
+ --lora-scaled FNAME S path to LoRA adapter with user defined scaling S (can be repeated to use multiple adapters)
+ -t, --threads N number of threads to use during computation (default: 4)
+ -o, --output FNAME output file (default: 'ggml-lora-merged-f16.gguf')
+```
+
+For example:
+
+```bash
+./bin/llama-export-lora \
+ -m open-llama-3b-v2.gguf \
+ -o open-llama-3b-v2-english2tokipona-chat.gguf \
+ --lora lora-open-llama-3b-v2-english2tokipona-chat-LATEST.gguf
+```
+
+Multiple LORA adapters can be applied by passing multiple `--lora FNAME` or `--lora-scaled FNAME S` command line parameters:
+
+```bash
+./bin/llama-export-lora \
+ -m your_base_model.gguf \
+ -o your_merged_model.gguf \
+ --lora-scaled lora_task_A.gguf 0.5 \
+ --lora-scaled lora_task_B.gguf 0.5
+```