summaryrefslogtreecommitdiff
path: root/llama.cpp/examples/embedding/README.md
diff options
context:
space:
mode:
authorMitja Felicijan <mitja.felicijan@gmail.com>2026-02-12 20:57:17 +0100
committerMitja Felicijan <mitja.felicijan@gmail.com>2026-02-12 20:57:17 +0100
commitb333b06772c89d96aacb5490d6a219fba7c09cc6 (patch)
tree211df60083a5946baa2ed61d33d8121b7e251b06 /llama.cpp/examples/embedding/README.md
downloadllmnpc-b333b06772c89d96aacb5490d6a219fba7c09cc6.tar.gz
Engage!
Diffstat (limited to 'llama.cpp/examples/embedding/README.md')
-rw-r--r--llama.cpp/examples/embedding/README.md61
1 files changed, 61 insertions, 0 deletions
diff --git a/llama.cpp/examples/embedding/README.md b/llama.cpp/examples/embedding/README.md
new file mode 100644
index 0000000..1684f36
--- /dev/null
+++ b/llama.cpp/examples/embedding/README.md
@@ -0,0 +1,61 @@
+# llama.cpp/example/embedding
+
+This example demonstrates generate high-dimensional embedding vector of a given text with llama.cpp.
+
+## Quick Start
+
+To get started right away, run the following command, making sure to use the correct path for the model you have:
+
+### Unix-based systems (Linux, macOS, etc.):
+
+```bash
+./llama-embedding -m ./path/to/model --pooling mean --log-disable -p "Hello World!" 2>/dev/null
+```
+
+### Windows:
+
+```powershell
+llama-embedding.exe -m ./path/to/model --pooling mean --log-disable -p "Hello World!" 2>$null
+```
+
+The above command will output space-separated float values.
+
+## extra parameters
+### --embd-normalize $integer$
+| $integer$ | description | formula |
+|-----------|---------------------|---------|
+| $-1$ | none |
+| $0$ | max absolute int16 | $\Large{{32760 * x_i} \over\max \lvert x_i\rvert}$
+| $1$ | taxicab | $\Large{x_i \over\sum \lvert x_i\rvert}$
+| $2$ | euclidean (default) | $\Large{x_i \over\sqrt{\sum x_i^2}}$
+| $>2$ | p-norm | $\Large{x_i \over\sqrt[p]{\sum \lvert x_i\rvert^p}}$
+
+### --embd-output-format $'string'$
+| $'string'$ | description | |
+|------------|------------------------------|--|
+| '' | same as before | (default)
+| 'array' | single embeddings | $[[x_1,...,x_n]]$
+| | multiple embeddings | $[[x_1,...,x_n],[x_1,...,x_n],...,[x_1,...,x_n]]$
+| 'json' | openai style |
+| 'json+' | add cosine similarity matrix |
+| 'raw' | plain text output |
+
+### --embd-separator $"string"$
+| $"string"$ | |
+|--------------|-|
+| "\n" | (default)
+| "<#embSep#>" | for example
+| "<#sep#>" | other example
+
+## examples
+### Unix-based systems (Linux, macOS, etc.):
+
+```bash
+./llama-embedding -p 'Castle<#sep#>Stronghold<#sep#>Dog<#sep#>Cat' --pooling mean --embd-separator '<#sep#>' --embd-normalize 2 --embd-output-format '' -m './path/to/model.gguf' --n-gpu-layers 99 --log-disable 2>/dev/null
+```
+
+### Windows:
+
+```powershell
+llama-embedding.exe -p 'Castle<#sep#>Stronghold<#sep#>Dog<#sep#>Cat' --pooling mean --embd-separator '<#sep#>' --embd-normalize 2 --embd-output-format '' -m './path/to/model.gguf' --n-gpu-layers 99 --log-disable 2>/dev/null
+```