summaryrefslogtreecommitdiff
path: root/llama.cpp/requirements
diff options
context:
space:
mode:
Diffstat (limited to 'llama.cpp/requirements')
-rw-r--r--llama.cpp/requirements/requirements-all.txt18
-rw-r--r--llama.cpp/requirements/requirements-compare-llama-bench.txt3
-rw-r--r--llama.cpp/requirements/requirements-convert_hf_to_gguf.txt9
-rw-r--r--llama.cpp/requirements/requirements-convert_hf_to_gguf_update.txt1
-rw-r--r--llama.cpp/requirements/requirements-convert_legacy_llama.txt7
-rw-r--r--llama.cpp/requirements/requirements-convert_llama_ggml_to_gguf.txt1
-rw-r--r--llama.cpp/requirements/requirements-convert_lora_to_gguf.txt4
-rw-r--r--llama.cpp/requirements/requirements-gguf_editor_gui.txt3
-rw-r--r--llama.cpp/requirements/requirements-pydantic.txt3
-rw-r--r--llama.cpp/requirements/requirements-server-bench.txt5
-rw-r--r--llama.cpp/requirements/requirements-test-tokenizer-random.txt1
-rw-r--r--llama.cpp/requirements/requirements-tool_bench.txt12
12 files changed, 67 insertions, 0 deletions
diff --git a/llama.cpp/requirements/requirements-all.txt b/llama.cpp/requirements/requirements-all.txt
new file mode 100644
index 0000000..6c6bea9
--- /dev/null
+++ b/llama.cpp/requirements/requirements-all.txt
@@ -0,0 +1,18 @@
+-r ../tools/mtmd/requirements.txt
+-r ../tools/server/bench/requirements.txt
+-r ../tools/server/tests/requirements.txt
+
+-r ./requirements-compare-llama-bench.txt
+-r ./requirements-server-bench.txt
+-r ./requirements-pydantic.txt
+-r ./requirements-test-tokenizer-random.txt
+
+-r ./requirements-convert_hf_to_gguf.txt
+-r ./requirements-convert_hf_to_gguf_update.txt
+-r ./requirements-convert_legacy_llama.txt
+-r ./requirements-convert_llama_ggml_to_gguf.txt
+-r ./requirements-tool_bench.txt
+
+-r ./requirements-gguf_editor_gui.txt
+
+-r ../examples/model-conversion/requirements.txt
diff --git a/llama.cpp/requirements/requirements-compare-llama-bench.txt b/llama.cpp/requirements/requirements-compare-llama-bench.txt
new file mode 100644
index 0000000..d87e897
--- /dev/null
+++ b/llama.cpp/requirements/requirements-compare-llama-bench.txt
@@ -0,0 +1,3 @@
+tabulate~=0.9.0
+GitPython~=3.1.43
+matplotlib~=3.10.0
diff --git a/llama.cpp/requirements/requirements-convert_hf_to_gguf.txt b/llama.cpp/requirements/requirements-convert_hf_to_gguf.txt
new file mode 100644
index 0000000..122b478
--- /dev/null
+++ b/llama.cpp/requirements/requirements-convert_hf_to_gguf.txt
@@ -0,0 +1,9 @@
+-r ./requirements-convert_legacy_llama.txt
+--extra-index-url https://download.pytorch.org/whl/cpu
+
+## Embedding Gemma requires PyTorch 2.6.0 or later
+torch~=2.6.0; platform_machine != "s390x"
+
+# torch s390x packages can only be found from nightly builds
+--extra-index-url https://download.pytorch.org/whl/nightly
+torch>=0.0.0.dev0; platform_machine == "s390x"
diff --git a/llama.cpp/requirements/requirements-convert_hf_to_gguf_update.txt b/llama.cpp/requirements/requirements-convert_hf_to_gguf_update.txt
new file mode 100644
index 0000000..afe2747
--- /dev/null
+++ b/llama.cpp/requirements/requirements-convert_hf_to_gguf_update.txt
@@ -0,0 +1 @@
+-r ./requirements-convert_legacy_llama.txt
diff --git a/llama.cpp/requirements/requirements-convert_legacy_llama.txt b/llama.cpp/requirements/requirements-convert_legacy_llama.txt
new file mode 100644
index 0000000..4898bf7
--- /dev/null
+++ b/llama.cpp/requirements/requirements-convert_legacy_llama.txt
@@ -0,0 +1,7 @@
+numpy~=1.26.4
+sentencepiece>=0.1.98,<0.3.0
+
+transformers>=4.57.1,<5.0.0
+
+gguf>=0.1.0
+protobuf>=4.21.0,<5.0.0
diff --git a/llama.cpp/requirements/requirements-convert_llama_ggml_to_gguf.txt b/llama.cpp/requirements/requirements-convert_llama_ggml_to_gguf.txt
new file mode 100644
index 0000000..afe2747
--- /dev/null
+++ b/llama.cpp/requirements/requirements-convert_llama_ggml_to_gguf.txt
@@ -0,0 +1 @@
+-r ./requirements-convert_legacy_llama.txt
diff --git a/llama.cpp/requirements/requirements-convert_lora_to_gguf.txt b/llama.cpp/requirements/requirements-convert_lora_to_gguf.txt
new file mode 100644
index 0000000..d091d56
--- /dev/null
+++ b/llama.cpp/requirements/requirements-convert_lora_to_gguf.txt
@@ -0,0 +1,4 @@
+-r ./requirements-convert_hf_to_gguf.txt
+--extra-index-url https://download.pytorch.org/whl/cpu
+# torch s390x packages can only be found from nightly builds
+--extra-index-url https://download.pytorch.org/whl/nightly
diff --git a/llama.cpp/requirements/requirements-gguf_editor_gui.txt b/llama.cpp/requirements/requirements-gguf_editor_gui.txt
new file mode 100644
index 0000000..fd25336
--- /dev/null
+++ b/llama.cpp/requirements/requirements-gguf_editor_gui.txt
@@ -0,0 +1,3 @@
+numpy~=1.26.4
+PySide6~=6.9.0
+gguf>=0.17.0
diff --git a/llama.cpp/requirements/requirements-pydantic.txt b/llama.cpp/requirements/requirements-pydantic.txt
new file mode 100644
index 0000000..67d4c1e
--- /dev/null
+++ b/llama.cpp/requirements/requirements-pydantic.txt
@@ -0,0 +1,3 @@
+docstring_parser~=0.15
+pydantic~=2.11.7
+requests
diff --git a/llama.cpp/requirements/requirements-server-bench.txt b/llama.cpp/requirements/requirements-server-bench.txt
new file mode 100644
index 0000000..ea5849f
--- /dev/null
+++ b/llama.cpp/requirements/requirements-server-bench.txt
@@ -0,0 +1,5 @@
+datasets~=3.2.0
+matplotlib~=3.10.0
+numpy~=1.26.4
+requests~=2.32.3
+tqdm~=4.67.1
diff --git a/llama.cpp/requirements/requirements-test-tokenizer-random.txt b/llama.cpp/requirements/requirements-test-tokenizer-random.txt
new file mode 100644
index 0000000..2785e71
--- /dev/null
+++ b/llama.cpp/requirements/requirements-test-tokenizer-random.txt
@@ -0,0 +1 @@
+cffi~=1.16.0
diff --git a/llama.cpp/requirements/requirements-tool_bench.txt b/llama.cpp/requirements/requirements-tool_bench.txt
new file mode 100644
index 0000000..3bb74fb
--- /dev/null
+++ b/llama.cpp/requirements/requirements-tool_bench.txt
@@ -0,0 +1,12 @@
+aiohttp~=3.9.3
+pytest~=8.3.3
+huggingface_hub>=0.34.0,<1.0
+matplotlib~=3.10.0
+numpy~=1.26.4
+openai~=2.14.0
+pandas~=2.2.3
+prometheus-client~=0.20.0
+requests~=2.32.3
+wget~=3.2
+typer~=0.15.1
+seaborn~=0.13.2