diff options
| author | Mitja Felicijan <mitja.felicijan@gmail.com> | 2026-02-12 20:57:17 +0100 |
|---|---|---|
| committer | Mitja Felicijan <mitja.felicijan@gmail.com> | 2026-02-12 20:57:17 +0100 |
| commit | b333b06772c89d96aacb5490d6a219fba7c09cc6 (patch) | |
| tree | 211df60083a5946baa2ed61d33d8121b7e251b06 /llama.cpp/requirements | |
| download | llmnpc-b333b06772c89d96aacb5490d6a219fba7c09cc6.tar.gz | |
Engage!
Diffstat (limited to 'llama.cpp/requirements')
12 files changed, 67 insertions, 0 deletions
diff --git a/llama.cpp/requirements/requirements-all.txt b/llama.cpp/requirements/requirements-all.txt new file mode 100644 index 0000000..6c6bea9 --- /dev/null +++ b/llama.cpp/requirements/requirements-all.txt @@ -0,0 +1,18 @@ +-r ../tools/mtmd/requirements.txt +-r ../tools/server/bench/requirements.txt +-r ../tools/server/tests/requirements.txt + +-r ./requirements-compare-llama-bench.txt +-r ./requirements-server-bench.txt +-r ./requirements-pydantic.txt +-r ./requirements-test-tokenizer-random.txt + +-r ./requirements-convert_hf_to_gguf.txt +-r ./requirements-convert_hf_to_gguf_update.txt +-r ./requirements-convert_legacy_llama.txt +-r ./requirements-convert_llama_ggml_to_gguf.txt +-r ./requirements-tool_bench.txt + +-r ./requirements-gguf_editor_gui.txt + +-r ../examples/model-conversion/requirements.txt diff --git a/llama.cpp/requirements/requirements-compare-llama-bench.txt b/llama.cpp/requirements/requirements-compare-llama-bench.txt new file mode 100644 index 0000000..d87e897 --- /dev/null +++ b/llama.cpp/requirements/requirements-compare-llama-bench.txt @@ -0,0 +1,3 @@ +tabulate~=0.9.0 +GitPython~=3.1.43 +matplotlib~=3.10.0 diff --git a/llama.cpp/requirements/requirements-convert_hf_to_gguf.txt b/llama.cpp/requirements/requirements-convert_hf_to_gguf.txt new file mode 100644 index 0000000..122b478 --- /dev/null +++ b/llama.cpp/requirements/requirements-convert_hf_to_gguf.txt @@ -0,0 +1,9 @@ +-r ./requirements-convert_legacy_llama.txt +--extra-index-url https://download.pytorch.org/whl/cpu + +## Embedding Gemma requires PyTorch 2.6.0 or later +torch~=2.6.0; platform_machine != "s390x" + +# torch s390x packages can only be found from nightly builds +--extra-index-url https://download.pytorch.org/whl/nightly +torch>=0.0.0.dev0; platform_machine == "s390x" diff --git a/llama.cpp/requirements/requirements-convert_hf_to_gguf_update.txt b/llama.cpp/requirements/requirements-convert_hf_to_gguf_update.txt new file mode 100644 index 0000000..afe2747 --- /dev/null +++ b/llama.cpp/requirements/requirements-convert_hf_to_gguf_update.txt @@ -0,0 +1 @@ +-r ./requirements-convert_legacy_llama.txt diff --git a/llama.cpp/requirements/requirements-convert_legacy_llama.txt b/llama.cpp/requirements/requirements-convert_legacy_llama.txt new file mode 100644 index 0000000..4898bf7 --- /dev/null +++ b/llama.cpp/requirements/requirements-convert_legacy_llama.txt @@ -0,0 +1,7 @@ +numpy~=1.26.4 +sentencepiece>=0.1.98,<0.3.0 + +transformers>=4.57.1,<5.0.0 + +gguf>=0.1.0 +protobuf>=4.21.0,<5.0.0 diff --git a/llama.cpp/requirements/requirements-convert_llama_ggml_to_gguf.txt b/llama.cpp/requirements/requirements-convert_llama_ggml_to_gguf.txt new file mode 100644 index 0000000..afe2747 --- /dev/null +++ b/llama.cpp/requirements/requirements-convert_llama_ggml_to_gguf.txt @@ -0,0 +1 @@ +-r ./requirements-convert_legacy_llama.txt diff --git a/llama.cpp/requirements/requirements-convert_lora_to_gguf.txt b/llama.cpp/requirements/requirements-convert_lora_to_gguf.txt new file mode 100644 index 0000000..d091d56 --- /dev/null +++ b/llama.cpp/requirements/requirements-convert_lora_to_gguf.txt @@ -0,0 +1,4 @@ +-r ./requirements-convert_hf_to_gguf.txt +--extra-index-url https://download.pytorch.org/whl/cpu +# torch s390x packages can only be found from nightly builds +--extra-index-url https://download.pytorch.org/whl/nightly diff --git a/llama.cpp/requirements/requirements-gguf_editor_gui.txt b/llama.cpp/requirements/requirements-gguf_editor_gui.txt new file mode 100644 index 0000000..fd25336 --- /dev/null +++ b/llama.cpp/requirements/requirements-gguf_editor_gui.txt @@ -0,0 +1,3 @@ +numpy~=1.26.4 +PySide6~=6.9.0 +gguf>=0.17.0 diff --git a/llama.cpp/requirements/requirements-pydantic.txt b/llama.cpp/requirements/requirements-pydantic.txt new file mode 100644 index 0000000..67d4c1e --- /dev/null +++ b/llama.cpp/requirements/requirements-pydantic.txt @@ -0,0 +1,3 @@ +docstring_parser~=0.15 +pydantic~=2.11.7 +requests diff --git a/llama.cpp/requirements/requirements-server-bench.txt b/llama.cpp/requirements/requirements-server-bench.txt new file mode 100644 index 0000000..ea5849f --- /dev/null +++ b/llama.cpp/requirements/requirements-server-bench.txt @@ -0,0 +1,5 @@ +datasets~=3.2.0 +matplotlib~=3.10.0 +numpy~=1.26.4 +requests~=2.32.3 +tqdm~=4.67.1 diff --git a/llama.cpp/requirements/requirements-test-tokenizer-random.txt b/llama.cpp/requirements/requirements-test-tokenizer-random.txt new file mode 100644 index 0000000..2785e71 --- /dev/null +++ b/llama.cpp/requirements/requirements-test-tokenizer-random.txt @@ -0,0 +1 @@ +cffi~=1.16.0 diff --git a/llama.cpp/requirements/requirements-tool_bench.txt b/llama.cpp/requirements/requirements-tool_bench.txt new file mode 100644 index 0000000..3bb74fb --- /dev/null +++ b/llama.cpp/requirements/requirements-tool_bench.txt @@ -0,0 +1,12 @@ +aiohttp~=3.9.3 +pytest~=8.3.3 +huggingface_hub>=0.34.0,<1.0 +matplotlib~=3.10.0 +numpy~=1.26.4 +openai~=2.14.0 +pandas~=2.2.3 +prometheus-client~=0.20.0 +requests~=2.32.3 +wget~=3.2 +typer~=0.15.1 +seaborn~=0.13.2 |
