From b333b06772c89d96aacb5490d6a219fba7c09cc6 Mon Sep 17 00:00:00 2001 From: Mitja Felicijan Date: Thu, 12 Feb 2026 20:57:17 +0100 Subject: Engage! --- llama.cpp/requirements/requirements-convert_hf_to_gguf.txt | 9 +++++++++ 1 file changed, 9 insertions(+) create mode 100644 llama.cpp/requirements/requirements-convert_hf_to_gguf.txt (limited to 'llama.cpp/requirements/requirements-convert_hf_to_gguf.txt') diff --git a/llama.cpp/requirements/requirements-convert_hf_to_gguf.txt b/llama.cpp/requirements/requirements-convert_hf_to_gguf.txt new file mode 100644 index 0000000..122b478 --- /dev/null +++ b/llama.cpp/requirements/requirements-convert_hf_to_gguf.txt @@ -0,0 +1,9 @@ +-r ./requirements-convert_legacy_llama.txt +--extra-index-url https://download.pytorch.org/whl/cpu + +## Embedding Gemma requires PyTorch 2.6.0 or later +torch~=2.6.0; platform_machine != "s390x" + +# torch s390x packages can only be found from nightly builds +--extra-index-url https://download.pytorch.org/whl/nightly +torch>=0.0.0.dev0; platform_machine == "s390x" -- cgit v1.2.3