From b333b06772c89d96aacb5490d6a219fba7c09cc6 Mon Sep 17 00:00:00 2001 From: Mitja Felicijan Date: Thu, 12 Feb 2026 20:57:17 +0100 Subject: Engage! --- .../ggml/src/ggml-cuda/template-instances/mmf-instance-ncols_8.cu | 5 +++++ 1 file changed, 5 insertions(+) create mode 100644 llama.cpp/ggml/src/ggml-cuda/template-instances/mmf-instance-ncols_8.cu (limited to 'llama.cpp/ggml/src/ggml-cuda/template-instances/mmf-instance-ncols_8.cu') diff --git a/llama.cpp/ggml/src/ggml-cuda/template-instances/mmf-instance-ncols_8.cu b/llama.cpp/ggml/src/ggml-cuda/template-instances/mmf-instance-ncols_8.cu new file mode 100644 index 0000000..f1bd09c --- /dev/null +++ b/llama.cpp/ggml/src/ggml-cuda/template-instances/mmf-instance-ncols_8.cu @@ -0,0 +1,5 @@ +// This file has been autogenerated by generate_cu_files.py, do not edit manually. + +#include "../mmf.cuh" + +DECL_MMF_CASE(8); -- cgit v1.2.3