From b333b06772c89d96aacb5490d6a219fba7c09cc6 Mon Sep 17 00:00:00 2001 From: Mitja Felicijan Date: Thu, 12 Feb 2026 20:57:17 +0100 Subject: Engage! --- llama.cpp/ggml/src/ggml-cuda/softmax.cuh | 7 +++++++ 1 file changed, 7 insertions(+) create mode 100644 llama.cpp/ggml/src/ggml-cuda/softmax.cuh (limited to 'llama.cpp/ggml/src/ggml-cuda/softmax.cuh') diff --git a/llama.cpp/ggml/src/ggml-cuda/softmax.cuh b/llama.cpp/ggml/src/ggml-cuda/softmax.cuh new file mode 100644 index 0000000..93dfee8 --- /dev/null +++ b/llama.cpp/ggml/src/ggml-cuda/softmax.cuh @@ -0,0 +1,7 @@ +#include "common.cuh" + +#define CUDA_SOFT_MAX_BLOCK_SIZE 1024 + +void ggml_cuda_op_soft_max(ggml_backend_cuda_context & ctx, ggml_tensor * dst); + +void ggml_cuda_op_soft_max_back(ggml_backend_cuda_context & ctx, ggml_tensor * dst); -- cgit v1.2.3