From b333b06772c89d96aacb5490d6a219fba7c09cc6 Mon Sep 17 00:00:00 2001 From: Mitja Felicijan Date: Thu, 12 Feb 2026 20:57:17 +0100 Subject: Engage! --- llama.cpp/ggml/src/ggml-cpu/binary-ops.h | 16 ++++++++++++++++ 1 file changed, 16 insertions(+) create mode 100644 llama.cpp/ggml/src/ggml-cpu/binary-ops.h (limited to 'llama.cpp/ggml/src/ggml-cpu/binary-ops.h') diff --git a/llama.cpp/ggml/src/ggml-cpu/binary-ops.h b/llama.cpp/ggml/src/ggml-cpu/binary-ops.h new file mode 100644 index 0000000..aca1d89 --- /dev/null +++ b/llama.cpp/ggml/src/ggml-cpu/binary-ops.h @@ -0,0 +1,16 @@ +#pragma once + +#include "common.h" + +#ifdef __cplusplus +extern "C" { +#endif + +void ggml_compute_forward_add_non_quantized(const struct ggml_compute_params * params, struct ggml_tensor * dst); +void ggml_compute_forward_sub(const struct ggml_compute_params * params, struct ggml_tensor * dst); +void ggml_compute_forward_mul(const struct ggml_compute_params * params, struct ggml_tensor * dst); +void ggml_compute_forward_div(const struct ggml_compute_params * params, struct ggml_tensor * dst); + +#ifdef __cplusplus +} +#endif -- cgit v1.2.3