From b333b06772c89d96aacb5490d6a219fba7c09cc6 Mon Sep 17 00:00:00 2001 From: Mitja Felicijan Date: Thu, 12 Feb 2026 20:57:17 +0100 Subject: Engage! --- llama.cpp/ggml/src/ggml-cpu/amx/mmq.h | 10 ++++++++++ 1 file changed, 10 insertions(+) create mode 100644 llama.cpp/ggml/src/ggml-cpu/amx/mmq.h (limited to 'llama.cpp/ggml/src/ggml-cpu/amx/mmq.h') diff --git a/llama.cpp/ggml/src/ggml-cpu/amx/mmq.h b/llama.cpp/ggml/src/ggml-cpu/amx/mmq.h new file mode 100644 index 0000000..baf7684 --- /dev/null +++ b/llama.cpp/ggml/src/ggml-cpu/amx/mmq.h @@ -0,0 +1,10 @@ +#pragma once +#include "common.h" + +size_t ggml_backend_amx_desired_wsize(const struct ggml_tensor * dst); + +size_t ggml_backend_amx_get_alloc_size(const struct ggml_tensor * tensor); + +void ggml_backend_amx_convert_weight(struct ggml_tensor * tensor, const void * data, size_t offset, size_t size); + +void ggml_backend_amx_mul_mat(const struct ggml_compute_params * params, struct ggml_tensor * dst); -- cgit v1.2.3