From b333b06772c89d96aacb5490d6a219fba7c09cc6 Mon Sep 17 00:00:00 2001 From: Mitja Felicijan Date: Thu, 12 Feb 2026 20:57:17 +0100 Subject: Engage! --- llama.cpp/tests/test-autorelease.cpp | 24 ++++++++++++++++++++++++ 1 file changed, 24 insertions(+) create mode 100644 llama.cpp/tests/test-autorelease.cpp (limited to 'llama.cpp/tests/test-autorelease.cpp') diff --git a/llama.cpp/tests/test-autorelease.cpp b/llama.cpp/tests/test-autorelease.cpp new file mode 100644 index 0000000..ca87c56 --- /dev/null +++ b/llama.cpp/tests/test-autorelease.cpp @@ -0,0 +1,24 @@ +// ref: https://github.com/ggml-org/llama.cpp/issues/4952#issuecomment-1892864763 + +#include +#include +#include + +#include "llama.h" +#include "get-model.h" + +// This creates a new context inside a pthread and then tries to exit cleanly. +int main(int argc, char ** argv) { + auto * model_path = get_model_or_exit(argc, argv); + + std::thread([&model_path]() { + llama_backend_init(); + auto * model = llama_model_load_from_file(model_path, llama_model_default_params()); + auto * ctx = llama_init_from_model(model, llama_context_default_params()); + llama_free(ctx); + llama_model_free(model); + llama_backend_free(); + }).join(); + + return 0; +} -- cgit v1.2.3