summaryrefslogtreecommitdiff
path: root/context.c
diff options
context:
space:
mode:
authorMitja Felicijan <mitja.felicijan@gmail.com>2026-02-20 13:54:21 +0100
committerMitja Felicijan <mitja.felicijan@gmail.com>2026-02-20 13:54:21 +0100
commit306c3cb6924c6231c102ff7d75aa3f68e3618ca2 (patch)
tree1a41c8c4b70b43796cc3fc14f0c9e52b39651e2f /context.c
parent201bbf3e917066fb05ff1f10f7166d262b8ed2cf (diff)
downloadllmnpc-306c3cb6924c6231c102ff7d75aa3f68e3618ca2.tar.gz
Update to multi model for embeddings and prompting
Diffstat (limited to 'context.c')
-rw-r--r--context.c7
1 files changed, 6 insertions, 1 deletions
diff --git a/context.c b/context.c
index e7aa0cf..66b8cc2 100644
--- a/context.c
+++ b/context.c
@@ -115,7 +115,10 @@ int main(int argc, char **argv) {
cfg = &models[0];
}
- struct llama_model *model = llama_model_load_from_file(cfg->filepath, llama_model_default_params());
+ struct llama_model_params model_params = llama_model_default_params();
+ model_params.n_gpu_layers = cfg->n_gpu_layers;
+ model_params.use_mmap = cfg->use_mmap;
+ struct llama_model *model = llama_model_load_from_file(cfg->filepath, model_params);
if (model == NULL) {
log_message(stderr, LOG_ERROR, "Unable to load embedding model");
llama_backend_free();
@@ -123,6 +126,8 @@ int main(int argc, char **argv) {
}
struct llama_context_params cparams = llama_context_default_params();
+ cparams.n_ctx = cfg->n_ctx;
+ cparams.n_batch = cfg->n_batch;
cparams.embeddings = true;
struct llama_context *embed_ctx = llama_init_from_model(model, cparams);