diff options
| author | Mitja Felicijan <mitja.felicijan@gmail.com> | 2026-02-20 13:54:21 +0100 |
|---|---|---|
| committer | Mitja Felicijan <mitja.felicijan@gmail.com> | 2026-02-20 13:54:21 +0100 |
| commit | 306c3cb6924c6231c102ff7d75aa3f68e3618ca2 (patch) | |
| tree | 1a41c8c4b70b43796cc3fc14f0c9e52b39651e2f /context.c | |
| parent | 201bbf3e917066fb05ff1f10f7166d262b8ed2cf (diff) | |
| download | llmnpc-306c3cb6924c6231c102ff7d75aa3f68e3618ca2.tar.gz | |
Update to multi model for embeddings and prompting
Diffstat (limited to 'context.c')
| -rw-r--r-- | context.c | 7 |
1 files changed, 6 insertions, 1 deletions
@@ -115,7 +115,10 @@ int main(int argc, char **argv) { cfg = &models[0]; } - struct llama_model *model = llama_model_load_from_file(cfg->filepath, llama_model_default_params()); + struct llama_model_params model_params = llama_model_default_params(); + model_params.n_gpu_layers = cfg->n_gpu_layers; + model_params.use_mmap = cfg->use_mmap; + struct llama_model *model = llama_model_load_from_file(cfg->filepath, model_params); if (model == NULL) { log_message(stderr, LOG_ERROR, "Unable to load embedding model"); llama_backend_free(); @@ -123,6 +126,8 @@ int main(int argc, char **argv) { } struct llama_context_params cparams = llama_context_default_params(); + cparams.n_ctx = cfg->n_ctx; + cparams.n_batch = cfg->n_batch; cparams.embeddings = true; struct llama_context *embed_ctx = llama_init_from_model(model, cparams); |
