Skip to content

Commit

Permalink
Use params when loading models in llava-cli (ggerganov#3976)
Browse files Browse the repository at this point in the history
llava-cli was loading models with default params and ignoring settings
from the cli. This switches to a generic function to load the params
from the cli options.
  • Loading branch information
tejom authored Nov 7, 2023
1 parent 46876d2 commit 54b4df8
Showing 1 changed file with 3 additions and 5 deletions.
8 changes: 3 additions & 5 deletions examples/llava/llava-cli.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -242,18 +242,16 @@ static struct llava_context * llava_init(gpt_params * params) {

llama_backend_init(params->numa);

llama_model_params model_params = llama_model_default_params();
llama_model_params model_params = llama_model_params_from_gpt_params(*params);

llama_model * model = llama_load_model_from_file(params->model.c_str(), model_params);
if (model == NULL) {
fprintf(stderr , "%s: error: unable to load model\n" , __func__);
return NULL;
}

llama_context_params ctx_params = llama_context_default_params();

llama_context_params ctx_params = llama_context_params_from_gpt_params(*params);
ctx_params.n_ctx = params->n_ctx < 2048 ? 2048 : params->n_ctx; // we need a longer context size to process image embeddings
ctx_params.n_threads = params->n_threads;
ctx_params.n_threads_batch = params->n_threads_batch == -1 ? params->n_threads : params->n_threads_batch;

llama_context * ctx_llama = llama_new_context_with_model(model, ctx_params);

Expand Down

0 comments on commit 54b4df8

Please sign in to comment.