@@ -571,7 +571,7 @@ struct llama_model_loader {
571
571
struct ggml_context * ggml_ctx = NULL ;
572
572
std::unique_ptr<llama_mmap> mapping;
573
573
574
- llama_model_loader (const std::string & fname_base, bool use_mmap, bool vocab_only ) {
574
+ llama_model_loader (const std::string & fname_base, bool use_mmap) {
575
575
file_loader = std::unique_ptr<llama_file_loader>(new llama_file_loader (fname_base.c_str (), tensors_map));
576
576
if (!llama_mmap::SUPPORTED) {
577
577
use_mmap = false ;
@@ -916,7 +916,7 @@ static void llama_model_load_internal(
916
916
917
917
model.t_start_us = ggml_time_us ();
918
918
919
- std::unique_ptr<llama_model_loader> ml (new llama_model_loader (fname, use_mmap, vocab_only ));
919
+ std::unique_ptr<llama_model_loader> ml (new llama_model_loader (fname, use_mmap));
920
920
921
921
vocab = std::move (ml->file_loader ->vocab );
922
922
model.hparams = ml->file_loader ->hparams ;
@@ -2299,8 +2299,7 @@ static void llama_model_quantize_internal(const std::string & fname_inp, const s
2299
2299
nthread = std::thread::hardware_concurrency ();
2300
2300
}
2301
2301
2302
- std::unique_ptr<llama_model_loader> model_loader (new llama_model_loader (fname_inp, /* use_mmap*/ false ,
2303
- /* vocab_only*/ false ));
2302
+ std::unique_ptr<llama_model_loader> model_loader (new llama_model_loader (fname_inp, /* use_mmap*/ false ));
2304
2303
llama_file_saver file_saver (fname_out.c_str (), model_loader->file_loader .get (), params->ftype );
2305
2304
2306
2305
#ifdef GGML_USE_K_QUANTS
@@ -2733,7 +2732,7 @@ int llama_apply_lora_from_file_internal(const struct llama_model & model, const
2733
2732
llama_buffer base_buf;
2734
2733
if (path_base_model) {
2735
2734
fprintf (stderr, " %s: loading base model from '%s'\n " , __func__, path_base_model);
2736
- model_loader.reset (new llama_model_loader (path_base_model, /* use_mmap*/ true , /* vocab_only */ false ));
2735
+ model_loader.reset (new llama_model_loader (path_base_model, /* use_mmap*/ true ));
2737
2736
2738
2737
size_t ctx_size;
2739
2738
size_t mmapped_size;
0 commit comments