|
|
@@ -1275,7 +1275,7 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
|
|
|
|
|
|
const bool use_mmap_buffer = true;
|
|
|
|
|
|
- LLAMA_LOG_INFO("%s: loading model tensors, this can take a while... (mmap = %s)\n", __func__, use_mmap_buffer ? "true" : "false");
|
|
|
+ LLAMA_LOG_INFO("%s: loading model tensors, this can take a while... (mmap = %s)\n", __func__, ml.use_mmap ? "true" : "false");
|
|
|
|
|
|
// build a list of buffer types for the CPU and GPU devices
|
|
|
pimpl->cpu_buft_list = make_cpu_buft_list(devices);
|