File tree 1 file changed +6
-3
lines changed 1 file changed +6
-3
lines changed Original file line number Diff line number Diff line change @@ -691,6 +691,7 @@ struct llama_model_loader {
691
691
void load_all_data (llama_progress_callback progress_callback, void * progress_callback_user_data, llama_mlock * lmlock) {
692
692
size_t data_size = 0 ;
693
693
size_t prefetch_size = 0 ;
694
+ size_t lock_size = 0 ;
694
695
for (const llama_load_tensor & lt : tensors_map.tensors ) {
695
696
data_size += lt.size ;
696
697
if (lt.ggml_tensor ->backend == GGML_BACKEND_CPU) {
@@ -716,6 +717,11 @@ struct llama_model_loader {
716
717
switch (lt.ggml_tensor ->backend ) {
717
718
case GGML_BACKEND_CPU:
718
719
lt.ggml_tensor ->data = lt.data ;
720
+ if (use_mmap && lmlock)
721
+ {
722
+ lock_size += lt.size ;
723
+ lmlock->grow_to (lock_size);
724
+ }
719
725
break ;
720
726
#ifdef GGML_USE_CUBLAS
721
727
case GGML_BACKEND_CUDA:
@@ -731,9 +737,6 @@ struct llama_model_loader {
731
737
continue ;
732
738
}
733
739
done_size += lt.size ;
734
- if (use_mmap && lmlock) {
735
- lmlock->grow_to (done_size);
736
- }
737
740
}
738
741
}
739
742
You can’t perform that action at this time.
0 commit comments