Skip to content

Commit 07a1b05

Browse files
committed
llama : on Metal, by default offload the full model
ggml-ci
1 parent 3cb1c1f commit 07a1b05

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

llama.cpp

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9069,7 +9069,8 @@ struct llama_model_params llama_model_default_params() {
90699069
};
90709070

90719071
#ifdef GGML_USE_METAL
9072-
result.n_gpu_layers = 1;
9072+
// note: we usually have plenty of VRAM, so by default offload all layers to the GPU
9073+
result.n_gpu_layers = 999;
90739074
#endif
90749075

90759076
return result;

0 commit comments

Comments
 (0)