diff options
-rw-r--r-- | llama.cpp | 2 |
1 files changed, 2 insertions, 0 deletions
@@ -1053,6 +1053,8 @@ static void llama_model_load_internal( fprintf(stderr, "%s: [cublas] total VRAM used: %zu MB\n", __func__, vram_total / 1024 / 1024); } +#else + (void) n_gpu_layers; #endif // loading time will be recalculate after the first eval, so |