aboutsummaryrefslogtreecommitdiff
path: root/llama.cpp
diff options
context:
space:
mode:
authorGeorgi Gerganov <ggerganov@gmail.com>2023-07-21 13:10:51 +0300
committerGitHub <noreply@github.com>2023-07-21 13:10:51 +0300
commitae178ab46bfd6ecb2422da5dad441a4e2fef8b7e (patch)
tree064a13d048ecd596bbd57bd081c9615aa91ebbf6 /llama.cpp
parent54e3bc76fed914f8d4a30a7a50c19867cccb1338 (diff)
llama : make tensor_split ptr instead of array (#2272)
Diffstat (limited to 'llama.cpp')
-rw-r--r--llama.cpp4
1 files changed, 2 insertions, 2 deletions
diff --git a/llama.cpp b/llama.cpp
index 796dfda..23e746d 100644
--- a/llama.cpp
+++ b/llama.cpp
@@ -849,7 +849,7 @@ struct llama_context_params llama_context_default_params() {
/*.n_batch =*/ 512,
/*.gpu_layers =*/ 0,
/*.main_gpu =*/ 0,
- /*.tensor_split =*/ {0},
+ /*.tensor_split =*/ nullptr,
/*.rope_freq_base =*/ 10000.0f,
/*.rope_freq_scale =*/ 1.0f,
/*.progress_callback =*/ nullptr,
@@ -1289,7 +1289,7 @@ static bool llama_model_load(
int n_batch,
int n_gpu_layers,
int main_gpu,
- float * tensor_split,
+ const float * tensor_split,
float rope_freq_base,
float rope_freq_scale,
bool low_vram,