8000
We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4a06b01 commit dbf17f0Copy full SHA for dbf17f0
llama.cpp
@@ -3763,6 +3763,7 @@ static bool llm_load_tensors(
3763
model.buft_layer[i] = llama_default_buffer_type_cpu(true);
3764
}
3765
3766
+#ifndef GGML_USE_KOMPUTE
3767
if (split_mode == LLAMA_SPLIT_MODE_LAYER) {
3768
// calculate the split points
3769
int device_count = llama_get_device_count();
@@ -3800,7 +3801,9 @@ static bool llm_load_tensors(
3800
3801
} else {
3802
model.buft_output = llama_default_buffer_type_cpu(true);
3803
- } else {
3804
+ } else
3805
+#endif
3806
+ {
3807
ggml_backend_buffer_type_t split_buft;
3808
if (split_mode == LLAMA_SPLIT_MODE_ROW) {
3809
split_buft = llama_default_buffer_type_split(main_gpu, tensor_split);
0 commit comments