8000
We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 5a648da commit fad3795Copy full SHA for fad3795
llama.cpp
@@ -4741,6 +4741,7 @@ static bool llm_load_tensors(
4741
model.buft_layer[i] = llama_default_buffer_type_cpu(true);
4742
}
4743
4744
+#ifndef GGML_USE_KOMPUTE
4745
if (split_mode == LLAMA_SPLIT_MODE_LAYER) {
4746
// calculate the split points
4747
int device_count = llama_get_device_count();
@@ -4778,7 +4779,9 @@ static bool llm_load_tensors(
4778
4779
} else {
4780
model.buft_output = llama_default_buffer_type_cpu(true);
4781
- } else {
4782
+ } else
4783
+#endif
4784
+ {
4785
ggml_backend_buffer_type_t split_buft;
4786
if (split_mode == LLAMA_SPLIT_MODE_ROW) {
4787
split_buft = llama_default_buffer_type_split(main_gpu, tensor_split);
0 commit comments