8000
We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 800e8fd commit 7cd9602Copy full SHA for 7cd9602
src/llama.cpp
@@ -5903,6 +5903,7 @@ static bool llm_load_tensors(
5903
model.buft_layer[i] = llama_default_buffer_type_cpu(true);
5904
}
5905
5906
+#ifndef GGML_USE_KOMPUTE
5907
if (split_mode == LLAMA_SPLIT_MODE_LAYER) {
5908
// calculate the split points
5909
int device_count = llama_get_device_count(model);
@@ -5940,7 +5941,9 @@ static bool llm_load_tensors(
5940
5941
} else {
5942
model.buft_output = llama_default_buffer_type_cpu(true);
5943
- } else {
5944
+ } else
5945
+#endif
5946
+ {
5947
ggml_backend_buffer_type_t split_buft;
5948
if (split_mode == LLAMA_SPLIT_MODE_ROW) {
5949
split_buft = llama_default_buffer_type_split(model, main_gpu, tensor_split);
0 commit comments