8000 fix(server): Use split_mode from model settings (#1594) · CISC/llama-cpp-python@66d5cdd · GitHub
[go: up one dir, main page]

Skip to content

Commit 66d5cdd

Browse files
fix(server): Use split_mode from model settings (abetlen#1594)
Co-authored-by: Andrei <abetlen@gmail.com>
1 parent 7613d23 commit 66d5cdd

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

llama_cpp/server/model.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -223,6 +223,7 @@ def load_llama_from_model_settings(settings: ModelSettings) -> llama_cpp.Llama:
223223
**kwargs,
224224
# Model Params
225225
n_gpu_layers=settings.n_gpu_layers,
226+
split_mode=settings.split_mode,
226227
main_gpu=settings.main_gpu,
227228
tensor_split=settings.tensor_split,
228229
vocab_only=settings.vocab_only,

0 commit comments

Comments
 (0)
0