You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: llama_cpp/llama_cpp.py
+6Lines changed: 6 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -648,6 +648,9 @@ class llama_model_kv_override(ctypes.Structure):
648
648
# // proportion of the model (layers or rows) to offload to each GPU, size: llama_max_devices()
649
649
# const float * tensor_split;
650
650
651
+
# // comma separated list of RPC servers to use for offloading
652
+
# const char * rpc_servers;
653
+
651
654
# // Called with a progress value between 0.0 and 1.0. Pass NULL to disable.
652
655
# // If the provided progress_callback returns true, model loading continues.
653
656
# // If it returns false, model loading is immediately aborted.
@@ -674,6 +677,7 @@ class llama_model_params(ctypes.Structure):
674
677
split_mode (int): how to split the model across multiple GPUs
675
678
main_gpu (int): the GPU that is used for the entire model. main_gpu interpretation depends on split_mode: LLAMA_SPLIT_NONE: the GPU that is used for the entire model LLAMA_SPLIT_ROW: the GPU that is used for small tensors and intermediate results LLAMA_SPLIT_LAYER: ignored
676
679
tensor_split (ctypes.Array[ctypes.ctypes.c_float]): proportion of the model (layers or rows) to offload to each GPU, size: llama_max_devices()
680
+
rpc_servers (ctypes.c_char_p): comma separated list of RPC servers to use for offloading
677
681
progress_callback (llama_progress_callback): called with a progress value between 0.0 and 1.0. Pass NULL to disable. If the provided progress_callback returns true, model loading continues. If it returns false, model loading is immediately aborted.
678
682
progress_callback_user_data (ctypes.ctypes.c_void_p): context pointer passed to the progress callback
679
683
kv_overrides (ctypes.Array[llama_model_kv_override]): override key-value pairs of the model meta data
@@ -687,6 +691,7 @@ class llama_model_params(ctypes.Structure):
0 commit comments