8000 Update llama.cpp · johnwick123f/llama-cpp-python@3553b14 · GitHub
[go: up one dir, main page]

Skip to content

Commit 3553b14

Browse files
committed
Update llama.cpp
1 parent 7467f12 commit 3553b14

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

llama_cpp/llama_cpp.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -445,7 +445,7 @@ class llama_model_params(Structure):
445445
# uint32_t n_batch; // prompt processing maximum batch size
446446
# uint32_t n_threads; // number of threads to use for generation
447447
# uint32_t n_threads_batch; // number of threads to use for batch processing
< 9831 /td>
448-
# int8_t rope_scaling_type; // RoPE scaling type, from `enum llama_rope_scaling_type`
448+
# int32_t rope_scaling_type; // RoPE scaling type, from `enum llama_rope_scaling_type`
449449

450450
# // ref: https://github.com/ggerganov/llama.cpp/pull/2054
451451
# float rope_freq_base; // RoPE base frequency, 0 = from model
@@ -502,7 +502,7 @@ class llama_context_params(Structure):
502502
("n_batch", c_uint32),
503503
("n_threads", c_uint32),
504504
("n_threads_batch", c_uint32),
505-
("rope_scaling_type", c_int8),
505+
("rope_scaling_type", c_int32),
506506
("rope_freq_base", c_float),
507507
("rope_freq_scale", c_float),
508508
("yarn_ext_factor", c_float),

vendor/llama.cpp

0 commit comments

Comments
 (0)
0