10000 Update llama.cpp · Dpaste20/llama-cpp-python@7c3743f · GitHub
[go: up one dir, main page]

Skip to content

Commit 7c3743f

Browse files
committed
Update llama.cpp
1 parent bc853e3 commit 7c3743f

File tree

3 files changed

+9
-5
lines changed

3 files changed

+9
-5
lines changed

llama_cpp/llama.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -297,21 +297,25 @@ def _sample_top_p_top_k(
297297
ctx=self.ctx,
298298
candidates=llama_cpp.ctypes.pointer(candidates),
299299
k=top_k,
300+
min_keep=llama_cpp.c_size_t(1),
300301
)
301302
llama_cpp.llama_sample_tail_free(
302303
ctx=self.ctx,
303304
candidates=llama_cpp.ctypes.pointer(candidates),
304305
z=llama_cpp.c_float(1.0),
306+
min_keep=llama_cpp.c_size_t(1),
305307
)
306308
llama_cpp.llama_sample_typical(
307309
ctx=self.ctx,
308310
candidates=llama_cpp.ctypes.pointer(candidates),
309311
p=llama_cpp.c_float(1.0),
312+
min_keep=llama_cpp.c_size_t(1),
310313
)
311314
llama_cpp.llama_sample_top_p(
312315
ctx=self.ctx,
313316
candidates=llama_cpp.ctypes.pointer(candidates),
314317
p=top_p,
318+
min_keep=llama_cpp.c_size_t(1),
315319
)
316320
llama_cpp.llama_sample_temperature(
317321
ctx=self.ctx,

llama_cpp/llama_cpp.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -515,7 +515,7 @@ def llama_sample_top_k(
515515
ctx: llama_context_p,
516516
candidates, # type: _Pointer[llama_token_data_array]
517517
k: c_int,
518-
min_keep: c_size_t = c_size_t(1),
518+
min_keep: c_size_t,
519519
):
520520
return _lib.llama_sample_top_k(ctx, candidates, k, min_keep)
521521

@@ -534,7 +534,7 @@ def llama_sample_top_p(
534534
ctx: llama_context_p,
535535
candidates, # type: _Pointer[llama_token_data_array]
536536
p: c_float,
537-
min_keep: c_size_t = c_size_t(1),
537+
min_keep: c_size_t,
538538
):
539539
return _lib.llama_sample_top_p(ctx, candidates, p, min_keep)
540540

@@ -553,7 +553,7 @@ def llama_sample_tail_free(
553553
ctx: llama_context_p,
554554
candidates, # type: _Pointer[llama_token_data_array]
555555
z: c_float,
556-
min_keep: c_size_t = c_size_t(1),
556+
min_keep: c_size_t,
557557
):
558558
return _lib.llama_sample_tail_free(ctx, candidates, z, min_keep)
559559

@@ -572,7 +572,7 @@ def llama_sample_typical(
572572
ctx: llama_context_p,
573573
candidates, # type: _Pointer[llama_token_data_array]
574574
p: c_float,
575-
min_keep: c_size_t = c_size_t(1),
575+
min_keep: c_size_t,
576576
):
577577
return _lib.llama_sample_typical(ctx, candidates, p, min_keep)
578578

vendor/llama.cpp

0 commit comments

Comments
 (0)
0