8000 Update low level api example · coderonion/llama-cpp-python@5e01114 · GitHub
[go: up one dir, main page]

Skip to content

Commit 5e01114

Browse files
committed
Update low level api example
1 parent 5f2e822 commit 5e01114

File tree

1 file changed

+5
-5
lines changed

1 file changed

+5
-5
lines changed

examples/low_level_api_llama_cpp.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@
3535

3636
embd = []
3737
last_n_size = 64
38-
last_n_tokens = [0] * last_n_size
38+
last_n_tokens_data = [0] * last_n_size
3939
n_batch = 24
4040

4141
while remaining_tokens > 0:
@@ -49,21 +49,21 @@
4949
if len(embd_inp) <= input_consumed:
5050
id = llama_cpp.llama_sample_top_p_top_k(
5151
ctx,
52-
(llama_cpp.c_int * len(last_n_tokens))(*last_n_tokens),
53-
len(last_n_tokens),
52+
(llama_cpp.c_int * len(last_n_tokens_data))(*last_n_tokens_data),
53+
len(last_n_tokens_data),
5454
40,
5555
0.8,
5656
0.2,
5757
1.0 / 0.85,
5858
)
59-
last_n_tokens = last_n_tokens[1:] + [id]
59+
last_n_tokens_data = last_n_tokens_data[1:] + [id]
6060
embd.append(id)
6161
input_noecho = False
6262
remaining_tokens -= 1
6363
else:
6464
while len(embd_inp) > input_consumed:
6565
embd.append(embd_inp[input_consumed])
66-
last_n_tokens = last_n_tokens[1:] + [embd_inp[input_consumed]]
66+
last_n_tokens_data = last_n_tokens_data[1:] + [embd_inp[input_consumed]]
6767
input_consumed += 1
6868
if len(embd) >= n_batch:
6969
break

0 commit comments

Comments
 (0)
0