8000 Bugfix: only eval new tokens · coderonion/llama-cpp-python@d8487de · GitHub
[go: up one dir, main page]

Skip to content

Commit d8487de

Browse files
abetlenpull[bot]
authored andcommitted
Bugfix: only eval new tokens
1 parent f33363d commit d8487de

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

llama_cpp/llama.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -280,6 +280,7 @@ def generate(
280280
if self.verbose:
281281
print("generate cache hit", file=sys.stderr)
282282
reset = False
283+
tokens = tokens[len(self.tokens) :]
283284
###
284285
if reset:
285286
self.reset()

0 commit comments

Comments
 (0)
0