mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
Bugfix: only eval new tokens
This commit is contained in:
@@ -280,6 +280,7 @@ class Llama:
|
|||||||
if self.verbose:
|
if self.verbose:
|
||||||
print("generate cache hit", file=sys.stderr)
|
print("generate cache hit", file=sys.stderr)
|
||||||
reset = False
|
reset = False
|
||||||
|
tokens = tokens[len(self.tokens) :]
|
||||||
###
|
###
|
||||||
if reset:
|
if reset:
|
||||||
self.reset()
|
self.reset()
|
||||||
|
|||||||
Reference in New Issue
Block a user