mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
Bugfix: only eval new tokens
This commit is contained in:
@@ -280,6 +280,7 @@ class Llama:
|
||||
if self.verbose:
|
||||
print("generate cache hit", file=sys.stderr)
|
||||
reset = False
|
||||
tokens = tokens[len(self.tokens) :]
|
||||
###
|
||||
if reset:
|
||||
self.reset()
|
||||
|
||||
Reference in New Issue
Block a user