mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
Add field to disable reseting between generations
This commit is contained in:
@@ -218,6 +218,7 @@ class Llama:
|
||||
top_p: float,
|
||||
temp: float,
|
||||
repeat_penalty: float,
|
||||
reset: bool = True,
|
||||
) -> Generator[
|
||||
llama_cpp.llama_token, Optional[Sequence[llama_cpp.llama_token]], None
|
||||
]:
|
||||
@@ -235,12 +236,14 @@ class Llama:
|
||||
top_p: The top-p sampling parameter.
|
||||
temp: The temperature parameter.
|
||||
repeat_penalty: The repeat penalty parameter.
|
||||
reset: Whether to reset the model state.
|
||||
|
||||
Yields:
|
||||
The generated tokens.
|
||||
"""
|
||||
assert self.ctx is not None
|
||||
self.reset()
|
||||
if reset:
|
||||
self.reset()
|
||||
while True:
|
||||
self.eval(tokens)
|
||||
token = self.sample(
|
||||
|
||||
Reference in New Issue
Block a user