From c2585b68894102ace0cb0c54dc812e27c36482b9 Mon Sep 17 00:00:00 2001 From: Maximilian-Winter Date: Thu, 25 May 2023 10:54:08 +0200 Subject: [PATCH] Fixed list elements typing --- llama_cpp/llama.py | 14 +++++++------- 1 file changed, 7 insertions(+), 7 deletions(-) diff --git a/llama_cpp/llama.py b/llama_cpp/llama.py index 8176136..144671b 100644 --- a/llama_cpp/llama.py +++ b/llama_cpp/llama.py @@ -316,7 +316,7 @@ class Llama: mirostat_tau: llama_cpp.c_float, mirostat_eta: llama_cpp.c_float, penalize_nl: bool = True, - logits_processors: List[Callable[[List[llama_cpp.c_int], List[llama_cpp.c_float]], List[float]]] = None + logits_processors: List[Callable[[List[int], List[float]], List[float]]] = None ): assert self.ctx is not None assert len(self.eval_logits) > 0 @@ -444,7 +444,7 @@ class Llama: mirostat_eta: float = 0.1, mirostat_tau: float = 5.0, penalize_nl: bool = True, - logits_processors: List[Callable[[List[llama_cpp.c_int], List[llama_cpp.c_float]], List[float]]] = None + logits_processors: List[Callable[[List[int], List[float]], List[float]]] = None ): """Sample a token from the model. @@ -496,7 +496,7 @@ class Llama: mirostat_mode: int = 0, mirostat_tau: float = 5.0, mirostat_eta: float = 0.1, - logits_processors: List[Callable[[List[llama_cpp.c_int], List[llama_cpp.c_float]], List[float]]] = None + logits_processors: List[Callable[[List[int], List[float]], List[float]]] = None ) -> Generator[int, Optional[Sequence[int]], None]: """Create a generator of tokens from a prompt. @@ -651,8 +651,8 @@ class Llama: mirostat_tau: float = 5.0, mirostat_eta: float = 0.1, model: Optional[str] = None, - logits_processors: List[Callable[[List[llama_cpp.c_int], List[llama_cpp.c_float]], List[float]]] = None, - stopping_criterias: List[Callable[[List[int], List[llama_cpp.c_float]], bool]] = None, + logits_processors: List[Callable[[List[int], List[float]], List[float]]] = None, + stopping_criterias: List[Callable[[List[int], List[float]], bool]] = None, ) -> Union[Iterator[Completion], Iterator[CompletionChunk]]: assert self.ctx is not None @@ -1035,8 +1035,8 @@ class Llama: mirostat_tau: float = 5.0, mirostat_eta: float = 0.1, model: Optional[str] = None, - logits_processors: List[Callable[[List[llama_cpp.c_int], List[llama_cpp.c_float]], List[float]]] = None, - stopping_criterias: List[Callable[[List[int], List[llama_cpp.c_float]], bool]] = None + logits_processors: List[Callable[[List[int], List[float]], List[float]]] = None, + stopping_criterias: List[Callable[[List[int], List[float]], bool]] = None ) -> Union[Completion, Iterator[CompletionChunk]]: """Generate text from a prompt.