mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
Fixed list elements typing
This commit is contained in:
@@ -316,7 +316,7 @@ class Llama:
|
||||
mirostat_tau: llama_cpp.c_float,
|
||||
mirostat_eta: llama_cpp.c_float,
|
||||
penalize_nl: bool = True,
|
||||
logits_processors: List[Callable[[List[llama_cpp.c_int], List[llama_cpp.c_float]], List[float]]] = None
|
||||
logits_processors: List[Callable[[List[int], List[float]], List[float]]] = None
|
||||
):
|
||||
assert self.ctx is not None
|
||||
assert len(self.eval_logits) > 0
|
||||
@@ -444,7 +444,7 @@ class Llama:
|
||||
mirostat_eta: float = 0.1,
|
||||
mirostat_tau: float = 5.0,
|
||||
penalize_nl: bool = True,
|
||||
logits_processors: List[Callable[[List[llama_cpp.c_int], List[llama_cpp.c_float]], List[float]]] = None
|
||||
logits_processors: List[Callable[[List[int], List[float]], List[float]]] = None
|
||||
|
||||
):
|
||||
"""Sample a token from the model.
|
||||
@@ -496,7 +496,7 @@ class Llama:
|
||||
mirostat_mode: int = 0,
|
||||
mirostat_tau: float = 5.0,
|
||||
mirostat_eta: float = 0.1,
|
||||
logits_processors: List[Callable[[List[llama_cpp.c_int], List[llama_cpp.c_float]], List[float]]] = None
|
||||
logits_processors: List[Callable[[List[int], List[float]], List[float]]] = None
|
||||
) -> Generator[int, Optional[Sequence[int]], None]:
|
||||
"""Create a generator of tokens from a prompt.
|
||||
|
||||
@@ -651,8 +651,8 @@ class Llama:
|
||||
mirostat_tau: float = 5.0,
|
||||
mirostat_eta: float = 0.1,
|
||||
model: Optional[str] = None,
|
||||
logits_processors: List[Callable[[List[llama_cpp.c_int], List[llama_cpp.c_float]], List[float]]] = None,
|
||||
stopping_criterias: List[Callable[[List[int], List[llama_cpp.c_float]], bool]] = None,
|
||||
logits_processors: List[Callable[[List[int], List[float]], List[float]]] = None,
|
||||
stopping_criterias: List[Callable[[List[int], List[float]], bool]] = None,
|
||||
) -> Union[Iterator[Completion], Iterator[CompletionChunk]]:
|
||||
assert self.ctx is not None
|
||||
|
||||
@@ -1035,8 +1035,8 @@ class Llama:
|
||||
mirostat_tau: float = 5.0,
|
||||
mirostat_eta: float = 0.1,
|
||||
model: Optional[str] = None,
|
||||
logits_processors: List[Callable[[List[llama_cpp.c_int], List[llama_cpp.c_float]], List[float]]] = None,
|
||||
stopping_criterias: List[Callable[[List[int], List[llama_cpp.c_float]], bool]] = None
|
||||
logits_processors: List[Callable[[List[int], List[float]], List[float]]] = None,
|
||||
stopping_criterias: List[Callable[[List[int], List[float]], bool]] = None
|
||||
) -> Union[Completion, Iterator[CompletionChunk]]:
|
||||
"""Generate text from a prompt.
|
||||
|
||||
|
||||
Reference in New Issue
Block a user