From 6c7cec0c65373d2892dbb23581af27ab407669d9 Mon Sep 17 00:00:00 2001 From: Andrei Betlen Date: Fri, 14 Apr 2023 10:01:15 -0400 Subject: [PATCH] Fix completion request --- llama_cpp/server/__main__.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/llama_cpp/server/__main__.py b/llama_cpp/server/__main__.py index c54d91b..7fc3c57 100644 --- a/llama_cpp/server/__main__.py +++ b/llama_cpp/server/__main__.py @@ -76,7 +76,7 @@ class CreateCompletionRequest(BaseModel): temperature: float = 0.8 top_p: float = 0.95 echo: bool = False - stop: List[str] = [] + stop: Optional[List[str]] = [] stream: bool = False # ignored or currently unsupported @@ -173,7 +173,7 @@ class CreateChatCompletionRequest(BaseModel): temperature: float = 0.8 top_p: float = 0.95 stream: bool = False - stop: List[str] = [] + stop: Optional[List[str]] = [] max_tokens: int = 128 # ignored or currently unsupported