mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
Fix completion request
This commit is contained in:
@@ -76,7 +76,7 @@ class CreateCompletionRequest(BaseModel):
|
|||||||
temperature: float = 0.8
|
temperature: float = 0.8
|
||||||
top_p: float = 0.95
|
top_p: float = 0.95
|
||||||
echo: bool = False
|
echo: bool = False
|
||||||
stop: List[str] = []
|
stop: Optional[List[str]] = []
|
||||||
stream: bool = False
|
stream: bool = False
|
||||||
|
|
||||||
# ignored or currently unsupported
|
# ignored or currently unsupported
|
||||||
@@ -173,7 +173,7 @@ class CreateChatCompletionRequest(BaseModel):
|
|||||||
temperature: float = 0.8
|
temperature: float = 0.8
|
||||||
top_p: float = 0.95
|
top_p: float = 0.95
|
||||||
stream: bool = False
|
stream: bool = False
|
||||||
stop: List[str] = []
|
stop: Optional[List[str]] = []
|
||||||
max_tokens: int = 128
|
max_tokens: int = 128
|
||||||
|
|
||||||
# ignored or currently unsupported
|
# ignored or currently unsupported
|
||||||
|
|||||||
Reference in New Issue
Block a user