mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
Set n_batch to the default value of 8. I think this is leftover from when n_ctx was missing and n_batch was 2048.
This commit is contained in:
@@ -27,7 +27,7 @@ from sse_starlette.sse import EventSourceResponse
|
|||||||
class Settings(BaseSettings):
|
class Settings(BaseSettings):
|
||||||
model: str
|
model: str
|
||||||
n_ctx: int = 2048
|
n_ctx: int = 2048
|
||||||
n_batch: int = 2048
|
n_batch: int = 8
|
||||||
n_threads: int = os.cpu_count() or 1
|
n_threads: int = os.cpu_count() or 1
|
||||||
f16_kv: bool = True
|
f16_kv: bool = True
|
||||||
use_mlock: bool = True
|
use_mlock: bool = True
|
||||||
|
|||||||
Reference in New Issue
Block a user