mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
tests: simple test for server module
This commit is contained in:
@@ -24,6 +24,7 @@ class Settings(BaseSettings):
|
||||
last_n_tokens_size: int = 64
|
||||
logits_all: bool = False
|
||||
cache: bool = False # WARNING: This is an experimental feature
|
||||
vocab_only: bool = False
|
||||
|
||||
|
||||
app = FastAPI(
|
||||
@@ -49,6 +50,7 @@ llama = llama_cpp.Llama(
|
||||
n_batch=settings.n_batch,
|
||||
n_ctx=settings.n_ctx,
|
||||
last_n_tokens_size=settings.last_n_tokens_size,
|
||||
vocab_only=settings.vocab_only,
|
||||
)
|
||||
if settings.cache:
|
||||
cache = llama_cpp.LlamaCache()
|
||||
|
||||
Reference in New Issue
Block a user