mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
12 lines
337 B
Bash
Executable File
12 lines
337 B
Bash
Executable File
#!/bin/sh
|
|
|
|
# For mmap support
|
|
ulimit -l unlimited
|
|
|
|
if [ "$IMAGE" = "python:3-slim-bullseye" ]; then
|
|
python3 -B -m llama_cpp.server --model /app/model.bin
|
|
else
|
|
# You may have to reduce --n_gpu_layers=1000 to 20 or less if you don't have enough VRAM
|
|
python3 -B -m llama_cpp.server --model /app/model.bin --n_gpu_layers=1000
|
|
fi
|