mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
Slim-Bullseye based docker image
ends up at ~669MB
This commit is contained in:
@@ -1,4 +1,4 @@
|
||||
FROM python:3-bullseye
|
||||
FROM python:3-slim-bullseye
|
||||
|
||||
# We need to set the host to 0.0.0.0 to allow outside access
|
||||
ENV HOST 0.0.0.0
|
||||
@@ -6,10 +6,10 @@ ENV HOST 0.0.0.0
|
||||
COPY . .
|
||||
|
||||
# Install the package
|
||||
RUN apt update && apt install -y libopenblas-dev
|
||||
RUN apt update && apt install -y libopenblas-dev ninja-build build-essential
|
||||
RUN python -m pip install --upgrade pip pytest cmake scikit-build setuptools fastapi uvicorn sse-starlette
|
||||
|
||||
RUN LLAMA_OPENBLAS=1 python3 setup.py develop
|
||||
|
||||
# Run the server
|
||||
CMD python3 -m llama_cpp.server
|
||||
CMD python3 -m llama_cpp.server
|
||||
|
||||
Reference in New Issue
Block a user