mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
Install from local directory
This commit is contained in:
@@ -3,8 +3,13 @@ FROM python:3-bullseye
|
|||||||
# We need to set the host to 0.0.0.0 to allow outside access
|
# We need to set the host to 0.0.0.0 to allow outside access
|
||||||
ENV HOST 0.0.0.0
|
ENV HOST 0.0.0.0
|
||||||
|
|
||||||
|
COPY . .
|
||||||
|
|
||||||
# Install the package
|
# Install the package
|
||||||
RUN apt update && apt install -y libopenblas-dev && LLAMA_OPENBLAS=1 pip install llama-cpp-python[server]
|
RUN apt update && apt install -y libopenblas-dev
|
||||||
|
RUN python -m pip install --upgrade pip pytest cmake scikit-build setuptools fastapi uvicorn sse-starlette
|
||||||
|
|
||||||
|
RUN LLAMA_OPENBLAS=1 python3 setup.py develop
|
||||||
|
|
||||||
# Run the server
|
# Run the server
|
||||||
CMD python3 -m llama_cpp.server
|
CMD python3 -m llama_cpp.server
|
||||||
Reference in New Issue
Block a user