llama-cpp-api / Dockerfile
imperialwool's picture
Update Dockerfile
be3d3fd
raw
history blame
673 Bytes
FROM python:3.11.1-bullseye
USER root
ENV PROTOCOL_BUFFERS_PYTHON_IMPLEMENTATION=python
RUN apt install g++ -y
RUN git clone https://github.com/ggerganov/llama.cpp.git
RUN cd llama.cpp
RUN make
RUN wget https://huggingface.co/OpenBuddy/openbuddy-ggml/resolve/main/openbuddy-openllama-3b-v10-q5_0.bin
COPY . ./
RUN chmod -R 777 ./
WORKDIR ./
RUN python3 -m pip install -U --no-cache-dir pip setuptools wheel
RUN pip install --no-cache-dir --upgrade -r /app/requirements.txt
RUN mkdir -p /.cache/huggingface/hub
RUN chown -R root:root /.cache/huggingface/hub
RUN chmod -R 777 /.cache/huggingface/hub
CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"]