Upgrade to the upstream fixes
Browse files- Dockerfile +12 -7
Dockerfile
CHANGED
@@ -1,13 +1,18 @@
|
|
1 |
-
FROM nvidia/cuda:
|
2 |
ARG MODEL
|
|
|
|
|
|
|
3 |
ARG MODEL_NAME
|
4 |
ARG ADDITIONAL
|
5 |
RUN mkdir /opt/koboldcpp
|
6 |
-
RUN apt update && apt install
|
7 |
-
RUN git clone https://github.com/lostruins/koboldcpp /opt/koboldcpp
|
8 |
WORKDIR /opt/koboldcpp
|
9 |
COPY default.json /opt/koboldcpp/default.json
|
10 |
-
RUN
|
11 |
-
RUN
|
12 |
-
|
13 |
-
|
|
|
|
|
|
|
|
1 |
+
FROM nvidia/cuda:12.1.1-devel-ubuntu22.04
|
2 |
ARG MODEL
|
3 |
+
ARG IMGMODEL
|
4 |
+
ARG WHISPERMODEL
|
5 |
+
ARG MMPROJ
|
6 |
ARG MODEL_NAME
|
7 |
ARG ADDITIONAL
|
8 |
RUN mkdir /opt/koboldcpp
|
9 |
+
RUN apt update && apt install curl -y
|
|
|
10 |
WORKDIR /opt/koboldcpp
|
11 |
COPY default.json /opt/koboldcpp/default.json
|
12 |
+
RUN curl -fLO koboldcpp https://koboldai.org/cpplinuxcu12
|
13 |
+
RUN chmod +x ./koboldcpp
|
14 |
+
RUN curl -fLO model.ggml $MODEL || true
|
15 |
+
RUN curl -fLO imgmodel.ggml $IMGMODEL || true
|
16 |
+
RUN curl -fLO mmproj.ggml $MMPROJ || true
|
17 |
+
RUN curl -fLO whispermodel.ggml $WHISPERMODEL || true
|
18 |
+
CMD ./koboldcpp --model model.ggml --whispermodel whispermodel.ggml --sdmodel imgmodel.ggml --sdthreads 4 --sdquant --sdclamped --mmproj mmproj.ggml $ADDITIONAL --port 7860 --hordemodelname $MODEL_NAME --hordemaxctx 1 --hordegenlen 1 --preloadstory default.json --ignoremissing
|