Spaces:
Runtime error
Runtime error
fix
Browse files- Dockerfile +11 -4
- README.md +1 -1
- start.sh +3 -3
Dockerfile
CHANGED
@@ -1,23 +1,30 @@
|
|
1 |
FROM intel/oneapi-runtime:latest
|
2 |
|
3 |
-
WORKDIR /app
|
4 |
-
|
5 |
RUN apt-get update && apt-get install -y --no-install-recommends \
|
6 |
aria2 \
|
7 |
unzip \
|
8 |
&& rm -rf /var/lib/apt/lists/*
|
9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
RUN aria2c -c -x16 https://github.com/MZWNET/actions/releases/download/llama_cpp-b2690/llama-b2690-bin-linux-avx2-intel-mkl-x64.zip \
|
11 |
&& unzip llama-b2690-bin-linux-avx2-intel-mkl-x64.zip -d /app/llama.cpp \
|
12 |
&& mv /app/llama.cpp/server /app/ \
|
13 |
&& chmod +x /app/server \
|
14 |
&& rm -rf llama-b2690-bin-linux-avx2-intel-mkl-x64.zip llama.cpp
|
15 |
|
16 |
-
COPY start.sh start.sh
|
17 |
|
18 |
RUN chmod +x start.sh
|
19 |
|
20 |
-
ENV PATH
|
21 |
|
22 |
EXPOSE 8080
|
23 |
|
|
|
1 |
FROM intel/oneapi-runtime:latest
|
2 |
|
|
|
|
|
3 |
RUN apt-get update && apt-get install -y --no-install-recommends \
|
4 |
aria2 \
|
5 |
unzip \
|
6 |
&& rm -rf /var/lib/apt/lists/*
|
7 |
|
8 |
+
RUN useradd -m -u 1000 user
|
9 |
+
|
10 |
+
USER user
|
11 |
+
|
12 |
+
ENV HOME=/home/user \
|
13 |
+
PATH=/home/user/.local/bin:$PATH
|
14 |
+
|
15 |
+
WORKDIR $HOME/app
|
16 |
+
|
17 |
RUN aria2c -c -x16 https://github.com/MZWNET/actions/releases/download/llama_cpp-b2690/llama-b2690-bin-linux-avx2-intel-mkl-x64.zip \
|
18 |
&& unzip llama-b2690-bin-linux-avx2-intel-mkl-x64.zip -d /app/llama.cpp \
|
19 |
&& mv /app/llama.cpp/server /app/ \
|
20 |
&& chmod +x /app/server \
|
21 |
&& rm -rf llama-b2690-bin-linux-avx2-intel-mkl-x64.zip llama.cpp
|
22 |
|
23 |
+
COPY --chown=user start.sh start.sh
|
24 |
|
25 |
RUN chmod +x start.sh
|
26 |
|
27 |
+
ENV PATH=$HOME/app:$PATH
|
28 |
|
29 |
EXPOSE 8080
|
30 |
|
README.md
CHANGED
@@ -15,7 +15,7 @@ custom_headers:
|
|
15 |
models:
|
16 |
- mzwing/AquilaChat2-7B-16K-GGUF
|
17 |
preload_from_hub:
|
18 |
-
- mzwing/AquilaChat2-7B-16K-GGUF AquilaChat2-7B-16K.Q8_0.gguf
|
19 |
---
|
20 |
|
21 |
AquilaChat2-7B-16K-GGUF, loaded by llama.cpp server.
|
|
|
15 |
models:
|
16 |
- mzwing/AquilaChat2-7B-16K-GGUF
|
17 |
preload_from_hub:
|
18 |
+
- mzwing/AquilaChat2-7B-16K-GGUF AquilaChat2-7B-16K.Q8_0.gguf 3bff96b3dbb6a2386f71199502786739ef71b246
|
19 |
---
|
20 |
|
21 |
AquilaChat2-7B-16K-GGUF, loaded by llama.cpp server.
|
start.sh
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
-
ls -alh
|
2 |
-
ls -alh
|
3 |
-
server -m
|
|
|
1 |
+
ls -alh ~/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/
|
2 |
+
ls -alh ~/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/snapshots/3bff96b3dbb6a2386f71199502786739ef71b246/
|
3 |
+
server -m ~/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/snapshots/3bff96b3dbb6a2386f71199502786739ef71b246/AquilaChat2-7B-16K.Q8_0.gguf -t $(nproc) -c 16384 -a AquilaChat2-7B-16K --port 8080 --api-key sk-1145141919810
|