mzwing commited on
Commit
e820cbb
1 Parent(s): cf4b1f9
Files changed (3) hide show
  1. Dockerfile +11 -4
  2. README.md +1 -1
  3. start.sh +3 -3
Dockerfile CHANGED
@@ -1,23 +1,30 @@
1
  FROM intel/oneapi-runtime:latest
2
 
3
- WORKDIR /app
4
-
5
  RUN apt-get update && apt-get install -y --no-install-recommends \
6
  aria2 \
7
  unzip \
8
  && rm -rf /var/lib/apt/lists/*
9
 
 
 
 
 
 
 
 
 
 
10
  RUN aria2c -c -x16 https://github.com/MZWNET/actions/releases/download/llama_cpp-b2690/llama-b2690-bin-linux-avx2-intel-mkl-x64.zip \
11
  && unzip llama-b2690-bin-linux-avx2-intel-mkl-x64.zip -d /app/llama.cpp \
12
  && mv /app/llama.cpp/server /app/ \
13
  && chmod +x /app/server \
14
  && rm -rf llama-b2690-bin-linux-avx2-intel-mkl-x64.zip llama.cpp
15
 
16
- COPY start.sh start.sh
17
 
18
  RUN chmod +x start.sh
19
 
20
- ENV PATH="/app:${PATH}"
21
 
22
  EXPOSE 8080
23
 
 
1
  FROM intel/oneapi-runtime:latest
2
 
 
 
3
  RUN apt-get update && apt-get install -y --no-install-recommends \
4
  aria2 \
5
  unzip \
6
  && rm -rf /var/lib/apt/lists/*
7
 
8
+ RUN useradd -m -u 1000 user
9
+
10
+ USER user
11
+
12
+ ENV HOME=/home/user \
13
+ PATH=/home/user/.local/bin:$PATH
14
+
15
+ WORKDIR $HOME/app
16
+
17
  RUN aria2c -c -x16 https://github.com/MZWNET/actions/releases/download/llama_cpp-b2690/llama-b2690-bin-linux-avx2-intel-mkl-x64.zip \
18
  && unzip llama-b2690-bin-linux-avx2-intel-mkl-x64.zip -d /app/llama.cpp \
19
  && mv /app/llama.cpp/server /app/ \
20
  && chmod +x /app/server \
21
  && rm -rf llama-b2690-bin-linux-avx2-intel-mkl-x64.zip llama.cpp
22
 
23
+ COPY --chown=user start.sh start.sh
24
 
25
  RUN chmod +x start.sh
26
 
27
+ ENV PATH=$HOME/app:$PATH
28
 
29
  EXPOSE 8080
30
 
README.md CHANGED
@@ -15,7 +15,7 @@ custom_headers:
15
  models:
16
  - mzwing/AquilaChat2-7B-16K-GGUF
17
  preload_from_hub:
18
- - mzwing/AquilaChat2-7B-16K-GGUF AquilaChat2-7B-16K.Q8_0.gguf
19
  ---
20
 
21
  AquilaChat2-7B-16K-GGUF, loaded by llama.cpp server.
 
15
  models:
16
  - mzwing/AquilaChat2-7B-16K-GGUF
17
  preload_from_hub:
18
+ - mzwing/AquilaChat2-7B-16K-GGUF AquilaChat2-7B-16K.Q8_0.gguf 3bff96b3dbb6a2386f71199502786739ef71b246
19
  ---
20
 
21
  AquilaChat2-7B-16K-GGUF, loaded by llama.cpp server.
start.sh CHANGED
@@ -1,3 +1,3 @@
1
- ls -alh /root/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/
2
- ls -alh /root/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/snapshots/
3
- server -m /root/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/AquilaChat2-7B-16K.Q8_0.gguf -t $(nproc) -c 16384 -a AquilaChat2-7B-16K --port 8080 --api-key sk-1145141919810
 
1
+ ls -alh ~/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/
2
+ ls -alh ~/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/snapshots/3bff96b3dbb6a2386f71199502786739ef71b246/
3
+ server -m ~/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/snapshots/3bff96b3dbb6a2386f71199502786739ef71b246/AquilaChat2-7B-16K.Q8_0.gguf -t $(nproc) -c 16384 -a AquilaChat2-7B-16K --port 8080 --api-key sk-1145141919810