mzwing commited on
Commit
0f6e189
·
verified ·
1 Parent(s): cb86435
Files changed (3) hide show
  1. Dockerfile +14 -10
  2. README.md +0 -2
  3. start.sh +1 -5
Dockerfile CHANGED
@@ -1,24 +1,28 @@
1
  FROM intel/oneapi-runtime:latest
2
 
3
- RUN useradd -m -u 1000 user
4
-
5
- USER root
6
-
7
  RUN apt-get update && apt-get install -y --no-install-recommends \
8
  aria2 \
9
  unzip \
10
- clinfo \
11
- sudo \
12
  && rm -rf /var/lib/apt/lists/*
13
 
14
- RUN getent passwd | wc -l
 
 
15
 
16
- ENV HOME=/root \
17
- PATH=/root/.local/bin:$PATH
18
 
19
  WORKDIR $HOME/app
20
 
21
- RUN aria2c -c -x16 https://github.com/MZWNET/actions/releases/download/llama_cpp-b2690/llama-b2690-bin-linux-avx2-intel-mkl-x64.zip \
 
 
 
 
 
 
 
 
22
  && unzip llama-b2690-bin-linux-avx2-intel-mkl-x64.zip -d llama.cpp \
23
  && mv llama.cpp/server . \
24
  && chmod +x ./server \
 
1
  FROM intel/oneapi-runtime:latest
2
 
 
 
 
 
3
  RUN apt-get update && apt-get install -y --no-install-recommends \
4
  aria2 \
5
  unzip \
 
 
6
  && rm -rf /var/lib/apt/lists/*
7
 
8
+ RUN useradd -m -u 1000 user
9
+
10
+ USER user
11
 
12
+ ENV HOME=/home/user \
13
+ PATH=/home/user/.local/bin:$PATH
14
 
15
  WORKDIR $HOME/app
16
 
17
+ RUN cat << EOF > download.txt
18
+ https://github.com/MZWNET/actions/releases/download/llama_cpp-b2690/llama-b2690-bin-linux-avx2-intel-mkl-x64.zip
19
+ out=llama-b2690-bin-linux-avx2-intel-mkl-x64.zip
20
+ https://huggingface.co/mzwing/AquilaChat2-7B-16K-GGUF/resolve/main/AquilaChat2-7B-16K.Q8_0.gguf?download=true
21
+ out=AquilaChat2-7B-16K.Q8_0.gguf
22
+ EOF \
23
+ && cat download.txt \
24
+ && df -ha
25
+ && aria2c -c -x16 --input-file=download.txt \
26
  && unzip llama-b2690-bin-linux-avx2-intel-mkl-x64.zip -d llama.cpp \
27
  && mv llama.cpp/server . \
28
  && chmod +x ./server \
README.md CHANGED
@@ -14,8 +14,6 @@ custom_headers:
14
  cross-origin-resource-policy: cross-origin
15
  models:
16
  - mzwing/AquilaChat2-7B-16K-GGUF
17
- preload_from_hub:
18
- - mzwing/AquilaChat2-7B-16K-GGUF AquilaChat2-7B-16K.Q8_0.gguf 3bff96b3dbb6a2386f71199502786739ef71b246
19
  ---
20
 
21
  AquilaChat2-7B-16K-GGUF, loaded by llama.cpp server.
 
14
  cross-origin-resource-policy: cross-origin
15
  models:
16
  - mzwing/AquilaChat2-7B-16K-GGUF
 
 
17
  ---
18
 
19
  AquilaChat2-7B-16K-GGUF, loaded by llama.cpp server.
start.sh CHANGED
@@ -1,5 +1 @@
1
- clinfo -l
2
- find / -type f -name "AquilaChat2-7B-16K.Q8_0.gguf" -print
3
- sudo clinfo -l
4
- sudo find / -type f -name "AquilaChat2-7B-16K.Q8_0.gguf" -print
5
- server -m ~/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/snapshots/3bff96b3dbb6a2386f71199502786739ef71b246/AquilaChat2-7B-16K.Q8_0.gguf -t $(nproc) -c 16384 -a AquilaChat2-7B-16K --port 8080 --api-key sk-1145141919810
 
1
+ server -m ./AquilaChat2-7B-16K.Q8_0.gguf -t $(nproc) -c 16384 -a AquilaChat2-7B-16K --port 8080 --api-key sk-1145141919810