Spaces:
Runtime error
Runtime error
debug
Browse files- Dockerfile +14 -10
- README.md +0 -2
- start.sh +1 -5
Dockerfile
CHANGED
@@ -1,24 +1,28 @@
|
|
1 |
FROM intel/oneapi-runtime:latest
|
2 |
|
3 |
-
RUN useradd -m -u 1000 user
|
4 |
-
|
5 |
-
USER root
|
6 |
-
|
7 |
RUN apt-get update && apt-get install -y --no-install-recommends \
|
8 |
aria2 \
|
9 |
unzip \
|
10 |
-
clinfo \
|
11 |
-
sudo \
|
12 |
&& rm -rf /var/lib/apt/lists/*
|
13 |
|
14 |
-
RUN
|
|
|
|
|
15 |
|
16 |
-
ENV HOME=/
|
17 |
-
PATH=/
|
18 |
|
19 |
WORKDIR $HOME/app
|
20 |
|
21 |
-
RUN
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
&& unzip llama-b2690-bin-linux-avx2-intel-mkl-x64.zip -d llama.cpp \
|
23 |
&& mv llama.cpp/server . \
|
24 |
&& chmod +x ./server \
|
|
|
1 |
FROM intel/oneapi-runtime:latest
|
2 |
|
|
|
|
|
|
|
|
|
3 |
RUN apt-get update && apt-get install -y --no-install-recommends \
|
4 |
aria2 \
|
5 |
unzip \
|
|
|
|
|
6 |
&& rm -rf /var/lib/apt/lists/*
|
7 |
|
8 |
+
RUN useradd -m -u 1000 user
|
9 |
+
|
10 |
+
USER user
|
11 |
|
12 |
+
ENV HOME=/home/user \
|
13 |
+
PATH=/home/user/.local/bin:$PATH
|
14 |
|
15 |
WORKDIR $HOME/app
|
16 |
|
17 |
+
RUN cat << EOF > download.txt
|
18 |
+
https://github.com/MZWNET/actions/releases/download/llama_cpp-b2690/llama-b2690-bin-linux-avx2-intel-mkl-x64.zip
|
19 |
+
out=llama-b2690-bin-linux-avx2-intel-mkl-x64.zip
|
20 |
+
https://huggingface.co/mzwing/AquilaChat2-7B-16K-GGUF/resolve/main/AquilaChat2-7B-16K.Q8_0.gguf?download=true
|
21 |
+
out=AquilaChat2-7B-16K.Q8_0.gguf
|
22 |
+
EOF \
|
23 |
+
&& cat download.txt \
|
24 |
+
&& df -ha
|
25 |
+
&& aria2c -c -x16 --input-file=download.txt \
|
26 |
&& unzip llama-b2690-bin-linux-avx2-intel-mkl-x64.zip -d llama.cpp \
|
27 |
&& mv llama.cpp/server . \
|
28 |
&& chmod +x ./server \
|
README.md
CHANGED
@@ -14,8 +14,6 @@ custom_headers:
|
|
14 |
cross-origin-resource-policy: cross-origin
|
15 |
models:
|
16 |
- mzwing/AquilaChat2-7B-16K-GGUF
|
17 |
-
preload_from_hub:
|
18 |
-
- mzwing/AquilaChat2-7B-16K-GGUF AquilaChat2-7B-16K.Q8_0.gguf 3bff96b3dbb6a2386f71199502786739ef71b246
|
19 |
---
|
20 |
|
21 |
AquilaChat2-7B-16K-GGUF, loaded by llama.cpp server.
|
|
|
14 |
cross-origin-resource-policy: cross-origin
|
15 |
models:
|
16 |
- mzwing/AquilaChat2-7B-16K-GGUF
|
|
|
|
|
17 |
---
|
18 |
|
19 |
AquilaChat2-7B-16K-GGUF, loaded by llama.cpp server.
|
start.sh
CHANGED
@@ -1,5 +1 @@
|
|
1 |
-
|
2 |
-
find / -type f -name "AquilaChat2-7B-16K.Q8_0.gguf" -print
|
3 |
-
sudo clinfo -l
|
4 |
-
sudo find / -type f -name "AquilaChat2-7B-16K.Q8_0.gguf" -print
|
5 |
-
server -m ~/.cache/huggingface/hub/models--mzwing--AquilaChat2-7B-16K-GGUF/snapshots/3bff96b3dbb6a2386f71199502786739ef71b246/AquilaChat2-7B-16K.Q8_0.gguf -t $(nproc) -c 16384 -a AquilaChat2-7B-16K --port 8080 --api-key sk-1145141919810
|
|
|
1 |
+
server -m ./AquilaChat2-7B-16K.Q8_0.gguf -t $(nproc) -c 16384 -a AquilaChat2-7B-16K --port 8080 --api-key sk-1145141919810
|
|
|
|
|
|
|
|