mzwing commited on
Commit
9486900
Β·
verified Β·
1 Parent(s): 220aaa6

feat: init

Browse files
Files changed (4) hide show
  1. Dockerfile +37 -0
  2. README.md +14 -5
  3. loadmodel.json +4 -0
  4. nginx.conf +41 -0
Dockerfile ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ FROM debian:bookworm-slim
2
+
3
+ WORKDIR /app
4
+
5
+ RUN apt-get update && apt-get install -y --no-install-recommends \
6
+ aria2 \
7
+ nginx \
8
+ curl \
9
+ unzip \
10
+ && rm -rf /var/lib/apt/lists/*
11
+
12
+ RUN aria2c -c -x16 https://github.com/MZWNET/actions/releases/download/nitro-v0.3.18/nitro-0.3.18-linux-amd64-intel-mkl-avx2.zip \
13
+ && unzip nitro-0.3.18-linux-amd64-intel-mkl-avx2.zip \
14
+ && mv nitro-0.3.18-linux-amd64-intel-mkl-avx2/nitro /app/nitro \
15
+ && chmod +x /app/nitro \
16
+ && rm -rf nitro-0.3.18-linux-amd64-intel-mkl-avx2.zip nitro-0.3.18-linux-amd64-intel-mkl-avx2
17
+
18
+ COPY nginx.conf /etc/nginx/nginx.conf --chown=user
19
+
20
+ RUN echo '#!/bin/sh\n\
21
+ /app/nitro &\n\
22
+ nginx\n\
23
+ sleep 5\n\
24
+ curl -X POST http://127.0.0.1:3928/inferences/llamacpp/loadmodel \\\n\
25
+ -H "Content-Type: application/json" \\\n\
26
+ -d @/app/model/loadmodel.json\n\
27
+ wait' > /start.sh && chmod +x /start.sh
28
+
29
+ COPY loadmodel.json /app/model/loadmodel.json --chown=user
30
+
31
+ ENV PATH="/app:${PATH}"
32
+
33
+ ENV HF_HOME="/data/.huggingface"
34
+
35
+ EXPOSE 80
36
+
37
+ CMD ["/start.sh"]
README.md CHANGED
@@ -1,11 +1,20 @@
1
  ---
2
- title: Nitro AquilaChat2 7B 16K GGUF
3
- emoji: πŸ“Š
4
- colorFrom: yellow
5
- colorTo: green
6
  sdk: docker
7
  pinned: false
8
  license: mit
 
 
 
 
 
 
 
 
 
9
  ---
10
 
11
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
1
  ---
2
+ title: Nitro AquilaChat2-7B-16K-GGUF
3
+ emoji: πŸ‘€
4
+ colorFrom: blue
5
+ colorTo: blue
6
  sdk: docker
7
  pinned: false
8
  license: mit
9
+ short_description: AquilaChat2-7B-16K-GGUF, loaded by nitro.
10
+ app_port: 80
11
+ models: mzwing/AquilaChat2-7B-16K-GGUF
12
+ custom_headers:
13
+ cross-origin-embedder-policy: cross-origin
14
+ cross-origin-opener-policy: cross-origin
15
+ cross-origin-resource-policy: cross-origin
16
+ preload_from_hub:
17
+ - mzwing/AquilaChat2-7B-16K-GGUF AquilaChat2-7B-16K.Q8_0.gguf
18
  ---
19
 
20
+ AquilaChat2-7B-16K-GGUF, loaded by nitro.
loadmodel.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "llama_model_path": "~/.cache/huggingface/hub/mzwing/AquilaChat2-7B-16K-GGUF/AquilaChat2-7B-16K.Q8_0.gguf",
3
+ "ctx_len": 16384
4
+ }
nginx.conf ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ worker_processes auto;
2
+
3
+ events {
4
+ worker_connections 1024;
5
+ }
6
+
7
+ http {
8
+ include mime.types;
9
+ default_type application/octet-stream;
10
+
11
+ sendfile on;
12
+ keepalive_timeout 65;
13
+
14
+ proxy_connect_timeout 600;
15
+ proxy_send_timeout 600;
16
+ proxy_read_timeout 600;
17
+ send_timeout 600;
18
+
19
+ map $http_upgrade $connection_upgrade {
20
+ default upgrade;
21
+ '' close;
22
+ }
23
+
24
+ server {
25
+ listen 80;
26
+
27
+ location / {
28
+ proxy_pass http://127.0.0.1:3928;
29
+ proxy_set_header Host $host;
30
+ proxy_set_header X-Real-IP $remote_addr;
31
+ proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
32
+ proxy_set_header Upgrade $http_upgrade;
33
+ proxy_set_header Connection $connection_upgrade;
34
+ }
35
+
36
+ location ~ ^/(inferences/llamacpp/loadmodel/|inferences/llamacpp/unloadmodel/) {
37
+ deny all;
38
+ return 403;
39
+ }
40
+ }
41
+ }