gururise commited on
Commit
fb173e2
·
1 Parent(s): 66d1c99

14b non-quantized float32 model

Browse files
Files changed (1) hide show
  1. config.py +18 -3
config.py CHANGED
@@ -6,7 +6,7 @@ quantized = {
6
  "runtimedtype": torch.bfloat16,
7
  "useGPU": torch.cuda.is_available(),
8
  "chunksize": 32, # larger = more accurate, but more memory (and slower)
9
- "target": 24 # your gpu max size, excess vram offloaded to cpu
10
  }
11
 
12
  # UNCOMMENT TO SELECT OPTIONS
@@ -69,8 +69,23 @@ quantized = {
69
  # [Vram usage: 15.0GB]
70
  # [File size: 14.4GB]
71
 
 
 
 
 
 
 
 
 
 
 
 
72
  config = {
73
- "path": "https://huggingface.co/Hazzzardous/RWKV-8Bit/resolve/main/RWKV-4-Pile-14B-20230204-7324.pqth"
 
 
 
 
74
  }
75
 
76
- title = "RWKV-4 (14b Quantized)"
 
6
  "runtimedtype": torch.bfloat16,
7
  "useGPU": torch.cuda.is_available(),
8
  "chunksize": 32, # larger = more accurate, but more memory (and slower)
9
+ "target": 100 # your gpu max size, excess vram offloaded to cpu
10
  }
11
 
12
  # UNCOMMENT TO SELECT OPTIONS
 
69
  # [Vram usage: 15.0GB]
70
  # [File size: 14.4GB]
71
 
72
+ # config = {
73
+ # "path": "https://huggingface.co/Hazzzardous/RWKV-8Bit/resolve/main/RWKV-4-Pile-14B-20230204-7324.pqth"
74
+ # }
75
+
76
+ # title = "RWKV-4 (14b Quantized)"
77
+
78
+ # RWKV 14B (latest as of feb 9)
79
+ # Approximate
80
+ # [Vram usage: 27.0GB]
81
+ # [File size: 28.4GB]
82
+
83
  config = {
84
+ "path": "https://huggingface.co/BlinkDL/rwkv-4-pile-14b/resolve/main/RWKV-4-Pile-14B-20230204-7324.pth",
85
+ "mode": TORCH,
86
+ "runtimedtype": torch.float32,
87
+ "useGPU": torch.cuda.is_available(),
88
+ "dtype": torch.float32
89
  }
90
 
91
+ title = "RWKV-4 (14b Feb 4 Snapshot)"