vivicai commited on
Commit
1000fe4
·
1 Parent(s): 3751f85

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -16
README.md CHANGED
@@ -13,9 +13,9 @@ license: apache-2.0
13
 
14
 
15
 
16
- This is a 4-bit GPTQ version of the [Tigerbot 70b chat v2](https://huggingface.co/TigerResearch/tigerbot-70b-chat).
17
 
18
- It was quantized to 4bit using: https://github.com/PanQiWei/AutoGPTQ
19
 
20
  ## How to download and use this model in github: https://github.com/TigerResearch/TigerBot
21
 
@@ -33,20 +33,13 @@ pip install -r requirements.txt
33
 
34
  Inference with command line interface
35
 
36
- infer with exllama
37
  ```
38
- # 安装exllama_lib
39
- pip install exllama_lib@git+https://github.com/taprosoft/exllama.git
40
-
41
- # 启动推理
42
- CUDA_VISIBLE_DEVICES=0 python other_infer/exllama_infer.py --model_path TigerResearch/tigerbot-70b-chat-4bit
43
- ```
44
-
45
- infer with auto-gptq
46
- ```
47
- # 安装auto-gptq
48
- pip install auto-gptq
49
 
50
- # 启动推理
51
- CUDA_VISIBLE_DEVICES=0 python other_infer/gptq_infer.py --model_path TigerResearch/tigerbot-70b-chat-4bit
52
  ```
 
13
 
14
 
15
 
16
+ This is a 4-bit EXL2 version of the [tigerbot-70b-chat-v2](https://huggingface.co/TigerResearch/tigerbot-70b-chat-v2).
17
 
18
+ It was quantized to 4bit using: https://github.com/turboderp/exllamav2
19
 
20
  ## How to download and use this model in github: https://github.com/TigerResearch/TigerBot
21
 
 
33
 
34
  Inference with command line interface
35
 
36
+ infer with exllamav2
37
  ```
38
+ # install exllamav2
39
+ git clone https://github.com/turboderp/exllamav2
40
+ cd exllamav2
41
+ pip install -r requirements.txt
 
 
 
 
 
 
 
42
 
43
+ # infer command
44
+ CUDA_VISIBLE_DEVICES=0 python other_infer/exllamav2_hf_infer.py --model_path TigerResearch/tigerbot-70b-chat-v2-4bit-exl2
45
  ```