aromal commited on
Commit
97173f2
·
1 Parent(s): c43ab43

Create config.json

Browse files
Files changed (1) hide show
  1. config.json +43 -0
config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model_name: "bn22/Mistral-7B-Instruct-v0.1-sharded"
2
+ adapters_name: "potato101/mistralengft"
3
+ device: "cuda"
4
+
5
+ # BitsAndBytes Config
6
+ bits_and_bytes_config:
7
+ load_in_4bit: true
8
+ bnb_4bit_use_double_quant: true
9
+ bnb_4bit_quant_type: "nf4"
10
+ bnb_4bit_compute_dtype: torch.bfloat16
11
+
12
+ # AutoModelForCausalLM Config
13
+ auto_model_config:
14
+ load_in_4bit: true
15
+ torch_dtype: torch.bfloat16
16
+
17
+ # PeftModel Config
18
+ peft_model_config:
19
+ # Add PeftModel-specific configuration if needed
20
+
21
+ # AutoTokenizer Config
22
+ auto_tokenizer_config:
23
+ bos_token_id: 1
24
+
25
+ # Inference Settings
26
+ max_new_tokens: 200
27
+ do_sample: true
28
+
29
+ # Miscellaneous
30
+ prompt_prefix: "[INST]"
31
+ exit_command: "exit"
32
+
33
+ # Logging
34
+ log_success_message: "Successfully loaded the model {model_name} into memory"
35
+
36
+ # Model Input Processing
37
+ model_input_processing:
38
+ add_special_tokens: false
39
+
40
+ # Output Display
41
+ output_display:
42
+ generated_output_message: "Generated Output:"
43
+ separator_line: "=" * 50