abhijeet06793 commited on
Commit
4e268d1
·
1 Parent(s): 0846fcd

Upload model

Browse files
Files changed (1) hide show
  1. README.md +21 -0
README.md CHANGED
@@ -6,6 +6,7 @@ tags:
6
  model-index:
7
  - name: mistral-finetuned-samsum_5epochs
8
  results: []
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -29,6 +30,25 @@ More information needed
29
 
30
  ## Training procedure
31
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
@@ -47,6 +67,7 @@ The following hyperparameters were used during training:
47
 
48
  ### Framework versions
49
 
 
50
  - Transformers 4.35.0.dev0
51
  - Pytorch 2.0.1+cu118
52
  - Datasets 2.14.5
 
6
  model-index:
7
  - name: mistral-finetuned-samsum_5epochs
8
  results: []
9
+ library_name: peft
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
30
 
31
  ## Training procedure
32
 
33
+
34
+ The following `bitsandbytes` quantization config was used during training:
35
+ - quant_method: gptq
36
+ - bits: 4
37
+ - tokenizer: None
38
+ - dataset: None
39
+ - group_size: 128
40
+ - damp_percent: 0.1
41
+ - desc_act: True
42
+ - sym: True
43
+ - true_sequential: True
44
+ - use_cuda_fp16: False
45
+ - model_seqlen: None
46
+ - block_name_to_quantize: None
47
+ - module_name_preceding_first_block: None
48
+ - batch_size: 1
49
+ - pad_token_id: None
50
+ - disable_exllama: True
51
+ - max_input_length: None
52
  ### Training hyperparameters
53
 
54
  The following hyperparameters were used during training:
 
67
 
68
  ### Framework versions
69
 
70
+ - PEFT 0.5.0
71
  - Transformers 4.35.0.dev0
72
  - Pytorch 2.0.1+cu118
73
  - Datasets 2.14.5