VERSIL91 commited on
Commit
538983e
·
verified ·
1 Parent(s): 05ab626

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,161 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ base_model: peft-internal-testing/tiny-dummy-qwen2
4
+ tags:
5
+ - axolotl
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: 284f2d0e-bfda-4337-926c-cea681a504a8
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
16
+ <details><summary>See axolotl config</summary>
17
+
18
+ axolotl version: `0.4.1`
19
+ ```yaml
20
+ adapter: lora
21
+ base_model: peft-internal-testing/tiny-dummy-qwen2
22
+ bf16: auto
23
+ chat_template: llama3
24
+ dataset_prepared_path: null
25
+ datasets:
26
+ - data_files:
27
+ - cb723d44360708a8_train_data.json
28
+ ds_type: json
29
+ format: custom
30
+ path: /workspace/input_data/cb723d44360708a8_train_data.json
31
+ type:
32
+ field_instruction: title
33
+ field_output: content
34
+ format: '{instruction}'
35
+ no_input_format: '{instruction}'
36
+ system_format: '{system}'
37
+ system_prompt: ''
38
+ debug: null
39
+ deepspeed: null
40
+ early_stopping_patience: null
41
+ eval_max_new_tokens: 128
42
+ eval_table_size: null
43
+ evals_per_epoch: 4
44
+ flash_attention: false
45
+ fp16: true
46
+ fsdp: null
47
+ fsdp_config: null
48
+ gradient_accumulation_steps: 4
49
+ gradient_checkpointing: true
50
+ group_by_length: false
51
+ hub_model_id: null
52
+ hub_repo: null
53
+ hub_strategy: checkpoint
54
+ hub_token: null
55
+ learning_rate: 0.0001
56
+ load_in_4bit: false
57
+ load_in_8bit: false
58
+ local_rank: 0
59
+ logging_steps: 3
60
+ lora_alpha: 128
61
+ lora_dropout: 0.1
62
+ lora_fan_in_fan_out: true
63
+ lora_model_dir: null
64
+ lora_r: 64
65
+ lora_target_linear: true
66
+ lr_scheduler: cosine
67
+ max_steps: 500
68
+ micro_batch_size: 8
69
+ mlflow_experiment_name: /tmp/cb723d44360708a8_train_data.json
70
+ model_type: AutoModelForCausalLM
71
+ num_epochs: 3
72
+ optimizer: adamw_bnb_8bit
73
+ output_dir: miner_id_24
74
+ pad_to_sequence_len: true
75
+ resume_from_checkpoint: null
76
+ s2_attention: false
77
+ sample_packing: false
78
+ saves_per_epoch: 4
79
+ sequence_len: 1024
80
+ strict: false
81
+ tf32: false
82
+ tokenizer_type: AutoTokenizer
83
+ train_on_inputs: false
84
+ trust_remote_code: true
85
+ val_set_size: 0.05
86
+ wandb_entity: sn56-miner
87
+ wandb_mode: disabled
88
+ wandb_name: null
89
+ wandb_project: god
90
+ wandb_run: wb6u
91
+ wandb_runid: null
92
+ warmup_steps: 10
93
+ weight_decay: 0.01
94
+ xformers_attention: null
95
+
96
+ ```
97
+
98
+ </details><br>
99
+
100
+ # 284f2d0e-bfda-4337-926c-cea681a504a8
101
+
102
+ This model is a fine-tuned version of [peft-internal-testing/tiny-dummy-qwen2](https://huggingface.co/peft-internal-testing/tiny-dummy-qwen2) on the None dataset.
103
+ It achieves the following results on the evaluation set:
104
+ - Loss: 11.9207
105
+
106
+ ## Model description
107
+
108
+ More information needed
109
+
110
+ ## Intended uses & limitations
111
+
112
+ More information needed
113
+
114
+ ## Training and evaluation data
115
+
116
+ More information needed
117
+
118
+ ## Training procedure
119
+
120
+ ### Training hyperparameters
121
+
122
+ The following hyperparameters were used during training:
123
+ - learning_rate: 0.0001
124
+ - train_batch_size: 8
125
+ - eval_batch_size: 8
126
+ - seed: 42
127
+ - distributed_type: multi-GPU
128
+ - num_devices: 4
129
+ - gradient_accumulation_steps: 4
130
+ - total_train_batch_size: 128
131
+ - total_eval_batch_size: 32
132
+ - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
133
+ - lr_scheduler_type: cosine
134
+ - lr_scheduler_warmup_steps: 10
135
+ - training_steps: 441
136
+
137
+ ### Training results
138
+
139
+ | Training Loss | Epoch | Step | Validation Loss |
140
+ |:-------------:|:------:|:----:|:---------------:|
141
+ | No log | 0.0068 | 1 | 11.9315 |
142
+ | 11.9261 | 0.2517 | 37 | 11.9249 |
143
+ | 11.9229 | 0.5034 | 74 | 11.9225 |
144
+ | 11.9227 | 0.7551 | 111 | 11.9223 |
145
+ | 11.9227 | 1.0068 | 148 | 11.9219 |
146
+ | 11.9226 | 1.2585 | 185 | 11.9214 |
147
+ | 11.9217 | 1.5102 | 222 | 11.9211 |
148
+ | 11.9217 | 1.7619 | 259 | 11.9209 |
149
+ | 11.9215 | 2.0136 | 296 | 11.9208 |
150
+ | 11.9214 | 2.2653 | 333 | 11.9207 |
151
+ | 11.9211 | 2.5170 | 370 | 11.9207 |
152
+ | 11.9211 | 2.7687 | 407 | 11.9207 |
153
+
154
+
155
+ ### Framework versions
156
+
157
+ - PEFT 0.13.2
158
+ - Transformers 4.46.0
159
+ - Pytorch 2.5.0+cu124
160
+ - Datasets 3.0.1
161
+ - Tokenizers 0.20.1
adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "peft-internal-testing/tiny-dummy-qwen2",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 128,
14
+ "lora_dropout": 0.1,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 64,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "o_proj",
24
+ "k_proj",
25
+ "v_proj",
26
+ "gate_proj",
27
+ "q_proj",
28
+ "down_proj",
29
+ "up_proj"
30
+ ],
31
+ "task_type": "CAUSAL_LM",
32
+ "use_dora": false,
33
+ "use_rslora": false
34
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35f684d884f32dff3e2f3b26576cbd9fbfc93df14a585cae65b64dd6e2249345
3
+ size 100226
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d85d034a32597d1b00c1577c94520a1eedbf1b3b031a431832226ddcef425bc4
3
+ size 93608
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "peft-internal-testing/tiny-dummy-qwen2",
4
+ "architectures": [
5
+ "Qwen2ForCausalLM"
6
+ ],
7
+ "attention_dropout": 0.0,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 8,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 32,
13
+ "max_position_embeddings": 32768,
14
+ "max_window_layers": 21,
15
+ "model_type": "qwen2",
16
+ "num_attention_heads": 4,
17
+ "num_hidden_layers": 2,
18
+ "num_key_value_heads": 2,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_scaling": null,
21
+ "rope_theta": 1000000.0,
22
+ "sliding_window": null,
23
+ "tie_word_embeddings": true,
24
+ "torch_dtype": "float32",
25
+ "transformers_version": "4.46.0",
26
+ "use_cache": false,
27
+ "use_sliding_window": false,
28
+ "vocab_size": 151936
29
+ }
last-checkpoint/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: peft-internal-testing/tiny-dummy-qwen2
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.13.2
last-checkpoint/adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "peft-internal-testing/tiny-dummy-qwen2",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 128,
14
+ "lora_dropout": 0.1,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 64,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "o_proj",
24
+ "k_proj",
25
+ "v_proj",
26
+ "gate_proj",
27
+ "q_proj",
28
+ "down_proj",
29
+ "up_proj"
30
+ ],
31
+ "task_type": "CAUSAL_LM",
32
+ "use_dora": false,
33
+ "use_rslora": false
34
+ }
last-checkpoint/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d85d034a32597d1b00c1577c94520a1eedbf1b3b031a431832226ddcef425bc4
3
+ size 93608
last-checkpoint/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
last-checkpoint/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49e4cc7bd3625ffd6d4a1921c6492defea3dcacb643d18fdecd1652217ba134c
3
+ size 197158
last-checkpoint/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d3acd41253109b861c5c48977cff6b2a1d8e9e5bcad10432ac1e3be0575ee7f
3
+ size 14960
last-checkpoint/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cf7a09346e979e993ba58256f5fda0e7cfb2721276d013ddfa9e6716b6197f3
3
+ size 15024
last-checkpoint/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c523ab57d8c44012691e0fa8da2e68b2546e13077a80a608be1d54527997c79
3
+ size 15024
last-checkpoint/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:027ba72f371decbec68b311970696ea39d299ff1cf16a9eb9f8b566a331f0ad8
3
+ size 15024
last-checkpoint/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08c76042fb6618d1d31186487d78be5a9adc9f6fd4e9f0432d42ae90c7689878
3
+ size 1064
last-checkpoint/special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|endoftext|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
last-checkpoint/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
last-checkpoint/tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|endoftext|>",
37
+ "errors": "replace",
38
+ "model_max_length": 32768,
39
+ "pad_token": "<|endoftext|>",
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "Qwen2Tokenizer",
42
+ "unk_token": null
43
+ }
last-checkpoint/trainer_state.json ADDED
@@ -0,0 +1,1158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "eval_steps": 37,
6
+ "global_step": 441,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.006802721088435374,
13
+ "eval_loss": 11.931451797485352,
14
+ "eval_runtime": 2.6548,
15
+ "eval_samples_per_second": 372.912,
16
+ "eval_steps_per_second": 11.677,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.02040816326530612,
21
+ "grad_norm": 0.01636049710214138,
22
+ "learning_rate": 3e-05,
23
+ "loss": 11.9317,
24
+ "step": 3
25
+ },
26
+ {
27
+ "epoch": 0.04081632653061224,
28
+ "grad_norm": 0.018158216029405594,
29
+ "learning_rate": 6e-05,
30
+ "loss": 11.9316,
31
+ "step": 6
32
+ },
33
+ {
34
+ "epoch": 0.061224489795918366,
35
+ "grad_norm": 0.017022155225276947,
36
+ "learning_rate": 9e-05,
37
+ "loss": 11.9314,
38
+ "step": 9
39
+ },
40
+ {
41
+ "epoch": 0.08163265306122448,
42
+ "grad_norm": 0.020832879468798637,
43
+ "learning_rate": 9.999468702816552e-05,
44
+ "loss": 11.9312,
45
+ "step": 12
46
+ },
47
+ {
48
+ "epoch": 0.10204081632653061,
49
+ "grad_norm": 0.03084990754723549,
50
+ "learning_rate": 9.996679701338662e-05,
51
+ "loss": 11.9309,
52
+ "step": 15
53
+ },
54
+ {
55
+ "epoch": 0.12244897959183673,
56
+ "grad_norm": 0.034195929765701294,
57
+ "learning_rate": 9.991501503086435e-05,
58
+ "loss": 11.9308,
59
+ "step": 18
60
+ },
61
+ {
62
+ "epoch": 0.14285714285714285,
63
+ "grad_norm": 0.0448000393807888,
64
+ "learning_rate": 9.983936584050992e-05,
65
+ "loss": 11.9304,
66
+ "step": 21
67
+ },
68
+ {
69
+ "epoch": 0.16326530612244897,
70
+ "grad_norm": 0.05750146135687828,
71
+ "learning_rate": 9.973988561450411e-05,
72
+ "loss": 11.9299,
73
+ "step": 24
74
+ },
75
+ {
76
+ "epoch": 0.1836734693877551,
77
+ "grad_norm": 0.06743118911981583,
78
+ "learning_rate": 9.961662192000138e-05,
79
+ "loss": 11.9293,
80
+ "step": 27
81
+ },
82
+ {
83
+ "epoch": 0.20408163265306123,
84
+ "grad_norm": 0.07145452499389648,
85
+ "learning_rate": 9.946963369638525e-05,
86
+ "loss": 11.928,
87
+ "step": 30
88
+ },
89
+ {
90
+ "epoch": 0.22448979591836735,
91
+ "grad_norm": 0.07963183522224426,
92
+ "learning_rate": 9.929899122708609e-05,
93
+ "loss": 11.9275,
94
+ "step": 33
95
+ },
96
+ {
97
+ "epoch": 0.24489795918367346,
98
+ "grad_norm": 0.06796199083328247,
99
+ "learning_rate": 9.91047761059745e-05,
100
+ "loss": 11.9261,
101
+ "step": 36
102
+ },
103
+ {
104
+ "epoch": 0.25170068027210885,
105
+ "eval_loss": 11.9249267578125,
106
+ "eval_runtime": 2.8686,
107
+ "eval_samples_per_second": 345.112,
108
+ "eval_steps_per_second": 10.807,
109
+ "step": 37
110
+ },
111
+ {
112
+ "epoch": 0.2653061224489796,
113
+ "grad_norm": 0.04962217062711716,
114
+ "learning_rate": 9.88870811983468e-05,
115
+ "loss": 11.925,
116
+ "step": 39
117
+ },
118
+ {
119
+ "epoch": 0.2857142857142857,
120
+ "grad_norm": 0.042071931064128876,
121
+ "learning_rate": 9.86460105965207e-05,
122
+ "loss": 11.9244,
123
+ "step": 42
124
+ },
125
+ {
126
+ "epoch": 0.30612244897959184,
127
+ "grad_norm": 0.03425384312868118,
128
+ "learning_rate": 9.838167957006295e-05,
129
+ "loss": 11.9248,
130
+ "step": 45
131
+ },
132
+ {
133
+ "epoch": 0.32653061224489793,
134
+ "grad_norm": 0.035133086144924164,
135
+ "learning_rate": 9.809421451067233e-05,
136
+ "loss": 11.9241,
137
+ "step": 48
138
+ },
139
+ {
140
+ "epoch": 0.3469387755102041,
141
+ "grad_norm": 0.02793348953127861,
142
+ "learning_rate": 9.778375287174465e-05,
143
+ "loss": 11.9238,
144
+ "step": 51
145
+ },
146
+ {
147
+ "epoch": 0.3673469387755102,
148
+ "grad_norm": 0.02287479303777218,
149
+ "learning_rate": 9.745044310264839e-05,
150
+ "loss": 11.9234,
151
+ "step": 54
152
+ },
153
+ {
154
+ "epoch": 0.3877551020408163,
155
+ "grad_norm": 0.016594573855400085,
156
+ "learning_rate": 9.70944445777427e-05,
157
+ "loss": 11.9232,
158
+ "step": 57
159
+ },
160
+ {
161
+ "epoch": 0.40816326530612246,
162
+ "grad_norm": 0.013930396176874638,
163
+ "learning_rate": 9.671592752017138e-05,
164
+ "loss": 11.9233,
165
+ "step": 60
166
+ },
167
+ {
168
+ "epoch": 0.42857142857142855,
169
+ "grad_norm": 0.015639444813132286,
170
+ "learning_rate": 9.631507292046955e-05,
171
+ "loss": 11.9231,
172
+ "step": 63
173
+ },
174
+ {
175
+ "epoch": 0.4489795918367347,
176
+ "grad_norm": 0.013282309286296368,
177
+ "learning_rate": 9.589207245002178e-05,
178
+ "loss": 11.9229,
179
+ "step": 66
180
+ },
181
+ {
182
+ "epoch": 0.46938775510204084,
183
+ "grad_norm": 0.011825870722532272,
184
+ "learning_rate": 9.544712836941308e-05,
185
+ "loss": 11.9232,
186
+ "step": 69
187
+ },
188
+ {
189
+ "epoch": 0.4897959183673469,
190
+ "grad_norm": 0.014082890003919601,
191
+ "learning_rate": 9.498045343171664e-05,
192
+ "loss": 11.9229,
193
+ "step": 72
194
+ },
195
+ {
196
+ "epoch": 0.5034013605442177,
197
+ "eval_loss": 11.922492027282715,
198
+ "eval_runtime": 2.7156,
199
+ "eval_samples_per_second": 364.561,
200
+ "eval_steps_per_second": 11.416,
201
+ "step": 74
202
+ },
203
+ {
204
+ "epoch": 0.5102040816326531,
205
+ "grad_norm": 0.012438306584954262,
206
+ "learning_rate": 9.449227078076443e-05,
207
+ "loss": 11.9226,
208
+ "step": 75
209
+ },
210
+ {
211
+ "epoch": 0.5306122448979592,
212
+ "grad_norm": 0.008214130997657776,
213
+ "learning_rate": 9.398281384444956e-05,
214
+ "loss": 11.9228,
215
+ "step": 78
216
+ },
217
+ {
218
+ "epoch": 0.5510204081632653,
219
+ "grad_norm": 0.008153177797794342,
220
+ "learning_rate": 9.345232622311092e-05,
221
+ "loss": 11.9228,
222
+ "step": 81
223
+ },
224
+ {
225
+ "epoch": 0.5714285714285714,
226
+ "grad_norm": 0.00989192258566618,
227
+ "learning_rate": 9.290106157305423e-05,
228
+ "loss": 11.9225,
229
+ "step": 84
230
+ },
231
+ {
232
+ "epoch": 0.5918367346938775,
233
+ "grad_norm": 0.008665865287184715,
234
+ "learning_rate": 9.23292834852644e-05,
235
+ "loss": 11.9233,
236
+ "step": 87
237
+ },
238
+ {
239
+ "epoch": 0.6122448979591837,
240
+ "grad_norm": 0.0082166139036417,
241
+ "learning_rate": 9.173726535936768e-05,
242
+ "loss": 11.9228,
243
+ "step": 90
244
+ },
245
+ {
246
+ "epoch": 0.6326530612244898,
247
+ "grad_norm": 0.0089117381721735,
248
+ "learning_rate": 9.112529027290382e-05,
249
+ "loss": 11.9229,
250
+ "step": 93
251
+ },
252
+ {
253
+ "epoch": 0.6530612244897959,
254
+ "grad_norm": 0.009443915449082851,
255
+ "learning_rate": 9.049365084597059e-05,
256
+ "loss": 11.9229,
257
+ "step": 96
258
+ },
259
+ {
260
+ "epoch": 0.673469387755102,
261
+ "grad_norm": 0.007972587831318378,
262
+ "learning_rate": 8.984264910130535e-05,
263
+ "loss": 11.9226,
264
+ "step": 99
265
+ },
266
+ {
267
+ "epoch": 0.6938775510204082,
268
+ "grad_norm": 0.008118909783661366,
269
+ "learning_rate": 8.917259631987098e-05,
270
+ "loss": 11.923,
271
+ "step": 102
272
+ },
273
+ {
274
+ "epoch": 0.7142857142857143,
275
+ "grad_norm": 0.009298138320446014,
276
+ "learning_rate": 8.848381289201459e-05,
277
+ "loss": 11.9228,
278
+ "step": 105
279
+ },
280
+ {
281
+ "epoch": 0.7346938775510204,
282
+ "grad_norm": 0.008324788883328438,
283
+ "learning_rate": 8.77766281642708e-05,
284
+ "loss": 11.9224,
285
+ "step": 108
286
+ },
287
+ {
288
+ "epoch": 0.7551020408163265,
289
+ "grad_norm": 0.008361688815057278,
290
+ "learning_rate": 8.705138028188228e-05,
291
+ "loss": 11.9227,
292
+ "step": 111
293
+ },
294
+ {
295
+ "epoch": 0.7551020408163265,
296
+ "eval_loss": 11.922290802001953,
297
+ "eval_runtime": 2.7455,
298
+ "eval_samples_per_second": 360.59,
299
+ "eval_steps_per_second": 11.291,
300
+ "step": 111
301
+ },
302
+ {
303
+ "epoch": 0.7755102040816326,
304
+ "grad_norm": 0.012222371064126492,
305
+ "learning_rate": 8.630841602711344e-05,
306
+ "loss": 11.9227,
307
+ "step": 114
308
+ },
309
+ {
310
+ "epoch": 0.7959183673469388,
311
+ "grad_norm": 0.009516862221062183,
312
+ "learning_rate": 8.554809065343397e-05,
313
+ "loss": 11.9227,
314
+ "step": 117
315
+ },
316
+ {
317
+ "epoch": 0.8163265306122449,
318
+ "grad_norm": 0.01142488420009613,
319
+ "learning_rate": 8.477076771565203e-05,
320
+ "loss": 11.9224,
321
+ "step": 120
322
+ },
323
+ {
324
+ "epoch": 0.8367346938775511,
325
+ "grad_norm": 0.01184010412544012,
326
+ "learning_rate": 8.3976818896078e-05,
327
+ "loss": 11.9227,
328
+ "step": 123
329
+ },
330
+ {
331
+ "epoch": 0.8571428571428571,
332
+ "grad_norm": 0.010025064460933208,
333
+ "learning_rate": 8.316662382680202e-05,
334
+ "loss": 11.9226,
335
+ "step": 126
336
+ },
337
+ {
338
+ "epoch": 0.8775510204081632,
339
+ "grad_norm": 0.007526579778641462,
340
+ "learning_rate": 8.234056990817024e-05,
341
+ "loss": 11.9226,
342
+ "step": 129
343
+ },
344
+ {
345
+ "epoch": 0.8979591836734694,
346
+ "grad_norm": 0.010171515867114067,
347
+ "learning_rate": 8.14990521235468e-05,
348
+ "loss": 11.9228,
349
+ "step": 132
350
+ },
351
+ {
352
+ "epoch": 0.9183673469387755,
353
+ "grad_norm": 0.010467471554875374,
354
+ "learning_rate": 8.064247285044972e-05,
355
+ "loss": 11.9226,
356
+ "step": 135
357
+ },
358
+ {
359
+ "epoch": 0.9387755102040817,
360
+ "grad_norm": 0.008835887536406517,
361
+ "learning_rate": 7.977124166815133e-05,
362
+ "loss": 11.9228,
363
+ "step": 138
364
+ },
365
+ {
366
+ "epoch": 0.9591836734693877,
367
+ "grad_norm": 0.010331097058951855,
368
+ "learning_rate": 7.88857751618352e-05,
369
+ "loss": 11.9224,
370
+ "step": 141
371
+ },
372
+ {
373
+ "epoch": 0.9795918367346939,
374
+ "grad_norm": 0.012341076508164406,
375
+ "learning_rate": 7.798649672340295e-05,
376
+ "loss": 11.9229,
377
+ "step": 144
378
+ },
379
+ {
380
+ "epoch": 1.0,
381
+ "grad_norm": 0.014008646830916405,
382
+ "learning_rate": 7.707383634902659e-05,
383
+ "loss": 11.9227,
384
+ "step": 147
385
+ },
386
+ {
387
+ "epoch": 1.0068027210884354,
388
+ "eval_loss": 11.921862602233887,
389
+ "eval_runtime": 2.7724,
390
+ "eval_samples_per_second": 357.086,
391
+ "eval_steps_per_second": 11.181,
392
+ "step": 148
393
+ },
394
+ {
395
+ "epoch": 1.0204081632653061,
396
+ "grad_norm": 0.011518532410264015,
397
+ "learning_rate": 7.614823043354287e-05,
398
+ "loss": 11.9224,
399
+ "step": 150
400
+ },
401
+ {
402
+ "epoch": 1.0408163265306123,
403
+ "grad_norm": 0.010834796354174614,
404
+ "learning_rate": 7.521012156178808e-05,
405
+ "loss": 11.9222,
406
+ "step": 153
407
+ },
408
+ {
409
+ "epoch": 1.0612244897959184,
410
+ "grad_norm": 0.010119671002030373,
411
+ "learning_rate": 7.425995829697305e-05,
412
+ "loss": 11.9226,
413
+ "step": 156
414
+ },
415
+ {
416
+ "epoch": 1.0816326530612246,
417
+ "grad_norm": 0.008356979116797447,
418
+ "learning_rate": 7.32981949661996e-05,
419
+ "loss": 11.9221,
420
+ "step": 159
421
+ },
422
+ {
423
+ "epoch": 1.1020408163265305,
424
+ "grad_norm": 0.011819167993962765,
425
+ "learning_rate": 7.232529144322087e-05,
426
+ "loss": 11.9226,
427
+ "step": 162
428
+ },
429
+ {
430
+ "epoch": 1.1224489795918366,
431
+ "grad_norm": 0.0105531495064497,
432
+ "learning_rate": 7.134171292854956e-05,
433
+ "loss": 11.9227,
434
+ "step": 165
435
+ },
436
+ {
437
+ "epoch": 1.1428571428571428,
438
+ "grad_norm": 0.01237469632178545,
439
+ "learning_rate": 7.034792972701911e-05,
440
+ "loss": 11.9224,
441
+ "step": 168
442
+ },
443
+ {
444
+ "epoch": 1.163265306122449,
445
+ "grad_norm": 0.013601982034742832,
446
+ "learning_rate": 6.934441702290412e-05,
447
+ "loss": 11.9221,
448
+ "step": 171
449
+ },
450
+ {
451
+ "epoch": 1.183673469387755,
452
+ "grad_norm": 0.008788252249360085,
453
+ "learning_rate": 6.833165465270785e-05,
454
+ "loss": 11.922,
455
+ "step": 174
456
+ },
457
+ {
458
+ "epoch": 1.2040816326530612,
459
+ "grad_norm": 0.010858463123440742,
460
+ "learning_rate": 6.7310126875725e-05,
461
+ "loss": 11.9218,
462
+ "step": 177
463
+ },
464
+ {
465
+ "epoch": 1.2244897959183674,
466
+ "grad_norm": 0.00929215457290411,
467
+ "learning_rate": 6.628032214248983e-05,
468
+ "loss": 11.9219,
469
+ "step": 180
470
+ },
471
+ {
472
+ "epoch": 1.2448979591836735,
473
+ "grad_norm": 0.01050562970340252,
474
+ "learning_rate": 6.524273286122018e-05,
475
+ "loss": 11.9226,
476
+ "step": 183
477
+ },
478
+ {
479
+ "epoch": 1.2585034013605443,
480
+ "eval_loss": 11.921433448791504,
481
+ "eval_runtime": 2.6878,
482
+ "eval_samples_per_second": 368.332,
483
+ "eval_steps_per_second": 11.534,
484
+ "step": 185
485
+ },
486
+ {
487
+ "epoch": 1.2653061224489797,
488
+ "grad_norm": 0.00961016584187746,
489
+ "learning_rate": 6.419785516236907e-05,
490
+ "loss": 11.9219,
491
+ "step": 186
492
+ },
493
+ {
494
+ "epoch": 1.2857142857142856,
495
+ "grad_norm": 0.009320979937911034,
496
+ "learning_rate": 6.314618866139657e-05,
497
+ "loss": 11.922,
498
+ "step": 189
499
+ },
500
+ {
501
+ "epoch": 1.306122448979592,
502
+ "grad_norm": 0.009897406212985516,
503
+ "learning_rate": 6.208823621987516e-05,
504
+ "loss": 11.922,
505
+ "step": 192
506
+ },
507
+ {
508
+ "epoch": 1.3265306122448979,
509
+ "grad_norm": 0.00750324223190546,
510
+ "learning_rate": 6.1024503705043e-05,
511
+ "loss": 11.9219,
512
+ "step": 195
513
+ },
514
+ {
515
+ "epoch": 1.346938775510204,
516
+ "grad_norm": 0.01272102352231741,
517
+ "learning_rate": 5.995549974792016e-05,
518
+ "loss": 11.9221,
519
+ "step": 198
520
+ },
521
+ {
522
+ "epoch": 1.3673469387755102,
523
+ "grad_norm": 0.013028099201619625,
524
+ "learning_rate": 5.888173550010301e-05,
525
+ "loss": 11.9213,
526
+ "step": 201
527
+ },
528
+ {
529
+ "epoch": 1.3877551020408163,
530
+ "grad_norm": 0.008812615647912025,
531
+ "learning_rate": 5.780372438935374e-05,
532
+ "loss": 11.9219,
533
+ "step": 204
534
+ },
535
+ {
536
+ "epoch": 1.4081632653061225,
537
+ "grad_norm": 0.00870923325419426,
538
+ "learning_rate": 5.672198187410122e-05,
539
+ "loss": 11.922,
540
+ "step": 207
541
+ },
542
+ {
543
+ "epoch": 1.4285714285714286,
544
+ "grad_norm": 0.007657302543520927,
545
+ "learning_rate": 5.563702519697108e-05,
546
+ "loss": 11.9218,
547
+ "step": 210
548
+ },
549
+ {
550
+ "epoch": 1.4489795918367347,
551
+ "grad_norm": 0.009517359547317028,
552
+ "learning_rate": 5.454937313746255e-05,
553
+ "loss": 11.9218,
554
+ "step": 213
555
+ },
556
+ {
557
+ "epoch": 1.469387755102041,
558
+ "grad_norm": 0.010530628263950348,
559
+ "learning_rate": 5.345954576389047e-05,
560
+ "loss": 11.9215,
561
+ "step": 216
562
+ },
563
+ {
564
+ "epoch": 1.489795918367347,
565
+ "grad_norm": 0.007918890565633774,
566
+ "learning_rate": 5.236806418471113e-05,
567
+ "loss": 11.9217,
568
+ "step": 219
569
+ },
570
+ {
571
+ "epoch": 1.510204081632653,
572
+ "grad_norm": 0.009183686226606369,
573
+ "learning_rate": 5.1275450299350534e-05,
574
+ "loss": 11.9217,
575
+ "step": 222
576
+ },
577
+ {
578
+ "epoch": 1.510204081632653,
579
+ "eval_loss": 11.921103477478027,
580
+ "eval_runtime": 2.7526,
581
+ "eval_samples_per_second": 359.663,
582
+ "eval_steps_per_second": 11.262,
583
+ "step": 222
584
+ },
585
+ {
586
+ "epoch": 1.5306122448979593,
587
+ "grad_norm": 0.013725336641073227,
588
+ "learning_rate": 5.018222654865472e-05,
589
+ "loss": 11.9218,
590
+ "step": 225
591
+ },
592
+ {
593
+ "epoch": 1.5510204081632653,
594
+ "grad_norm": 0.01041232980787754,
595
+ "learning_rate": 4.908891566508103e-05,
596
+ "loss": 11.922,
597
+ "step": 228
598
+ },
599
+ {
600
+ "epoch": 1.5714285714285714,
601
+ "grad_norm": 0.013697491958737373,
602
+ "learning_rate": 4.7996040422749965e-05,
603
+ "loss": 11.9217,
604
+ "step": 231
605
+ },
606
+ {
607
+ "epoch": 1.5918367346938775,
608
+ "grad_norm": 0.013910558074712753,
609
+ "learning_rate": 4.690412338747721e-05,
610
+ "loss": 11.9217,
611
+ "step": 234
612
+ },
613
+ {
614
+ "epoch": 1.6122448979591837,
615
+ "grad_norm": 0.009510298259556293,
616
+ "learning_rate": 4.5813686666905096e-05,
617
+ "loss": 11.9216,
618
+ "step": 237
619
+ },
620
+ {
621
+ "epoch": 1.6326530612244898,
622
+ "grad_norm": 0.010002524591982365,
623
+ "learning_rate": 4.472525166085336e-05,
624
+ "loss": 11.9219,
625
+ "step": 240
626
+ },
627
+ {
628
+ "epoch": 1.6530612244897958,
629
+ "grad_norm": 0.01105476450175047,
630
+ "learning_rate": 4.363933881200811e-05,
631
+ "loss": 11.9217,
632
+ "step": 243
633
+ },
634
+ {
635
+ "epoch": 1.6734693877551021,
636
+ "grad_norm": 0.008844634518027306,
637
+ "learning_rate": 4.25564673570687e-05,
638
+ "loss": 11.9216,
639
+ "step": 246
640
+ },
641
+ {
642
+ "epoch": 1.693877551020408,
643
+ "grad_norm": 0.012191870249807835,
644
+ "learning_rate": 4.147715507847101e-05,
645
+ "loss": 11.9218,
646
+ "step": 249
647
+ },
648
+ {
649
+ "epoch": 1.7142857142857144,
650
+ "grad_norm": 0.01220732368528843,
651
+ "learning_rate": 4.040191805680627e-05,
652
+ "loss": 11.9215,
653
+ "step": 252
654
+ },
655
+ {
656
+ "epoch": 1.7346938775510203,
657
+ "grad_norm": 0.00849160086363554,
658
+ "learning_rate": 3.933127042405362e-05,
659
+ "loss": 11.9218,
660
+ "step": 255
661
+ },
662
+ {
663
+ "epoch": 1.7551020408163265,
664
+ "grad_norm": 0.009470686316490173,
665
+ "learning_rate": 3.826572411774426e-05,
666
+ "loss": 11.9217,
667
+ "step": 258
668
+ },
669
+ {
670
+ "epoch": 1.7619047619047619,
671
+ "eval_loss": 11.920886039733887,
672
+ "eval_runtime": 2.9807,
673
+ "eval_samples_per_second": 332.134,
674
+ "eval_steps_per_second": 10.4,
675
+ "step": 259
676
+ },
677
+ {
678
+ "epoch": 1.7755102040816326,
679
+ "grad_norm": 0.008847080171108246,
680
+ "learning_rate": 3.7205788636175075e-05,
681
+ "loss": 11.9214,
682
+ "step": 261
683
+ },
684
+ {
685
+ "epoch": 1.7959183673469388,
686
+ "grad_norm": 0.014345753937959671,
687
+ "learning_rate": 3.6151970794788524e-05,
688
+ "loss": 11.9214,
689
+ "step": 264
690
+ },
691
+ {
692
+ "epoch": 1.816326530612245,
693
+ "grad_norm": 0.012002954259514809,
694
+ "learning_rate": 3.510477448383529e-05,
695
+ "loss": 11.9217,
696
+ "step": 267
697
+ },
698
+ {
699
+ "epoch": 1.836734693877551,
700
+ "grad_norm": 0.008899074047803879,
701
+ "learning_rate": 3.406470042743575e-05,
702
+ "loss": 11.9212,
703
+ "step": 270
704
+ },
705
+ {
706
+ "epoch": 1.8571428571428572,
707
+ "grad_norm": 0.010487317107617855,
708
+ "learning_rate": 3.3032245944155285e-05,
709
+ "loss": 11.9216,
710
+ "step": 273
711
+ },
712
+ {
713
+ "epoch": 1.8775510204081631,
714
+ "grad_norm": 0.014448858797550201,
715
+ "learning_rate": 3.20079047092079e-05,
716
+ "loss": 11.9215,
717
+ "step": 276
718
+ },
719
+ {
720
+ "epoch": 1.8979591836734695,
721
+ "grad_norm": 0.015154534950852394,
722
+ "learning_rate": 3.0992166518402193e-05,
723
+ "loss": 11.9215,
724
+ "step": 279
725
+ },
726
+ {
727
+ "epoch": 1.9183673469387754,
728
+ "grad_norm": 0.011302746832370758,
729
+ "learning_rate": 2.9985517053941924e-05,
730
+ "loss": 11.9212,
731
+ "step": 282
732
+ },
733
+ {
734
+ "epoch": 1.9387755102040818,
735
+ "grad_norm": 0.008427857421338558,
736
+ "learning_rate": 2.8988437652193878e-05,
737
+ "loss": 11.9213,
738
+ "step": 285
739
+ },
740
+ {
741
+ "epoch": 1.9591836734693877,
742
+ "grad_norm": 0.009195243939757347,
743
+ "learning_rate": 2.8001405073533515e-05,
744
+ "loss": 11.9216,
745
+ "step": 288
746
+ },
747
+ {
748
+ "epoch": 1.9795918367346939,
749
+ "grad_norm": 0.009580681100487709,
750
+ "learning_rate": 2.7024891274378693e-05,
751
+ "loss": 11.9215,
752
+ "step": 291
753
+ },
754
+ {
755
+ "epoch": 2.0,
756
+ "grad_norm": 0.012584543786942959,
757
+ "learning_rate": 2.6059363181520518e-05,
758
+ "loss": 11.9215,
759
+ "step": 294
760
+ },
761
+ {
762
+ "epoch": 2.0136054421768708,
763
+ "eval_loss": 11.920774459838867,
764
+ "eval_runtime": 2.705,
765
+ "eval_samples_per_second": 365.99,
766
+ "eval_steps_per_second": 11.46,
767
+ "step": 296
768
+ },
769
+ {
770
+ "epoch": 2.020408163265306,
771
+ "grad_norm": 0.011345138773322105,
772
+ "learning_rate": 2.5105282468859027e-05,
773
+ "loss": 11.9217,
774
+ "step": 297
775
+ },
776
+ {
777
+ "epoch": 2.0408163265306123,
778
+ "grad_norm": 0.010026050731539726,
779
+ "learning_rate": 2.4163105336650642e-05,
780
+ "loss": 11.9212,
781
+ "step": 300
782
+ },
783
+ {
784
+ "epoch": 2.061224489795918,
785
+ "grad_norm": 0.012251692824065685,
786
+ "learning_rate": 2.3233282293372976e-05,
787
+ "loss": 11.9216,
788
+ "step": 303
789
+ },
790
+ {
791
+ "epoch": 2.0816326530612246,
792
+ "grad_norm": 0.009570994414389133,
793
+ "learning_rate": 2.2316257940310947e-05,
794
+ "loss": 11.9215,
795
+ "step": 306
796
+ },
797
+ {
798
+ "epoch": 2.1020408163265305,
799
+ "grad_norm": 0.011543840169906616,
800
+ "learning_rate": 2.141247075896774e-05,
801
+ "loss": 11.9216,
802
+ "step": 309
803
+ },
804
+ {
805
+ "epoch": 2.122448979591837,
806
+ "grad_norm": 0.008260723203420639,
807
+ "learning_rate": 2.0522352901401787e-05,
808
+ "loss": 11.9213,
809
+ "step": 312
810
+ },
811
+ {
812
+ "epoch": 2.142857142857143,
813
+ "grad_norm": 0.01163262128829956,
814
+ "learning_rate": 1.964632998359036e-05,
815
+ "loss": 11.9213,
816
+ "step": 315
817
+ },
818
+ {
819
+ "epoch": 2.163265306122449,
820
+ "grad_norm": 0.008361991494894028,
821
+ "learning_rate": 1.8784820881918275e-05,
822
+ "loss": 11.9215,
823
+ "step": 318
824
+ },
825
+ {
826
+ "epoch": 2.183673469387755,
827
+ "grad_norm": 0.009338167496025562,
828
+ "learning_rate": 1.7938237532889357e-05,
829
+ "loss": 11.9213,
830
+ "step": 321
831
+ },
832
+ {
833
+ "epoch": 2.204081632653061,
834
+ "grad_norm": 0.012479112483561039,
835
+ "learning_rate": 1.7106984736156077e-05,
836
+ "loss": 11.9212,
837
+ "step": 324
838
+ },
839
+ {
840
+ "epoch": 2.2244897959183674,
841
+ "grad_norm": 0.010746627114713192,
842
+ "learning_rate": 1.6291459960961887e-05,
843
+ "loss": 11.9212,
844
+ "step": 327
845
+ },
846
+ {
847
+ "epoch": 2.2448979591836733,
848
+ "grad_norm": 0.008642462082207203,
849
+ "learning_rate": 1.54920531560885e-05,
850
+ "loss": 11.9218,
851
+ "step": 330
852
+ },
853
+ {
854
+ "epoch": 2.2653061224489797,
855
+ "grad_norm": 0.009854198433458805,
856
+ "learning_rate": 1.4709146563399301e-05,
857
+ "loss": 11.9214,
858
+ "step": 333
859
+ },
860
+ {
861
+ "epoch": 2.2653061224489797,
862
+ "eval_loss": 11.920734405517578,
863
+ "eval_runtime": 2.8232,
864
+ "eval_samples_per_second": 350.662,
865
+ "eval_steps_per_second": 10.98,
866
+ "step": 333
867
+ },
868
+ {
869
+ "epoch": 2.2857142857142856,
870
+ "grad_norm": 0.01021676603704691,
871
+ "learning_rate": 1.394311453506763e-05,
872
+ "loss": 11.9214,
873
+ "step": 336
874
+ },
875
+ {
876
+ "epoch": 2.306122448979592,
877
+ "grad_norm": 0.011460648849606514,
878
+ "learning_rate": 1.319432335457786e-05,
879
+ "loss": 11.9212,
880
+ "step": 339
881
+ },
882
+ {
883
+ "epoch": 2.326530612244898,
884
+ "grad_norm": 0.010614064522087574,
885
+ "learning_rate": 1.246313106158431e-05,
886
+ "loss": 11.9213,
887
+ "step": 342
888
+ },
889
+ {
890
+ "epoch": 2.3469387755102042,
891
+ "grad_norm": 0.012491398490965366,
892
+ "learning_rate": 1.1749887280712164e-05,
893
+ "loss": 11.9209,
894
+ "step": 345
895
+ },
896
+ {
897
+ "epoch": 2.36734693877551,
898
+ "grad_norm": 0.011822333559393883,
899
+ "learning_rate": 1.1054933054382106e-05,
900
+ "loss": 11.9218,
901
+ "step": 348
902
+ },
903
+ {
904
+ "epoch": 2.387755102040816,
905
+ "grad_norm": 0.009794401936233044,
906
+ "learning_rate": 1.0378600679738442e-05,
907
+ "loss": 11.9217,
908
+ "step": 351
909
+ },
910
+ {
911
+ "epoch": 2.4081632653061225,
912
+ "grad_norm": 0.007897097617387772,
913
+ "learning_rate": 9.721213549759011e-06,
914
+ "loss": 11.9214,
915
+ "step": 354
916
+ },
917
+ {
918
+ "epoch": 2.4285714285714284,
919
+ "grad_norm": 0.010740007273852825,
920
+ "learning_rate": 9.08308599862256e-06,
921
+ "loss": 11.9216,
922
+ "step": 357
923
+ },
924
+ {
925
+ "epoch": 2.4489795918367347,
926
+ "grad_norm": 0.009365779347717762,
927
+ "learning_rate": 8.464523151407722e-06,
928
+ "loss": 11.9217,
929
+ "step": 360
930
+ },
931
+ {
932
+ "epoch": 2.4693877551020407,
933
+ "grad_norm": 0.008345544338226318,
934
+ "learning_rate": 7.865820778195366e-06,
935
+ "loss": 11.9213,
936
+ "step": 363
937
+ },
938
+ {
939
+ "epoch": 2.489795918367347,
940
+ "grad_norm": 0.00913438480347395,
941
+ "learning_rate": 7.287265152644079e-06,
942
+ "loss": 11.9213,
943
+ "step": 366
944
+ },
945
+ {
946
+ "epoch": 2.510204081632653,
947
+ "grad_norm": 0.010115435346961021,
948
+ "learning_rate": 6.7291329151064455e-06,
949
+ "loss": 11.9211,
950
+ "step": 369
951
+ },
952
+ {
953
+ "epoch": 2.5170068027210886,
954
+ "eval_loss": 11.920697212219238,
955
+ "eval_runtime": 2.9612,
956
+ "eval_samples_per_second": 334.32,
957
+ "eval_steps_per_second": 10.469,
958
+ "step": 370
959
+ },
960
+ {
961
+ "epoch": 2.5306122448979593,
962
+ "grad_norm": 0.011003920808434486,
963
+ "learning_rate": 6.191690940351569e-06,
964
+ "loss": 11.9213,
965
+ "step": 372
966
+ },
967
+ {
968
+ "epoch": 2.5510204081632653,
969
+ "grad_norm": 0.010234669782221317,
970
+ "learning_rate": 5.67519620995704e-06,
971
+ "loss": 11.9212,
972
+ "step": 375
973
+ },
974
+ {
975
+ "epoch": 2.571428571428571,
976
+ "grad_norm": 0.008843546733260155,
977
+ "learning_rate": 5.179895689431502e-06,
978
+ "loss": 11.921,
979
+ "step": 378
980
+ },
981
+ {
982
+ "epoch": 2.5918367346938775,
983
+ "grad_norm": 0.010195941664278507,
984
+ "learning_rate": 4.706026210126302e-06,
985
+ "loss": 11.9212,
986
+ "step": 381
987
+ },
988
+ {
989
+ "epoch": 2.612244897959184,
990
+ "grad_norm": 0.008211387321352959,
991
+ "learning_rate": 4.2538143559929915e-06,
992
+ "loss": 11.9214,
993
+ "step": 384
994
+ },
995
+ {
996
+ "epoch": 2.63265306122449,
997
+ "grad_norm": 0.008034368976950645,
998
+ "learning_rate": 3.823476355240668e-06,
999
+ "loss": 11.9217,
1000
+ "step": 387
1001
+ },
1002
+ {
1003
+ "epoch": 2.6530612244897958,
1004
+ "grad_norm": 0.010721351020038128,
1005
+ "learning_rate": 3.41521797694494e-06,
1006
+ "loss": 11.9214,
1007
+ "step": 390
1008
+ },
1009
+ {
1010
+ "epoch": 2.673469387755102,
1011
+ "grad_norm": 0.012811296619474888,
1012
+ "learning_rate": 3.029234432658051e-06,
1013
+ "loss": 11.9213,
1014
+ "step": 393
1015
+ },
1016
+ {
1017
+ "epoch": 2.693877551020408,
1018
+ "grad_norm": 0.008711621165275574,
1019
+ "learning_rate": 2.6657102830671078e-06,
1020
+ "loss": 11.9216,
1021
+ "step": 396
1022
+ },
1023
+ {
1024
+ "epoch": 2.7142857142857144,
1025
+ "grad_norm": 0.0077378894202411175,
1026
+ "learning_rate": 2.324819349745133e-06,
1027
+ "loss": 11.9212,
1028
+ "step": 399
1029
+ },
1030
+ {
1031
+ "epoch": 2.7346938775510203,
1032
+ "grad_norm": 0.008991695940494537,
1033
+ "learning_rate": 2.0067246320370935e-06,
1034
+ "loss": 11.9217,
1035
+ "step": 402
1036
+ },
1037
+ {
1038
+ "epoch": 2.7551020408163263,
1039
+ "grad_norm": 0.012436670251190662,
1040
+ "learning_rate": 1.7115782291206083e-06,
1041
+ "loss": 11.9211,
1042
+ "step": 405
1043
+ },
1044
+ {
1045
+ "epoch": 2.7687074829931975,
1046
+ "eval_loss": 11.92069149017334,
1047
+ "eval_runtime": 2.6971,
1048
+ "eval_samples_per_second": 367.061,
1049
+ "eval_steps_per_second": 11.494,
1050
+ "step": 407
1051
+ },
1052
+ {
1053
+ "epoch": 2.7755102040816326,
1054
+ "grad_norm": 0.009732716716825962,
1055
+ "learning_rate": 1.4395212672787373e-06,
1056
+ "loss": 11.9218,
1057
+ "step": 408
1058
+ },
1059
+ {
1060
+ "epoch": 2.795918367346939,
1061
+ "grad_norm": 0.015267143957316875,
1062
+ "learning_rate": 1.1906838324194258e-06,
1063
+ "loss": 11.9214,
1064
+ "step": 411
1065
+ },
1066
+ {
1067
+ "epoch": 2.816326530612245,
1068
+ "grad_norm": 0.00914175808429718,
1069
+ "learning_rate": 9.65184907874045e-07,
1070
+ "loss": 11.9214,
1071
+ "step": 414
1072
+ },
1073
+ {
1074
+ "epoch": 2.836734693877551,
1075
+ "grad_norm": 0.007872041314840317,
1076
+ "learning_rate": 7.631323175047167e-07,
1077
+ "loss": 11.9209,
1078
+ "step": 417
1079
+ },
1080
+ {
1081
+ "epoch": 2.857142857142857,
1082
+ "grad_norm": 0.010440999642014503,
1083
+ "learning_rate": 5.846226741475557e-07,
1084
+ "loss": 11.9219,
1085
+ "step": 420
1086
+ },
1087
+ {
1088
+ "epoch": 2.877551020408163,
1089
+ "grad_norm": 0.0071489280089735985,
1090
+ "learning_rate": 4.2974133341662226e-07,
1091
+ "loss": 11.9211,
1092
+ "step": 423
1093
+ },
1094
+ {
1095
+ "epoch": 2.8979591836734695,
1096
+ "grad_norm": 0.011952205561101437,
1097
+ "learning_rate": 2.985623528904913e-07,
1098
+ "loss": 11.9216,
1099
+ "step": 426
1100
+ },
1101
+ {
1102
+ "epoch": 2.9183673469387754,
1103
+ "grad_norm": 0.010003963485360146,
1104
+ "learning_rate": 1.9114845670117342e-07,
1105
+ "loss": 11.9211,
1106
+ "step": 429
1107
+ },
1108
+ {
1109
+ "epoch": 2.938775510204082,
1110
+ "grad_norm": 0.012489134445786476,
1111
+ "learning_rate": 1.0755100554208364e-07,
1112
+ "loss": 11.9213,
1113
+ "step": 432
1114
+ },
1115
+ {
1116
+ "epoch": 2.9591836734693877,
1117
+ "grad_norm": 0.007483236957341433,
1118
+ "learning_rate": 4.780997210962479e-08,
1119
+ "loss": 11.9215,
1120
+ "step": 435
1121
+ },
1122
+ {
1123
+ "epoch": 2.979591836734694,
1124
+ "grad_norm": 0.008097127079963684,
1125
+ "learning_rate": 1.1953921989915628e-08,
1126
+ "loss": 11.9215,
1127
+ "step": 438
1128
+ },
1129
+ {
1130
+ "epoch": 3.0,
1131
+ "grad_norm": 0.009465374983847141,
1132
+ "learning_rate": 0.0,
1133
+ "loss": 11.9213,
1134
+ "step": 441
1135
+ }
1136
+ ],
1137
+ "logging_steps": 3,
1138
+ "max_steps": 441,
1139
+ "num_input_tokens_seen": 0,
1140
+ "num_train_epochs": 3,
1141
+ "save_steps": 37,
1142
+ "stateful_callbacks": {
1143
+ "TrainerControl": {
1144
+ "args": {
1145
+ "should_epoch_stop": false,
1146
+ "should_evaluate": false,
1147
+ "should_log": false,
1148
+ "should_save": true,
1149
+ "should_training_stop": true
1150
+ },
1151
+ "attributes": {}
1152
+ }
1153
+ },
1154
+ "total_flos": 16600521768960.0,
1155
+ "train_batch_size": 8,
1156
+ "trial_name": null,
1157
+ "trial_params": null
1158
+ }
last-checkpoint/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5b5db478b32acb23e06e22cfb0e8772065e7601806c043041ff117dd456594b
3
+ size 6712
last-checkpoint/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|endoftext|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|endoftext|>",
37
+ "errors": "replace",
38
+ "model_max_length": 32768,
39
+ "pad_token": "<|endoftext|>",
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "Qwen2Tokenizer",
42
+ "unk_token": null
43
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5b5db478b32acb23e06e22cfb0e8772065e7601806c043041ff117dd456594b
3
+ size 6712
vocab.json ADDED
The diff for this file is too large to render. See raw diff