sedrickkeh commited on
Commit
203cdb0
·
verified ·
1 Parent(s): baf7bb3

Upload configs.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. configs.yaml +44 -0
configs.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bf16: 'True'
2
+ chosen: chosen
3
+ cutoff_len: '2048'
4
+ dataset: mlfoundations-dev/gemma2-ultrafeedback-armorm
5
+ dataset_dir: ONLINE
6
+ ddp_timeout: '180000000'
7
+ deepspeed: /opt/ml/code/zero3.json
8
+ do_train: 'True'
9
+ enable_liger_kernel: 'False'
10
+ eval_strategy: epoch
11
+ finetuning_type: full
12
+ formatting: sharegpt
13
+ global_batch_size: '128'
14
+ gradient_accumulation_steps: '8'
15
+ gradient_checkpointing: 'True'
16
+ hub_model_id: mlfoundations-dev/simpo-stackoverflow_25000tasks_1p
17
+ learning_rate: 8e-07
18
+ logging_steps: '1'
19
+ lr_scheduler_type: cosine
20
+ messages: conversations
21
+ model_name_or_path: mlfoundations-dev/stackoverflow_25000tasks_1p
22
+ num_train_epochs: '1.0'
23
+ optim: adamw_torch
24
+ output_dir: /opt/ml/model
25
+ overwrite_cache: 'True'
26
+ overwrite_output_dir: 'True'
27
+ per_device_eval_batch_size: '2'
28
+ per_device_train_batch_size: '2'
29
+ plot_loss: 'True'
30
+ pref_beta: '10'
31
+ pref_loss: simpo
32
+ preprocessing_num_workers: '16'
33
+ push_to_db: 'True'
34
+ push_to_hub: 'True'
35
+ ranking: 'True'
36
+ rejected: rejected
37
+ report_to: wandb
38
+ run_name: simpo-stackoverflow_25000tasks_1p
39
+ save_strategy: epoch
40
+ simpo_gamma: '5'
41
+ stage: dpo
42
+ template: gemma
43
+ val_size: '0.05'
44
+ warmup_ratio: '0.1'