File size: 1,516 Bytes
ec2e79f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 |
{
"data_root_dir": "/hai/scratch/belkhale/datasets",
"hf_token": ".hf_token",
"image_aug": false,
"is_resume": true,
"pretrained_checkpoint": "runs/prism-qwen25-dinosiglip-224px+0_5b+mx-libero-90+n1+b32+x7--extra_tokens/checkpoints/step-060000-epoch-26-loss=0.1209.pt",
"resume_epoch": 26,
"resume_step": 60000,
"run_id": "prism-qwen25-dinosiglip-224px+0_5b+mx-libero-90+n1+b32+x7--extra_tokens",
"run_id_note": "extra_tokens",
"run_root_dir": "runs",
"save_interval": 2500,
"seed": 7,
"trackers": [
"jsonl",
"wandb"
],
"vla": {
"action_tokenizer": "extra_action_tokenizer",
"base_vlm": "prism-qwen25-extra-dinosiglip-224px+0_5b",
"data_mix": "libero_90",
"enable_gradient_checkpointing": true,
"enable_mixed_precision_training": true,
"epochs": 1000,
"expected_world_size": 8,
"freeze_llm_backbone": false,
"freeze_vision_backbone": false,
"global_batch_size": 256,
"learning_rate": 2e-05,
"lr_scheduler_type": "constant",
"max_grad_norm": 1.0,
"max_steps": null,
"per_device_batch_size": 32,
"reduce_in_full_precision": true,
"save_every_n_steps": 25000,
"shuffle_buffer_size": 256000,
"train_strategy": "fsdp-full-shard",
"type": "prism-qwen25-dinosiglip-224px+0_5b+mx-libero-90",
"unfreeze_last_llm_layer": false,
"vla_id": "prism-qwen25-dinosiglip-224px+0_5b+mx-libero-90",
"warmup_ratio": 0.0,
"weight_decay": 0.0
},
"wandb_entity": null,
"wandb_project": "prismatic"
}
|