{ | |
"text_config": { | |
"float32_logits": true, | |
"dtype": "bfloat16", | |
"activations": [ | |
"gelu", | |
"linear" | |
], | |
"normalize_qk": false, | |
"use_bias": false, | |
"force_scale": false, | |
"attention_dropout": 0.0, | |
"mlp_dropout_rate": 0.0, | |
"unroll": 100, | |
"remat_policy": "none", | |
"eos_token_id": 2, | |
"mask_token_id": 4, | |
"pad_token_id": 3, | |
"bos_token_id": 1, | |
"masked_pred_prob": 0.75, | |
"is_decoder": true, | |
"pool_type": null, | |
"num_queries": 1, | |
"vocab_size": 65536, | |
"hidden_size": 1024, | |
"max_length": 64, | |
"num_layers": 12, | |
"use_rmsnorm": true, | |
"ln_type": "normformer", | |
"num_heads": 16, | |
"position_embedding_type": "rotary", | |
"use_causal_mask": true, | |
"mlp_dim": 3072 | |
}, | |
"vision_config": { | |
"float32_logits": true, | |
"position_embedding_type": "learnt", | |
"position_embedding_shape": null, | |
"position_embedding_factorized": false, | |
"dtype": "bfloat16", | |
"activations": [ | |
"gelu", | |
"linear" | |
], | |
"normalize_qk": false, | |
"use_bias": false, | |
"force_scale": false, | |
"attention_dropout": 0.0, | |
"mlp_dropout_rate": 0.0, | |
"pool_type": null, | |
"unroll": 100, | |
"registers": 8, | |
"keep_registers": true, | |
"remat_policy": "none", | |
"num_queries": 1, | |
"image_size": 256, | |
"hidden_size": 1024, | |
"patch_size": 16, | |
"num_layers": 24, | |
"use_rmsnorm": true, | |
"ln_type": "normformer", | |
"num_heads": 16, | |
"use_causal_mask": false, | |
"mlp_dim": 3072 | |
}, | |
"projection_dim": 1024, | |
"logit_scale_init_value": 2.3, | |
"logit_bias_init_value": -10.0, | |
"dtype": "bfloat16" | |
} |