|
{ |
|
|
|
"cache_dir":"/tmp", |
|
"downstream_fusion":false, |
|
"downstream_fusion_layers":1, |
|
"downstream_fusion_method":"elmo", |
|
"drop_rate":0.1, |
|
"freeze_RoBERTa":false, |
|
"freeze_ViT":false, |
|
"freeze_layer_count_roberta":false, |
|
"freeze_layer_count_vit":false, |
|
"head_hidden_scale":2, |
|
"hidden_size":768, |
|
"image_size":288, |
|
"input_text_embed_size":768, |
|
"link_tower_shared":false, |
|
"link_tower_type":"add", |
|
"log_dir":"log_dir", |
|
"loss_names":{"contras": 0, |
|
"irtr": 0, |
|
"itm": 0, |
|
"mlm": 0, |
|
"mpp": 0, |
|
"nlvr2": 0, |
|
"snli": 0, |
|
"vcr": 0, |
|
"vcr_qar": 0, |
|
"vqa": 1}, |
|
"max_text_len":50, |
|
"mlp_ratio":4, |
|
"model_type":"bridgetower", |
|
"num_heads":12, |
|
"num_layers":6, |
|
"num_nodes":1, |
|
"only_load_cross_modal_from_meter":false, |
|
"patch_size":16, |
|
"resolution_before":224, |
|
"stop_gradient":false, |
|
"task_head_layers":2, |
|
"test_only":false, |
|
"tokenizer":"roberta-base", |
|
"unfreeze_RoBERTa_attention":false, |
|
"unfreeze_RoBERTa_embeddings":false, |
|
"unfreeze_RoBERTa_encoder":false, |
|
"unfreeze_RoBERTa_layernorm":false, |
|
"unfreeze_ViT_attention":false, |
|
"unfreeze_ViT_layernorm":false, |
|
"vit":"ViT-B/16", |
|
"vit_layernorm_init_from_vit":false, |
|
"vit_layernorm_shared":true, |
|
"vit_remove_last":false, |
|
"vocab_size":50265 |
|
} |
|
|