ZhangYuhan commited on
Commit
94f0dfc
1 Parent(s): 2a4294a

Upload config.yaml

Browse files
Files changed (1) hide show
  1. config.yaml +180 -0
config.yaml ADDED
@@ -0,0 +1,180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ accelerator:
2
+ _target_: trainer.accelerators.deepspeed_accelerator.DeepSpeedAccelerator
3
+ output_dir: outputs/Multi-CLIP_ViT_H_14-clean/from_con_image/stage2_eval_from_latest
4
+ mixed_precision: BF16
5
+ gradient_accumulation_steps: 1
6
+ log_with: WANDB
7
+ debug:
8
+ activate: false
9
+ port: 5900
10
+ seed: 42
11
+ resume_from_latest: true
12
+ resume_from_name_or_path: outputs/Multi-CLIP_ViT_H_14-clean/from_con_image/stage1_neg/checkpoint-latest
13
+ train_from_scratch: true
14
+ max_steps: 4000
15
+ num_epochs: 10
16
+ validate_steps: 100
17
+ eval_on_start: true
18
+ project_name: Multi-CLIP_ViT_H_14-clean
19
+ max_grad_norm: 1.0
20
+ save_steps: 100
21
+ metric_name: loss
22
+ metric_mode: MIN
23
+ limit_num_checkpoints: 3
24
+ save_only_if_best: true
25
+ dynamo_backend: 'NO'
26
+ keep_best_ckpts: true
27
+ deepspeed:
28
+ fp16:
29
+ enabled: false
30
+ bf16:
31
+ enabled: true
32
+ optimizer:
33
+ type: AdamW
34
+ params:
35
+ lr: auto
36
+ weight_decay: auto
37
+ torch_adam: true
38
+ adam_w_mode: true
39
+ scheduler:
40
+ type: WarmupDecayLR
41
+ params:
42
+ warmup_min_lr: auto
43
+ warmup_max_lr: auto
44
+ warmup_num_steps: auto
45
+ total_num_steps: auto
46
+ zero_optimization:
47
+ stage: 2
48
+ allgather_partitions: true
49
+ allgather_bucket_size: 200000000.0
50
+ overlap_comm: true
51
+ reduce_scatter: true
52
+ reduce_bucket_size: 500000000
53
+ contiguous_gradients: true
54
+ gradient_accumulation_steps: 16
55
+ gradient_clipping: 1.0
56
+ steps_per_print: 1
57
+ train_batch_size: auto
58
+ train_micro_batch_size_per_gpu: auto
59
+ wall_clock_breakdown: false
60
+ deepspeed_final:
61
+ fp16:
62
+ enabled: false
63
+ bf16:
64
+ enabled: true
65
+ optimizer:
66
+ type: AdamW
67
+ params:
68
+ lr: auto
69
+ weight_decay: auto
70
+ torch_adam: true
71
+ adam_w_mode: true
72
+ scheduler:
73
+ type: WarmupDecayLR
74
+ params:
75
+ warmup_min_lr: auto
76
+ warmup_max_lr: auto
77
+ warmup_num_steps: auto
78
+ total_num_steps: auto
79
+ zero_optimization:
80
+ stage: 2
81
+ allgather_partitions: true
82
+ allgather_bucket_size: 200000000.0
83
+ overlap_comm: true
84
+ reduce_scatter: true
85
+ reduce_bucket_size: 500000000
86
+ contiguous_gradients: true
87
+ gradient_accumulation_steps: 16
88
+ gradient_clipping: 1.0
89
+ steps_per_print: .inf
90
+ train_batch_size: auto
91
+ train_micro_batch_size_per_gpu: auto
92
+ wall_clock_breakdown: false
93
+ task:
94
+ limit_examples_to_wandb: 50
95
+ _target_: trainer.tasks.mvclip_task.MVCLIPTask
96
+ pretrained_clip_model_name_or_path: laion/CLIP-ViT-H-14-laion2B-s32B-b79K
97
+ label_0_column_name: label_0
98
+ label_1_column_name: label_1
99
+ reference_type_column_name: reference_type
100
+ reference_input_column_name: reference_input
101
+ reference_idx_column_name: reference_idx
102
+ normal_pixels_0_column_name: normal_pixel_values_0
103
+ normal_pixels_1_column_name: normal_pixel_values_1
104
+ rgb_pixels_0_column_name: rgb_pixel_values_0
105
+ rgb_pixels_1_column_name: rgb_pixel_values_1
106
+ model:
107
+ _target_: trainer.models.mvclip_model.MVCLIPModel
108
+ pretrained_clip_model_name_or_path: laion/CLIP-ViT-H-14-laion2B-s32B-b79K
109
+ freeze_clip: true
110
+ freeze_vision: false
111
+ freeze_logit_scale: false
112
+ freeze_logit_proj: false
113
+ criterion:
114
+ _target_: trainer.criterions.criterion.MVCriterion
115
+ is_distributed: true
116
+ reference_type_column_name: reference_type
117
+ reference_input_column_name: reference_input
118
+ label_0_column_name: label_0
119
+ label_1_column_name: label_1
120
+ normal_pixels_0_column_name: normal_pixel_values_0
121
+ normal_pixels_1_column_name: normal_pixel_values_1
122
+ rgb_pixels_0_column_name: rgb_pixel_values_0
123
+ rgb_pixels_1_column_name: rgb_pixel_values_1
124
+ num_examples_per_prompt_column_name: num_example_per_prompt
125
+ in_batch_negatives: false
126
+ dataset:
127
+ train_split_name: train_valid
128
+ valid_split_name: valid
129
+ test_split_name: test
130
+ batch_size: 16
131
+ num_workers: 2
132
+ drop_last: true
133
+ _target_: trainer.datasetss.dataset.MVDataset
134
+ records_dir: /mnt/petrelfs/zhangyuhan/data/evaluation/records/battle/clean/splits
135
+ gallery_dir: /mnt/petrelfs/zhangyuhan/data/evaluation/gallery510
136
+ text_gallery_path: /mnt/petrelfs/zhangyuhan/data/evaluation/gallery510/prompts_510.json
137
+ image_gallery_dir: /mnt/petrelfs/zhangyuhan/data/evaluation/gallery510/rgba
138
+ con_images_dir: /mnt/petrelfs/zhangyuhan/data/evaluation/con_images
139
+ split_con_image: false
140
+ cache_dir: null
141
+ reference_type_column_name: reference_type
142
+ reference_idx_column_name: reference_idx
143
+ reference_column_name: reference
144
+ normal_image_0_column_name: normal_image_0
145
+ normal_image_1_column_name: normal_image_1
146
+ rgb_image_0_column_name: rgb_image_0
147
+ rgb_image_1_column_name: rgb_image_1
148
+ eval_dims_column_name: eval_dims
149
+ label_0_column_name: label_0
150
+ label_1_column_name: label_1
151
+ are_different_column_name: are_different
152
+ has_label_column_name: has_label
153
+ reference_input_column_name: reference_input
154
+ normal_pixels_0_column_name: normal_pixel_values_0
155
+ normal_pixels_1_column_name: normal_pixel_values_1
156
+ rgb_pixels_0_column_name: rgb_pixel_values_0
157
+ rgb_pixels_1_column_name: rgb_pixel_values_1
158
+ num_examples_per_prompt_column_name: num_example_per_prompt
159
+ shuffle: true
160
+ keep_only_different: false
161
+ keep_only_with_label: false
162
+ keep_only_with_label_in_non_train: true
163
+ processor:
164
+ _target_: transformers.AutoProcessor.from_pretrained
165
+ pretrained_model_name_or_path: laion/CLIP-ViT-H-14-laion2B-s32B-b79K
166
+ limit_examples_per_prompt: -1
167
+ only_on_best: false
168
+ optimizer:
169
+ _target_: trainer.optimizers.dummy_optimizer.BaseDummyOptim
170
+ lr: 3.0e-06
171
+ weight_decay: 0.3
172
+ lr_scheduler:
173
+ _target_: trainer.lr_schedulers.dummy_lr_scheduler.instantiate_dummy_lr_scheduler
174
+ lr: 3.0e-06
175
+ lr_warmup_steps: 500
176
+ total_num_steps: 4000
177
+ debug:
178
+ activate: false
179
+ port: 5900
180
+ output_dir: outputs/Multi-CLIP_ViT_H_14-clean/from_con_image/stage2_eval_from_latest