|
#!/bin/bash |
|
|
|
SCRIPT_DIR=. |
|
MODEL_DIR=~/models/training_v3_new |
|
|
|
IMAGE_ENCODER="openai/clip-vit-base-patch32" |
|
TEXT_ENCODER="indobenchmark/indobert-base-p2" |
|
|
|
python ${SCRIPT_DIR}/run_hybrid_clip.py \ |
|
--output_dir ${MODEL_DIR} \ |
|
--overwrite_output_dir \ |
|
--tokenizer_name=${TEXT_ENCODER} \ |
|
--train_file="../data/train_dataset_v3.json" \ |
|
--validation_file="../data/val_dataset_v3.json" \ |
|
--do_train --do_eval \ |
|
--num_train_epochs="10" --max_seq_length 96 \ |
|
--per_device_train_batch_size="64" \ |
|
--per_device_eval_batch_size="64" \ |
|
--learning_rate="0.00005" --warmup_ratio 0.1 --weight_decay 0.0 \ |
|
--preprocessing_num_workers 16 \ |
|
--exp_name training_v3 \ |
|
--text_model_name_or_path=${TEXT_ENCODER} \ |
|
--vision_model_name_or_path=${IMAGE_ENCODER} \ |
|
--eval_steps 2500 \ |
|
--logging_steps 200 \ |
|
--save_steps 2500 \ |
|
--save_total_limit 5 \ |
|
--log_wandb \ |
|
--freeze_backbones |
|
|
|
|