Spaces:
Runtime error
Runtime error
# Guide: | |
# This script supports distributed training on multi-gpu workers (as well as single-worker training). | |
# Please set the options below according to the comments. | |
# For multi-gpu workers training, these options should be manually set for each worker. | |
# After setting the options, please run the script on each worker. | |
# To use the shuffled data (if exists), please uncomment the Line 24. | |
# Number of GPUs per GPU worker | |
GPUS_PER_NODE=8 | |
# Number of GPU workers, for single-worker training, please set to 1 | |
WORKER_CNT=2 | |
# The ip address of the rank-0 worker, for single-worker training, please set to localhost | |
export MASTER_ADDR=XX.XX.XX.XX | |
# The port for communication | |
export MASTER_PORT=8514 | |
# The rank of this worker, should be in {0, ..., WORKER_CNT-1}, for single-worker training, please set to 0 | |
export RANK=0 | |
data_dir=../../dataset/imagenet_1k_data | |
data=${data_dir}/imagenet_1k_train.tsv,${data_dir}/imagenet_1k_val_subset.tsv | |
# Note: If you have shuffled the data in advance, please uncomment the line below. | |
# data=${data_dir}/imagenet_1k_train_1.tsv,${data_dir}/imagenet_1k_train_2.tsv,${data_dir}/imagenet_1k_train_3.tsv,${data_dir}/imagenet_1k_train_4.tsv,${data_dir}/imagenet_1k_train_5.tsv,${data_dir}/imagenet_1k_train_6.tsv,${data_dir}/imagenet_1k_train_7.tsv,${data_dir}/imagenet_1k_train_8.tsv,${data_dir}/imagenet_1k_train_9.tsv,${data_dir}/imagenet_1k_train_10.tsv,${data_dir}/imagenet_1k_val_subset.tsv | |
ans2label_file=../../dataset/imagenet_1k_data/class2label_new.pkl | |
restore_file=../../checkpoints/ofa_large.pt | |
selected_cols=0,2 | |
log_dir=./imagenet_1k_logs | |
save_dir=./imagenet_1k_checkpoints | |
mkdir -p $log_dir $save_dir | |
bpe_dir=../../utils/BPE | |
user_dir=../../ofa_module | |
task=image_classify | |
arch=ofa_large | |
criterion=adjust_label_smoothed_cross_entropy | |
label_smoothing=0.1 | |
batch_size=4 | |
update_freq=4 | |
resnet_drop_path_rate=0.0 | |
encoder_drop_path_rate=0.1 | |
decoder_drop_path_rate=0.1 | |
dropout=0.1 | |
attention_dropout=0.0 | |
max_src_length=128 | |
max_tgt_length=30 | |
num_bins=1000 | |
patch_image_size=480 | |
for total_num_updates in {160000,}; do | |
echo "total_num_updates "${total_num_updates} | |
for warmup_updates in {1000,}; do | |
echo "warmup_updates "${warmup_updates} | |
for lr in {5e-5,}; do | |
echo "lr "${lr} | |
for patch_image_size in {480,}; do | |
echo "patch_image_size "${patch_image_size} | |
log_file=${log_dir}/${total_num_updates}"_"${warmup_updates}"_"${lr}"_"${patch_image_size}"_rank"${RANK}".log" | |
save_path=${save_dir}/${total_num_updates}"_"${warmup_updates}"_"${lr}"_"${patch_image_size} | |
mkdir -p $save_path | |
python3 -m torch.distributed.launch --nproc_per_node=${GPUS_PER_NODE} --nnodes=${WORKER_CNT} --node_rank=${RANK} --master_addr=${MASTER_ADDR} --master_port=${MASTER_PORT} ../../train.py \ | |
${data} \ | |
--selected-cols=${selected_cols} \ | |
--bpe-dir=${bpe_dir} \ | |
--user-dir=${user_dir} \ | |
--restore-file=${restore_file} \ | |
--reset-optimizer --reset-dataloader --reset-meters \ | |
--save-dir=${save_path} \ | |
--task=${task} \ | |
--arch=${arch} \ | |
--criterion=${criterion} \ | |
--label-smoothing=${label_smoothing} \ | |
--batch-size=${batch_size} \ | |
--update-freq=${update_freq} \ | |
--encoder-normalize-before \ | |
--decoder-normalize-before \ | |
--share-decoder-input-output-embed \ | |
--share-all-embeddings \ | |
--layernorm-embedding \ | |
--patch-layernorm-embedding \ | |
--code-layernorm-embedding \ | |
--resnet-drop-path-rate=${resnet_drop_path_rate} \ | |
--encoder-drop-path-rate=${encoder_drop_path_rate} \ | |
--decoder-drop-path-rate=${decoder_drop_path_rate} \ | |
--dropout=${dropout} \ | |
--attention-dropout=${attention_dropout} \ | |
--weight-decay=0.01 \ | |
--optimizer=adam \ | |
--adam-betas="(0.9,0.999)" \ | |
--adam-eps=1e-08 \ | |
--clip-norm=1.0 \ | |
--lr-scheduler=polynomial_decay \ | |
--lr=${lr} \ | |
--total-num-update=${total_num_updates} \ | |
--warmup-updates=${warmup_updates} \ | |
--log-format=simple \ | |
--log-interval=10 \ | |
--fixed-validation-seed=7 \ | |
--keep-last-epochs=15 \ | |
--save-interval=1 --validate-interval=1 \ | |
--max-update=${total_num_updates} \ | |
--best-checkpoint-metric=score --maximize-best-checkpoint-metric \ | |
--max-src-length=${max_src_length} \ | |
--max-tgt-length=${max_tgt_length} \ | |
--find-unused-parameters \ | |
--freeze-encoder-embedding \ | |
--freeze-decoder-embedding \ | |
--ans2label-file=${ans2label_file} \ | |
--valid-batch-size=20 \ | |
--add-type-embedding \ | |
--scale-attn \ | |
--scale-fc \ | |
--scale-heads \ | |
--disable-entangle \ | |
--num-bins=${num_bins} \ | |
--patch-image-size=${patch_image_size} \ | |
--fp16 \ | |
--fp16-scale-window=512 \ | |
--imagenet-default-mean-and-std \ | |
--num-workers=0 > ${log_file} 2>&1 | |
done | |
done | |
done | |
done | |