3outeille's picture
3outeille HF staff
Upload llama-1B/64_GPUS/dp-1_tp-8_pp-8_mbz-16
8990860 verified
raw
history blame
134 kB
========================
START TIME: Sat Jul 6 09:43:27 UTC 2024
python3 version = Python 3.10.14
========================
The token has not been saved to the git credentials helper. Pass `add_to_git_credential=True` in this function directly or `--add-to-git-credential` if using via `huggingface-cli` if you want to set the git credential as well.
Token is valid (permission: write).
Your token has been saved to /admin/home/ferdinand_mom/.cache/huggingface/token
Login successful
Already on 'bench_cluster'
M examples/config_tiny_llama.py
M examples/config_tiny_llama.yaml
M examples/train_tiny_llama.sh
Your branch is up to date with 'origin/bench_cluster'.
Job status: RUNNING
[2024-07-06 09:43:29,611] torch.distributed.run: [WARNING]
[2024-07-06 09:43:29,611] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,611] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
[2024-07-06 09:43:29,611] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,612] torch.distributed.run: [WARNING]
[2024-07-06 09:43:29,612] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,612] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
[2024-07-06 09:43:29,612] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,621] torch.distributed.run: [WARNING]
[2024-07-06 09:43:29,621] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,621] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
[2024-07-06 09:43:29,621] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,640] torch.distributed.run: [WARNING]
[2024-07-06 09:43:29,640] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,640] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
[2024-07-06 09:43:29,640] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,642] torch.distributed.run: [WARNING]
[2024-07-06 09:43:29,642] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,642] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
[2024-07-06 09:43:29,642] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,678] torch.distributed.run: [WARNING]
[2024-07-06 09:43:29,678] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,678] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
[2024-07-06 09:43:29,678] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,681] torch.distributed.run: [WARNING]
[2024-07-06 09:43:29,681] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,681] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
[2024-07-06 09:43:29,681] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,712] torch.distributed.run: [WARNING]
[2024-07-06 09:43:29,712] torch.distributed.run: [WARNING] *****************************************
[2024-07-06 09:43:29,712] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
[2024-07-06 09:43:29,712] torch.distributed.run: [WARNING] *****************************************
[default0]:07/06/2024 09:43:49 [WARNING|DP=0|PP=0|TP=0|ip-26-0-160-103]: [Vocab Size Padding] Padded vocab (size: 50257) with 7 dummy tokens (new size: 50264)
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Config:
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Config(general=GeneralArgs(project='bench_cluster',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: run='%date_%jobid',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: seed=42,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: step=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: consumed_train_samples=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: benchmark_csv_path=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: ignore_sanity_checks=True),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: parallelism=ParallelismArgs(dp=1,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: pp=8,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: tp=8,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: pp_engine=<nanotron.parallel.pipeline_parallel.engine.AllForwardAllBackwardPipelineEngine object at 0x7fd62db78730>,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: tp_mode=<TensorParallelLinearMode.REDUCE_SCATTER: 2>,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: tp_linear_async_communication=False,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: expert_parallel_size=1),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: model=ModelArgs(model_config=LlamaConfig(bos_token_id=1,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: eos_token_id=2,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: hidden_act='silu',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: hidden_size=2048,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: initializer_range=0.02,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: intermediate_size=4096,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: is_llama_config=True,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: max_position_embeddings=4096,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: num_attention_heads=32,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: num_hidden_layers=24,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: num_key_value_heads=32,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: pad_token_id=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: pretraining_tp=1,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: rms_norm_eps=1e-05,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: rope_scaling=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: rope_theta=10000.0,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: tie_word_embeddings=True,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: use_cache=True,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: vocab_size=50264),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: init_method=RandomInit(std=0.025),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: dtype=torch.bfloat16,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: make_vocab_size_divisible_by=1,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: ddp_bucket_cap_mb=25),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: tokenizer=TokenizerArgs(tokenizer_name_or_path='openai-community/gpt2',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: tokenizer_revision=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: tokenizer_max_length=None),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: checkpoints=CheckpointsArgs(checkpoints_path=PosixPath('/dev/null'),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: checkpoint_interval=100000,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: save_initial_state=False,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: resume_checkpoint_path=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: checkpoints_path_is_shared_file_system=False),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: logging=LoggingArgs(log_level='info',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: log_level_replica='info',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: iteration_step_info_interval=1),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: tokens=TokensArgs(sequence_length=4096,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: train_steps=20,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: micro_batch_size=16,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: batch_accumulation_per_replica=64,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: val_check_interval=-1,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: limit_val_batches=0,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: limit_test_batches=0),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: optimizer=OptimizerArgs(optimizer_factory=AdamWOptimizerArgs(adam_eps=1e-08,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: adam_beta1=0.9,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: adam_beta2=0.95,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: torch_adam_is_fused=True,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: name='adamW'),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: zero_stage=1,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: weight_decay=0.01,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: clip_grad=1.0,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: accumulate_grad_in_fp32=True,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: learning_rate_scheduler=LRSchedulerArgs(learning_rate=0.0001,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: lr_warmup_steps=1,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: lr_warmup_style='linear',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: lr_decay_style='linear',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: lr_decay_steps=19,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: lr_decay_starting_step=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: min_decay_lr=1e-05)),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: data_stages=[DatasetStageArgs(name='Training Stage',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: start_training_step=1,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: data=DataArgs(dataset=PretrainDatasetsArgs(hf_dataset_or_datasets='roneneldan/TinyStories',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: hf_dataset_splits='train',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: hf_dataset_config_name=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: dataset_processing_num_proc_per_process=64,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: dataset_overwrite_cache=False,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: text_column_name='text'),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: seed=42,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: num_loading_workers=0))],
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: profiler=ProfilerArgs(profiler_export_path=PosixPath('/fsx/ferdinandmom/ferdinand-hf/bench_cluster/results/llama-1B/64_GPUS/dp-1_tp-8_pp-8_mbz-16')),
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: lighteval=None)
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Model Config:
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: LlamaConfig(bos_token_id=1,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: eos_token_id=2,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: hidden_act='silu',
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: hidden_size=2048,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: initializer_range=0.02,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: intermediate_size=4096,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: is_llama_config=True,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: max_position_embeddings=4096,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: num_attention_heads=32,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: num_hidden_layers=24,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: num_key_value_heads=32,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: pad_token_id=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: pretraining_tp=1,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: rms_norm_eps=1e-05,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: rope_scaling=None,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: rope_theta=10000.0,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: tie_word_embeddings=True,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: use_cache=True,
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: vocab_size=50264)
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Building model..
[default0]:07/06/2024 09:43:49 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Setting PP block ranks...
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=7|ip-26-0-161-78]: Local number of parameters: 15.7M (30.02MiB)
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=7|ip-26-0-161-78]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=0|ip-26-0-165-38]: Local number of parameters: 15.7M (30.02MiB)
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=0|ip-26-0-165-38]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=0|ip-26-0-165-38]: No checkpoint path provided.
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=6|ip-26-0-161-78]: Local number of parameters: 15.7M (30.02MiB)
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=7|ip-26-0-161-78]: No checkpoint path provided.
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=6|ip-26-0-161-78]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=6|ip-26-0-161-78]: No checkpoint path provided.
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=3|ip-26-0-165-38]: Local number of parameters: 15.7M (30.02MiB)
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=3|ip-26-0-165-38]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=3|ip-26-0-165-38]: No checkpoint path provided.
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=2|ip-26-0-161-78]: Local number of parameters: 15.7M (30.02MiB)
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=2|ip-26-0-161-78]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=2|ip-26-0-161-78]: No checkpoint path provided.
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=1|ip-26-0-165-38]: Local number of parameters: 15.7M (30.02MiB)
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=1|ip-26-0-165-38]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=1|ip-26-0-165-38]: No checkpoint path provided.
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=3|ip-26-0-161-78]: Local number of parameters: 15.7M (30.02MiB)
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=3|ip-26-0-161-78]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=3|ip-26-0-161-78]: No checkpoint path provided.
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=5|ip-26-0-165-38]: Local number of parameters: 15.7M (30.02MiB)
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=0|ip-26-0-161-78]: Local number of parameters: 15.7M (30.02MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=5|ip-26-0-165-38]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=7|ip-26-0-165-38]: Local number of parameters: 15.7M (30.02MiB)
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=0|ip-26-0-161-78]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=0|ip-26-0-161-78]: No checkpoint path provided.
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=3|ip-26-0-160-103]: Local number of parameters: 33.9M (64.57MiB)
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=3|ip-26-0-160-103]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=3|ip-26-0-160-103]: No checkpoint path provided.
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=7|ip-26-0-165-38]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=4|ip-26-0-161-78]: Local number of parameters: 15.7M (30.02MiB)
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=4|ip-26-0-161-78]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=4|ip-26-0-161-78]: No checkpoint path provided.
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=2|ip-26-0-160-103]: Local number of parameters: 33.9M (64.57MiB)
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=2|ip-26-0-160-103]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=0|ip-26-0-165-24]: Local number of parameters: 21M (40.03MiB)
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=0|ip-26-0-165-24]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=0|ip-26-0-165-24]: No checkpoint path provided.
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=6|ip-26-0-165-38]: Local number of parameters: 15.7M (30.02MiB)
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=0|ip-26-0-165-213]: Local number of parameters: 15.7M (30.02MiB)
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=0|ip-26-0-165-213]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=0|ip-26-0-165-213]: No checkpoint path provided.
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=1|ip-26-0-161-78]: Local number of parameters: 15.7M (30.02MiB)
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=1|ip-26-0-161-78]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=1|ip-26-0-161-78]: No checkpoint path provided.
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=5|ip-26-0-160-103]: Local number of parameters: 33.9M (64.57MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=5|ip-26-0-160-103]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=5|ip-26-0-165-24]: Local number of parameters: 21M (40.03MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=5|ip-26-0-165-24]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=5|ip-26-0-165-24]: No checkpoint path provided.
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=6|ip-26-0-165-38]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=5|ip-26-0-165-213]: Local number of parameters: 15.7M (30.02MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=5|ip-26-0-165-213]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=5|ip-26-0-161-78]: Local number of parameters: 15.7M (30.02MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=5|ip-26-0-161-78]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=1|TP=5|ip-26-0-161-78]: No checkpoint path provided.
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=4|ip-26-0-160-103]: Local number of parameters: 33.9M (64.57MiB)
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=4|ip-26-0-160-103]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=4|ip-26-0-165-24]: Local number of parameters: 21M (40.03MiB)
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=4|ip-26-0-165-24]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=4|ip-26-0-165-24]: No checkpoint path provided.
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=4|ip-26-0-165-38]: Local number of parameters: 15.7M (30.02MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=5|ip-26-0-165-213]: No checkpoint path provided.
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=4|ip-26-0-160-103]: No checkpoint path provided.
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=5|ip-26-0-160-103]: No checkpoint path provided.
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=4|ip-26-0-165-38]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=4|ip-26-0-165-38]: No checkpoint path provided.
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=4|ip-26-0-165-213]: Local number of parameters: 15.7M (30.02MiB)
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=4|ip-26-0-165-213]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=4|ip-26-0-165-213]: No checkpoint path provided.
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=6|ip-26-0-160-103]: Local number of parameters: 33.9M (64.57MiB)
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=6|ip-26-0-160-103]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=6|ip-26-0-160-103]: No checkpoint path provided.
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=6|ip-26-0-165-38]: No checkpoint path provided.
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=3|ip-26-0-165-213]: Local number of parameters: 15.7M (30.02MiB)
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=2|ip-26-0-160-103]: No checkpoint path provided.
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=2|ip-26-0-165-38]: Local number of parameters: 15.7M (30.02MiB)
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=3|ip-26-0-165-213]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=3|ip-26-0-165-213]: No checkpoint path provided.
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=1|ip-26-0-160-103]: Local number of parameters: 33.9M (64.57MiB)
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=1|ip-26-0-160-103]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=3|ip-26-0-165-24]: Local number of parameters: 21M (40.03MiB)
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=3|ip-26-0-165-24]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=3|ip-26-0-165-24]: No checkpoint path provided.
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=2|ip-26-0-165-38]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=7|ip-26-0-165-38]: No checkpoint path provided.
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=7|ip-26-0-165-213]: Local number of parameters: 15.7M (30.02MiB)
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=1|ip-26-0-160-103]: No checkpoint path provided.
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=5|ip-26-0-165-38]: No checkpoint path provided.
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=7|ip-26-0-165-213]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=7|ip-26-0-165-213]: No checkpoint path provided.
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=7|ip-26-0-160-103]: Local number of parameters: 33.9M (64.57MiB)
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=7|ip-26-0-160-103]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=4|TP=2|ip-26-0-165-38]: No checkpoint path provided.
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=6|ip-26-0-165-213]: Local number of parameters: 15.7M (30.02MiB)
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=7|ip-26-0-160-103]: No checkpoint path provided.
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=1|ip-26-0-165-24]: Local number of parameters: 21M (40.03MiB)
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=6|ip-26-0-165-213]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=6|ip-26-0-165-213]: No checkpoint path provided.
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=1|ip-26-0-165-24]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=2|ip-26-0-165-213]: Local number of parameters: 15.7M (30.02MiB)
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=6|ip-26-0-171-230]: Local number of parameters: 21M (40.03MiB)
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=6|ip-26-0-171-230]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=1|ip-26-0-165-24]: No checkpoint path provided.
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=2|ip-26-0-165-213]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=2|ip-26-0-165-213]: No checkpoint path provided.
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=6|ip-26-0-171-230]: No checkpoint path provided.
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=3|ip-26-0-171-230]: Local number of parameters: 21M (40.03MiB)
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=3|ip-26-0-171-230]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=6|ip-26-0-165-24]: Local number of parameters: 21M (40.03MiB)
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=1|ip-26-0-165-213]: Local number of parameters: 15.7M (30.02MiB)
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=1|ip-26-0-165-213]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=0|ip-26-0-171-230]: Local number of parameters: 21M (40.03MiB)
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=6|ip-26-0-165-24]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=6|ip-26-0-165-24]: No checkpoint path provided.
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=2|TP=1|ip-26-0-165-213]: No checkpoint path provided.
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=0|ip-26-0-171-230]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=3|ip-26-0-171-230]: No checkpoint path provided.
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=2|ip-26-0-165-24]: Local number of parameters: 21M (40.03MiB)
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=2|ip-26-0-165-24]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=0|ip-26-0-171-230]: No checkpoint path provided.
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Total number of parameters: 1.21G (2314.22MiB)
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Local number of parameters: 33.9M (64.57MiB)
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: No checkpoint path provided.
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Parametrizing model parameters using StandardParametrizator
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=2|ip-26-0-165-24]: No checkpoint path provided.
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=7|ip-26-0-171-230]: Local number of parameters: 21M (40.03MiB)
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=7|ip-26-0-171-230]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=7|ip-26-0-171-230]: No checkpoint path provided.
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=7|ip-26-0-165-24]: Local number of parameters: 21M (40.03MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=5|ip-26-0-171-249]: Local number of parameters: 12.9M (24.55MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=5|ip-26-0-171-249]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=5|ip-26-0-171-249]: No checkpoint path provided.
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=6|ip-26-0-171-249]: Local number of parameters: 12.9M (24.55MiB)
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=6|ip-26-0-171-249]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=6|ip-26-0-171-249]: No checkpoint path provided.
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=4|ip-26-0-171-230]: Local number of parameters: 21M (40.03MiB)
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=4|ip-26-0-171-230]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=4|ip-26-0-171-230]: No checkpoint path provided.
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=7|ip-26-0-165-24]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=3|TP=7|ip-26-0-165-24]: No checkpoint path provided.
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=3|ip-26-0-171-249]: Local number of parameters: 12.9M (24.55MiB)
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=3|ip-26-0-171-249]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=1|ip-26-0-171-230]: Local number of parameters: 21M (40.03MiB)
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=1|ip-26-0-171-230]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=3|ip-26-0-171-249]: No checkpoint path provided.
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=6|ip-26-0-166-15]: Local number of parameters: 15.7M (30.02MiB)
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=6|ip-26-0-166-15]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default6]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=6|ip-26-0-166-15]: No checkpoint path provided.
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=1|ip-26-0-171-230]: No checkpoint path provided.
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=2|ip-26-0-171-249]: Local number of parameters: 12.9M (24.55MiB)
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=2|ip-26-0-171-249]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=2|ip-26-0-171-249]: No checkpoint path provided.
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=3|ip-26-0-166-15]: Local number of parameters: 15.7M (30.02MiB)
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=3|ip-26-0-166-15]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=2|ip-26-0-166-15]: Local number of parameters: 15.7M (30.02MiB)
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=2|ip-26-0-171-230]: Local number of parameters: 21M (40.03MiB)
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=0|ip-26-0-171-249]: Local number of parameters: 12.9M (24.55MiB)
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=2|ip-26-0-166-15]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=1|ip-26-0-166-15]: Local number of parameters: 15.7M (30.02MiB)
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=1|ip-26-0-166-15]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=2|ip-26-0-171-230]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=2|ip-26-0-171-230]: No checkpoint path provided.
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=0|ip-26-0-171-249]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB
[default3]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=3|ip-26-0-166-15]: No checkpoint path provided.
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=1|ip-26-0-166-15]: No checkpoint path provided.
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=5|ip-26-0-171-230]: Local number of parameters: 21M (40.03MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=5|ip-26-0-171-230]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=6|TP=5|ip-26-0-171-230]: No checkpoint path provided.
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=1|ip-26-0-171-249]: Local number of parameters: 12.9M (24.55MiB)
[default2]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=2|ip-26-0-166-15]: No checkpoint path provided.
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=1|ip-26-0-171-249]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=0|ip-26-0-171-249]: No checkpoint path provided.
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=4|ip-26-0-166-15]: Local number of parameters: 15.7M (30.02MiB)
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=4|ip-26-0-166-15]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=4|ip-26-0-166-15]: No checkpoint path provided.
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=7|ip-26-0-171-249]: Local number of parameters: 12.9M (24.55MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=5|ip-26-0-166-15]: Local number of parameters: 15.7M (30.02MiB)
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=5|ip-26-0-166-15]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default5]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=5|ip-26-0-166-15]: No checkpoint path provided.
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=7|ip-26-0-171-249]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=7|ip-26-0-171-249]: No checkpoint path provided.
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=0|ip-26-0-166-15]: Local number of parameters: 15.7M (30.02MiB)
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=0|ip-26-0-166-15]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default1]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=1|ip-26-0-171-249]: No checkpoint path provided.
[default0]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=0|ip-26-0-166-15]: No checkpoint path provided.
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=4|ip-26-0-171-249]: Local number of parameters: 12.9M (24.55MiB)
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=4|ip-26-0-171-249]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB
[default4]:07/06/2024 09:44:08 [INFO|DP=0|PP=7|TP=4|ip-26-0-171-249]: No checkpoint path provided.
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=7|ip-26-0-166-15]: Local number of parameters: 15.7M (30.02MiB)
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=7|ip-26-0-166-15]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB
[default7]:07/06/2024 09:44:08 [INFO|DP=0|PP=5|TP=7|ip-26-0-166-15]: No checkpoint path provided.
[default0]:07/06/2024 09:44:09 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: [Optimizer Building] Using LearningRateForSP as learning rate
[default0]:07/06/2024 09:44:09 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: [ZeRO sharding] Size of optimizer params per rank:
[default0]:07/06/2024 09:44:09 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: [ZeRO sharding] DP Rank 0 has 33.9M out of 33.9M (100.00%) params' optimizer states
[default0]:07/06/2024 09:44:10 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: [Training Plan] Stage Training Stage has 19 remaining training steps and has consumed 0 samples
[default0]:07/06/2024 09:44:10 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Using `datasets` library
[default0]:07/06/2024 09:44:10 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Loading tokenizer from openai-community/gpt2 and transformers/hf_hub versions ('4.41.2', '0.23.4')
[default0]:Repo card metadata block was not found. Setting CardData to empty.
[default0]:07/06/2024 09:44:11 [WARNING|DP=0|PP=0|TP=0|ip-26-0-160-103]: Repo card metadata block was not found. Setting CardData to empty.
[default0]:07/06/2024 09:44:11 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: [Training Plan] There are 1 training stages
[default0]:07/06/2024 09:44:11 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: [Stage Training Stage] start from step 1
[default0]:07/06/2024 09:44:11 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]:
[default0]:07/06/2024 09:44:11 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: [Start training] datetime: 2024-07-06 09:44:11.968074 | mbs: 16 | grad_accum: 64 | global_batch_size: 1024 | sequence_length: 4096 | train_steps: 20 | start_iteration_step: 0 | consumed_train_samples: 0
[default6]:Repo card metadata block was not found. Setting CardData to empty.
[default4]:Repo card metadata block was not found. Setting CardData to empty.
[default0]:Repo card metadata block was not found. Setting CardData to empty.
[default7]:07/06/2024 09:44:12 [WARNING|DP=0|PP=1|TP=7|ip-26-0-161-78]: Repo card metadata block was not found. Setting CardData to empty.
[default5]:Repo card metadata block was not found. Setting CardData to empty.
[default3]:07/06/2024 09:44:12 [WARNING|DP=0|PP=1|TP=3|ip-26-0-161-78]: Repo card metadata block was not found. Setting CardData to empty.
[default3]:Repo card metadata block was not found. Setting CardData to empty.
[default0]:07/06/2024 09:44:12 [WARNING|DP=0|PP=4|TP=0|ip-26-0-165-38]: Repo card metadata block was not found. Setting CardData to empty.
[default1]:07/06/2024 09:44:12 [WARNING|DP=0|PP=4|TP=1|ip-26-0-165-38]: Repo card metadata block was not found. Setting CardData to empty.
[default1]:Repo card metadata block was not found. Setting CardData to empty.
[default5]:07/06/2024 09:44:12 [WARNING|DP=0|PP=4|TP=5|ip-26-0-165-38]: Repo card metadata block was not found. Setting CardData to empty.
[default4]:07/06/2024 09:44:12 [WARNING|DP=0|PP=4|TP=4|ip-26-0-165-38]: Repo card metadata block was not found. Setting CardData to empty.
[default6]:07/06/2024 09:44:12 [WARNING|DP=0|PP=4|TP=6|ip-26-0-165-38]: Repo card metadata block was not found. Setting CardData to empty.
[default4]:07/06/2024 09:44:12 [WARNING|DP=0|PP=1|TP=4|ip-26-0-161-78]: Repo card metadata block was not found. Setting CardData to empty.
[default0]:07/06/2024 09:44:12 [WARNING|DP=0|PP=1|TP=0|ip-26-0-161-78]: Repo card metadata block was not found. Setting CardData to empty.
[default0]:Repo card metadata block was not found. Setting CardData to empty.
[default4]:Repo card metadata block was not found. Setting CardData to empty.
[default6]:07/06/2024 09:44:12 [WARNING|DP=0|PP=0|TP=6|ip-26-0-160-103]: Repo card metadata block was not found. Setting CardData to empty.
[default2]:07/06/2024 09:44:12 [WARNING|DP=0|PP=0|TP=2|ip-26-0-160-103]: Repo card metadata block was not found. Setting CardData to empty.
[default2]:Repo card metadata block was not found. Setting CardData to empty.
[default7]:Repo card metadata block was not found. Setting CardData to empty.
[default0]:07/06/2024 09:44:12 [WARNING|DP=0|PP=2|TP=0|ip-26-0-165-213]: Repo card metadata block was not found. Setting CardData to empty.
[default1]:Repo card metadata block was not found. Setting CardData to empty.
[default5]:07/06/2024 09:44:12 [WARNING|DP=0|PP=0|TP=5|ip-26-0-160-103]: Repo card metadata block was not found. Setting CardData to empty.
[default2]:Repo card metadata block was not found. Setting CardData to empty.
[default5]:07/06/2024 09:44:12 [WARNING|DP=0|PP=2|TP=5|ip-26-0-165-213]: Repo card metadata block was not found. Setting CardData to empty.
[default6]:Repo card metadata block was not found. Setting CardData to empty.
[default4]:07/06/2024 09:44:12 [WARNING|DP=0|PP=0|TP=4|ip-26-0-160-103]: Repo card metadata block was not found. Setting CardData to empty.
[default5]:Repo card metadata block was not found. Setting CardData to empty.
[default4]:07/06/2024 09:44:12 [WARNING|DP=0|PP=3|TP=4|ip-26-0-165-24]: Repo card metadata block was not found. Setting CardData to empty.
[default2]:Repo card metadata block was not found. Setting CardData to empty.
[default1]:07/06/2024 09:44:12 [WARNING|DP=0|PP=0|TP=1|ip-26-0-160-103]: Repo card metadata block was not found. Setting CardData to empty.
[default7]:Repo card metadata block was not found. Setting CardData to empty.
[default3]:07/06/2024 09:44:12 [WARNING|DP=0|PP=3|TP=3|ip-26-0-165-24]: Repo card metadata block was not found. Setting CardData to empty.
[default7]:Repo card metadata block was not found. Setting CardData to empty.
[default7]:07/06/2024 09:44:12 [WARNING|DP=0|PP=0|TP=7|ip-26-0-160-103]: Repo card metadata block was not found. Setting CardData to empty.
[default1]:Repo card metadata block was not found. Setting CardData to empty.
[default6]:07/06/2024 09:44:12 [WARNING|DP=0|PP=2|TP=6|ip-26-0-165-213]: Repo card metadata block was not found. Setting CardData to empty.
[default6]:Repo card metadata block was not found. Setting CardData to empty.
[default7]:07/06/2024 09:44:12 [WARNING|DP=0|PP=2|TP=7|ip-26-0-165-213]: Repo card metadata block was not found. Setting CardData to empty.
[default0]:Repo card metadata block was not found. Setting CardData to empty.
[default1]:07/06/2024 09:44:12 [WARNING|DP=0|PP=2|TP=1|ip-26-0-165-213]: Repo card metadata block was not found. Setting CardData to empty.
[default3]:Repo card metadata block was not found. Setting CardData to empty.
[default2]:07/06/2024 09:44:12 [WARNING|DP=0|PP=3|TP=2|ip-26-0-165-24]: Repo card metadata block was not found. Setting CardData to empty.
[default4]:Repo card metadata block was not found. Setting CardData to empty.
[default2]:07/06/2024 09:44:12 [WARNING|DP=0|PP=2|TP=2|ip-26-0-165-213]: Repo card metadata block was not found. Setting CardData to empty.
[default7]:Repo card metadata block was not found. Setting CardData to empty.
[default6]:07/06/2024 09:44:12 [WARNING|DP=0|PP=3|TP=6|ip-26-0-165-24]: Repo card metadata block was not found. Setting CardData to empty.
[default7]:07/06/2024 09:44:12 [WARNING|DP=0|PP=3|TP=7|ip-26-0-165-24]: Repo card metadata block was not found. Setting CardData to empty.
[default0]:07/06/2024 09:44:12 [WARNING|DP=0|PP=6|TP=0|ip-26-0-171-230]: Repo card metadata block was not found. Setting CardData to empty.
[default6]:Repo card metadata block was not found. Setting CardData to empty.
[default6]:07/06/2024 09:44:12 [WARNING|DP=0|PP=6|TP=6|ip-26-0-171-230]: Repo card metadata block was not found. Setting CardData to empty.
[default0]:Repo card metadata block was not found. Setting CardData to empty.
[default6]:Repo card metadata block was not found. Setting CardData to empty.
[default3]:07/06/2024 09:44:12 [WARNING|DP=0|PP=6|TP=3|ip-26-0-171-230]: Repo card metadata block was not found. Setting CardData to empty.
[default0]:Repo card metadata block was not found. Setting CardData to empty.
[default7]:07/06/2024 09:44:12 [WARNING|DP=0|PP=6|TP=7|ip-26-0-171-230]: Repo card metadata block was not found. Setting CardData to empty.
[default7]:Repo card metadata block was not found. Setting CardData to empty.
[default1]:07/06/2024 09:44:12 [WARNING|DP=0|PP=6|TP=1|ip-26-0-171-230]: Repo card metadata block was not found. Setting CardData to empty.
[default5]:Repo card metadata block was not found. Setting CardData to empty.
[default4]:07/06/2024 09:44:12 [WARNING|DP=0|PP=6|TP=4|ip-26-0-171-230]: Repo card metadata block was not found. Setting CardData to empty.
[default2]:Repo card metadata block was not found. Setting CardData to empty.
[default6]:07/06/2024 09:44:12 [WARNING|DP=0|PP=7|TP=6|ip-26-0-171-249]: Repo card metadata block was not found. Setting CardData to empty.
[default1]:Repo card metadata block was not found. Setting CardData to empty.
[default3]:07/06/2024 09:44:12 [WARNING|DP=0|PP=5|TP=3|ip-26-0-166-15]: Repo card metadata block was not found. Setting CardData to empty.
[default6]:07/06/2024 09:44:12 [WARNING|DP=0|PP=5|TP=6|ip-26-0-166-15]: Repo card metadata block was not found. Setting CardData to empty.
[default4]:Repo card metadata block was not found. Setting CardData to empty.
[default5]:07/06/2024 09:44:12 [WARNING|DP=0|PP=6|TP=5|ip-26-0-171-230]: Repo card metadata block was not found. Setting CardData to empty.
[default2]:Repo card metadata block was not found. Setting CardData to empty.
[default2]:07/06/2024 09:44:12 [WARNING|DP=0|PP=7|TP=2|ip-26-0-171-249]: Repo card metadata block was not found. Setting CardData to empty.
[default6]:Repo card metadata block was not found. Setting CardData to empty.
[default2]:07/06/2024 09:44:12 [WARNING|DP=0|PP=6|TP=2|ip-26-0-171-230]: Repo card metadata block was not found. Setting CardData to empty.
[default4]:Repo card metadata block was not found. Setting CardData to empty.
[default4]:07/06/2024 09:44:12 [WARNING|DP=0|PP=7|TP=4|ip-26-0-171-249]: Repo card metadata block was not found. Setting CardData to empty.
[default7]:Repo card metadata block was not found. Setting CardData to empty.
[default1]:07/06/2024 09:44:12 [WARNING|DP=0|PP=7|TP=1|ip-26-0-171-249]: Repo card metadata block was not found. Setting CardData to empty.
[default3]:Repo card metadata block was not found. Setting CardData to empty.
[default7]:07/06/2024 09:44:12 [WARNING|DP=0|PP=7|TP=7|ip-26-0-171-249]: Repo card metadata block was not found. Setting CardData to empty.
[default3]:Repo card metadata block was not found. Setting CardData to empty.
[default0]:07/06/2024 09:44:12 [WARNING|DP=0|PP=7|TP=0|ip-26-0-171-249]: Repo card metadata block was not found. Setting CardData to empty.
[default0]:Repo card metadata block was not found. Setting CardData to empty.
[default5]:07/06/2024 09:44:12 [WARNING|DP=0|PP=5|TP=5|ip-26-0-166-15]: Repo card metadata block was not found. Setting CardData to empty.
[default7]:Repo card metadata block was not found. Setting CardData to empty.
[default0]:07/06/2024 09:44:12 [WARNING|DP=0|PP=5|TP=0|ip-26-0-166-15]: Repo card metadata block was not found. Setting CardData to empty.
[default1]:Repo card metadata block was not found. Setting CardData to empty.
[default7]:07/06/2024 09:44:12 [WARNING|DP=0|PP=5|TP=7|ip-26-0-166-15]: Repo card metadata block was not found. Setting CardData to empty.
[default5]:Repo card metadata block was not found. Setting CardData to empty.
[default5]:Repo card metadata block was not found. Setting CardData to empty.
[default4]:Repo card metadata block was not found. Setting CardData to empty.
[default6]:Repo card metadata block was not found. Setting CardData to empty.
[default6]:07/06/2024 09:44:12 [WARNING|DP=0|PP=1|TP=6|ip-26-0-161-78]: Repo card metadata block was not found. Setting CardData to empty.
[default2]:07/06/2024 09:44:12 [WARNING|DP=0|PP=1|TP=2|ip-26-0-161-78]: Repo card metadata block was not found. Setting CardData to empty.
[default5]:Repo card metadata block was not found. Setting CardData to empty.
[default2]:Repo card metadata block was not found. Setting CardData to empty.
[default3]:07/06/2024 09:44:12 [WARNING|DP=0|PP=4|TP=3|ip-26-0-165-38]: Repo card metadata block was not found. Setting CardData to empty.
[default7]:07/06/2024 09:44:12 [WARNING|DP=0|PP=4|TP=7|ip-26-0-165-38]: Repo card metadata block was not found. Setting CardData to empty.
[default1]:07/06/2024 09:44:12 [WARNING|DP=0|PP=1|TP=1|ip-26-0-161-78]: Repo card metadata block was not found. Setting CardData to empty.
[default5]:07/06/2024 09:44:12 [WARNING|DP=0|PP=1|TP=5|ip-26-0-161-78]: Repo card metadata block was not found. Setting CardData to empty.
[default3]:Repo card metadata block was not found. Setting CardData to empty.
[default3]:07/06/2024 09:44:12 [WARNING|DP=0|PP=0|TP=3|ip-26-0-160-103]: Repo card metadata block was not found. Setting CardData to empty.
[default1]:Repo card metadata block was not found. Setting CardData to empty.
[default5]:07/06/2024 09:44:12 [WARNING|DP=0|PP=3|TP=5|ip-26-0-165-24]: Repo card metadata block was not found. Setting CardData to empty.
[default4]:07/06/2024 09:44:12 [WARNING|DP=0|PP=2|TP=4|ip-26-0-165-213]: Repo card metadata block was not found. Setting CardData to empty.
[default3]:Repo card metadata block was not found. Setting CardData to empty.
[default3]:07/06/2024 09:44:12 [WARNING|DP=0|PP=2|TP=3|ip-26-0-165-213]: Repo card metadata block was not found. Setting CardData to empty.
[default7]:Repo card metadata block was not found. Setting CardData to empty.
[default6]:Repo card metadata block was not found. Setting CardData to empty.
[default3]:Repo card metadata block was not found. Setting CardData to empty.
[default1]:Repo card metadata block was not found. Setting CardData to empty.
[default1]:07/06/2024 09:44:12 [WARNING|DP=0|PP=3|TP=1|ip-26-0-165-24]: Repo card metadata block was not found. Setting CardData to empty.
[default5]:Repo card metadata block was not found. Setting CardData to empty.
[default4]:Repo card metadata block was not found. Setting CardData to empty.
[default5]:07/06/2024 09:44:12 [WARNING|DP=0|PP=7|TP=5|ip-26-0-171-249]: Repo card metadata block was not found. Setting CardData to empty.
[default3]:07/06/2024 09:44:12 [WARNING|DP=0|PP=7|TP=3|ip-26-0-171-249]: Repo card metadata block was not found. Setting CardData to empty.
[default1]:07/06/2024 09:44:12 [WARNING|DP=0|PP=5|TP=1|ip-26-0-166-15]: Repo card metadata block was not found. Setting CardData to empty.
[default1]:Repo card metadata block was not found. Setting CardData to empty.
[default4]:Repo card metadata block was not found. Setting CardData to empty.
[default4]:07/06/2024 09:44:12 [WARNING|DP=0|PP=5|TP=4|ip-26-0-166-15]: Repo card metadata block was not found. Setting CardData to empty.
[default3]:Repo card metadata block was not found. Setting CardData to empty.
[default5]:Repo card metadata block was not found. Setting CardData to empty.
[default2]:07/06/2024 09:44:12 [WARNING|DP=0|PP=4|TP=2|ip-26-0-165-38]: Repo card metadata block was not found. Setting CardData to empty.
[default2]:Repo card metadata block was not found. Setting CardData to empty.
[default0]:07/06/2024 09:44:17 [WARNING|DP=0|PP=3|TP=0|ip-26-0-165-24]: Repo card metadata block was not found. Setting CardData to empty.
[default0]:Repo card metadata block was not found. Setting CardData to empty.
[default2]:07/06/2024 09:44:17 [WARNING|DP=0|PP=5|TP=2|ip-26-0-166-15]: Repo card metadata block was not found. Setting CardData to empty.
[default2]:Repo card metadata block was not found. Setting CardData to empty.
[default0]:07/06/2024 09:44:18 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Resuming training from stage Training Stage, it has trained for 0 samples and has 19 remaining train steps
[default0]:07/06/2024 09:44:18 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-103]: Memory usage: 328.58MiB. Peak allocated 328.59MiB. Peak reserved: 338.00MiB
[default2]:Traceback (most recent call last):
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in <module>
[default1]:Traceback (most recent call last):
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in <module>
[default5]:Traceback (most recent call last):
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in <module>
[default5]: trainer.train(dataloader)
[default1]: trainer.train(dataloader)
[default7]:Traceback (most recent call last):
[default2]: trainer.train(dataloader)
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in <module>
[default3]:Traceback (most recent call last):
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in <module>
[default3]: trainer.train(dataloader)
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train
[default7]: trainer.train(dataloader)
[default6]:Traceback (most recent call last):
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in <module>
[default5]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader)
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train
[default3]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader)
[default1]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader)
[default7]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader)
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step
[default6]: trainer.train(dataloader)
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step
[default1]: outputs = self.pipeline_engine.train_batch_iter(
[default7]: outputs = self.pipeline_engine.train_batch_iter(
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter
[default5]: outputs = self.pipeline_engine.train_batch_iter(
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train
[default7]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model)
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter
[default1]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model)
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step
[default7]: output = model(**micro_batch)
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward
[default2]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader)
[default6]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader)
[default3]: outputs = self.pipeline_engine.train_batch_iter(
[default1]: output = model(**micro_batch)
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter
[default1]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step
[default7]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default7]: return self._call_impl(*args, **kwargs)
[default5]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model)
[default3]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model)
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward
[default7]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default7]: return forward_call(*args, **kwargs)
[default1]: return self._call_impl(*args, **kwargs)
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward
[default2]: outputs = self.pipeline_engine.train_batch_iter(
[default1]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default1]: return forward_call(*args, **kwargs)
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step
[default6]: outputs = self.pipeline_engine.train_batch_iter(
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward
[default7]: sharded_logits = self.model(
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter
[default6]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model)
[default5]: output = model(**micro_batch)
[default1]: sharded_logits = self.model(
[default1]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default3]: output = model(**micro_batch)
[default7]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default1]: return self._call_impl(*args, **kwargs)
[default2]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model)
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward
[default5]: return self._call_impl(*args, **kwargs)
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default6]: output = model(**micro_batch)
[default1]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default5]: return forward_call(*args, **kwargs)
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default7]: return self._call_impl(*args, **kwargs)
[default7]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default2]: output = model(**micro_batch)
[default1]: return forward_call(*args, **kwargs)
[default3]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward
[default5]: sharded_logits = self.model(
[default6]: return self._call_impl(*args, **kwargs)
[default3]: return self._call_impl(*args, **kwargs)
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default7]: return forward_call(*args, **kwargs)
[default3]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward
[default3]: return forward_call(*args, **kwargs)
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward
[default5]: return self._call_impl(*args, **kwargs)
[default3]: sharded_logits = self.model(
[default3]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default6]: return forward_call(*args, **kwargs)
[default2]: return self._call_impl(*args, **kwargs)
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default1]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0]
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward
[default6]: sharded_logits = self.model(
[default7]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0]
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states
[default3]: return self._call_impl(*args, **kwargs)
[default5]: return forward_call(*args, **kwargs)
[default7]: hidden_encoder_states = encoder_block(**hidden_encoder_states)
[default7]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default3]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default3]: return forward_call(*args, **kwargs)
[default2]: return forward_call(*args, **kwargs)
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward
[default7]: return self._call_impl(*args, **kwargs)
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward
[default2]: sharded_logits = self.model(
[default7]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default5]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0]
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default6]: return self._call_impl(*args, **kwargs)
[default7]: return forward_call(*args, **kwargs)
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states
[default1]: hidden_encoder_states = encoder_block(**hidden_encoder_states)
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default2]: return self._call_impl(*args, **kwargs)
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward
[default7]: output = self.pp_block(**new_kwargs)
[default7]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default1]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default1]: return self._call_impl(*args, **kwargs)
[default6]: return forward_call(*args, **kwargs)
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default5]: hidden_encoder_states = encoder_block(**hidden_encoder_states)
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default7]: return self._call_impl(*args, **kwargs)
[default1]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward
[default7]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default1]: return forward_call(*args, **kwargs)
[default3]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0]
[default2]: return forward_call(*args, **kwargs)
[default7]: return forward_call(*args, **kwargs)
[default6]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0]
[default5]: return self._call_impl(*args, **kwargs)
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 630, in forward
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states
[default6]: hidden_encoder_states = encoder_block(**hidden_encoder_states)
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states
[default3]: hidden_encoder_states = encoder_block(**hidden_encoder_states)
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default7]: output = self.attn(hidden_states=hidden_states, sequence_mask=sequence_mask)
[default7]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default2]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0]
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward
[default1]: output = self.pp_block(**new_kwargs)
[default1]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default7]: return self._call_impl(*args, **kwargs)
[default3]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default3]: return self._call_impl(*args, **kwargs)
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default5]: return forward_call(*args, **kwargs)
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward
[default3]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default7]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default7]: return forward_call(*args, **kwargs)
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states
[default2]: hidden_encoder_states = encoder_block(**hidden_encoder_states)
[default3]: return forward_call(*args, **kwargs)
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward
[default5]: output = self.pp_block(**new_kwargs)
[default7]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 562, in forward
[default6]: return self._call_impl(*args, **kwargs)
[default3]: output = self.pp_block(**new_kwargs)
[default3]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default7]: key_value_states = torch.cat([key_states.unsqueeze(0), value_states.unsqueeze(0)], dim=0)
[default7]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 64.00 MiB. GPU 7 has a total capacity of 79.33 GiB of which 23.94 MiB is free. Including non-PyTorch memory, this process has 79.29 GiB memory in use. Of the allocated memory 70.03 GiB is allocated by PyTorch, and 37.91 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default1]: return self._call_impl(*args, **kwargs)
[default5]: return self._call_impl(*args, **kwargs)
[default3]: return self._call_impl(*args, **kwargs)
[default3]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default6]: return forward_call(*args, **kwargs)
[default1]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default3]: return forward_call(*args, **kwargs)
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 630, in forward
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default2]: return self._call_impl(*args, **kwargs)
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default3]: output = self.attn(hidden_states=hidden_states, sequence_mask=sequence_mask)
[default1]: return forward_call(*args, **kwargs)
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward
[default2]: return forward_call(*args, **kwargs)
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward
[default3]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default5]: return forward_call(*args, **kwargs)
[default6]: output = self.pp_block(**new_kwargs)
[default2]: output = self.pp_block(**new_kwargs)
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default3]: return self._call_impl(*args, **kwargs)
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default6]: return self._call_impl(*args, **kwargs)
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 630, in forward
[default1]: output = self.attn(hidden_states=hidden_states, sequence_mask=sequence_mask)
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 636, in forward
[default5]: hidden_states = self.mlp(hidden_states=hidden_states)["hidden_states"]
[default3]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default3]: return forward_call(*args, **kwargs)
[default1]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default3]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 564, in forward
[default5]: return self._call_impl(*args, **kwargs)
[default6]: return forward_call(*args, **kwargs)
[default1]: return self._call_impl(*args, **kwargs)
[default1]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default2]: return self._call_impl(*args, **kwargs)
[default3]: key_value_states = key_value_states.permute(1, 2, 0, 3, 4).contiguous()
[default1]: return forward_call(*args, **kwargs)
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 636, in forward
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default5]: return forward_call(*args, **kwargs)
[default1]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 562, in forward
[default6]: hidden_states = self.mlp(hidden_states=hidden_states)["hidden_states"]
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default2]: return forward_call(*args, **kwargs)
[default1]: key_value_states = torch.cat([key_states.unsqueeze(0), value_states.unsqueeze(0)], dim=0)
[default1]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 64.00 MiB. GPU 1 has a total capacity of 79.33 GiB of which 5.94 MiB is free. Including non-PyTorch memory, this process has 79.31 GiB memory in use. Of the allocated memory 70.03 GiB is allocated by PyTorch, and 37.91 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 171, in forward
[default5]: hidden_states = self.down_proj(self.split_silu_mul(merged_states))
[default6]: return self._call_impl(*args, **kwargs)
[default3]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 64.00 MiB. GPU 3 has a total capacity of 79.33 GiB of which 57.94 MiB is free. Including non-PyTorch memory, this process has 79.26 GiB memory in use. Of the allocated memory 70.09 GiB is allocated by PyTorch, and 37.91 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default5]: return self._call_impl(*args, **kwargs)
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 630, in forward
[default6]: return forward_call(*args, **kwargs)
[default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 171, in forward
[default5]: return forward_call(*args, **kwargs)
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/nn.py", line 159, in forward
[default2]: output = self.attn(hidden_states=hidden_states, sequence_mask=sequence_mask)
[default6]: hidden_states = self.down_proj(self.split_silu_mul(merged_states))
[default5]: return row_linear(
[default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/functional.py", line 474, in row_linear
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default0]:STAGE:2024-07-06 09:44:36 226671:226671 ActivityProfilerController.cpp:314] Completed Stage: Warm Up
[default5]: out = F.linear(input, weight, bias)
[default6]: return self._call_impl(*args, **kwargs)
[default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default6]: return forward_call(*args, **kwargs)
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/nn.py", line 159, in forward
[default6]: return row_linear(
[default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/functional.py", line 474, in row_linear
[default6]: out = F.linear(input, weight, bias)
[default6]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 MiB. GPU 6 has a total capacity of 79.33 GiB of which 215.94 MiB is free. Including non-PyTorch memory, this process has 79.10 GiB memory in use. Of the allocated memory 69.59 GiB is allocated by PyTorch, and 5.94 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
[default5]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 MiB. GPU 5 has a total capacity of 79.33 GiB of which 215.94 MiB is free. Including non-PyTorch memory, this process has 79.10 GiB memory in use. Of the allocated memory 69.59 GiB is allocated by PyTorch, and 5.94 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default2]: return self._call_impl(*args, **kwargs)
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default2]: return forward_call(*args, **kwargs)
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 359, in forward
[default2]: qkv_states = self.qkv_proj(
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default2]: return self._call_impl(*args, **kwargs)
[default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default2]: return forward_call(*args, **kwargs)
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/nn.py", line 87, in forward
[default2]: return column_linear(
[default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/functional.py", line 359, in column_linear
[default2]: return F.linear(input, weight, bias)
[default2]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 96.00 MiB. GPU 2 has a total capacity of 79.33 GiB of which 87.94 MiB is free. Including non-PyTorch memory, this process has 79.23 GiB memory in use. Of the allocated memory 69.84 GiB is allocated by PyTorch, and 37.91 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
[default4]:Traceback (most recent call last):
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in <module>
[default4]: trainer.train(dataloader)
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train
[default4]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader)
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step
[default4]: outputs = self.pipeline_engine.train_batch_iter(
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter
[default4]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model)
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward
[default4]: output = model(**micro_batch)
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default4]: return self._call_impl(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default4]: return forward_call(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward
[default4]: sharded_logits = self.model(
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default4]: return self._call_impl(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default4]: return forward_call(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward
[default4]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0]
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states
[default4]: hidden_encoder_states = encoder_block(**hidden_encoder_states)
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default4]: return self._call_impl(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default4]: return forward_call(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward
[default4]: output = self.pp_block(**new_kwargs)
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default4]: return self._call_impl(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default4]: return forward_call(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 636, in forward
[default4]: hidden_states = self.mlp(hidden_states=hidden_states)["hidden_states"]
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default4]: return self._call_impl(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default4]: return forward_call(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 171, in forward
[default4]: hidden_states = self.down_proj(self.split_silu_mul(merged_states))
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl
[default4]: return self._call_impl(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl
[default4]: return forward_call(*args, **kwargs)
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/nn.py", line 159, in forward
[default4]: return row_linear(
[default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/functional.py", line 474, in row_linear
[default4]: out = F.linear(input, weight, bias)
[default4]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 MiB. GPU 4 has a total capacity of 79.33 GiB of which 215.94 MiB is free. Including non-PyTorch memory, this process has 79.10 GiB memory in use. Of the allocated memory 69.59 GiB is allocated by PyTorch, and 5.94 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
[2024-07-06 09:44:41,892] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 226671 closing signal SIGTERM
[2024-07-06 09:44:43,024] torch.distributed.elastic.multiprocessing.api: [ERROR] failed (exitcode: 1) local_rank: 1 (pid: 226672) of binary: /fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/python3.10
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in <module>
sys.exit(main())
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper
return f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main
run(args)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run
elastic_launch(
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__
return launch_agent(self._config, self._entrypoint, list(args))
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 268, in launch_agent
raise ChildFailedError(
torch.distributed.elastic.multiprocessing.errors.ChildFailedError:
============================================================
/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py FAILED
------------------------------------------------------------
Failures:
[1]:
time : 2024-07-06_09:44:41
host : ip-26-0-160-103.ec2.internal
rank : 2 (local_rank: 2)
exitcode : 1 (pid: 226673)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[2]:
time : 2024-07-06_09:44:41
host : ip-26-0-160-103.ec2.internal
rank : 3 (local_rank: 3)
exitcode : 1 (pid: 226674)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[3]:
time : 2024-07-06_09:44:41
host : ip-26-0-160-103.ec2.internal
rank : 4 (local_rank: 4)
exitcode : 1 (pid: 226675)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[4]:
time : 2024-07-06_09:44:41
host : ip-26-0-160-103.ec2.internal
rank : 5 (local_rank: 5)
exitcode : 1 (pid: 226676)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[5]:
time : 2024-07-06_09:44:41
host : ip-26-0-160-103.ec2.internal
rank : 6 (local_rank: 6)
exitcode : 1 (pid: 226677)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[6]:
time : 2024-07-06_09:44:41
host : ip-26-0-160-103.ec2.internal
rank : 7 (local_rank: 7)
exitcode : 1 (pid: 226678)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
------------------------------------------------------------
Root Cause (first observed failure):
[0]:
time : 2024-07-06_09:44:41
host : ip-26-0-160-103.ec2.internal
rank : 1 (local_rank: 1)
exitcode : 1 (pid: 226672)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
============================================================
srun: error: ip-26-0-160-103: task 0: Exited with exit code 1
[2024-07-06 09:44:45,887] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-165-38.ec2.internal_130790_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError.
[2024-07-06 09:44:45,917] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-166-15.ec2.internal_88800_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError.
[2024-07-06 09:44:46,745] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-161-78.ec2.internal_236947_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError.
[2024-07-06 09:44:46,773] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-165-24.ec2.internal_1566894_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError.
[2024-07-06 09:44:46,801] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-165-213.ec2.internal_165069_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError.
[2024-07-06 09:44:46,808] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-171-230.ec2.internal_3251332_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError.
[2024-07-06 09:44:46,863] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-171-249.ec2.internal_2919626_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError.
[2024-07-06 09:44:46,898] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 237018 closing signal SIGTERM
[2024-07-06 09:44:46,898] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 237019 closing signal SIGTERM
[2024-07-06 09:44:46,899] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 237020 closing signal SIGTERM
[2024-07-06 09:44:46,900] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 237021 closing signal SIGTERM
[2024-07-06 09:44:46,900] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 237022 closing signal SIGTERM
[2024-07-06 09:44:46,901] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 237023 closing signal SIGTERM
[2024-07-06 09:44:46,902] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 237024 closing signal SIGTERM
[2024-07-06 09:44:46,904] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 130860 closing signal SIGTERM
[2024-07-06 09:44:46,903] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 165138 closing signal SIGTERM
[2024-07-06 09:44:46,905] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 130861 closing signal SIGTERM
[2024-07-06 09:44:46,903] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 237025 closing signal SIGTERM
[2024-07-06 09:44:46,906] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 130862 closing signal SIGTERM
[2024-07-06 09:44:46,904] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 165139 closing signal SIGTERM
[2024-07-06 09:44:46,905] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 165140 closing signal SIGTERM
[2024-07-06 09:44:46,905] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 165141 closing signal SIGTERM
[2024-07-06 09:44:46,905] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 88871 closing signal SIGTERM
[2024-07-06 09:44:46,906] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 165142 closing signal SIGTERM
[2024-07-06 09:44:46,906] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 88872 closing signal SIGTERM
[2024-07-06 09:44:46,906] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 165143 closing signal SIGTERM
[2024-07-06 09:44:46,907] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2919696 closing signal SIGTERM
[2024-07-06 09:44:46,909] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 130863 closing signal SIGTERM
[2024-07-06 09:44:46,909] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 130864 closing signal SIGTERM
[2024-07-06 09:44:46,907] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 3251402 closing signal SIGTERM
[2024-07-06 09:44:46,908] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1566963 closing signal SIGTERM
[2024-07-06 09:44:46,908] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 165144 closing signal SIGTERM
[2024-07-06 09:44:46,907] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 88873 closing signal SIGTERM
[2024-07-06 09:44:46,907] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 88874 closing signal SIGTERM
[2024-07-06 09:44:46,909] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1566964 closing signal SIGTERM
[2024-07-06 09:44:46,909] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2919697 closing signal SIGTERM
[2024-07-06 09:44:46,909] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 165145 closing signal SIGTERM
[2024-07-06 09:44:46,908] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 88875 closing signal SIGTERM
[2024-07-06 09:44:46,908] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 3251403 closing signal SIGTERM
[2024-07-06 09:44:46,909] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 3251404 closing signal SIGTERM
[2024-07-06 09:44:46,910] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 130865 closing signal SIGTERM
[2024-07-06 09:44:46,910] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1566965 closing signal SIGTERM
[2024-07-06 09:44:46,910] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 130866 closing signal SIGTERM
[2024-07-06 09:44:46,910] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1566966 closing signal SIGTERM
[2024-07-06 09:44:46,910] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1566967 closing signal SIGTERM
[2024-07-06 09:44:46,911] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1566968 closing signal SIGTERM
[2024-07-06 09:44:46,912] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 130867 closing signal SIGTERM
[2024-07-06 09:44:46,911] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2919698 closing signal SIGTERM
[2024-07-06 09:44:46,910] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 88876 closing signal SIGTERM
[2024-07-06 09:44:46,912] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1566969 closing signal SIGTERM
[2024-07-06 09:44:46,911] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 88877 closing signal SIGTERM
[2024-07-06 09:44:46,911] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 88878 closing signal SIGTERM
[2024-07-06 09:44:46,911] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 3251405 closing signal SIGTERM
[2024-07-06 09:44:46,911] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 3251406 closing signal SIGTERM
[2024-07-06 09:44:46,912] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2919699 closing signal SIGTERM
[2024-07-06 09:44:46,912] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 3251407 closing signal SIGTERM
[2024-07-06 09:44:46,913] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1566970 closing signal SIGTERM
[2024-07-06 09:44:46,913] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 3251408 closing signal SIGTERM
[2024-07-06 09:44:46,914] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2919700 closing signal SIGTERM
[2024-07-06 09:44:46,915] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2919701 closing signal SIGTERM
[2024-07-06 09:44:46,915] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 3251409 closing signal SIGTERM
[2024-07-06 09:44:46,916] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2919702 closing signal SIGTERM
[2024-07-06 09:44:46,917] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2919703 closing signal SIGTERM
[2024-07-06 09:44:50,651] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-171-249.ec2.internal_2919626_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError.
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store
return getattr(self._store, store_op)(*args, **kwargs)
torch.distributed.DistNetworkError: Broken pipe
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in <module>
sys.exit(main())
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper
return f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main
run(args)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run
elastic_launch(
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__
return launch_agent(self._config, self._entrypoint, list(args))
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent
result = agent.run()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper
result = f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run
result = self._invoke_run(role)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run
num_nodes_waiting = rdzv_handler.num_nodes_waiting()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting
self._state_holder.sync()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync
get_response = self._backend.get_state()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state
base64_state: bytes = self._call_store("get", self._key)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store
raise RendezvousConnectionError(
torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details.
[2024-07-06 09:44:50,835] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-161-78.ec2.internal_236947_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError.
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store
return getattr(self._store, store_op)(*args, **kwargs)
torch.distributed.DistNetworkError: Broken pipe
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in <module>
sys.exit(main())
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper
return f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main
run(args)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run
elastic_launch(
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__
return launch_agent(self._config, self._entrypoint, list(args))
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent
result = agent.run()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper
result = f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run
result = self._invoke_run(role)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run
num_nodes_waiting = rdzv_handler.num_nodes_waiting()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting
self._state_holder.sync()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync
get_response = self._backend.get_state()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state
base64_state: bytes = self._call_store("get", self._key)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store
raise RendezvousConnectionError(
torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details.
[2024-07-06 09:44:50,892] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-165-38.ec2.internal_130790_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError.
srun: error: ip-26-0-171-249: task 7: Exited with exit code 1
[2024-07-06 09:44:50,922] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-166-15.ec2.internal_88800_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError.
[2024-07-06 09:44:51,043] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-165-213.ec2.internal_165069_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError.
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store
return getattr(self._store, store_op)(*args, **kwargs)
torch.distributed.DistNetworkError: Broken pipe
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in <module>
sys.exit(main())
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper
return f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main
run(args)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run
elastic_launch(
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__
return launch_agent(self._config, self._entrypoint, list(args))
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent
result = agent.run()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper
result = f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run
result = self._invoke_run(role)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run
num_nodes_waiting = rdzv_handler.num_nodes_waiting()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting
self._state_holder.sync()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync
get_response = self._backend.get_state()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state
base64_state: bytes = self._call_store("get", self._key)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store
raise RendezvousConnectionError(
torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details.
srun: error: ip-26-0-161-78: task 1: Exited with exit code 1
[2024-07-06 09:44:51,236] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-165-38.ec2.internal_130790_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError.
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store
return getattr(self._store, store_op)(*args, **kwargs)
torch.distributed.DistNetworkError: Broken pipe
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in <module>
sys.exit(main())
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper
return f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main
run(args)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run
elastic_launch(
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__
return launch_agent(self._config, self._entrypoint, list(args))
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent
result = agent.run()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper
result = f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run
result = self._invoke_run(role)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run
num_nodes_waiting = rdzv_handler.num_nodes_waiting()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting
self._state_holder.sync()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync
get_response = self._backend.get_state()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state
base64_state: bytes = self._call_store("get", self._key)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store
raise RendezvousConnectionError(
torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details.
[2024-07-06 09:44:51,251] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-171-230.ec2.internal_3251332_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError.
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store
return getattr(self._store, store_op)(*args, **kwargs)
torch.distributed.DistNetworkError: Broken pipe
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in <module>
sys.exit(main())
[2024-07-06 09:44:51,252] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-166-15.ec2.internal_88800_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError.
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store
return f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main
return getattr(self._store, store_op)(*args, **kwargs)
run(args)
torch.distributed.DistNetworkError: Broken pipe
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in <module>
elastic_launch(
sys.exit(main())
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__
return f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main
return launch_agent(self._config, self._entrypoint, list(args))
run(args)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent
result = agent.run()
elastic_launch(
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper
result = f(*args, **kwargs)
return launch_agent(self._config, self._entrypoint, list(args))
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run
result = agent.run()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper
result = self._invoke_run(role)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run
num_nodes_waiting = rdzv_handler.num_nodes_waiting()
result = f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting
result = self._invoke_run(role)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run
self._state_holder.sync()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync
get_response = self._backend.get_state()
num_nodes_waiting = rdzv_handler.num_nodes_waiting()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state
base64_state: bytes = self._call_store("get", self._key)
self._state_holder.sync()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store
get_response = self._backend.get_state()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state
raise RendezvousConnectionError(
torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details.
base64_state: bytes = self._call_store("get", self._key)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store
raise RendezvousConnectionError(
torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details.
srun: error: ip-26-0-165-213: task 4: Exited with exit code 1
srun: error: ip-26-0-165-38: task 3: Exited with exit code 1
[2024-07-06 09:44:51,545] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-165-24.ec2.internal_1566894_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError.
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store
return getattr(self._store, store_op)(*args, **kwargs)
torch.distributed.DistNetworkError: Broken pipe
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in <module>
sys.exit(main())
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper
return f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main
run(args)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run
elastic_launch(
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__
return launch_agent(self._config, self._entrypoint, list(args))
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent
srun: error: ip-26-0-171-230: task 6: Exited with exit code 1
result = agent.run()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper
result = f(*args, **kwargs)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run
result = self._invoke_run(role)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run
num_nodes_waiting = rdzv_handler.num_nodes_waiting()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting
self._state_holder.sync()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync
get_response = self._backend.get_state()
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state
base64_state: bytes = self._call_store("get", self._key)
File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store
raise RendezvousConnectionError(
torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details.
srun: error: ip-26-0-166-15: task 5: Exited with exit code 1
srun: error: ip-26-0-165-24: task 2: Exited with exit code 1
Consider using `hf_transfer` for faster uploads. This solution comes with some limitations. See https://huggingface.co/docs/huggingface_hub/hf_transfer for more details.