======================== START TIME: Sat Jul 6 09:37:43 UTC 2024 python3 version = Python 3.10.14 ======================== The token has not been saved to the git credentials helper. Pass `add_to_git_credential=True` in this function directly or `--add-to-git-credential` if using via `huggingface-cli` if you want to set the git credential as well. Token is valid (permission: write). Your token has been saved to /admin/home/ferdinand_mom/.cache/huggingface/token Login successful Already on 'bench_cluster' M examples/config_tiny_llama.py M examples/config_tiny_llama.yaml M examples/train_tiny_llama.sh Your branch is up to date with 'origin/bench_cluster'. Job status: RUNNING [2024-07-06 09:37:46,109] torch.distributed.run: [WARNING] [2024-07-06 09:37:46,109] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,109] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. [2024-07-06 09:37:46,109] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,113] torch.distributed.run: [WARNING] [2024-07-06 09:37:46,113] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,113] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. [2024-07-06 09:37:46,113] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,135] torch.distributed.run: [WARNING] [2024-07-06 09:37:46,135] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,135] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. [2024-07-06 09:37:46,135] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,163] torch.distributed.run: [WARNING] [2024-07-06 09:37:46,163] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,163] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. [2024-07-06 09:37:46,163] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,170] torch.distributed.run: [WARNING] [2024-07-06 09:37:46,170] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,170] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. [2024-07-06 09:37:46,170] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,188] torch.distributed.run: [WARNING] [2024-07-06 09:37:46,188] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,188] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. [2024-07-06 09:37:46,188] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,215] torch.distributed.run: [WARNING] [2024-07-06 09:37:46,215] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,215] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. [2024-07-06 09:37:46,215] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,239] torch.distributed.run: [WARNING] [2024-07-06 09:37:46,239] torch.distributed.run: [WARNING] ***************************************** [2024-07-06 09:37:46,239] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. [2024-07-06 09:37:46,239] torch.distributed.run: [WARNING] ***************************************** [default0]:07/06/2024 09:38:05 [WARNING|DP=0|PP=0|TP=0|ip-26-0-160-100]: [Vocab Size Padding] Padded vocab (size: 50257) with 7 dummy tokens (new size: 50264) [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Config: [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Config(general=GeneralArgs(project='bench_cluster', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: run='%date_%jobid', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: seed=42, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: step=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: consumed_train_samples=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: benchmark_csv_path=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: ignore_sanity_checks=True), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: parallelism=ParallelismArgs(dp=1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: pp=8, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: tp=8, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: pp_engine=, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: tp_mode=, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: tp_linear_async_communication=False, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: expert_parallel_size=1), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: model=ModelArgs(model_config=LlamaConfig(bos_token_id=1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: eos_token_id=2, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: hidden_act='silu', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: hidden_size=2048, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: initializer_range=0.02, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: intermediate_size=4096, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: is_llama_config=True, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: max_position_embeddings=4096, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: num_attention_heads=32, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: num_hidden_layers=24, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: num_key_value_heads=32, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: pad_token_id=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: pretraining_tp=1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: rms_norm_eps=1e-05, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: rope_scaling=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: rope_theta=10000.0, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: tie_word_embeddings=True, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: use_cache=True, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: vocab_size=50264), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: init_method=RandomInit(std=0.025), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: dtype=torch.bfloat16, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: make_vocab_size_divisible_by=1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: ddp_bucket_cap_mb=25), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: tokenizer=TokenizerArgs(tokenizer_name_or_path='openai-community/gpt2', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: tokenizer_revision=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: tokenizer_max_length=None), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: checkpoints=CheckpointsArgs(checkpoints_path=PosixPath('/dev/null'), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: checkpoint_interval=100000, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: save_initial_state=False, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: resume_checkpoint_path=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: checkpoints_path_is_shared_file_system=False), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: logging=LoggingArgs(log_level='info', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: log_level_replica='info', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: iteration_step_info_interval=1), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: tokens=TokensArgs(sequence_length=4096, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: train_steps=20, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: micro_batch_size=1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: batch_accumulation_per_replica=1024, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: val_check_interval=-1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: limit_val_batches=0, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: limit_test_batches=0), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: optimizer=OptimizerArgs(optimizer_factory=AdamWOptimizerArgs(adam_eps=1e-08, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: adam_beta1=0.9, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: adam_beta2=0.95, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: torch_adam_is_fused=True, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: name='adamW'), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: zero_stage=1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: weight_decay=0.01, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: clip_grad=1.0, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: accumulate_grad_in_fp32=True, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: learning_rate_scheduler=LRSchedulerArgs(learning_rate=0.0001, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: lr_warmup_steps=1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: lr_warmup_style='linear', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: lr_decay_style='linear', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: lr_decay_steps=19, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: lr_decay_starting_step=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: min_decay_lr=1e-05)), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: data_stages=[DatasetStageArgs(name='Training Stage', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: start_training_step=1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: data=DataArgs(dataset=PretrainDatasetsArgs(hf_dataset_or_datasets='roneneldan/TinyStories', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: hf_dataset_splits='train', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: hf_dataset_config_name=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: dataset_processing_num_proc_per_process=64, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: dataset_overwrite_cache=False, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: text_column_name='text'), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: seed=42, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: num_loading_workers=0))], [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: profiler=ProfilerArgs(profiler_export_path=PosixPath('/fsx/ferdinandmom/ferdinand-hf/bench_cluster/results/llama-1B/64_GPUS/dp-1_tp-8_pp-8_mbz-1')), [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: lighteval=None) [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Model Config: [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: LlamaConfig(bos_token_id=1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: eos_token_id=2, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: hidden_act='silu', [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: hidden_size=2048, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: initializer_range=0.02, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: intermediate_size=4096, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: is_llama_config=True, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: max_position_embeddings=4096, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: num_attention_heads=32, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: num_hidden_layers=24, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: num_key_value_heads=32, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: pad_token_id=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: pretraining_tp=1, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: rms_norm_eps=1e-05, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: rope_scaling=None, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: rope_theta=10000.0, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: tie_word_embeddings=True, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: use_cache=True, [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: vocab_size=50264) [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Building model.. [default0]:07/06/2024 09:38:05 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Setting PP block ranks... [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=1|ip-26-0-164-236]: Local number of parameters: 15.7M (30.02MiB) [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=3|ip-26-0-164-236]: Local number of parameters: 15.7M (30.02MiB) [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=3|ip-26-0-164-236]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=6|ip-26-0-164-236]: Local number of parameters: 15.7M (30.02MiB) [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=1|ip-26-0-164-207]: Local number of parameters: 21M (40.03MiB) [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=1|ip-26-0-164-207]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=6|ip-26-0-168-120]: Local number of parameters: 21M (40.03MiB) [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=6|ip-26-0-168-120]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=1|ip-26-0-164-236]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=1|ip-26-0-164-236]: No checkpoint path provided. [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=6|ip-26-0-164-207]: Local number of parameters: 21M (40.03MiB) [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=6|ip-26-0-164-207]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=4|ip-26-0-168-120]: Local number of parameters: 21M (40.03MiB) [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=4|ip-26-0-168-120]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=4|ip-26-0-164-207]: Local number of parameters: 21M (40.03MiB) [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=4|ip-26-0-164-207]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=4|ip-26-0-164-207]: No checkpoint path provided. [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=4|ip-26-0-168-120]: No checkpoint path provided. [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=3|ip-26-0-167-9]: Local number of parameters: 15.7M (30.02MiB) [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=3|ip-26-0-167-9]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=3|ip-26-0-167-9]: No checkpoint path provided. [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=0|ip-26-0-164-207]: Local number of parameters: 21M (40.03MiB) [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=0|ip-26-0-164-207]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=0|ip-26-0-164-207]: No checkpoint path provided. [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=7|ip-26-0-164-207]: Local number of parameters: 21M (40.03MiB) [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=2|ip-26-0-170-132]: Local number of parameters: 12.9M (24.55MiB) [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=2|ip-26-0-170-132]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=2|ip-26-0-170-132]: No checkpoint path provided. [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=7|ip-26-0-164-207]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=7|ip-26-0-164-207]: No checkpoint path provided. [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=1|ip-26-0-170-132]: Local number of parameters: 12.9M (24.55MiB) [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=1|ip-26-0-170-132]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=5|ip-26-0-164-207]: Local number of parameters: 21M (40.03MiB) [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=5|ip-26-0-164-207]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=1|ip-26-0-170-132]: No checkpoint path provided. [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=7|ip-26-0-170-132]: Local number of parameters: 12.9M (24.55MiB) [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=7|ip-26-0-170-132]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=7|ip-26-0-170-132]: No checkpoint path provided. [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=5|ip-26-0-164-207]: No checkpoint path provided. [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=5|ip-26-0-170-132]: Local number of parameters: 12.9M (24.55MiB) [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=3|ip-26-0-164-207]: Local number of parameters: 21M (40.03MiB) [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=3|ip-26-0-164-207]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=3|ip-26-0-164-207]: No checkpoint path provided. [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=3|ip-26-0-170-132]: Local number of parameters: 12.9M (24.55MiB) [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=2|ip-26-0-164-207]: Local number of parameters: 21M (40.03MiB) [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=2|ip-26-0-164-207]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=2|ip-26-0-164-207]: No checkpoint path provided. [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=4|ip-26-0-170-132]: Local number of parameters: 12.9M (24.55MiB) [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=5|ip-26-0-170-132]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=5|ip-26-0-170-132]: No checkpoint path provided. [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=0|ip-26-0-170-132]: Local number of parameters: 12.9M (24.55MiB) [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=0|ip-26-0-170-132]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=3|ip-26-0-170-132]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=3|ip-26-0-170-132]: No checkpoint path provided. [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=4|ip-26-0-170-132]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=4|ip-26-0-170-132]: No checkpoint path provided. [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=0|ip-26-0-170-132]: No checkpoint path provided. [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=6|ip-26-0-170-132]: Local number of parameters: 12.9M (24.55MiB) [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=6|ip-26-0-170-132]: [After model building] Memory usage: 24.56MiB. Peak allocated: 24.58MiB Peak reserved: 28.00MiB [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=7|TP=6|ip-26-0-170-132]: No checkpoint path provided. [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=7|ip-26-0-160-100]: Local number of parameters: 33.9M (64.57MiB) [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=7|ip-26-0-160-100]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=2|ip-26-0-160-100]: Local number of parameters: 33.9M (64.57MiB) [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=7|ip-26-0-160-100]: No checkpoint path provided. [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=2|ip-26-0-160-100]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=2|ip-26-0-160-100]: No checkpoint path provided. [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=7|ip-26-0-164-0]: Local number of parameters: 15.7M (30.02MiB) [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=7|ip-26-0-164-0]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=7|ip-26-0-164-0]: No checkpoint path provided. [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=6|ip-26-0-160-100]: Local number of parameters: 33.9M (64.57MiB) [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=0|ip-26-0-164-0]: Local number of parameters: 15.7M (30.02MiB) [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=0|ip-26-0-164-0]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=0|ip-26-0-164-0]: No checkpoint path provided. [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=6|ip-26-0-160-100]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=3|ip-26-0-160-100]: Local number of parameters: 33.9M (64.57MiB) [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=3|ip-26-0-160-100]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=6|ip-26-0-160-100]: No checkpoint path provided. [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=4|ip-26-0-160-100]: Local number of parameters: 33.9M (64.57MiB) [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=4|ip-26-0-160-100]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=4|ip-26-0-160-100]: No checkpoint path provided. [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=3|ip-26-0-160-100]: No checkpoint path provided. [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Total number of parameters: 1.21G (2314.22MiB) [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Local number of parameters: 33.9M (64.57MiB) [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: No checkpoint path provided. [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Parametrizing model parameters using StandardParametrizator [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=2|ip-26-0-168-120]: Local number of parameters: 21M (40.03MiB) [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=3|ip-26-0-164-0]: Local number of parameters: 15.7M (30.02MiB) [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=3|ip-26-0-164-0]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=5|ip-26-0-160-100]: Local number of parameters: 33.9M (64.57MiB) [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=5|ip-26-0-160-100]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=5|ip-26-0-160-100]: No checkpoint path provided. [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=1|ip-26-0-168-120]: Local number of parameters: 21M (40.03MiB) [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=2|ip-26-0-168-120]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=2|ip-26-0-168-120]: No checkpoint path provided. [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=3|ip-26-0-164-0]: No checkpoint path provided. [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=1|ip-26-0-160-100]: Local number of parameters: 33.9M (64.57MiB) [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=1|ip-26-0-160-100]: [After model building] Memory usage: 68.59MiB. Peak allocated: 70.62MiB Peak reserved: 78.00MiB [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=1|ip-26-0-168-120]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=1|ip-26-0-168-120]: No checkpoint path provided. [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=5|ip-26-0-164-0]: Local number of parameters: 15.7M (30.02MiB) [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=5|ip-26-0-164-0]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=5|ip-26-0-164-0]: No checkpoint path provided. [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=0|TP=1|ip-26-0-160-100]: No checkpoint path provided. [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=6|ip-26-0-164-0]: Local number of parameters: 15.7M (30.02MiB) [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=6|ip-26-0-164-0]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=6|ip-26-0-164-0]: No checkpoint path provided. [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=0|ip-26-0-163-158]: Local number of parameters: 15.7M (30.02MiB) [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=0|ip-26-0-163-158]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=0|ip-26-0-163-158]: No checkpoint path provided. [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=6|ip-26-0-163-158]: Local number of parameters: 15.7M (30.02MiB) [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=4|ip-26-0-164-0]: Local number of parameters: 15.7M (30.02MiB) [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=6|ip-26-0-163-158]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=2|ip-26-0-163-158]: Local number of parameters: 15.7M (30.02MiB) [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=4|ip-26-0-164-0]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=2|ip-26-0-163-158]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=6|ip-26-0-163-158]: No checkpoint path provided. [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=2|ip-26-0-163-158]: No checkpoint path provided. [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=7|ip-26-0-164-236]: Local number of parameters: 15.7M (30.02MiB) [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=1|ip-26-0-167-9]: Local number of parameters: 15.7M (30.02MiB) [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=4|ip-26-0-164-0]: No checkpoint path provided. [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=1|ip-26-0-163-158]: Local number of parameters: 15.7M (30.02MiB) [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=1|ip-26-0-163-158]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=1|ip-26-0-163-158]: No checkpoint path provided. [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=0|ip-26-0-164-236]: Local number of parameters: 15.7M (30.02MiB) [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=0|ip-26-0-168-120]: Local number of parameters: 21M (40.03MiB) [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=6|ip-26-0-167-9]: Local number of parameters: 15.7M (30.02MiB) [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=2|ip-26-0-164-0]: Local number of parameters: 15.7M (30.02MiB) [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=2|ip-26-0-164-0]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=5|ip-26-0-163-158]: Local number of parameters: 15.7M (30.02MiB) [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=0|ip-26-0-164-236]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=0|ip-26-0-168-120]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=6|ip-26-0-167-9]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=6|ip-26-0-167-9]: No checkpoint path provided. [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=2|ip-26-0-164-0]: No checkpoint path provided. [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=5|ip-26-0-163-158]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=5|ip-26-0-163-158]: No checkpoint path provided. [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=0|ip-26-0-164-236]: No checkpoint path provided. [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=5|ip-26-0-164-236]: Local number of parameters: 15.7M (30.02MiB) [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=5|ip-26-0-164-236]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=0|ip-26-0-168-120]: No checkpoint path provided. [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=1|ip-26-0-167-9]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=1|ip-26-0-167-9]: No checkpoint path provided. [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=1|ip-26-0-164-0]: Local number of parameters: 15.7M (30.02MiB) [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=4|ip-26-0-163-158]: Local number of parameters: 15.7M (30.02MiB) [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=2|ip-26-0-164-236]: Local number of parameters: 15.7M (30.02MiB) [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=7|ip-26-0-164-236]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=7|ip-26-0-164-236]: No checkpoint path provided. [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=3|ip-26-0-168-120]: Local number of parameters: 21M (40.03MiB) [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=5|ip-26-0-167-9]: Local number of parameters: 15.7M (30.02MiB) [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=1|ip-26-0-164-0]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=2|TP=1|ip-26-0-164-0]: No checkpoint path provided. [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=4|ip-26-0-163-158]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=4|ip-26-0-163-158]: No checkpoint path provided. [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=2|ip-26-0-164-236]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=5|ip-26-0-164-236]: No checkpoint path provided. [default1]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=1|ip-26-0-164-207]: No checkpoint path provided. [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=3|ip-26-0-168-120]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=3|ip-26-0-168-120]: No checkpoint path provided. [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=5|ip-26-0-167-9]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=5|ip-26-0-167-9]: No checkpoint path provided. [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=3|ip-26-0-163-158]: Local number of parameters: 15.7M (30.02MiB) [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=2|ip-26-0-164-236]: No checkpoint path provided. [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=4|ip-26-0-164-236]: Local number of parameters: 15.7M (30.02MiB) [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=3|TP=6|ip-26-0-164-207]: No checkpoint path provided. [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=7|ip-26-0-168-120]: Local number of parameters: 21M (40.03MiB) [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=7|ip-26-0-168-120]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=2|ip-26-0-167-9]: Local number of parameters: 15.7M (30.02MiB) [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=2|ip-26-0-167-9]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=3|ip-26-0-163-158]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=4|ip-26-0-164-236]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=4|ip-26-0-164-236]: No checkpoint path provided. [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=7|ip-26-0-168-120]: No checkpoint path provided. [default2]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=2|ip-26-0-167-9]: No checkpoint path provided. [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=3|ip-26-0-163-158]: No checkpoint path provided. [default3]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=3|ip-26-0-164-236]: No checkpoint path provided. [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=5|ip-26-0-168-120]: Local number of parameters: 21M (40.03MiB) [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=5|ip-26-0-168-120]: [After model building] Memory usage: 44.04MiB. Peak allocated: 46.07MiB Peak reserved: 52.00MiB [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=0|ip-26-0-167-9]: Local number of parameters: 15.7M (30.02MiB) [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=7|ip-26-0-163-158]: Local number of parameters: 15.7M (30.02MiB) [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=7|ip-26-0-163-158]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=6|ip-26-0-164-236]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default5]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=5|ip-26-0-168-120]: No checkpoint path provided. [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=0|ip-26-0-167-9]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=1|TP=7|ip-26-0-163-158]: No checkpoint path provided. [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=4|TP=6|ip-26-0-164-236]: No checkpoint path provided. [default6]:07/06/2024 09:38:23 [INFO|DP=0|PP=6|TP=6|ip-26-0-168-120]: No checkpoint path provided. [default0]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=0|ip-26-0-167-9]: No checkpoint path provided. [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=7|ip-26-0-167-9]: Local number of parameters: 15.7M (30.02MiB) [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=4|ip-26-0-167-9]: Local number of parameters: 15.7M (30.02MiB) [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=7|ip-26-0-167-9]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default7]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=7|ip-26-0-167-9]: No checkpoint path provided. [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=4|ip-26-0-167-9]: [After model building] Memory usage: 33.03MiB. Peak allocated: 35.06MiB Peak reserved: 50.00MiB [default4]:07/06/2024 09:38:23 [INFO|DP=0|PP=5|TP=4|ip-26-0-167-9]: No checkpoint path provided. [default0]:07/06/2024 09:38:25 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: [Optimizer Building] Using LearningRateForSP as learning rate [default0]:07/06/2024 09:38:25 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: [ZeRO sharding] Size of optimizer params per rank: [default0]:07/06/2024 09:38:25 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: [ZeRO sharding] DP Rank 0 has 33.9M out of 33.9M (100.00%) params' optimizer states [default0]:07/06/2024 09:38:26 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: [Training Plan] Stage Training Stage has 19 remaining training steps and has consumed 0 samples [default0]:07/06/2024 09:38:26 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Using `datasets` library [default0]:07/06/2024 09:38:26 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Loading tokenizer from openai-community/gpt2 and transformers/hf_hub versions ('4.41.2', '0.23.4') [default0]:07/06/2024 09:38:26 [WARNING|DP=0|PP=0|TP=0|ip-26-0-160-100]: Repo card metadata block was not found. Setting CardData to empty. [default0]:Repo card metadata block was not found. Setting CardData to empty. [default0]:07/06/2024 09:38:27 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: [Training Plan] There are 1 training stages [default0]:07/06/2024 09:38:27 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: [Stage Training Stage] start from step 1 [default0]:07/06/2024 09:38:27 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: [default0]:07/06/2024 09:38:27 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: [Start training] datetime: 2024-07-06 09:38:27.269095 | mbs: 1 | grad_accum: 1024 | global_batch_size: 1024 | sequence_length: 4096 | train_steps: 20 | start_iteration_step: 0 | consumed_train_samples: 0 [default2]:07/06/2024 09:38:27 [WARNING|DP=0|PP=0|TP=2|ip-26-0-160-100]: Repo card metadata block was not found. Setting CardData to empty. [default7]:07/06/2024 09:38:27 [WARNING|DP=0|PP=0|TP=7|ip-26-0-160-100]: Repo card metadata block was not found. Setting CardData to empty. [default3]:Repo card metadata block was not found. Setting CardData to empty. [default6]:07/06/2024 09:38:27 [WARNING|DP=0|PP=0|TP=6|ip-26-0-160-100]: Repo card metadata block was not found. Setting CardData to empty. [default4]:07/06/2024 09:38:27 [WARNING|DP=0|PP=0|TP=4|ip-26-0-160-100]: Repo card metadata block was not found. Setting CardData to empty. [default3]:07/06/2024 09:38:27 [WARNING|DP=0|PP=0|TP=3|ip-26-0-160-100]: Repo card metadata block was not found. Setting CardData to empty. [default4]:Repo card metadata block was not found. Setting CardData to empty. [default2]:Repo card metadata block was not found. Setting CardData to empty. [default1]:07/06/2024 09:38:27 [WARNING|DP=0|PP=0|TP=1|ip-26-0-160-100]: Repo card metadata block was not found. Setting CardData to empty. [default2]:Repo card metadata block was not found. Setting CardData to empty. [default1]:Repo card metadata block was not found. Setting CardData to empty. [default1]:Repo card metadata block was not found. Setting CardData to empty. [default2]:07/06/2024 09:38:27 [WARNING|DP=0|PP=2|TP=2|ip-26-0-164-0]: Repo card metadata block was not found. Setting CardData to empty. [default1]:07/06/2024 09:38:27 [WARNING|DP=0|PP=2|TP=1|ip-26-0-164-0]: Repo card metadata block was not found. Setting CardData to empty. [default6]:07/06/2024 09:38:27 [WARNING|DP=0|PP=5|TP=6|ip-26-0-167-9]: Repo card metadata block was not found. Setting CardData to empty. [default0]:07/06/2024 09:38:27 [WARNING|DP=0|PP=4|TP=0|ip-26-0-164-236]: Repo card metadata block was not found. Setting CardData to empty. [default5]:07/06/2024 09:38:27 [WARNING|DP=0|PP=4|TP=5|ip-26-0-164-236]: Repo card metadata block was not found. Setting CardData to empty. [default0]:07/06/2024 09:38:27 [WARNING|DP=0|PP=6|TP=0|ip-26-0-168-120]: Repo card metadata block was not found. Setting CardData to empty. [default3]:Repo card metadata block was not found. Setting CardData to empty. [default0]:Repo card metadata block was not found. Setting CardData to empty. [default0]:Repo card metadata block was not found. Setting CardData to empty. [default1]:07/06/2024 09:38:27 [WARNING|DP=0|PP=3|TP=1|ip-26-0-164-207]: Repo card metadata block was not found. Setting CardData to empty. [default6]:Repo card metadata block was not found. Setting CardData to empty. [default0]:Repo card metadata block was not found. Setting CardData to empty. [default7]:Repo card metadata block was not found. Setting CardData to empty. [default2]:Repo card metadata block was not found. Setting CardData to empty. [default1]:Repo card metadata block was not found. Setting CardData to empty. [default3]:07/06/2024 09:38:27 [WARNING|DP=0|PP=5|TP=3|ip-26-0-167-9]: Repo card metadata block was not found. Setting CardData to empty. [default6]:Repo card metadata block was not found. Setting CardData to empty. [default7]:07/06/2024 09:38:27 [WARNING|DP=0|PP=3|TP=7|ip-26-0-164-207]: Repo card metadata block was not found. Setting CardData to empty. [default5]:Repo card metadata block was not found. Setting CardData to empty. [default4]:07/06/2024 09:38:27 [WARNING|DP=0|PP=3|TP=4|ip-26-0-164-207]: Repo card metadata block was not found. Setting CardData to empty. [default4]:Repo card metadata block was not found. Setting CardData to empty. [default1]:Repo card metadata block was not found. Setting CardData to empty. [default5]:Repo card metadata block was not found. Setting CardData to empty. [default5]:07/06/2024 09:38:27 [WARNING|DP=0|PP=3|TP=5|ip-26-0-164-207]: Repo card metadata block was not found. Setting CardData to empty. [default5]:07/06/2024 09:38:27 [WARNING|DP=0|PP=7|TP=5|ip-26-0-170-132]: Repo card metadata block was not found. Setting CardData to empty. [default3]:07/06/2024 09:38:27 [WARNING|DP=0|PP=7|TP=3|ip-26-0-170-132]: Repo card metadata block was not found. Setting CardData to empty. [default6]:Repo card metadata block was not found. Setting CardData to empty. [default5]:Repo card metadata block was not found. Setting CardData to empty. [default3]:Repo card metadata block was not found. Setting CardData to empty. [default6]:07/06/2024 09:38:27 [WARNING|DP=0|PP=7|TP=6|ip-26-0-170-132]: Repo card metadata block was not found. Setting CardData to empty. [default7]:07/06/2024 09:38:27 [WARNING|DP=0|PP=2|TP=7|ip-26-0-164-0]: Repo card metadata block was not found. Setting CardData to empty. [default0]:07/06/2024 09:38:27 [WARNING|DP=0|PP=2|TP=0|ip-26-0-164-0]: Repo card metadata block was not found. Setting CardData to empty. [default6]:Repo card metadata block was not found. Setting CardData to empty. [default5]:07/06/2024 09:38:27 [WARNING|DP=0|PP=0|TP=5|ip-26-0-160-100]: Repo card metadata block was not found. Setting CardData to empty. [default4]:Repo card metadata block was not found. Setting CardData to empty. [default3]:Repo card metadata block was not found. Setting CardData to empty. [default5]:Repo card metadata block was not found. Setting CardData to empty. [default0]:Repo card metadata block was not found. Setting CardData to empty. [default5]:07/06/2024 09:38:27 [WARNING|DP=0|PP=2|TP=5|ip-26-0-164-0]: Repo card metadata block was not found. Setting CardData to empty. [default5]:Repo card metadata block was not found. Setting CardData to empty. [default1]:07/06/2024 09:38:27 [WARNING|DP=0|PP=6|TP=1|ip-26-0-168-120]: Repo card metadata block was not found. Setting CardData to empty. [default5]:Repo card metadata block was not found. Setting CardData to empty. [default2]:07/06/2024 09:38:27 [WARNING|DP=0|PP=6|TP=2|ip-26-0-168-120]: Repo card metadata block was not found. Setting CardData to empty. [default7]:Repo card metadata block was not found. Setting CardData to empty. [default0]:07/06/2024 09:38:27 [WARNING|DP=0|PP=1|TP=0|ip-26-0-163-158]: Repo card metadata block was not found. Setting CardData to empty. [default6]:Repo card metadata block was not found. Setting CardData to empty. [default7]:Repo card metadata block was not found. Setting CardData to empty. [default3]:07/06/2024 09:38:27 [WARNING|DP=0|PP=2|TP=3|ip-26-0-164-0]: Repo card metadata block was not found. Setting CardData to empty. [default6]:07/06/2024 09:38:27 [WARNING|DP=0|PP=2|TP=6|ip-26-0-164-0]: Repo card metadata block was not found. Setting CardData to empty. [default0]:Repo card metadata block was not found. Setting CardData to empty. [default2]:07/06/2024 09:38:27 [WARNING|DP=0|PP=1|TP=2|ip-26-0-163-158]: Repo card metadata block was not found. Setting CardData to empty. [default2]:Repo card metadata block was not found. Setting CardData to empty. [default4]:07/06/2024 09:38:27 [WARNING|DP=0|PP=2|TP=4|ip-26-0-164-0]: Repo card metadata block was not found. Setting CardData to empty. [default6]:07/06/2024 09:38:27 [WARNING|DP=0|PP=1|TP=6|ip-26-0-163-158]: Repo card metadata block was not found. Setting CardData to empty. [default5]:07/06/2024 09:38:27 [WARNING|DP=0|PP=5|TP=5|ip-26-0-167-9]: Repo card metadata block was not found. Setting CardData to empty. [default5]:07/06/2024 09:38:27 [WARNING|DP=0|PP=1|TP=5|ip-26-0-163-158]: Repo card metadata block was not found. Setting CardData to empty. [default7]:07/06/2024 09:38:27 [WARNING|DP=0|PP=4|TP=7|ip-26-0-164-236]: Repo card metadata block was not found. Setting CardData to empty. [default3]:Repo card metadata block was not found. Setting CardData to empty. [default5]:07/06/2024 09:38:27 [WARNING|DP=0|PP=6|TP=5|ip-26-0-168-120]: Repo card metadata block was not found. Setting CardData to empty. [default4]:Repo card metadata block was not found. Setting CardData to empty. [default1]:07/06/2024 09:38:27 [WARNING|DP=0|PP=5|TP=1|ip-26-0-167-9]: Repo card metadata block was not found. Setting CardData to empty. [default6]:Repo card metadata block was not found. Setting CardData to empty. [default4]:07/06/2024 09:38:27 [WARNING|DP=0|PP=1|TP=4|ip-26-0-163-158]: Repo card metadata block was not found. Setting CardData to empty. [default3]:Repo card metadata block was not found. Setting CardData to empty. [default4]:07/06/2024 09:38:27 [WARNING|DP=0|PP=4|TP=4|ip-26-0-164-236]: Repo card metadata block was not found. Setting CardData to empty. [default7]:Repo card metadata block was not found. Setting CardData to empty. [default7]:07/06/2024 09:38:27 [WARNING|DP=0|PP=6|TP=7|ip-26-0-168-120]: Repo card metadata block was not found. Setting CardData to empty. [default2]:Repo card metadata block was not found. Setting CardData to empty. [default4]:07/06/2024 09:38:27 [WARNING|DP=0|PP=5|TP=4|ip-26-0-167-9]: Repo card metadata block was not found. Setting CardData to empty. [default6]:Repo card metadata block was not found. Setting CardData to empty. [default3]:07/06/2024 09:38:27 [WARNING|DP=0|PP=1|TP=3|ip-26-0-163-158]: Repo card metadata block was not found. Setting CardData to empty. [default4]:Repo card metadata block was not found. Setting CardData to empty. [default2]:07/06/2024 09:38:27 [WARNING|DP=0|PP=4|TP=2|ip-26-0-164-236]: Repo card metadata block was not found. Setting CardData to empty. [default4]:Repo card metadata block was not found. Setting CardData to empty. [default3]:07/06/2024 09:38:27 [WARNING|DP=0|PP=6|TP=3|ip-26-0-168-120]: Repo card metadata block was not found. Setting CardData to empty. [default2]:Repo card metadata block was not found. Setting CardData to empty. [default7]:07/06/2024 09:38:27 [WARNING|DP=0|PP=5|TP=7|ip-26-0-167-9]: Repo card metadata block was not found. Setting CardData to empty. [default6]:Repo card metadata block was not found. Setting CardData to empty. [default7]:07/06/2024 09:38:27 [WARNING|DP=0|PP=1|TP=7|ip-26-0-163-158]: Repo card metadata block was not found. Setting CardData to empty. [default1]:Repo card metadata block was not found. Setting CardData to empty. [default2]:07/06/2024 09:38:27 [WARNING|DP=0|PP=5|TP=2|ip-26-0-167-9]: Repo card metadata block was not found. Setting CardData to empty. [default0]:Repo card metadata block was not found. Setting CardData to empty. [default0]:07/06/2024 09:38:27 [WARNING|DP=0|PP=5|TP=0|ip-26-0-167-9]: Repo card metadata block was not found. Setting CardData to empty. [default4]:Repo card metadata block was not found. Setting CardData to empty. [default6]:07/06/2024 09:38:27 [WARNING|DP=0|PP=6|TP=6|ip-26-0-168-120]: Repo card metadata block was not found. Setting CardData to empty. [default7]:Repo card metadata block was not found. Setting CardData to empty. [default1]:07/06/2024 09:38:27 [WARNING|DP=0|PP=4|TP=1|ip-26-0-164-236]: Repo card metadata block was not found. Setting CardData to empty. [default1]:Repo card metadata block was not found. Setting CardData to empty. [default6]:07/06/2024 09:38:27 [WARNING|DP=0|PP=4|TP=6|ip-26-0-164-236]: Repo card metadata block was not found. Setting CardData to empty. [default0]:Repo card metadata block was not found. Setting CardData to empty. [default6]:07/06/2024 09:38:27 [WARNING|DP=0|PP=3|TP=6|ip-26-0-164-207]: Repo card metadata block was not found. Setting CardData to empty. [default2]:Repo card metadata block was not found. Setting CardData to empty. [default0]:07/06/2024 09:38:27 [WARNING|DP=0|PP=3|TP=0|ip-26-0-164-207]: Repo card metadata block was not found. Setting CardData to empty. [default5]:Repo card metadata block was not found. Setting CardData to empty. [default4]:07/06/2024 09:38:27 [WARNING|DP=0|PP=6|TP=4|ip-26-0-168-120]: Repo card metadata block was not found. Setting CardData to empty. [default7]:Repo card metadata block was not found. Setting CardData to empty. [default1]:07/06/2024 09:38:27 [WARNING|DP=0|PP=7|TP=1|ip-26-0-170-132]: Repo card metadata block was not found. Setting CardData to empty. [default3]:Repo card metadata block was not found. Setting CardData to empty. [default2]:07/06/2024 09:38:27 [WARNING|DP=0|PP=3|TP=2|ip-26-0-164-207]: Repo card metadata block was not found. Setting CardData to empty. [default5]:Repo card metadata block was not found. Setting CardData to empty. [default3]:07/06/2024 09:38:27 [WARNING|DP=0|PP=3|TP=3|ip-26-0-164-207]: Repo card metadata block was not found. Setting CardData to empty. [default7]:Repo card metadata block was not found. Setting CardData to empty. [default2]:07/06/2024 09:38:27 [WARNING|DP=0|PP=7|TP=2|ip-26-0-170-132]: Repo card metadata block was not found. Setting CardData to empty. [default2]:Repo card metadata block was not found. Setting CardData to empty. [default7]:07/06/2024 09:38:27 [WARNING|DP=0|PP=7|TP=7|ip-26-0-170-132]: Repo card metadata block was not found. Setting CardData to empty. [default7]:Repo card metadata block was not found. Setting CardData to empty. [default4]:07/06/2024 09:38:27 [WARNING|DP=0|PP=7|TP=4|ip-26-0-170-132]: Repo card metadata block was not found. Setting CardData to empty. [default1]:Repo card metadata block was not found. Setting CardData to empty. [default0]:07/06/2024 09:38:27 [WARNING|DP=0|PP=7|TP=0|ip-26-0-170-132]: Repo card metadata block was not found. Setting CardData to empty. [default4]:Repo card metadata block was not found. Setting CardData to empty. [default1]:07/06/2024 09:38:27 [WARNING|DP=0|PP=1|TP=1|ip-26-0-163-158]: Repo card metadata block was not found. Setting CardData to empty. [default1]:Repo card metadata block was not found. Setting CardData to empty. [default3]:Repo card metadata block was not found. Setting CardData to empty. [default3]:07/06/2024 09:38:27 [WARNING|DP=0|PP=4|TP=3|ip-26-0-164-236]: Repo card metadata block was not found. Setting CardData to empty. [default0]:07/06/2024 09:38:34 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Resuming training from stage Training Stage, it has trained for 0 samples and has 19 remaining train steps [default0]:07/06/2024 09:38:34 [INFO|DP=0|PP=0|TP=0|ip-26-0-160-100]: Memory usage: 328.58MiB. Peak allocated 328.59MiB. Peak reserved: 338.00MiB [default2]:Traceback (most recent call last): [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in [default2]: trainer.train(dataloader) [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train [default2]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader) [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step [default2]: outputs = self.pipeline_engine.train_batch_iter( [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter [default2]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model) [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward [default5]:Traceback (most recent call last): [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in [default4]:Traceback (most recent call last): [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in [default2]: output = model(**micro_batch) [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default4]: trainer.train(dataloader) [default5]: trainer.train(dataloader) [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train [default5]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader) [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train [default2]: return self._call_impl(*args, **kwargs) [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default2]: return forward_call(*args, **kwargs) [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step [default4]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader) [default5]: outputs = self.pipeline_engine.train_batch_iter( [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step [default4]: outputs = self.pipeline_engine.train_batch_iter( [default2]: sharded_logits = self.model( [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter [default5]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model) [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter [default4]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model) [default2]: return self._call_impl(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default4]: output = model(**micro_batch) [default5]: output = model(**micro_batch) [default2]: return forward_call(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default4]: return self._call_impl(*args, **kwargs) [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default5]: return self._call_impl(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward [default4]: return forward_call(*args, **kwargs) [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default5]: return forward_call(*args, **kwargs) [default2]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0] [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states [default2]: hidden_encoder_states = encoder_block(**hidden_encoder_states) [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default2]: return self._call_impl(*args, **kwargs) [default5]: sharded_logits = self.model( [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward [default4]: sharded_logits = self.model( [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default4]: return self._call_impl(*args, **kwargs) [default2]: return forward_call(*args, **kwargs) [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward [default2]: output = self.pp_block(**new_kwargs) [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default5]: return self._call_impl(*args, **kwargs) [default4]: return forward_call(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default5]: return forward_call(*args, **kwargs) [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default4]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0] [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward [default2]: return self._call_impl(*args, **kwargs) [default5]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0] [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states [default4]: hidden_encoder_states = encoder_block(**hidden_encoder_states) [default2]: return forward_call(*args, **kwargs) [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 636, in forward [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default5]: hidden_encoder_states = encoder_block(**hidden_encoder_states) [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default2]: hidden_states = self.mlp(hidden_states=hidden_states)["hidden_states"] [default4]: return self._call_impl(*args, **kwargs) [default5]: return self._call_impl(*args, **kwargs) [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default5]: return forward_call(*args, **kwargs) [default4]: return forward_call(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward [default5]: output = self.pp_block(**new_kwargs) [default4]: output = self.pp_block(**new_kwargs) [default2]: return self._call_impl(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default5]: return self._call_impl(*args, **kwargs) [default4]: return self._call_impl(*args, **kwargs) [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default5]: return forward_call(*args, **kwargs) [default4]: return forward_call(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 630, in forward [default2]: return forward_call(*args, **kwargs) [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 630, in forward [default5]: output = self.attn(hidden_states=hidden_states, sequence_mask=sequence_mask) [default4]: output = self.attn(hidden_states=hidden_states, sequence_mask=sequence_mask) [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default5]: return self._call_impl(*args, **kwargs) [default4]: return self._call_impl(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 171, in forward [default4]: return forward_call(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 597, in forward [default4]: output = self.o_proj(attention_output) [default2]: hidden_states = self.down_proj(self.split_silu_mul(merged_states)) [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default5]: return forward_call(*args, **kwargs) [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 597, in forward [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default2]: return self._call_impl(*args, **kwargs) [default4]: return self._call_impl(*args, **kwargs) [default2]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default2]: return forward_call(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default5]: output = self.o_proj(attention_output) [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default4]: return forward_call(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/nn.py", line 159, in forward [default5]: return self._call_impl(*args, **kwargs) [default5]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default4]: return row_linear( [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/nn.py", line 159, in forward [default2]: return row_linear( [default5]: return forward_call(*args, **kwargs) [default4]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/functional.py", line 474, in row_linear [default4]: out = F.linear(input, weight, bias) [default2]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/functional.py", line 474, in row_linear [default4]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 16.00 MiB. GPU 4 has a total capacity of 79.33 GiB of which 5.94 MiB is free. Including non-PyTorch memory, this process has 79.31 GiB memory in use. Of the allocated memory 69.79 GiB is allocated by PyTorch, and 14.27 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) [default2]: out = F.linear(input, weight, bias) [default2]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 16.00 MiB. GPU 2 has a total capacity of 79.33 GiB of which 5.94 MiB is free. Including non-PyTorch memory, this process has 79.31 GiB memory in use. Of the allocated memory 69.94 GiB is allocated by PyTorch, and 16.13 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/nn.py", line 159, in forward [default5]: return row_linear( [default5]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/functional.py", line 474, in row_linear [default5]: out = F.linear(input, weight, bias) [default5]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 16.00 MiB. GPU 5 has a total capacity of 79.33 GiB of which 5.94 MiB is free. Including non-PyTorch memory, this process has 79.31 GiB memory in use. Of the allocated memory 69.79 GiB is allocated by PyTorch, and 14.27 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) [default6]:Traceback (most recent call last): [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py", line 237, in [default6]: trainer.train(dataloader) [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 430, in train [default6]: outputs, loss_avg = self.training_step(dataloader=self.current_dataloader) [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/trainer.py", line 459, in training_step [default6]: outputs = self.pipeline_engine.train_batch_iter( [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 187, in train_batch_iter [default6]: output = self.forward(context=context, state=state, micro_batch=micro_batch, model=model) [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/engine.py", line 44, in forward [default6]: output = model(**micro_batch) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default6]: return self._call_impl(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default6]: return forward_call(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 890, in forward [default6]: sharded_logits = self.model( [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default6]: return self._call_impl(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default6]: return forward_call(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 764, in forward [default6]: return self.forward_with_hidden_states(input_ids=input_ids, input_mask=input_mask)[0] [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 780, in forward_with_hidden_states [default6]: hidden_encoder_states = encoder_block(**hidden_encoder_states) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default6]: return self._call_impl(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default6]: return forward_call(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/pipeline_parallel/block.py", line 151, in forward [default6]: output = self.pp_block(**new_kwargs) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default6]: return self._call_impl(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default6]: return forward_call(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 630, in forward [default6]: output = self.attn(hidden_states=hidden_states, sequence_mask=sequence_mask) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default6]: return self._call_impl(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default6]: return forward_call(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/models/llama.py", line 597, in forward [default6]: output = self.o_proj(attention_output) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl [default6]: return self._call_impl(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl [default6]: return forward_call(*args, **kwargs) [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/nn.py", line 159, in forward [default6]: return row_linear( [default6]: File "/fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/src/nanotron/parallel/tensor_parallel/functional.py", line 474, in row_linear [default6]: out = F.linear(input, weight, bias) [default6]:torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 16.00 MiB. GPU 6 has a total capacity of 79.33 GiB of which 5.94 MiB is free. Including non-PyTorch memory, this process has 79.31 GiB memory in use. Of the allocated memory 69.79 GiB is allocated by PyTorch, and 14.27 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) [2024-07-06 09:39:03,435] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 36808 closing signal SIGTERM [2024-07-06 09:39:03,436] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 36809 closing signal SIGTERM [2024-07-06 09:39:03,436] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 36811 closing signal SIGTERM [2024-07-06 09:39:03,437] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 36815 closing signal SIGTERM [2024-07-06 09:39:07,467] torch.distributed.elastic.multiprocessing.api: [ERROR] failed (exitcode: 1) local_rank: 2 (pid: 36810) of binary: /fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/python3.10 Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in sys.exit(main()) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper return f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main run(args) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run elastic_launch( File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__ return launch_agent(self._config, self._entrypoint, list(args)) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 268, in launch_agent raise ChildFailedError( torch.distributed.elastic.multiprocessing.errors.ChildFailedError: ============================================================ /fsx/ferdinandmom/ferdinand-hf/bench_cluster/nanotron/run_train.py FAILED ------------------------------------------------------------ Failures: [1]: time : 2024-07-06_09:39:03 host : ip-26-0-160-100.ec2.internal rank : 4 (local_rank: 4) exitcode : 1 (pid: 36812) error_file: traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html [2]: time : 2024-07-06_09:39:03 host : ip-26-0-160-100.ec2.internal rank : 5 (local_rank: 5) exitcode : 1 (pid: 36813) error_file: traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html [3]: time : 2024-07-06_09:39:03 host : ip-26-0-160-100.ec2.internal rank : 6 (local_rank: 6) exitcode : 1 (pid: 36814) error_file: traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html ------------------------------------------------------------ Root Cause (first observed failure): [0]: time : 2024-07-06_09:39:03 host : ip-26-0-160-100.ec2.internal rank : 2 (local_rank: 2) exitcode : 1 (pid: 36810) error_file: traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html ============================================================ srun: error: ip-26-0-160-100: task 0: Exited with exit code 1 [2024-07-06 09:39:08,233] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-164-207.ec2.internal_481642_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:08,299] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-164-0.ec2.internal_1425172_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:08,343] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-168-120.ec2.internal_83560_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:08,357] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-167-9.ec2.internal_2824651_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:08,407] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-170-132.ec2.internal_1350610_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:08,444] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 481712 closing signal SIGTERM [2024-07-06 09:39:08,444] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 481713 closing signal SIGTERM [2024-07-06 09:39:08,447] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 481714 closing signal SIGTERM [2024-07-06 09:39:08,447] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 481715 closing signal SIGTERM [2024-07-06 09:39:08,447] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 481716 closing signal SIGTERM [2024-07-06 09:39:08,449] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1425242 closing signal SIGTERM [2024-07-06 09:39:08,450] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1425243 closing signal SIGTERM [2024-07-06 09:39:08,448] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 83632 closing signal SIGTERM [2024-07-06 09:39:08,450] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1425244 closing signal SIGTERM [2024-07-06 09:39:08,448] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 83633 closing signal SIGTERM [2024-07-06 09:39:08,448] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2824721 closing signal SIGTERM [2024-07-06 09:39:08,448] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2824722 closing signal SIGTERM [2024-07-06 09:39:08,450] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 83634 closing signal SIGTERM [2024-07-06 09:39:08,450] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 83635 closing signal SIGTERM [2024-07-06 09:39:08,449] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 481717 closing signal SIGTERM [2024-07-06 09:39:08,450] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 481718 closing signal SIGTERM [2024-07-06 09:39:08,451] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1350681 closing signal SIGTERM [2024-07-06 09:39:08,450] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 481719 closing signal SIGTERM [2024-07-06 09:39:08,452] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1350682 closing signal SIGTERM [2024-07-06 09:39:08,449] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2824723 closing signal SIGTERM [2024-07-06 09:39:08,452] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 83636 closing signal SIGTERM [2024-07-06 09:39:08,454] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1425245 closing signal SIGTERM [2024-07-06 09:39:08,451] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2824724 closing signal SIGTERM [2024-07-06 09:39:08,454] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1425246 closing signal SIGTERM [2024-07-06 09:39:08,452] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 83637 closing signal SIGTERM [2024-07-06 09:39:08,452] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2824725 closing signal SIGTERM [2024-07-06 09:39:08,455] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1425247 closing signal SIGTERM [2024-07-06 09:39:08,454] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1350683 closing signal SIGTERM [2024-07-06 09:39:08,453] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2824726 closing signal SIGTERM [2024-07-06 09:39:08,455] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1350684 closing signal SIGTERM [2024-07-06 09:39:08,454] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 83638 closing signal SIGTERM [2024-07-06 09:39:08,454] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2824727 closing signal SIGTERM [2024-07-06 09:39:08,454] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2824728 closing signal SIGTERM [2024-07-06 09:39:08,455] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 83639 closing signal SIGTERM [2024-07-06 09:39:08,457] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1425248 closing signal SIGTERM [2024-07-06 09:39:08,457] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1425249 closing signal SIGTERM [2024-07-06 09:39:08,456] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1350685 closing signal SIGTERM [2024-07-06 09:39:08,457] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 296422 closing signal SIGTERM [2024-07-06 09:39:08,458] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1350686 closing signal SIGTERM [2024-07-06 09:39:08,457] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 296423 closing signal SIGTERM [2024-07-06 09:39:08,458] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 296424 closing signal SIGTERM [2024-07-06 09:39:08,460] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1350687 closing signal SIGTERM [2024-07-06 09:39:08,459] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 296425 closing signal SIGTERM [2024-07-06 09:39:08,460] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 1350688 closing signal SIGTERM [2024-07-06 09:39:08,459] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 296426 closing signal SIGTERM [2024-07-06 09:39:08,460] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 296427 closing signal SIGTERM [2024-07-06 09:39:08,461] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 296428 closing signal SIGTERM [2024-07-06 09:39:08,463] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 296429 closing signal SIGTERM [2024-07-06 09:39:12,299] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-170-132.ec2.internal_1350610_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError. Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store return getattr(self._store, store_op)(*args, **kwargs) torch.distributed.DistNetworkError: Broken pipe The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in sys.exit(main()) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper return f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main run(args) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run elastic_launch( File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__ return launch_agent(self._config, self._entrypoint, list(args)) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent result = agent.run() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper result = f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run result = self._invoke_run(role) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run num_nodes_waiting = rdzv_handler.num_nodes_waiting() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting self._state_holder.sync() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync get_response = self._backend.get_state() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state base64_state: bytes = self._call_store("get", self._key) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store raise RendezvousConnectionError( torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details. [2024-07-06 09:39:12,417] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-164-236.ec2.internal_296353_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:12,541] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-163-158.ec2.internal_2776739_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. srun: error: ip-26-0-170-132: task 7: Exited with exit code 1 [2024-07-06 09:39:13,237] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-164-207.ec2.internal_481642_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:13,304] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-164-0.ec2.internal_1425172_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:13,348] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-168-120.ec2.internal_83560_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:13,361] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-167-9.ec2.internal_2824651_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:13,454] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2776809 closing signal SIGTERM [2024-07-06 09:39:13,454] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2776810 closing signal SIGTERM [2024-07-06 09:39:13,455] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2776811 closing signal SIGTERM [2024-07-06 09:39:13,456] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2776812 closing signal SIGTERM [2024-07-06 09:39:13,457] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2776813 closing signal SIGTERM [2024-07-06 09:39:13,458] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2776814 closing signal SIGTERM [2024-07-06 09:39:13,460] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2776815 closing signal SIGTERM [2024-07-06 09:39:13,461] torch.distributed.elastic.multiprocessing.api: [WARNING] Sending process 2776816 closing signal SIGTERM [2024-07-06 09:39:14,292] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-164-236.ec2.internal_296353_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError. Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store return getattr(self._store, store_op)(*args, **kwargs) [2024-07-06 09:39:14,292] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-167-9.ec2.internal_2824651_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError. torch.distributed.DistNetworkError: Connection reset by peer The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store sys.exit(main()) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper return getattr(self._store, store_op)(*args, **kwargs) return f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main torch.distributed.DistNetworkError: Broken pipe The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in run(args) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run sys.exit(main()) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper return f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main run(args) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run elastic_launch( File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__ elastic_launch( File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__ return launch_agent(self._config, self._entrypoint, list(args)) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent return launch_agent(self._config, self._entrypoint, list(args)) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent result = agent.run() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper result = f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run result = agent.run() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper result = f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run result = self._invoke_run(role) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run num_nodes_waiting = rdzv_handler.num_nodes_waiting() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting result = self._invoke_run(role) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run self._state_holder.sync() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync num_nodes_waiting = rdzv_handler.num_nodes_waiting() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting get_response = self._backend.get_state() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state self._state_holder.sync() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync base64_state: bytes = self._call_store("get", self._key) get_response = self._backend.get_state() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store base64_state: bytes = self._call_store("get", self._key) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store raise RendezvousConnectionError( torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details. raise RendezvousConnectionError( torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details. [2024-07-06 09:39:14,494] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-164-0.ec2.internal_1425172_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError. Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store return getattr(self._store, store_op)(*args, **kwargs) torch.distributed.DistNetworkError: Broken pipe The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in sys.exit(main()) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper return f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main run(args) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run elastic_launch( File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__ return launch_agent(self._config, self._entrypoint, list(args)) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent result = agent.run() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper result = f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run result = self._invoke_run(role) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run num_nodes_waiting = rdzv_handler.num_nodes_waiting() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting self._state_holder.sync() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync get_response = self._backend.get_state() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state base64_state: bytes = self._call_store("get", self._key) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store raise RendezvousConnectionError( torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details. srun: error: ip-26-0-167-9: task 5: Exited with exit code 1 srun: error: ip-26-0-164-236: task 4: Exited with exit code 1 srun: error: ip-26-0-164-0: task 2: Exited with exit code 1 [2024-07-06 09:39:14,995] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-164-207.ec2.internal_481642_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError. Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store return getattr(self._store, store_op)(*args, **kwargs) torch.distributed.DistNetworkError: Broken pipe The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in sys.exit(main()) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper return f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main run(args) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run elastic_launch( File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__ return launch_agent(self._config, self._entrypoint, list(args)) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent result = agent.run() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper result = f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run result = self._invoke_run(role) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run num_nodes_waiting = rdzv_handler.num_nodes_waiting() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting self._state_holder.sync() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync get_response = self._backend.get_state() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state base64_state: bytes = self._call_store("get", self._key) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store raise RendezvousConnectionError( torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details. [2024-07-06 09:39:15,096] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-168-120.ec2.internal_83560_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError. Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store return getattr(self._store, store_op)(*args, **kwargs) torch.distributed.DistNetworkError: Broken pipe The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in sys.exit(main()) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper return f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main run(args) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run elastic_launch( File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__ return launch_agent(self._config, self._entrypoint, list(args)) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent result = agent.run() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper result = f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run result = self._invoke_run(role) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run num_nodes_waiting = rdzv_handler.num_nodes_waiting() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting self._state_holder.sync() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync get_response = self._backend.get_state() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state base64_state: bytes = self._call_store("get", self._key) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store raise RendezvousConnectionError( torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details. srun: error: ip-26-0-164-207: task 3: Exited with exit code 1 srun: error: ip-26-0-168-120: task 6: Exited with exit code 1 [2024-07-06 09:39:17,546] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-163-158.ec2.internal_2776739_0' has failed to send a keep-alive heartbeat to the rendezvous 'none' due to an error of type RendezvousConnectionError. [2024-07-06 09:39:20,404] torch.distributed.elastic.rendezvous.dynamic_rendezvous: [WARNING] The node 'ip-26-0-163-158.ec2.internal_2776739_0' has failed to shutdown the rendezvous 'none' due to an error of type RendezvousConnectionError. Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 113, in _call_store return getattr(self._store, store_op)(*args, **kwargs) torch.distributed.DistNetworkError: Broken pipe The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/bin/torchrun", line 8, in sys.exit(main()) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper return f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 812, in main run(args) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/run.py", line 803, in run elastic_launch( File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 135, in __call__ return launch_agent(self._config, self._entrypoint, list(args)) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 259, in launch_agent result = agent.run() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 123, in wrapper result = f(*args, **kwargs) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 727, in run result = self._invoke_run(role) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 900, in _invoke_run num_nodes_waiting = rdzv_handler.num_nodes_waiting() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 1083, in num_nodes_waiting self._state_holder.sync() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/dynamic_rendezvous.py", line 409, in sync get_response = self._backend.get_state() File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 73, in get_state base64_state: bytes = self._call_store("get", self._key) File "/fsx/ferdinandmom/miniforge3/envs/env-bench-cluster/lib/python3.10/site-packages/torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py", line 115, in _call_store raise RendezvousConnectionError( torch.distributed.elastic.rendezvous.api.RendezvousConnectionError: The connection to the C10d store has failed. See inner exception for details. srun: error: ip-26-0-163-158: task 1: Exited with exit code 1 Consider using `hf_transfer` for faster uploads. This solution comes with some limitations. See https://huggingface.co/docs/huggingface_hub/hf_transfer for more details.