File size: 3,528 Bytes
abf3d8c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 |
_name: null
common:
_name: null
no_progress_bar: false
log_interval: 10
log_format: csv
log_file: null
wandb_project: null
wandb_entity: null
seed: 1
fp16: false
memory_efficient_fp16: false
fp16_no_flatten_grads: false
fp16_init_scale: 128
fp16_scale_window: null
fp16_scale_tolerance: 0.0
on_cpu_convert_precision: false
min_loss_scale: 0.0001
threshold_loss_scale: null
empty_cache_freq: 0
all_gather_list_size: 16384
model_parallel_size: 1
profile: false
reset_logging: false
suppress_crashes: false
common_eval:
_name: null
path: null
quiet: false
model_overrides: '{}'
save_outputs: false
results_path: null
distributed_training:
_name: null
distributed_world_size: 4
distributed_rank: 0
distributed_backend: nccl
distributed_init_method: null
distributed_port: 12355
device_id: 0
ddp_comm_hook: none
bucket_cap_mb: 25
fix_batches_to_gpus: false
find_unused_parameters: false
heartbeat_timeout: -1
broadcast_buffers: false
fp16: ${common.fp16}
memory_efficient_fp16: ${common.memory_efficient_fp16}
dataset:
_name: null
num_workers: 10
skip_invalid_size_inputs_valid_test: false
max_tokens: null
batch_size: 64
required_batch_size_multiple: 8
data_buffer_size: 10
train_subset: train
valid_subset: valid
combine_valid_subsets: null
ignore_unused_valid_subsets: false
validate_interval: 1
validate_interval_updates: 0
validate_after_updates: 0
fixed_validation_seed: null
disable_validation: false
max_tokens_valid: ${dataset.max_tokens}
batch_size_valid: ${dataset.batch_size}
max_valid_steps: null
curriculum: 0
num_shards: 1
shard_id: 0
optimization:
_name: null
max_epoch: 200
max_update: 0
lr:
- 5.0e-05
stop_time_hours: 0.0
clip_norm: 0.0
update_freq:
- 2
stop_min_lr: -1.0
checkpoint:
_name: null
save_dir: <REDACTED>
restore_file: checkpoint_last.pt
finetune_from_model: null
reset_dataloader: false
reset_lr_scheduler: false
reset_meters: false
reset_optimizer: false
optimizer_overrides: '{}'
save_interval: 10
save_interval_updates: 0
keep_interval_updates: -1
keep_interval_updates_pattern: -1
keep_last_epochs: 0
keep_best_checkpoints: -1
no_save: false
no_epoch_checkpoints: false
no_last_checkpoints: false
no_save_optimizer_state: false
best_checkpoint_metric: loss
maximize_best_checkpoint_metric: false
patience: -1
checkpoint_suffix: ''
checkpoint_shard_count: 1
load_checkpoint_on_all_dp_ranks: false
model:
_name: wav2vec2_cmsc
apply_mask: true
mask_prob: 0.65
encoder_layers: 24
encoder_embed_dim: 1024
encoder_ffn_embed_dim: 4096
encoder_attention_heads: 16
quantize_targets: true
final_dim: 256
dropout_input: 0.1
dropout_features: 0.1
feature_grad_mult: 0.1
in_d: 12
task:
_name: ecg_pretraining
data: <REDACTED>/cmsc
perturbation_mode:
- random_leads_masking
p:
- 1.0
mask_leads_selection: random
mask_leads_prob: 0.5
normalize: false
enable_padding: true
enable_padding_leads: false
leads_to_load: null
criterion:
_name: wav2vec2_with_cmsc
infonce: true
log_keys:
- prob_perplexity
- code_perplexity
- temp
loss_weights:
- 0.1
- 10
lr_scheduler:
_name: fixed
warmup_updates: 0
optimizer:
_name: adam
adam_betas: (0.9, 0.98)
adam_eps: 1.0e-06
weight_decay: 0.01
|