jimregan commited on
Commit
bfc8182
1 Parent(s): 4891be6

copies to try matxa-hf

Browse files
Files changed (2) hide show
  1. config.yaml +129 -0
  2. pytorch_model.bin +3 -0
config.yaml ADDED
@@ -0,0 +1,129 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ task_name: train
2
+ run_name: ljspeech
3
+ tags:
4
+ - ljspeech
5
+ train: true
6
+ test: true
7
+ ckpt_path: null
8
+ seed: 1234
9
+ data:
10
+ _target_: matcha.data.text_mel_datamodule.TextMelDataModule
11
+ name: ljspeech
12
+ train_filelist_path: data/LJSpeech-1.1/train.txt
13
+ valid_filelist_path: data/LJSpeech-1.1/val.txt
14
+ batch_size: 32
15
+ num_workers: 20
16
+ pin_memory: true
17
+ cleaners:
18
+ - english_cleaners2
19
+ add_blank: true
20
+ n_spks: 1
21
+ n_fft: 1024
22
+ n_feats: 80
23
+ sample_rate: 22050
24
+ hop_length: 256
25
+ win_length: 1024
26
+ f_min: 0
27
+ f_max: 8000
28
+ data_statistics:
29
+ mel_mean: -5.536622
30
+ mel_std: 2.116101
31
+ seed: ${seed}
32
+ load_durations: false
33
+ model:
34
+ _target_: matcha.models.matcha_tts.MatchaTTS
35
+ n_vocab: 178
36
+ n_spks: ${data.n_spks}
37
+ spk_emb_dim: 64
38
+ n_feats: 80
39
+ data_statistics: ${data.data_statistics}
40
+ out_size: null
41
+ prior_loss: true
42
+ use_precomputed_durations: ${data.load_durations}
43
+ encoder:
44
+ encoder_type: RoPE Encoder
45
+ encoder_params:
46
+ n_feats: ${model.n_feats}
47
+ n_channels: 192
48
+ filter_channels: 768
49
+ filter_channels_dp: 256
50
+ n_heads: 2
51
+ n_layers: 6
52
+ kernel_size: 3
53
+ p_dropout: 0.1
54
+ spk_emb_dim: 64
55
+ n_spks: 1
56
+ prenet: true
57
+ duration_predictor_params:
58
+ filter_channels_dp: ${model.encoder.encoder_params.filter_channels_dp}
59
+ kernel_size: 3
60
+ p_dropout: ${model.encoder.encoder_params.p_dropout}
61
+ decoder:
62
+ channels:
63
+ - 256
64
+ - 256
65
+ dropout: 0.05
66
+ attention_head_dim: 64
67
+ n_blocks: 1
68
+ num_mid_blocks: 2
69
+ num_heads: 2
70
+ act_fn: snakebeta
71
+ cfm:
72
+ name: CFM
73
+ solver: euler
74
+ sigma_min: 0.0001
75
+ optimizer:
76
+ _target_: torch.optim.Adam
77
+ _partial_: true
78
+ lr: 0.0001
79
+ weight_decay: 0.0
80
+ callbacks:
81
+ model_checkpoint:
82
+ _target_: lightning.pytorch.callbacks.ModelCheckpoint
83
+ dirpath: ${paths.output_dir}/checkpoints
84
+ filename: checkpoint_{epoch:03d}
85
+ monitor: epoch
86
+ verbose: false
87
+ save_last: true
88
+ save_top_k: 10
89
+ mode: max
90
+ auto_insert_metric_name: true
91
+ save_weights_only: false
92
+ every_n_train_steps: null
93
+ train_time_interval: null
94
+ every_n_epochs: 100
95
+ save_on_train_epoch_end: null
96
+ model_summary:
97
+ _target_: lightning.pytorch.callbacks.RichModelSummary
98
+ max_depth: 3
99
+ rich_progress_bar:
100
+ _target_: lightning.pytorch.callbacks.RichProgressBar
101
+ logger:
102
+ tensorboard:
103
+ _target_: lightning.pytorch.loggers.tensorboard.TensorBoardLogger
104
+ save_dir: ${paths.output_dir}/tensorboard/
105
+ name: null
106
+ log_graph: false
107
+ default_hp_metric: true
108
+ prefix: ''
109
+ trainer:
110
+ _target_: lightning.pytorch.trainer.Trainer
111
+ default_root_dir: ${paths.output_dir}
112
+ max_epochs: -1
113
+ accelerator: gpu
114
+ devices:
115
+ - 0
116
+ precision: 16-mixed
117
+ check_val_every_n_epoch: 1
118
+ deterministic: false
119
+ gradient_clip_val: 5.0
120
+ paths:
121
+ root_dir: ${oc.env:PROJECT_ROOT}
122
+ data_dir: ${paths.root_dir}/data/
123
+ log_dir: ${paths.root_dir}/logs/
124
+ output_dir: ${hydra:runtime.output_dir}
125
+ work_dir: ${hydra:runtime.cwd}
126
+ extras:
127
+ ignore_warnings: false
128
+ enforce_tags: true
129
+ print_config: true
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:284eb05eb7e8339ede808586d0bf3befa6fa81e23f68ffcaeae71d4c9081fa23
3
+ size 218839944