kishorekashyap commited on
Commit
35669e1
1 Parent(s): da8a335

Update hyperparams.yaml

Browse files
Files changed (1) hide show
  1. hyperparams.yaml +12 -57
hyperparams.yaml CHANGED
@@ -1,65 +1,21 @@
1
 
2
  # ################################
3
- # Model: Tacotroon2 for TTS
4
  # Authors: Jonas Freiknecht
5
  # ################################
6
 
7
- mask_padding: True
8
- n_mel_channels: 80
9
- n_symbols: 148
10
- symbols_embedding_dim: 256
11
- encoder_kernel_size: 5
12
- encoder_n_convolutions: 3
13
- encoder_embedding_dim: 256
14
- attention_rnn_dim: 1024
15
- attention_dim: 128
16
- attention_location_n_filters: 32
17
- attention_location_kernel_size: 31
18
- n_frames_per_step: 1
19
- decoder_rnn_dim: 512
20
- prenet_dim: 128
21
- max_decoder_steps: 1000
22
- gate_threshold: 0.5
23
- p_attention_dropout: 0.1
24
- p_decoder_dropout: 0.1
25
- postnet_embedding_dim: 256
26
- postnet_kernel_size: 5
27
- postnet_n_convolutions: 5
28
- decoder_no_early_stopping: False
29
- sample_rate: 22050
30
 
31
- # Model
32
- model: !new:speechbrain.lobes.models.Tacotron2.Tacotron2
33
- mask_padding: !ref <mask_padding>
34
- n_mel_channels: !ref <n_mel_channels>
35
- # symbols
36
- n_symbols: !ref <n_symbols>
37
- symbols_embedding_dim: !ref <symbols_embedding_dim>
38
- # encoder
39
- encoder_kernel_size: !ref <encoder_kernel_size>
40
- encoder_n_convolutions: !ref <encoder_n_convolutions>
41
- encoder_embedding_dim: !ref <encoder_embedding_dim>
42
- # attention
43
- attention_rnn_dim: !ref <attention_rnn_dim>
44
- attention_dim: !ref <attention_dim>
45
- # attention location
46
- attention_location_n_filters: !ref <attention_location_n_filters>
47
- attention_location_kernel_size: !ref <attention_location_kernel_size>
48
- # decoder
49
- n_frames_per_step: !ref <n_frames_per_step>
50
- decoder_rnn_dim: !ref <decoder_rnn_dim>
51
- prenet_dim: !ref <prenet_dim>
52
- max_decoder_steps: !ref <max_decoder_steps>
53
- gate_threshold: !ref <gate_threshold>
54
- p_attention_dropout: !ref <p_attention_dropout>
55
- p_decoder_dropout: !ref <p_decoder_dropout>
56
- # postnet
57
- postnet_embedding_dim: !ref <postnet_embedding_dim>
58
- postnet_kernel_size: !ref <postnet_kernel_size>
59
- postnet_n_convolutions: !ref <postnet_n_convolutions>
60
- decoder_no_early_stopping: !ref <decoder_no_early_stopping>
61
-
62
- # Function that converts the text into a sequence of valid characters.
63
  generator: !new:speechbrain.lobes.models.HifiGAN.HifiganGenerator
64
  in_channels: !ref <in_channels>
65
  out_channels: !ref <out_channels>
@@ -81,4 +37,3 @@ pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
81
  generator: !ref <generator>
82
 
83
 
84
-
 
1
 
2
  # ################################
3
+ # Model: German HiFIGAN
4
  # Authors: Jonas Freiknecht
5
  # ################################
6
 
7
+ in_channels: 80
8
+ out_channels: 1
9
+ resblock_type: "1"
10
+ resblock_dilation_sizes: [[1, 3, 5], [1, 3, 5], [1, 3, 5]]
11
+ resblock_kernel_sizes: [3, 7, 11]
12
+ upsample_kernel_sizes: [16, 16, 4, 4]
13
+ upsample_initial_channel: 512
14
+ upsample_factors: [8, 8, 2, 2]
15
+ inference_padding: 5
16
+ cond_channels: 0
17
+ conv_post_bias: True
 
 
 
 
 
 
 
 
 
 
 
 
18
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  generator: !new:speechbrain.lobes.models.HifiGAN.HifiganGenerator
20
  in_channels: !ref <in_channels>
21
  out_channels: !ref <out_channels>
 
37
  generator: !ref <generator>
38
 
39