MatanP commited on
Commit
f50a58a
·
verified ·
1 Parent(s): df80935

Upload model

Browse files
README.md ADDED
@@ -0,0 +1,199 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags: []
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+ This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
config.json ADDED
@@ -0,0 +1,115 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "C:\\Users\\HP\\Work3\\TTS Working\\checkpoints\\best_model",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "relu",
5
+ "adaptor_dropout": 0.1,
6
+ "adaptor_kernel_size": 8,
7
+ "adaptor_stride": 8,
8
+ "add_adapter": true,
9
+ "architectures": [
10
+ "SeamlessM4TModel"
11
+ ],
12
+ "attention_dropout": 0.1,
13
+ "bos_token_id": 2,
14
+ "conv_depthwise_kernel_size": 31,
15
+ "decoder_attention_heads": 16,
16
+ "decoder_ffn_dim": 8192,
17
+ "decoder_layerdrop": 0.05,
18
+ "decoder_layers": 24,
19
+ "decoder_start_token_id": 3,
20
+ "dropout": 0.1,
21
+ "encoder_attention_heads": 16,
22
+ "encoder_ffn_dim": 8192,
23
+ "encoder_layerdrop": 0.05,
24
+ "encoder_layers": 24,
25
+ "eos_token_id": 3,
26
+ "feature_projection_input_dim": 160,
27
+ "hidden_size": 1024,
28
+ "initializer_range": 0.02,
29
+ "is_encoder_decoder": true,
30
+ "lang_embed_dim": 256,
31
+ "layer_norm_eps": 1e-05,
32
+ "leaky_relu_slope": 0.1,
33
+ "max_new_tokens": 256,
34
+ "max_position_embeddings": 1024,
35
+ "max_source_positions": 4096,
36
+ "model_type": "seamless_m4t",
37
+ "num_adapter_layers": 1,
38
+ "num_attention_heads": 16,
39
+ "num_conv_pos_embedding_groups": 16,
40
+ "num_conv_pos_embeddings": 128,
41
+ "num_hidden_layers": 24,
42
+ "pad_token_id": 0,
43
+ "position_embeddings_type": "relative",
44
+ "resblock_dilation_sizes": [
45
+ [
46
+ 1,
47
+ 3,
48
+ 5
49
+ ],
50
+ [
51
+ 1,
52
+ 3,
53
+ 5
54
+ ],
55
+ [
56
+ 1,
57
+ 3,
58
+ 5
59
+ ]
60
+ ],
61
+ "resblock_kernel_sizes": [
62
+ 3,
63
+ 7,
64
+ 11
65
+ ],
66
+ "rotary_embedding_base": 10000,
67
+ "sampling_rate": 16000,
68
+ "scale_embedding": true,
69
+ "speech_encoder_attention_heads": 16,
70
+ "speech_encoder_dropout": 0.0,
71
+ "speech_encoder_hidden_act": "swish",
72
+ "speech_encoder_intermediate_size": 4096,
73
+ "speech_encoder_layerdrop": 0.1,
74
+ "speech_encoder_layers": 24,
75
+ "spkr_embed_dim": 256,
76
+ "t2u_bos_token_id": 0,
77
+ "t2u_decoder_attention_heads": 16,
78
+ "t2u_decoder_ffn_dim": 8192,
79
+ "t2u_decoder_layers": 6,
80
+ "t2u_decoder_start_token_id": 2,
81
+ "t2u_encoder_attention_heads": 16,
82
+ "t2u_encoder_ffn_dim": 8192,
83
+ "t2u_encoder_layers": 6,
84
+ "t2u_eos_token_id": 2,
85
+ "t2u_max_new_tokens": 1024,
86
+ "t2u_max_position_embeddings": 2048,
87
+ "t2u_pad_token_id": 1,
88
+ "t2u_vocab_size": 10082,
89
+ "torch_dtype": "float32",
90
+ "transformers_version": "4.45.2",
91
+ "unit_embed_dim": 1280,
92
+ "unit_hifi_gan_vocab_size": 10000,
93
+ "upsample_initial_channel": 512,
94
+ "upsample_kernel_sizes": [
95
+ 11,
96
+ 8,
97
+ 8,
98
+ 4,
99
+ 4
100
+ ],
101
+ "upsample_rates": [
102
+ 5,
103
+ 4,
104
+ 4,
105
+ 2,
106
+ 2
107
+ ],
108
+ "use_cache": true,
109
+ "var_pred_dropout": 0.5,
110
+ "variance_predictor_kernel_size": 3,
111
+ "vocab_size": 256102,
112
+ "vocoder_num_langs": 36,
113
+ "vocoder_num_spkrs": 200,
114
+ "vocoder_offset": 4
115
+ }
generation_config.json ADDED
@@ -0,0 +1,186 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 2,
3
+ "decoder_start_token_id": 3,
4
+ "eos_token_id": 3,
5
+ "max_new_tokens": 256,
6
+ "pad_token_id": 0,
7
+ "t2u_lang_code_to_id": {
8
+ "arb": 10043,
9
+ "ben": 10044,
10
+ "cat": 10045,
11
+ "ces": 10046,
12
+ "cmn": 10047,
13
+ "cym": 10048,
14
+ "dan": 10049,
15
+ "deu": 10050,
16
+ "eng": 10051,
17
+ "est": 10052,
18
+ "fin": 10053,
19
+ "fra": 10054,
20
+ "hin": 10055,
21
+ "ind": 10056,
22
+ "ita": 10057,
23
+ "jpn": 10058,
24
+ "kan": 10059,
25
+ "kor": 10060,
26
+ "mlt": 10061,
27
+ "nld": 10062,
28
+ "pes": 10063,
29
+ "pol": 10064,
30
+ "por": 10065,
31
+ "ron": 10066,
32
+ "rus": 10067,
33
+ "slk": 10068,
34
+ "spa": 10069,
35
+ "swe": 10070,
36
+ "swh": 10071,
37
+ "tam": 10072,
38
+ "tel": 10073,
39
+ "tgl": 10074,
40
+ "tha": 10075,
41
+ "tur": 10076,
42
+ "ukr": 10077,
43
+ "urd": 10078,
44
+ "uzn": 10079,
45
+ "vie": 10080
46
+ },
47
+ "text_decoder_lang_to_code_id": {
48
+ "afr": 256001,
49
+ "amh": 256002,
50
+ "arb": 256003,
51
+ "ary": 256004,
52
+ "arz": 256005,
53
+ "asm": 256006,
54
+ "azj": 256007,
55
+ "bel": 256008,
56
+ "ben": 256009,
57
+ "bos": 256010,
58
+ "bul": 256011,
59
+ "cat": 256012,
60
+ "ceb": 256013,
61
+ "ces": 256014,
62
+ "ckb": 256015,
63
+ "cmn": 256016,
64
+ "cmn_Hant": 256017,
65
+ "cym": 256018,
66
+ "dan": 256019,
67
+ "deu": 256020,
68
+ "ell": 256021,
69
+ "eng": 256022,
70
+ "est": 256023,
71
+ "eus": 256024,
72
+ "fin": 256025,
73
+ "fra": 256026,
74
+ "fuv": 256027,
75
+ "gaz": 256028,
76
+ "gle": 256029,
77
+ "glg": 256030,
78
+ "guj": 256031,
79
+ "heb": 256032,
80
+ "hin": 256033,
81
+ "hrv": 256034,
82
+ "hun": 256035,
83
+ "hye": 256036,
84
+ "ibo": 256037,
85
+ "ind": 256038,
86
+ "isl": 256039,
87
+ "ita": 256040,
88
+ "jav": 256041,
89
+ "jpn": 256042,
90
+ "kan": 256043,
91
+ "kat": 256044,
92
+ "kaz": 256045,
93
+ "khk": 256046,
94
+ "khm": 256047,
95
+ "kir": 256048,
96
+ "kor": 256049,
97
+ "lao": 256050,
98
+ "lit": 256051,
99
+ "lug": 256052,
100
+ "luo": 256053,
101
+ "lvs": 256054,
102
+ "mai": 256055,
103
+ "mal": 256056,
104
+ "mar": 256057,
105
+ "mkd": 256058,
106
+ "mlt": 256059,
107
+ "mni": 256060,
108
+ "mya": 256061,
109
+ "nld": 256062,
110
+ "nno": 256063,
111
+ "nob": 256064,
112
+ "npi": 256065,
113
+ "nya": 256066,
114
+ "ory": 256067,
115
+ "pan": 256068,
116
+ "pbt": 256069,
117
+ "pes": 256070,
118
+ "pol": 256071,
119
+ "por": 256072,
120
+ "ron": 256073,
121
+ "rus": 256074,
122
+ "sat": 256075,
123
+ "slk": 256076,
124
+ "slv": 256077,
125
+ "sna": 256078,
126
+ "snd": 256079,
127
+ "som": 256080,
128
+ "spa": 256081,
129
+ "srp": 256082,
130
+ "swe": 256083,
131
+ "swh": 256084,
132
+ "tam": 256085,
133
+ "tel": 256086,
134
+ "tgk": 256087,
135
+ "tgl": 256088,
136
+ "tha": 256089,
137
+ "tur": 256090,
138
+ "ukr": 256091,
139
+ "urd": 256092,
140
+ "uzn": 256093,
141
+ "vie": 256094,
142
+ "yor": 256095,
143
+ "yue": 256096,
144
+ "zlm": 256097,
145
+ "zul": 256098
146
+ },
147
+ "transformers_version": "4.45.2",
148
+ "vocoder_lang_code_to_id": {
149
+ "arb": 0,
150
+ "ben": 1,
151
+ "cat": 2,
152
+ "ces": 3,
153
+ "cmn": 4,
154
+ "cym": 5,
155
+ "dan": 6,
156
+ "deu": 7,
157
+ "eng": 8,
158
+ "est": 9,
159
+ "fin": 10,
160
+ "fra": 11,
161
+ "hin": 12,
162
+ "ind": 13,
163
+ "ita": 14,
164
+ "jpn": 15,
165
+ "kor": 16,
166
+ "mlt": 17,
167
+ "nld": 18,
168
+ "pes": 19,
169
+ "pol": 20,
170
+ "por": 21,
171
+ "ron": 22,
172
+ "rus": 23,
173
+ "slk": 24,
174
+ "spa": 25,
175
+ "swe": 26,
176
+ "swh": 27,
177
+ "tel": 28,
178
+ "tgl": 29,
179
+ "tha": 30,
180
+ "tur": 31,
181
+ "ukr": 32,
182
+ "urd": 33,
183
+ "uzn": 34,
184
+ "vie": 35
185
+ }
186
+ }
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcb140228d2d55a4f6cc2551d624e3d3e3b10283cb7fff6df61506a535e80b42
3
+ size 4998891752
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3132753cac476b9aa2373cd14c73450595cca9c2e02e63d9e3b135356aa6ee12
3
+ size 4441568568
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff