jialinselenasong commited on
Commit
facb390
1 Parent(s): 885021f

wav2vec2-large-xls-r-300m-zhhk

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: apache-2.0
3
- base_model: ctl/wav2vec2-large-xlsr-cantonese
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # wav2vec2-large-xls-r-300m-zhhk
15
 
16
- This model is a fine-tuned version of [ctl/wav2vec2-large-xlsr-cantonese](https://huggingface.co/ctl/wav2vec2-large-xlsr-cantonese) on an unknown dataset.
17
 
18
  ## Model description
19
 
@@ -32,16 +32,17 @@ More information needed
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
- - learning_rate: 0.0003
36
- - train_batch_size: 16
37
  - eval_batch_size: 8
38
  - seed: 42
39
- - gradient_accumulation_steps: 2
40
- - total_train_batch_size: 32
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
  - lr_scheduler_warmup_steps: 500
44
- - num_epochs: 3
 
45
 
46
  ### Training results
47
 
@@ -50,6 +51,6 @@ The following hyperparameters were used during training:
50
  ### Framework versions
51
 
52
  - Transformers 4.35.2
53
- - Pytorch 2.1.0+cu118
54
  - Datasets 2.15.0
55
  - Tokenizers 0.15.0
 
1
  ---
2
  license: apache-2.0
3
+ base_model: w11wo/wav2vec2-xls-r-300m-zh-HK-lm-v2
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # wav2vec2-large-xls-r-300m-zhhk
15
 
16
+ This model is a fine-tuned version of [w11wo/wav2vec2-xls-r-300m-zh-HK-lm-v2](https://huggingface.co/w11wo/wav2vec2-xls-r-300m-zh-HK-lm-v2) on an unknown dataset.
17
 
18
  ## Model description
19
 
 
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
+ - learning_rate: 0.001
36
+ - train_batch_size: 8
37
  - eval_batch_size: 8
38
  - seed: 42
39
+ - gradient_accumulation_steps: 8
40
+ - total_train_batch_size: 64
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
  - lr_scheduler_warmup_steps: 500
44
+ - num_epochs: 5
45
+ - mixed_precision_training: Native AMP
46
 
47
  ### Training results
48
 
 
51
  ### Framework versions
52
 
53
  - Transformers 4.35.2
54
+ - Pytorch 2.1.0+cu121
55
  - Datasets 2.15.0
56
  - Tokenizers 0.15.0
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "_name_or_path": "ctl/wav2vec2-large-xlsr-cantonese",
3
- "activation_dropout": 0.0,
4
  "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
6
  "adapter_stride": 2,
@@ -12,7 +12,7 @@
12
  "attention_dropout": 0.0,
13
  "bos_token_id": 1,
14
  "classifier_proj_size": 256,
15
- "codevector_dim": 256,
16
  "contrastive_logits_temperature": 0.1,
17
  "conv_bias": true,
18
  "conv_dim": [
@@ -60,20 +60,12 @@
60
  "intermediate_size": 4096,
61
  "layer_norm_eps": 1e-05,
62
  "layerdrop": 0.0,
63
- "mask_channel_length": 10,
64
- "mask_channel_min_space": 1,
65
- "mask_channel_other": 0.0,
66
- "mask_channel_prob": 0.0,
67
- "mask_channel_selection": "static",
68
- "mask_feature_length": 10,
69
  "mask_feature_min_masks": 0,
70
- "mask_feature_prob": 0.0,
71
  "mask_time_length": 10,
72
  "mask_time_min_masks": 2,
73
- "mask_time_min_space": 1,
74
- "mask_time_other": 0.0,
75
- "mask_time_prob": 0.05,
76
- "mask_time_selection": "static",
77
  "model_type": "wav2vec2",
78
  "num_adapter_layers": 3,
79
  "num_attention_heads": 16,
@@ -86,7 +78,7 @@
86
  "num_negatives": 100,
87
  "output_hidden_size": 1024,
88
  "pad_token_id": 3580,
89
- "proj_codevector_dim": 256,
90
  "tdnn_dilation": [
91
  1,
92
  2,
 
1
  {
2
+ "_name_or_path": "w11wo/wav2vec2-xls-r-300m-zh-HK-lm-v2",
3
+ "activation_dropout": 0.1,
4
  "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
6
  "adapter_stride": 2,
 
12
  "attention_dropout": 0.0,
13
  "bos_token_id": 1,
14
  "classifier_proj_size": 256,
15
+ "codevector_dim": 768,
16
  "contrastive_logits_temperature": 0.1,
17
  "conv_bias": true,
18
  "conv_dim": [
 
60
  "intermediate_size": 4096,
61
  "layer_norm_eps": 1e-05,
62
  "layerdrop": 0.0,
63
+ "mask_feature_length": 64,
 
 
 
 
 
64
  "mask_feature_min_masks": 0,
65
+ "mask_feature_prob": 0.25,
66
  "mask_time_length": 10,
67
  "mask_time_min_masks": 2,
68
+ "mask_time_prob": 0.0,
 
 
 
69
  "model_type": "wav2vec2",
70
  "num_adapter_layers": 3,
71
  "num_attention_heads": 16,
 
78
  "num_negatives": 100,
79
  "output_hidden_size": 1024,
80
  "pad_token_id": 3580,
81
+ "proj_codevector_dim": 768,
82
  "tdnn_dilation": [
83
  1,
84
  2,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d548d1ae1fd6388d58311df44718c4967e7c1364a719cbff3d63f575d2830880
3
  size 1276497820
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:138922249ab18bee86ade0ea53c2d60f6e66a1dda527b02042014a32de9bd208
3
  size 1276497820
runs/Dec17_19-14-30_70668636b91a/events.out.tfevents.1702840787.70668636b91a.2568.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fa4606352634c6779c67337d7568266ddc0f5ae207f21c35821ed7eaa84feca
3
+ size 5968
runs/Dec17_19-47-55_70668636b91a/events.out.tfevents.1702842619.70668636b91a.2568.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc9036c4dbaa4920c99d0504b858fcd5a4f469b6c1476bc92828cd56050d2ced
3
+ size 5968
runs/Dec17_20-08-38_70668636b91a/events.out.tfevents.1702843862.70668636b91a.2568.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fee028a40b91e61412479afcc9dcc435ad9ae642b97fbf18fd00c39a972bc9cd
3
+ size 6164
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02359f91a1840d9316705763ba27b7a1c409d7a1934e4844d345d046ac32ea65
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f437f05240d6cabc870e307679be862f3c6e7e2e5c2f4afe923ef56011fb65f
3
  size 4728