gary109's picture
update model card README.md
0e4a2d2
|
raw
history blame
7.86 kB
metadata
tags:
  - generated_from_trainer
datasets:
  - ai_light_dance
metrics:
  - wer
model-index:
  - name: ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-2
    results:
      - task:
          name: Automatic Speech Recognition
          type: automatic-speech-recognition
        dataset:
          name: ai_light_dance
          type: ai_light_dance
          config: onset-idmt-2
          split: train
          args: onset-idmt-2
        metrics:
          - name: Wer
            type: wer
            value: 0.26

ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-2

This model is a fine-tuned version of gary109/ai-light-dance_drums_pretrain_wav2vec2-base on the ai_light_dance dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5174
  • Wer: 0.26

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0003
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • gradient_accumulation_steps: 4
  • total_train_batch_size: 16
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 30
  • num_epochs: 100.0
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Wer
No log 1.0 9 97.9319 1.0
17.1836 2.0 18 45.7229 1.0
13.2869 3.0 27 2.7579 1.0
2.6495 4.0 36 2.7427 1.0
1.7135 5.0 45 2.5477 1.0
1.4609 6.0 54 1.7126 1.0
1.374 7.0 63 1.3668 0.9967
1.2951 8.0 72 1.1274 0.9867
1.0493 9.0 81 0.7346 0.5178
0.8835 10.0 90 0.7664 0.4122
0.8835 11.0 99 0.5438 0.3867
0.7019 12.0 108 0.4876 0.3711
0.6906 13.0 117 0.5194 0.36
0.6535 14.0 126 0.4489 0.3556
0.6225 15.0 135 0.4383 0.3333
0.547 16.0 144 0.4521 0.3556
0.5525 17.0 153 0.5476 0.3344
0.6152 18.0 162 0.4466 0.36
0.5055 19.0 171 0.3981 0.3256
0.5204 20.0 180 0.4924 0.3078
0.5204 21.0 189 0.4085 0.32
0.4742 22.0 198 0.4255 0.3233
0.4774 23.0 207 0.4321 0.2889
0.5029 24.0 216 0.4412 0.3167
0.4889 25.0 225 0.4051 0.3044
0.4446 26.0 234 0.3918 0.3089
0.4255 27.0 243 0.4039 0.2956
0.4396 28.0 252 0.4113 0.2956
0.4265 29.0 261 0.5576 0.3022
0.4289 30.0 270 0.3558 0.3078
0.4289 31.0 279 0.3390 0.3167
0.3817 32.0 288 0.3739 0.3422
0.4192 33.0 297 0.3179 0.3056
0.3719 34.0 306 0.3622 0.3033
0.3685 35.0 315 0.4057 0.3256
0.3752 36.0 324 0.3950 0.31
0.378 37.0 333 0.3907 0.3567
0.4438 38.0 342 0.3376 0.31
0.3978 39.0 351 0.3395 0.2833
0.3639 40.0 360 0.3646 0.2856
0.3639 41.0 369 0.3546 0.3044
0.3535 42.0 378 0.3699 0.2889
0.3311 43.0 387 0.3882 0.3022
0.3475 44.0 396 0.4749 0.2889
0.4048 45.0 405 0.3437 0.2911
0.2984 46.0 414 0.3664 0.27
0.3535 47.0 423 0.3291 0.2889
0.3015 48.0 432 0.3538 0.2767
0.3628 49.0 441 0.4411 0.2733
0.3303 50.0 450 0.3425 0.29
0.3303 51.0 459 0.3162 0.3011
0.271 52.0 468 0.3685 0.2933
0.3299 53.0 477 0.4216 0.2933
0.2782 54.0 486 0.4713 0.3044
0.348 55.0 495 0.4310 0.3078
0.2969 56.0 504 0.4898 0.2767
0.2757 57.0 513 0.5195 0.2789
0.2662 58.0 522 0.4631 0.2911
0.2706 59.0 531 0.4275 0.2833
0.2684 60.0 540 0.5535 0.2789
0.2684 61.0 549 0.4733 0.2978
0.2819 62.0 558 0.4969 0.2833
0.2819 63.0 567 0.6202 0.2789
0.2889 64.0 576 0.3955 0.2733
0.2515 65.0 585 0.3806 0.2656
0.2468 66.0 594 0.3473 0.2722
0.2557 67.0 603 0.4170 0.2722
0.2477 68.0 612 0.4749 0.2678
0.2965 69.0 621 0.4387 0.2611
0.2606 70.0 630 0.4586 0.2656
0.2606 71.0 639 0.5755 0.2733
0.2442 72.0 648 0.5582 0.2656
0.347 73.0 657 0.3897 0.2711
0.2444 74.0 666 0.3369 0.2533
0.2811 75.0 675 0.3487 0.2578
0.24 76.0 684 0.3692 0.2589
0.2466 77.0 693 0.4567 0.2578
0.2769 78.0 702 0.4041 0.2633
0.2464 79.0 711 0.3813 0.2622
0.2791 80.0 720 0.3990 0.2556
0.2791 81.0 729 0.3997 0.2489
0.2365 82.0 738 0.4537 0.2533
0.2693 83.0 747 0.5943 0.2611
0.2285 84.0 756 0.5805 0.2656
0.2468 85.0 765 0.5609 0.2656
0.2226 86.0 774 0.5948 0.2667
0.2419 87.0 783 0.5910 0.2544
0.2254 88.0 792 0.5741 0.26
0.2083 89.0 801 0.4984 0.2611
0.2318 90.0 810 0.5093 0.26
0.2318 91.0 819 0.5284 0.2633
0.2458 92.0 828 0.4885 0.2656
0.2394 93.0 837 0.4818 0.2622
0.2018 94.0 846 0.5037 0.26
0.235 95.0 855 0.5011 0.2578
0.2252 96.0 864 0.4931 0.2611
0.2147 97.0 873 0.4881 0.2589
0.2227 98.0 882 0.4956 0.2589
0.2168 99.0 891 0.5097 0.2589
0.2282 100.0 900 0.5174 0.26

Framework versions

  • Transformers 4.25.0.dev0
  • Pytorch 1.8.1+cu111
  • Datasets 2.7.1.dev0
  • Tokenizers 0.13.2