--- tags: - generated_from_trainer datasets: - ai_light_dance metrics: - wer model-index: - name: ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-2 results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: ai_light_dance type: ai_light_dance config: onset-idmt-2 split: train args: onset-idmt-2 metrics: - name: Wer type: wer value: 0.26 --- # ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-2 This model is a fine-tuned version of [gary109/ai-light-dance_drums_pretrain_wav2vec2-base](https://huggingface.co/gary109/ai-light-dance_drums_pretrain_wav2vec2-base) on the ai_light_dance dataset. It achieves the following results on the evaluation set: - Loss: 0.5174 - Wer: 0.26 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 30 - num_epochs: 100.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 1.0 | 9 | 97.9319 | 1.0 | | 17.1836 | 2.0 | 18 | 45.7229 | 1.0 | | 13.2869 | 3.0 | 27 | 2.7579 | 1.0 | | 2.6495 | 4.0 | 36 | 2.7427 | 1.0 | | 1.7135 | 5.0 | 45 | 2.5477 | 1.0 | | 1.4609 | 6.0 | 54 | 1.7126 | 1.0 | | 1.374 | 7.0 | 63 | 1.3668 | 0.9967 | | 1.2951 | 8.0 | 72 | 1.1274 | 0.9867 | | 1.0493 | 9.0 | 81 | 0.7346 | 0.5178 | | 0.8835 | 10.0 | 90 | 0.7664 | 0.4122 | | 0.8835 | 11.0 | 99 | 0.5438 | 0.3867 | | 0.7019 | 12.0 | 108 | 0.4876 | 0.3711 | | 0.6906 | 13.0 | 117 | 0.5194 | 0.36 | | 0.6535 | 14.0 | 126 | 0.4489 | 0.3556 | | 0.6225 | 15.0 | 135 | 0.4383 | 0.3333 | | 0.547 | 16.0 | 144 | 0.4521 | 0.3556 | | 0.5525 | 17.0 | 153 | 0.5476 | 0.3344 | | 0.6152 | 18.0 | 162 | 0.4466 | 0.36 | | 0.5055 | 19.0 | 171 | 0.3981 | 0.3256 | | 0.5204 | 20.0 | 180 | 0.4924 | 0.3078 | | 0.5204 | 21.0 | 189 | 0.4085 | 0.32 | | 0.4742 | 22.0 | 198 | 0.4255 | 0.3233 | | 0.4774 | 23.0 | 207 | 0.4321 | 0.2889 | | 0.5029 | 24.0 | 216 | 0.4412 | 0.3167 | | 0.4889 | 25.0 | 225 | 0.4051 | 0.3044 | | 0.4446 | 26.0 | 234 | 0.3918 | 0.3089 | | 0.4255 | 27.0 | 243 | 0.4039 | 0.2956 | | 0.4396 | 28.0 | 252 | 0.4113 | 0.2956 | | 0.4265 | 29.0 | 261 | 0.5576 | 0.3022 | | 0.4289 | 30.0 | 270 | 0.3558 | 0.3078 | | 0.4289 | 31.0 | 279 | 0.3390 | 0.3167 | | 0.3817 | 32.0 | 288 | 0.3739 | 0.3422 | | 0.4192 | 33.0 | 297 | 0.3179 | 0.3056 | | 0.3719 | 34.0 | 306 | 0.3622 | 0.3033 | | 0.3685 | 35.0 | 315 | 0.4057 | 0.3256 | | 0.3752 | 36.0 | 324 | 0.3950 | 0.31 | | 0.378 | 37.0 | 333 | 0.3907 | 0.3567 | | 0.4438 | 38.0 | 342 | 0.3376 | 0.31 | | 0.3978 | 39.0 | 351 | 0.3395 | 0.2833 | | 0.3639 | 40.0 | 360 | 0.3646 | 0.2856 | | 0.3639 | 41.0 | 369 | 0.3546 | 0.3044 | | 0.3535 | 42.0 | 378 | 0.3699 | 0.2889 | | 0.3311 | 43.0 | 387 | 0.3882 | 0.3022 | | 0.3475 | 44.0 | 396 | 0.4749 | 0.2889 | | 0.4048 | 45.0 | 405 | 0.3437 | 0.2911 | | 0.2984 | 46.0 | 414 | 0.3664 | 0.27 | | 0.3535 | 47.0 | 423 | 0.3291 | 0.2889 | | 0.3015 | 48.0 | 432 | 0.3538 | 0.2767 | | 0.3628 | 49.0 | 441 | 0.4411 | 0.2733 | | 0.3303 | 50.0 | 450 | 0.3425 | 0.29 | | 0.3303 | 51.0 | 459 | 0.3162 | 0.3011 | | 0.271 | 52.0 | 468 | 0.3685 | 0.2933 | | 0.3299 | 53.0 | 477 | 0.4216 | 0.2933 | | 0.2782 | 54.0 | 486 | 0.4713 | 0.3044 | | 0.348 | 55.0 | 495 | 0.4310 | 0.3078 | | 0.2969 | 56.0 | 504 | 0.4898 | 0.2767 | | 0.2757 | 57.0 | 513 | 0.5195 | 0.2789 | | 0.2662 | 58.0 | 522 | 0.4631 | 0.2911 | | 0.2706 | 59.0 | 531 | 0.4275 | 0.2833 | | 0.2684 | 60.0 | 540 | 0.5535 | 0.2789 | | 0.2684 | 61.0 | 549 | 0.4733 | 0.2978 | | 0.2819 | 62.0 | 558 | 0.4969 | 0.2833 | | 0.2819 | 63.0 | 567 | 0.6202 | 0.2789 | | 0.2889 | 64.0 | 576 | 0.3955 | 0.2733 | | 0.2515 | 65.0 | 585 | 0.3806 | 0.2656 | | 0.2468 | 66.0 | 594 | 0.3473 | 0.2722 | | 0.2557 | 67.0 | 603 | 0.4170 | 0.2722 | | 0.2477 | 68.0 | 612 | 0.4749 | 0.2678 | | 0.2965 | 69.0 | 621 | 0.4387 | 0.2611 | | 0.2606 | 70.0 | 630 | 0.4586 | 0.2656 | | 0.2606 | 71.0 | 639 | 0.5755 | 0.2733 | | 0.2442 | 72.0 | 648 | 0.5582 | 0.2656 | | 0.347 | 73.0 | 657 | 0.3897 | 0.2711 | | 0.2444 | 74.0 | 666 | 0.3369 | 0.2533 | | 0.2811 | 75.0 | 675 | 0.3487 | 0.2578 | | 0.24 | 76.0 | 684 | 0.3692 | 0.2589 | | 0.2466 | 77.0 | 693 | 0.4567 | 0.2578 | | 0.2769 | 78.0 | 702 | 0.4041 | 0.2633 | | 0.2464 | 79.0 | 711 | 0.3813 | 0.2622 | | 0.2791 | 80.0 | 720 | 0.3990 | 0.2556 | | 0.2791 | 81.0 | 729 | 0.3997 | 0.2489 | | 0.2365 | 82.0 | 738 | 0.4537 | 0.2533 | | 0.2693 | 83.0 | 747 | 0.5943 | 0.2611 | | 0.2285 | 84.0 | 756 | 0.5805 | 0.2656 | | 0.2468 | 85.0 | 765 | 0.5609 | 0.2656 | | 0.2226 | 86.0 | 774 | 0.5948 | 0.2667 | | 0.2419 | 87.0 | 783 | 0.5910 | 0.2544 | | 0.2254 | 88.0 | 792 | 0.5741 | 0.26 | | 0.2083 | 89.0 | 801 | 0.4984 | 0.2611 | | 0.2318 | 90.0 | 810 | 0.5093 | 0.26 | | 0.2318 | 91.0 | 819 | 0.5284 | 0.2633 | | 0.2458 | 92.0 | 828 | 0.4885 | 0.2656 | | 0.2394 | 93.0 | 837 | 0.4818 | 0.2622 | | 0.2018 | 94.0 | 846 | 0.5037 | 0.26 | | 0.235 | 95.0 | 855 | 0.5011 | 0.2578 | | 0.2252 | 96.0 | 864 | 0.4931 | 0.2611 | | 0.2147 | 97.0 | 873 | 0.4881 | 0.2589 | | 0.2227 | 98.0 | 882 | 0.4956 | 0.2589 | | 0.2168 | 99.0 | 891 | 0.5097 | 0.2589 | | 0.2282 | 100.0 | 900 | 0.5174 | 0.26 | ### Framework versions - Transformers 4.25.0.dev0 - Pytorch 1.8.1+cu111 - Datasets 2.7.1.dev0 - Tokenizers 0.13.2