--- library_name: transformers language: - sn license: cc-by-nc-4.0 base_model: facebook/mms-300m tags: - generated_from_trainer datasets: - DigitalUmuganda_Afrivoice/Shona metrics: - wer model-index: - name: facebook/mms-300m results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: DigitalUmuganda type: DigitalUmuganda_Afrivoice/Shona metrics: - name: Wer type: wer value: 0.3497254482725108 --- # facebook/mms-300m This model is a fine-tuned version of [facebook/mms-300m](https://huggingface.co/facebook/mms-300m) on the DigitalUmuganda dataset. It achieves the following results on the evaluation set: - Loss: 0.6190 - Wer: 0.3497 - Cer: 0.0791 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 32 - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 150 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | Cer | |:-------------:|:-----:|:-----:|:---------------:|:------:|:------:| | 22.4497 | 1.0 | 273 | 2.9247 | 1.0 | 1.0 | | 8.283 | 2.0 | 546 | 0.5185 | 0.5978 | 0.1175 | | 1.5316 | 3.0 | 819 | 0.2757 | 0.3547 | 0.0632 | | 1.0241 | 4.0 | 1092 | 0.2358 | 0.3222 | 0.0570 | | 0.9072 | 5.0 | 1365 | 0.2283 | 0.3081 | 0.0560 | | 0.8161 | 6.0 | 1638 | 0.2201 | 0.2945 | 0.0527 | | 0.7469 | 7.0 | 1911 | 0.2122 | 0.2818 | 0.0487 | | 0.6981 | 8.0 | 2184 | 0.2191 | 0.2946 | 0.0519 | | 0.6454 | 9.0 | 2457 | 0.2122 | 0.2892 | 0.0504 | | 0.6094 | 10.0 | 2730 | 0.2206 | 0.2952 | 0.0498 | | 0.5649 | 11.0 | 3003 | 0.2164 | 0.2732 | 0.0482 | | 0.5197 | 12.0 | 3276 | 0.2143 | 0.2728 | 0.0479 | | 0.4793 | 13.0 | 3549 | 0.2344 | 0.2766 | 0.0485 | | 0.4368 | 14.0 | 3822 | 0.2266 | 0.2751 | 0.0486 | | 0.3977 | 15.0 | 4095 | 0.2358 | 0.2669 | 0.0458 | | 0.3667 | 16.0 | 4368 | 0.2447 | 0.2793 | 0.0474 | | 0.3335 | 17.0 | 4641 | 0.2547 | 0.2692 | 0.0461 | | 0.3044 | 18.0 | 4914 | 0.2566 | 0.2684 | 0.0463 | | 0.2801 | 19.0 | 5187 | 0.2698 | 0.2737 | 0.0473 | | 0.2514 | 20.0 | 5460 | 0.2967 | 0.2790 | 0.0480 | | 0.2314 | 21.0 | 5733 | 0.2855 | 0.2708 | 0.0462 | | 0.2092 | 22.0 | 6006 | 0.3074 | 0.2785 | 0.0476 | | 0.1991 | 23.0 | 6279 | 0.2953 | 0.2681 | 0.0449 | | 0.183 | 24.0 | 6552 | 0.3118 | 0.2712 | 0.0457 | | 0.1706 | 25.0 | 6825 | 0.3108 | 0.2654 | 0.0454 | | 0.1573 | 26.0 | 7098 | 0.3310 | 0.2730 | 0.0463 | | 0.1538 | 27.0 | 7371 | 0.3142 | 0.2767 | 0.0463 | | 0.1489 | 28.0 | 7644 | 0.3337 | 0.2666 | 0.0458 | | 0.1404 | 29.0 | 7917 | 0.3176 | 0.2728 | 0.0462 | | 0.1309 | 30.0 | 8190 | 0.3431 | 0.2707 | 0.0461 | | 0.1246 | 31.0 | 8463 | 0.3368 | 0.2725 | 0.0458 | | 0.1216 | 32.0 | 8736 | 0.3490 | 0.2648 | 0.0448 | | 0.1205 | 33.0 | 9009 | 0.3436 | 0.2627 | 0.0447 | | 0.1145 | 34.0 | 9282 | 0.3384 | 0.2693 | 0.0454 | | 0.1117 | 35.0 | 9555 | 0.3500 | 0.2610 | 0.0443 | | 0.1058 | 36.0 | 9828 | 0.3478 | 0.2608 | 0.0441 | | 0.1039 | 37.0 | 10101 | 0.3434 | 0.2590 | 0.0438 | | 0.1 | 38.0 | 10374 | 0.3507 | 0.2609 | 0.0438 | | 0.0987 | 39.0 | 10647 | 0.3553 | 0.2610 | 0.0443 | | 0.0977 | 40.0 | 10920 | 0.3691 | 0.2683 | 0.0451 | | 0.0972 | 41.0 | 11193 | 0.3508 | 0.2657 | 0.0450 | | 0.0946 | 42.0 | 11466 | 0.3445 | 0.2605 | 0.0445 | | 0.0934 | 43.0 | 11739 | 0.3541 | 0.2628 | 0.0445 | | 0.0896 | 44.0 | 12012 | 0.3424 | 0.2600 | 0.0437 | | 0.0844 | 45.0 | 12285 | 0.3504 | 0.2599 | 0.0439 | | 0.0847 | 46.0 | 12558 | 0.3630 | 0.2588 | 0.0433 | | 0.0809 | 47.0 | 12831 | 0.3650 | 0.2565 | 0.0435 | | 0.0812 | 48.0 | 13104 | 0.3632 | 0.2665 | 0.0446 | | 0.0783 | 49.0 | 13377 | 0.3585 | 0.2629 | 0.0435 | | 0.0739 | 50.0 | 13650 | 0.3722 | 0.2650 | 0.0435 | | 0.0772 | 51.0 | 13923 | 0.3627 | 0.2607 | 0.0433 | | 0.0746 | 52.0 | 14196 | 0.3712 | 0.2600 | 0.0439 | | 0.0713 | 53.0 | 14469 | 0.3540 | 0.2664 | 0.0445 | | 0.073 | 54.0 | 14742 | 0.3764 | 0.2571 | 0.0433 | | 0.0712 | 55.0 | 15015 | 0.3625 | 0.2533 | 0.0428 | | 0.0679 | 56.0 | 15288 | 0.3818 | 0.2603 | 0.0435 | | 0.0701 | 57.0 | 15561 | 0.3921 | 0.2565 | 0.0434 | | 0.0672 | 58.0 | 15834 | 0.3807 | 0.2584 | 0.0434 | | 0.0681 | 59.0 | 16107 | 0.3787 | 0.2575 | 0.0438 | | 0.063 | 60.0 | 16380 | 0.3818 | 0.2561 | 0.0428 | | 0.062 | 61.0 | 16653 | 0.3919 | 0.2567 | 0.0429 | | 0.0613 | 62.0 | 16926 | 0.4038 | 0.2598 | 0.0431 | | 0.0617 | 63.0 | 17199 | 0.3885 | 0.2566 | 0.0431 | | 0.0618 | 64.0 | 17472 | 0.3949 | 0.2557 | 0.0430 | | 0.0572 | 65.0 | 17745 | 0.3839 | 0.2529 | 0.0418 | ### Framework versions - Transformers 4.47.0.dev0 - Pytorch 2.1.0+cu118 - Datasets 3.0.2 - Tokenizers 0.20.1