t5-base-finetuned-ehealth
This model is a fine-tuned version of t5-base on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 1.3953
- Rouge1: 16.9989
- Rouge2: 4.8395
- Rougel: 13.1702
- Rougelsum: 15.6472
- Gen Len: 19.0
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 100
Training results
Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
---|---|---|---|---|---|---|---|---|
No log | 1.0 | 22 | 4.2413 | 9.137 | 1.2333 | 6.9806 | 8.1957 | 18.6901 |
No log | 2.0 | 44 | 3.5352 | 9.5584 | 1.2176 | 7.2081 | 8.5048 | 18.8187 |
No log | 3.0 | 66 | 3.3124 | 9.9504 | 1.2105 | 7.4652 | 8.7962 | 18.8187 |
No log | 4.0 | 88 | 3.2065 | 10.3375 | 1.1847 | 7.7904 | 9.1801 | 18.8947 |
No log | 5.0 | 110 | 3.1208 | 10.777 | 1.326 | 8.1305 | 9.6488 | 18.8947 |
No log | 6.0 | 132 | 3.0495 | 11.1502 | 1.4947 | 8.4386 | 9.9076 | 18.924 |
No log | 7.0 | 154 | 2.9851 | 11.1759 | 1.5744 | 8.4744 | 9.9534 | 18.924 |
No log | 8.0 | 176 | 2.9232 | 10.5745 | 1.5079 | 8.1888 | 9.4731 | 18.8363 |
No log | 9.0 | 198 | 2.8663 | 10.3156 | 1.452 | 8.1662 | 9.385 | 18.8947 |
No log | 10.0 | 220 | 2.8110 | 10.5445 | 1.6067 | 8.3821 | 9.6755 | 18.8538 |
No log | 11.0 | 242 | 2.7625 | 11.0628 | 1.6957 | 8.7832 | 10.1425 | 18.8947 |
No log | 12.0 | 264 | 2.7129 | 10.9152 | 1.8386 | 8.7865 | 10.0545 | 18.8538 |
No log | 13.0 | 286 | 2.6680 | 10.8689 | 1.9024 | 8.6892 | 9.883 | 18.8889 |
No log | 14.0 | 308 | 2.6235 | 10.4118 | 1.9101 | 8.2442 | 9.4505 | 18.8947 |
No log | 15.0 | 330 | 2.5810 | 11.2578 | 2.0742 | 8.7641 | 10.2349 | 18.8947 |
No log | 16.0 | 352 | 2.5412 | 11.815 | 2.1727 | 9.2403 | 10.6655 | 18.9591 |
No log | 17.0 | 374 | 2.5056 | 11.8324 | 2.1849 | 9.2089 | 10.7361 | 18.9649 |
No log | 18.0 | 396 | 2.4710 | 11.4611 | 2.1406 | 8.9329 | 10.4319 | 18.8246 |
No log | 19.0 | 418 | 2.4365 | 12.0309 | 2.4387 | 9.3966 | 11.0327 | 18.8655 |
No log | 20.0 | 440 | 2.4039 | 11.9636 | 2.4332 | 9.3448 | 11.0055 | 18.8363 |
No log | 21.0 | 462 | 2.3734 | 12.709 | 2.6945 | 9.8722 | 11.572 | 18.7602 |
No log | 22.0 | 484 | 2.3414 | 13.2227 | 2.6249 | 10.1069 | 11.968 | 18.7895 |
3.1829 | 23.0 | 506 | 2.3132 | 13.3682 | 2.6082 | 10.1546 | 12.0317 | 18.8246 |
3.1829 | 24.0 | 528 | 2.2861 | 14.3195 | 3.0288 | 10.8036 | 12.8973 | 18.8713 |
3.1829 | 25.0 | 550 | 2.2592 | 14.1227 | 2.6271 | 10.6826 | 12.7174 | 18.9064 |
3.1829 | 26.0 | 572 | 2.2324 | 14.3697 | 2.8314 | 10.9239 | 13.0199 | 18.9064 |
3.1829 | 27.0 | 594 | 2.2054 | 14.4512 | 2.9546 | 11.0853 | 13.1193 | 18.9474 |
3.1829 | 28.0 | 616 | 2.1810 | 15.12 | 3.3732 | 11.5842 | 13.6805 | 18.9474 |
3.1829 | 29.0 | 638 | 2.1563 | 14.8242 | 3.2998 | 11.2467 | 13.3076 | 18.9474 |
3.1829 | 30.0 | 660 | 2.1333 | 15.0384 | 3.3988 | 11.4676 | 13.6825 | 18.9123 |
3.1829 | 31.0 | 682 | 2.1102 | 14.9877 | 3.3844 | 11.4417 | 13.5657 | 18.9591 |
3.1829 | 32.0 | 704 | 2.0884 | 14.9699 | 3.4128 | 11.4893 | 13.6109 | 18.9591 |
3.1829 | 33.0 | 726 | 2.0646 | 14.7391 | 3.0552 | 11.2351 | 13.3809 | 18.9591 |
3.1829 | 34.0 | 748 | 2.0419 | 14.9203 | 3.1074 | 11.2239 | 13.4966 | 18.9591 |
3.1829 | 35.0 | 770 | 2.0203 | 15.1875 | 3.2249 | 11.3843 | 13.8011 | 18.9591 |
3.1829 | 36.0 | 792 | 1.9988 | 15.1457 | 3.1865 | 11.5238 | 13.7114 | 18.9591 |
3.1829 | 37.0 | 814 | 1.9786 | 15.2334 | 3.3739 | 11.6124 | 13.8956 | 18.9591 |
3.1829 | 38.0 | 836 | 1.9580 | 15.7105 | 3.4331 | 11.8577 | 14.2217 | 18.9474 |
3.1829 | 39.0 | 858 | 1.9387 | 15.6612 | 3.5588 | 12.0279 | 14.2183 | 18.9474 |
3.1829 | 40.0 | 880 | 1.9210 | 15.8692 | 3.5665 | 12.0078 | 14.3505 | 18.9591 |
3.1829 | 41.0 | 902 | 1.9041 | 15.9888 | 3.6914 | 12.0342 | 14.3375 | 18.9591 |
3.1829 | 42.0 | 924 | 1.8834 | 15.9551 | 3.6863 | 12.0562 | 14.5444 | 18.9591 |
3.1829 | 43.0 | 946 | 1.8648 | 15.9107 | 3.9128 | 12.1663 | 14.5029 | 18.9591 |
3.1829 | 44.0 | 968 | 1.8468 | 15.9831 | 3.8588 | 12.196 | 14.5114 | 18.9591 |
3.1829 | 45.0 | 990 | 1.8290 | 15.9072 | 3.6844 | 12.1007 | 14.5031 | 18.9591 |
2.4484 | 46.0 | 1012 | 1.8127 | 15.9918 | 3.792 | 12.2569 | 14.5287 | 18.9591 |
2.4484 | 47.0 | 1034 | 1.7959 | 15.9685 | 3.7664 | 12.1033 | 14.473 | 18.9591 |
2.4484 | 48.0 | 1056 | 1.7799 | 15.7128 | 3.505 | 11.9947 | 14.216 | 18.9591 |
2.4484 | 49.0 | 1078 | 1.7636 | 15.8033 | 3.6874 | 12.1043 | 14.37 | 18.9591 |
2.4484 | 50.0 | 1100 | 1.7487 | 15.914 | 3.758 | 12.1635 | 14.4603 | 18.9591 |
2.4484 | 51.0 | 1122 | 1.7338 | 15.7088 | 3.7272 | 11.951 | 14.2862 | 18.9591 |
2.4484 | 52.0 | 1144 | 1.7202 | 15.7231 | 3.6274 | 12.0492 | 14.3036 | 18.9591 |
2.4484 | 53.0 | 1166 | 1.7081 | 15.6734 | 3.5837 | 11.9265 | 14.2674 | 18.9591 |
2.4484 | 54.0 | 1188 | 1.6935 | 15.6501 | 3.5574 | 11.8579 | 14.2387 | 18.9591 |
2.4484 | 55.0 | 1210 | 1.6793 | 15.8984 | 3.8029 | 12.0981 | 14.3888 | 18.9591 |
2.4484 | 56.0 | 1232 | 1.6666 | 15.7263 | 3.6691 | 12.0325 | 14.3152 | 18.9591 |
2.4484 | 57.0 | 1254 | 1.6516 | 15.8016 | 3.6151 | 12.0349 | 14.3556 | 18.9591 |
2.4484 | 58.0 | 1276 | 1.6385 | 15.8773 | 3.7501 | 12.1887 | 14.456 | 18.9591 |
2.4484 | 59.0 | 1298 | 1.6266 | 16.0252 | 3.8027 | 12.3099 | 14.5017 | 18.9591 |
2.4484 | 60.0 | 1320 | 1.6151 | 16.29 | 3.9544 | 12.5391 | 14.7691 | 18.9591 |
2.4484 | 61.0 | 1342 | 1.6034 | 16.2891 | 4.0512 | 12.5053 | 14.8155 | 18.9591 |
2.4484 | 62.0 | 1364 | 1.5925 | 16.1871 | 4.0482 | 12.4821 | 14.6986 | 18.9591 |
2.4484 | 63.0 | 1386 | 1.5812 | 16.1774 | 3.9903 | 12.4861 | 14.7798 | 18.9591 |
2.4484 | 64.0 | 1408 | 1.5716 | 16.1663 | 3.9399 | 12.4316 | 14.7449 | 18.9591 |
2.4484 | 65.0 | 1430 | 1.5623 | 16.4455 | 4.2777 | 12.7206 | 14.9193 | 18.9591 |
2.4484 | 66.0 | 1452 | 1.5517 | 16.466 | 4.2148 | 12.7613 | 15.052 | 18.9591 |
2.4484 | 67.0 | 1474 | 1.5414 | 16.5696 | 4.193 | 12.6949 | 15.1064 | 18.9591 |
2.4484 | 68.0 | 1496 | 1.5347 | 16.7602 | 4.4803 | 12.938 | 15.3339 | 18.9649 |
2.1379 | 69.0 | 1518 | 1.5278 | 16.6684 | 4.3943 | 12.9152 | 15.2626 | 18.9649 |
2.1379 | 70.0 | 1540 | 1.5193 | 16.7462 | 4.4151 | 12.9251 | 15.3619 | 18.9649 |
2.1379 | 71.0 | 1562 | 1.5104 | 16.658 | 4.4187 | 12.8792 | 15.2538 | 18.9591 |
2.1379 | 72.0 | 1584 | 1.5026 | 16.8475 | 4.481 | 13.0381 | 15.4041 | 18.9591 |
2.1379 | 73.0 | 1606 | 1.4944 | 16.9066 | 4.6433 | 13.1838 | 15.489 | 18.9591 |
2.1379 | 74.0 | 1628 | 1.4864 | 16.9434 | 4.6401 | 13.0527 | 15.4966 | 18.9591 |
2.1379 | 75.0 | 1650 | 1.4801 | 16.9744 | 4.694 | 13.1585 | 15.5739 | 19.0 |
2.1379 | 76.0 | 1672 | 1.4733 | 17.0546 | 4.6971 | 13.0968 | 15.633 | 19.0 |
2.1379 | 77.0 | 1694 | 1.4668 | 17.1603 | 4.7771 | 13.2896 | 15.7112 | 19.0 |
2.1379 | 78.0 | 1716 | 1.4607 | 17.086 | 4.7411 | 13.2587 | 15.6842 | 19.0 |
2.1379 | 79.0 | 1738 | 1.4552 | 17.0322 | 4.7652 | 13.2693 | 15.711 | 19.0 |
2.1379 | 80.0 | 1760 | 1.4493 | 17.1045 | 4.8492 | 13.2752 | 15.7876 | 19.0 |
2.1379 | 81.0 | 1782 | 1.4445 | 17.0275 | 4.8688 | 13.2621 | 15.7825 | 19.0 |
2.1379 | 82.0 | 1804 | 1.4392 | 17.0985 | 4.8148 | 13.2498 | 15.7718 | 19.0 |
2.1379 | 83.0 | 1826 | 1.4337 | 17.1395 | 4.8482 | 13.357 | 15.8122 | 19.0 |
2.1379 | 84.0 | 1848 | 1.4294 | 17.0411 | 4.8237 | 13.3126 | 15.7736 | 19.0 |
2.1379 | 85.0 | 1870 | 1.4254 | 17.1265 | 4.8691 | 13.3033 | 15.81 | 19.0 |
2.1379 | 86.0 | 1892 | 1.4212 | 16.9899 | 4.7712 | 13.1785 | 15.6416 | 19.0 |
2.1379 | 87.0 | 1914 | 1.4176 | 17.0389 | 4.7936 | 13.219 | 15.7048 | 19.0 |
2.1379 | 88.0 | 1936 | 1.4141 | 17.2266 | 4.9339 | 13.3935 | 15.8629 | 19.0 |
2.1379 | 89.0 | 1958 | 1.4108 | 17.0176 | 4.8752 | 13.2829 | 15.7145 | 19.0 |
2.1379 | 90.0 | 1980 | 1.4084 | 17.154 | 4.9912 | 13.3718 | 15.8255 | 19.0 |
1.9718 | 91.0 | 2002 | 1.4061 | 17.0783 | 4.9171 | 13.2617 | 15.7864 | 19.0 |
1.9718 | 92.0 | 2024 | 1.4037 | 17.0967 | 4.9393 | 13.2608 | 15.8054 | 19.0 |
1.9718 | 93.0 | 2046 | 1.4020 | 17.1524 | 4.995 | 13.332 | 15.8315 | 19.0 |
1.9718 | 94.0 | 2068 | 1.4001 | 17.1357 | 4.9699 | 13.3064 | 15.7932 | 19.0 |
1.9718 | 95.0 | 2090 | 1.3988 | 17.0758 | 4.8899 | 13.2231 | 15.7124 | 19.0 |
1.9718 | 96.0 | 2112 | 1.3976 | 16.9842 | 4.8395 | 13.173 | 15.653 | 19.0 |
1.9718 | 97.0 | 2134 | 1.3967 | 17.0425 | 4.8395 | 13.2243 | 15.6976 | 19.0 |
1.9718 | 98.0 | 2156 | 1.3960 | 16.9842 | 4.8395 | 13.173 | 15.653 | 19.0 |
1.9718 | 99.0 | 2178 | 1.3955 | 16.9842 | 4.8395 | 13.173 | 15.653 | 19.0 |
1.9718 | 100.0 | 2200 | 1.3953 | 16.9989 | 4.8395 | 13.1702 | 15.6472 | 19.0 |
Framework versions
- Transformers 4.31.0
- Pytorch 2.0.1+cu118
- Datasets 2.14.1
- Tokenizers 0.13.3
- Downloads last month
- 8
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for GuysTrans/t5-base-finetuned-ehealth
Base model
google-t5/t5-base