kimsan0622 commited on
Commit
6a14d9e
1 Parent(s): 0602671

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +87 -0
README.md ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - ko
5
+ tags:
6
+ - generated_from_trainer
7
+ datasets:
8
+ - >-
9
+ KETI-AIR/aihub_koenzh_food_translation,KETI-AIR/aihub_scitech_translation,KETI-AIR/aihub_scitech20_translation,KETI-AIR/aihub_socialtech20_translation,KETI-AIR/aihub_spoken_language_translation
10
+ metrics:
11
+ - bleu
12
+ model-index:
13
+ - name: en2ko
14
+ results:
15
+ - task:
16
+ name: Translation
17
+ type: translation
18
+ dataset:
19
+ name: >-
20
+ KETI-AIR/aihub_koenzh_food_translation,KETI-AIR/aihub_scitech_translation,KETI-AIR/aihub_scitech20_translation,KETI-AIR/aihub_socialtech20_translation,KETI-AIR/aihub_spoken_language_translation
21
+ koen,none,none,none,none
22
+ type: >-
23
+ KETI-AIR/aihub_koenzh_food_translation,KETI-AIR/aihub_scitech_translation,KETI-AIR/aihub_scitech20_translation,KETI-AIR/aihub_socialtech20_translation,KETI-AIR/aihub_spoken_language_translation
24
+ args: koen,none,none,none,none
25
+ metrics:
26
+ - name: Bleu
27
+ type: bleu
28
+ value: 42.463
29
+ license: apache-2.0
30
+ pipeline_tag: translation
31
+ ---
32
+
33
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
34
+ should probably proofread and complete it, then remove this comment. -->
35
+
36
+ # en2ko
37
+
38
+ This model is a fine-tuned version of [KETI-AIR/long-ke-t5-base](https://huggingface.co/KETI-AIR/long-ke-t5-base) on the KETI-AIR/aihub_koenzh_food_translation,KETI-AIR/aihub_scitech_translation,KETI-AIR/aihub_scitech20_translation,KETI-AIR/aihub_socialtech20_translation,KETI-AIR/aihub_spoken_language_translation koen,none,none,none,none dataset.
39
+ It achieves the following results on the evaluation set:
40
+ - Loss: 0.6000
41
+ - Bleu: 42.463
42
+ - Gen Len: 30.6512
43
+
44
+ ## Model description
45
+
46
+ More information needed
47
+
48
+ ## Intended uses & limitations
49
+
50
+ More information needed
51
+
52
+ ## Training and evaluation data
53
+
54
+ More information needed
55
+
56
+ ## Training procedure
57
+
58
+ ### Training hyperparameters
59
+
60
+ The following hyperparameters were used during training:
61
+ - learning_rate: 0.001
62
+ - train_batch_size: 16
63
+ - eval_batch_size: 16
64
+ - seed: 42
65
+ - distributed_type: multi-GPU
66
+ - num_devices: 8
67
+ - total_train_batch_size: 128
68
+ - total_eval_batch_size: 128
69
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
70
+ - lr_scheduler_type: linear
71
+ - num_epochs: 3.0
72
+
73
+ ### Training results
74
+
75
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
76
+ |:-------------:|:-----:|:------:|:---------------:|:-------:|:-------:|
77
+ | 0.6989 | 1.0 | 93762 | 0.6666 | 20.3697 | 18.1258 |
78
+ | 0.6143 | 2.0 | 187524 | 0.6181 | 21.2903 | 18.1428 |
79
+ | 0.5544 | 3.0 | 281286 | 0.6000 | 21.9763 | 18.1424 |
80
+
81
+
82
+ ### Framework versions
83
+
84
+ - Transformers 4.25.1
85
+ - Pytorch 1.12.0
86
+ - Datasets 2.8.0
87
+ - Tokenizers 0.13.2