sumedh commited on
Commit
2d68f16
1 Parent(s): 7ed11eb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -6
README.md CHANGED
@@ -12,12 +12,28 @@ thumbnail: https://huggingface.co/front/thumbnails/distilbart_medium.png
12
 
13
  This checkpoint should be loaded into `BartForConditionalGeneration.from_pretrained`. See the [BART docs](https://huggingface.co/transformers/model_doc/bart.html?#transformers.BartForConditionalGeneration) for more information.
14
 
15
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
 
17
- ### Metrics for DistilBART models
18
 
19
- | Model Name | MM Params | Inference Time (MS) | Speedup | Rouge 1 | Rouge-L |
20
- |:---------------------------------|------------:|----------------------:|----------:|----------:|----------:|
21
- | distilbart-cnn-12-6-amazonreviews| 306 | 137 | 1.68 | 10.557| 33.37 |
22
 
23
- ---
 
 
 
 
12
 
13
  This checkpoint should be loaded into `BartForConditionalGeneration.from_pretrained`. See the [BART docs](https://huggingface.co/transformers/model_doc/bart.html?#transformers.BartForConditionalGeneration) for more information.
14
 
15
+ ### Training hyperparameters
16
+
17
+ The following hyperparameters were used during training:
18
+ - learning_rate: 2e-05
19
+ - train_batch_size: 4
20
+ - eval_batch_size: 4
21
+ - seed: 42
22
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
23
+ - lr_scheduler_type: linear
24
+ - num_epochs: 1
25
+ - mixed_precision_training: Native AMP
26
+
27
+ ### Training results
28
+
29
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
30
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:-------:|:---------:|:-------:|
31
+ | 1.2875 | 1.0 | 5754 | 1.6294 | 11.009 | 7.4618 | 10.5573 | 10.8087 | 58.3382 |
32
 
 
33
 
34
+ ### Framework versions
 
 
35
 
36
+ - Transformers 4.19.2
37
+ - Pytorch 1.11.0+cu113
38
+ - Datasets 2.2.2
39
+ - Tokenizers 0.12.1