File size: 1,828 Bytes
cd2cbd7
be43789
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4fb2b90
be43789
4fb2b90
be43789
4fb2b90
be43789
4fb2b90
be43789
4fb2b90
be43789
4fb2b90
be43789
4fb2b90
be43789
4fb2b90
be43789
4fb2b90
 
 
be43789
4fb2b90
 
 
be43789
cd2cbd7
 
 
6f77a9c
 
be43789
 
6f77a9c
 
90e2a29
cd2cbd7
be43789
 
 
 
 
 
cd2cbd7
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
---
language: en
tags:
- abstractive summarization
model-index:
- name: kubershahi/pegasus-inshorts
  results:
  - task:
      type: abstractitive summarization
      name: abstractive summarization
    dataset:
      name: inshorts 
      type: inshorts
      config: inshorts
      split: train
    metrics:
    - name: ROUGE-L-P
      type: rouge
      value: 0.01074
      verified: true
    - name: ROUGE-L-R
      type: rouge
      value: 0.08284
      verified: true
    - name: ROUGE-L-F
      type: rouge
      value: 0.08284
      verified: true
    - name: ROUGE-1-P
      type: rouge
      value: 0.01074
      verified: true
    - name: ROUGE-1-R
      type: rouge
      value: 0.08284
      verified: true
    - name: ROUGE-1-f
      type: rouge
      value: 0.08284
      verified: true
---


# Problem Statment: 

Given a news article, generate a summary of two-to-three sentences and a headline for the article. The summary should be abstractive rather than extractive.
In abstractive summarization, new sentences are generated as part of the summary and the sentences in the summary might not be present in the news article.


# Model Description

This model builds on the [google/pegasus-large](https://huggingface.co/google/pegasus-large) model by finetuning it on a custom summary-headline dataset called [inshorts](https://github.com/kubershahi/ashoka-aml/blob/master/dataset/news_headline.csv).
After finetuning, to generate an appropriate headline of an article, get the summary of the article first from the pegasus-large model and then pass the summary through this model. 
The two-way approach was taken to get apt headline from summary rather then generating the headline from the pegasus-large itself. 


For more details about the project, click [here](https://github.com/kubershahi/ashoka-aml).