Xuefeng Bai commited on
Commit
9eb10ce
1 Parent(s): 8597b67

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -0
README.md CHANGED
@@ -1,3 +1,63 @@
1
  ---
 
 
 
2
  license: mit
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language: en
3
+ tags:
4
+ - AMRBART
5
  license: mit
6
  ---
7
+
8
+ ## AMRBART (base-sized model)
9
+
10
+ AMRBART model is continually pre-trained on the English text and AMR Graphs based on the BART model. It was introduced in the paper: [Graph Pre-training for AMR Parsing and Generation](https://arxiv.org/pdf/2203.07836.pdf) by bai et al. in ACL 2022 and first released in [this repository](https://github.com/muyeby/AMRBART).
11
+
12
+ ## Model description
13
+
14
+ AMRBART follows the BART model which uses a transformer encoder-encoder architecture. AMRBART is pre-trained with 6 tasks:
15
+
16
+ + learning to reconstruct the text based on the corrupted text.
17
+ + learning to reconstruct AMR graphs based on the corrupted AMR graph.
18
+ + learning to reconstruct the text based on the corrupted text and its corresponding AMR graph.
19
+ + learning to reconstruct an AMR graph based on the corrupted AMR graph and its corresponding text.
20
+ + learning to reconstruct the text based on the corrupted text and its corresponding corrupted AMR graph.
21
+ + learning to reconstruct an AMR graph based on the corrupted AMR graph and its corresponding corrupted text.
22
+
23
+ AMRBART is particularly effective when fine-tuned for AMR parsing and AMR-to-text generation tasks.
24
+
25
+ ## Training data
26
+
27
+ The AMRBART model is pre-trained on [AMR3.0](https://catalog.ldc.upenn.edu/LDC2020T02), a dataset consisting of 55,635
28
+ training instances and [English Gigaword](https://catalog.ldc.upenn.edu/LDC2003T05) (we randomly sampled 200,000 sentences).
29
+
30
+ ## Intended uses & limitations
31
+
32
+ You can use the raw model for either AMR encoding or AMR parsing, but it's mostly intended to
33
+ be fine-tuned on a downstream task.
34
+
35
+ ## How to use
36
+ Here is how to initialize this model in PyTorch:
37
+
38
+ ```python
39
+ from transformers import BartModel
40
+ model = BartModel.from_pretrained("xfbai/AMRBART-base")
41
+ ```
42
+ Please refer to [this repository](https://github.com/muyeby/AMRBART) for tokenizer initialization and data preprocessing.
43
+
44
+
45
+ ## BibTeX entry and citation info
46
+ Please cite this paper if you find this model helpful
47
+
48
+ ```bibtex
49
+ @inproceedings{bai-etal-2022-graph,
50
+ title = "Graph Pre-training for {AMR} Parsing and Generation",
51
+ author = "Bai, Xuefeng and
52
+ Chen, Yulong and
53
+ Zhang, Yue",
54
+ booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
55
+ month = may,
56
+ year = "2022",
57
+ address = "Online",
58
+ publisher = "Association for Computational Linguistics",
59
+ url = "todo",
60
+ doi = "todo",
61
+ pages = "todo"
62
+ }
63
+ ```