andreiujica
commited on
Commit
•
2684b92
1
Parent(s):
de03f06
Update README.md
Browse files
README.md
CHANGED
@@ -60,15 +60,16 @@ It achieves the following results on the evaluation set:
|
|
60 |
|
61 |
## Model description
|
62 |
|
63 |
-
|
64 |
|
65 |
## Intended uses & limitations
|
66 |
|
67 |
-
|
|
|
|
|
68 |
|
69 |
-
|
70 |
-
|
71 |
-
More information needed
|
72 |
|
73 |
## Training procedure
|
74 |
|
|
|
60 |
|
61 |
## Model description
|
62 |
|
63 |
+
This model is a long-form text generation model based on the Longformer Encoder-Decoder (LED) architecture. The LED architecture extends the Transformer model to handle long documents by incorporating sparse attention mechanisms. This makes it suitable for tasks such as summarization of lengthy patent documents, where traditional models might struggle with context length limitations. The model has been fine-tuned on the BigPatent dataset, a large collection of patent documents, to enhance its performance in generating concise and informative summaries.
|
64 |
|
65 |
## Intended uses & limitations
|
66 |
|
67 |
+
Intended uses
|
68 |
+
- Patent summarization: Generate concise summaries of patent documents.
|
69 |
+
- Long document summarization: Useful for summarizing other types of long-form documents beyond patents.
|
70 |
|
71 |
+
Limitations
|
72 |
+
- Context length: Although LED handles long documents better than standard Transformers, extremely lengthy documents might still present challenges.
|
|
|
73 |
|
74 |
## Training procedure
|
75 |
|