AndreaSimeri
commited on
Commit
•
e265f10
1
Parent(s):
3f66ae2
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,42 @@
|
|
1 |
-
---
|
2 |
-
license: cc-by-nc-nd-4.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-nc-nd-4.0
|
3 |
+
pipeline_tag: text-classification
|
4 |
+
tags:
|
5 |
+
- deep learning
|
6 |
+
- law article retrieval
|
7 |
+
- natural language processing
|
8 |
+
- BERT
|
9 |
+
- information retrieval
|
10 |
+
- legal ai
|
11 |
+
- legal bert
|
12 |
+
- gdpr
|
13 |
+
- general data protection regulation
|
14 |
+
language:
|
15 |
+
- en
|
16 |
+
library_name: transformers
|
17 |
+
---
|
18 |
+
|
19 |
+
### Abstract
|
20 |
+
The General Data Protection Regulation (GDPR) is an European regulation on data protection and privacy for all individuals within the European Union (EU) and the European Economic Area (EEA), and for all foreign subjects dealing with European citizens data. Therefore, the GDPR has important legislation implications that hold beyond EU member states. In this paper, we address the problem of GDPR article retrieval through the use of pre-trained language models (PLMs). Our approach features several key aspects, which include both domain-general and domain-specific pre-trained BERT models, further powered by self-supervised task-adaptive pre-training stages, with or without data enrichment based on recitals. Our study endeavors to demonstrate the potential of PLMs in addressing the challenges posed by the GDPR’s intricate legal framework, thus ultimately facilitating efficient access to GDPR provisions for government agencies, law firms, legal professionals, and citizens alike.
|
21 |
+
### GDPR Article Retrieval based on Domain-adaptive and Task-adaptive Legal Pre-trained Language Models.
|
22 |
+
|
23 |
+
![image/webp](https://cdn-uploads.huggingface.co/production/uploads/62867cb4504d3770030ae173/Bn1qvPxZVLmM7tdyCYWzi.webp)
|
24 |
+
|
25 |
+
### BibTeX Entry and Citation Info
|
26 |
+
```
|
27 |
+
@article{Lamberta,
|
28 |
+
author = {Andrea Tagarelli and Andrea Simeri},
|
29 |
+
title = {{Unsupervised law article mining based on deep pre-trained language representation models with application to the Italian civil code}},
|
30 |
+
journal = {Artif. Intell. Law},
|
31 |
+
volume = {30(3)},
|
32 |
+
pages = {417--473. Published: 15 September 2021},
|
33 |
+
year = {2022},
|
34 |
+
doi ={10.1007/s10506-021-09301-8}
|
35 |
+
}
|
36 |
+
|
37 |
+
```
|
38 |
+
|
39 |
+
### References
|
40 |
+
- Tagarelli, A., Simeri, A. Unsupervised law article mining based on deep pre-trained language representation models with application to the Italian civil code. Artif Intell Law 30, 417–473 (2022). https://doi.org/10.1007/s10506-021-09301-8
|
41 |
+
|
42 |
+
---
|