File size: 6,653 Bytes
b222475
a33abfb
 
 
485e81e
a33abfb
 
 
 
 
 
 
 
 
 
 
 
485e81e
 
 
 
 
 
a62438a
 
 
 
909b679
 
 
 
a7f7c97
 
 
 
485e81e
909b679
485e81e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a62438a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
909b679
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a7f7c97
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b222475
 
a33abfb
 
 
 
 
 
 
 
 
 
 
 
e127746
 
48be4d3
e127746
 
 
 
a33abfb
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9560acf
a33abfb
 
 
 
 
 
 
 
 
 
 
 
 
 
e127746
 
f55a55b
4e4a865
f55a55b
 
 
a33abfb
689d11d
a33abfb
 
 
f55a55b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a33abfb
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
---
language:
- da
- sv
- 'no'
- is
license:
- cc-by-4.0
multilinguality:
- multilingual
size_categories:
- 10M<n<100M
task_categories:
- text-generation
- fill-mask
task_ids:
- language-modeling
pretty_name: ScandiReddit
configs:
- config_name: da
  data_files:
  - split: train
    path: da/train-*
- config_name: is
  data_files:
  - split: train
    path: is/train-*
- config_name: 'no'
  data_files:
  - split: train
    path: no/train-*
- config_name: sv
  data_files:
  - split: train
    path: sv/train-*
dataset_info:
- config_name: da
  features:
  - name: doc
    dtype: string
  - name: subreddit
    dtype: string
  - name: language
    dtype: string
  - name: language_confidence
    dtype: float64
  splits:
  - name: train
    num_bytes: 1386063758.1912868
    num_examples: 4965195
  download_size: 849659088
  dataset_size: 1386063758.1912868
- config_name: is
  features:
  - name: doc
    dtype: string
  - name: subreddit
    dtype: string
  - name: language
    dtype: string
  - name: language_confidence
    dtype: float64
  splits:
  - name: train
    num_bytes: 57698465.441296645
    num_examples: 206689
  download_size: 41270604
  dataset_size: 57698465.441296645
- config_name: 'no'
  features:
  - name: doc
    dtype: string
  - name: subreddit
    dtype: string
  - name: language
    dtype: string
  - name: language_confidence
    dtype: float64
  splits:
  - name: train
    num_bytes: 374200184.6740509
    num_examples: 1340470
  download_size: 286174352
  dataset_size: 374200184.6740509
- config_name: sv
  features:
  - name: doc
    dtype: string
  - name: subreddit
    dtype: string
  - name: language
    dtype: string
  - name: language_confidence
    dtype: float64
  splits:
  - name: train
    num_bytes: 1944996792.6933656
    num_examples: 6967420
  download_size: 1164301734
  dataset_size: 1944996792.6933656
---

# Dataset Card for ScandiReddit

## Dataset Description

- **Repository:** <https://github.com/alexandrainst/ScandiReddit>
- **Point of Contact:** [Dan Saattrup Nielsen](mailto:dan.nielsen@alexandra.dk)
- **Size of downloaded dataset files:** 2341 MB
- **Size of the generated dataset:** 3594 MB
- **Total amount of disk used:** 5935 MB

### Dataset Summary

ScandiReddit is a filtered and post-processed corpus consisting of comments from [Reddit](https://reddit.com/).

All Reddit comments from December 2005 up until October 2022 were downloaded through [PushShift](https://files.pushshift.io/reddit/comments/), after which these were filtered based on the FastText language detection model. Any comment which was classified as Danish (`da`), Norwegian (`no`), Swedish (`sv`) or Icelandic (`is`) with a confidence score above 70% was kept.

The resulting comments were then deduplicated, removing roughly 438,000 comments. 5,000 comments written by Reddit bots were removed, and roughly 189,000 comments belonging to inappropriate subreddits (explicit and drug-related) were also removed.

Lastly, we remove roughly 40,000 near-duplicate comments from the resulting corpus, where near-duplicate here means that the comments have more than 80% of their word 5-grams in common.


### Supported Tasks and Leaderboards

Training language models is the intended task for this dataset. No leaderboard is active at this point.


### Languages

The dataset is available in Danish (`da`), Swedish (`sv`), Norwegian (`no`) and Icelandic (`is`).


## Dataset Structure

### Data Instances

- **Size of downloaded dataset files:** 2341 MB
- **Size of the generated dataset:** 3594 MB
- **Total amount of disk used:** 5935 MB

An example from the dataset looks as follows.
```
{
    'doc': 'Bergen er ødelagt. Det er ikke moro mer.',
    'subreddit': 'Norway',
    'language': 'da',
    'language_confidence': 0.7472341656684875
}
```

### Data Fields

The data fields are the same among all splits.

- `doc`: a `string` feature.
- `subreddit`: a `string` feature.
- `language`: a `string` feature.
- `language_confidence`: a `float64` feature.

### Language Distribution

|   name   |  count   |
|----------|---------:|
| sv       | 6,967,420  |
| da       | 4,965,195  |
| no       | 1,340,470  |
| is       | 206,689    |
| total    | 13,479,774 |

### Top-50 Subreddit Distribution

|   name   |  count  |
|----------|--------:|
|sweden                  |4,881,483|
|Denmark                 |3,579,178|
|norge                   |1,281,655|
|svenskpolitik           | 771,960|
|InfluencergossipDK      | 649,910|
|swedishproblems         | 339,683|
|Iceland                 | 183,488|
|dkfinance               | 113,860|
|unket                   |  81,077|
|DanishEnts              |  69,055|
|dankmark                |  62,928|
|swedents                |  58,576|
|scandinavia             |  57,136|
|Allsvenskan             |  56,006|
|Gothenburg              |  54,395|
|stockholm               |  51,016|
|ISKbets                 |  47,944|
|Sverige                 |  39,552|
|SWARJE                  |  34,691|
|GossipDK                |  29,332|
|NorskFotball            |  28,571|
|Superligaen             |  23,641|
|Aarhus                  |  22,516|
|Svenska                 |  20,561|
|newsdk                  |  19,893|
|AskReddit               |  16,672|
|copenhagen              |  16,668|
|okpolarncp              |  16,583|
|SwedditUniversalis      |  15,990|
|Sveriges_politik        |  15,058|
|intresseklubben         |  13,246|
|Aktiemarknaden          |  13,202|
|soccer                  |  12,637|
|teenagers               |  10,845|
|Norway                  |  10,680|
|europe                  |  10,247|
|Matinbum                |   9,792|
|oslo                    |   9,650|
|iksdagen                |   9,232|
|Asksweddit              |   8,851|
|Forsvaret               |   8,641|
|Sverigesforsvarsmakt    |   8,469|
|memes                   |   8,299|
|Danish                  |   8,268|
|DANMAG                  |   8,214|
|PewdiepieSubmissions    |   7,800|
|sweddpolitik            |   7,646|
|pinsamt                 |   7,318|
|arbetarrorelsen         |   7,317|
|Ishockey                |   6,824|


## Dataset Creation

### Curation Rationale

The Scandinavian languages do not have many open source social media datasets.

### Source Data

The raw Reddit data was collected through [PushShift](https://files.pushshift.io/reddit/comments/).


## Additional Information

### Dataset Curators

[Dan Saattrup Nielsen](https://saattrupdan.github.io/) from the [The Alexandra
Institute](https://alexandra.dk/) curated this dataset.

### Licensing Information

The dataset is licensed under the [CC BY 4.0
license](https://creativecommons.org/licenses/by/4.0/).