Datasets:

Languages:
English
ArXiv:
License:
File size: 9,298 Bytes
ade652b
 
 
 
 
2551e87
ade652b
2551e87
833d07e
ade652b
 
 
 
 
 
 
91133d2
ade652b
91133d2
ade652b
 
1d6f5c3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b634532
 
 
1d6f5c3
 
 
 
 
ade652b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
833d07e
ade652b
 
 
 
 
 
 
 
 
 
 
 
 
1d6f5c3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
---
annotations_creators:
- expert-generated
language_creators:
- found
language:
- en
license:
- cc-by-nc-nd-4.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- extended|other-wider
task_categories:
- object-detection
task_ids:
- face-detection
paperswithcode_id: wider-face-1
pretty_name: WIDER FACE
dataset_info:
  features:
  - name: image
    dtype: image
  - name: faces
    sequence:
    - name: bbox
      sequence: float32
      length: 4
    - name: blur
      dtype:
        class_label:
          names:
            0: clear
            1: normal
            2: heavy
    - name: expression
      dtype:
        class_label:
          names:
            0: typical
            1: exaggerate
    - name: illumination
      dtype:
        class_label:
          names:
            0: normal
            1: 'exaggerate '
    - name: occlusion
      dtype:
        class_label:
          names:
            0: 'no'
            1: partial
            2: heavy
    - name: pose
      dtype:
        class_label:
          names:
            0: typical
            1: atypical
    - name: invalid
      dtype: bool
  splits:
  - name: train
    num_bytes: 12049881
    num_examples: 12880
  - name: test
    num_bytes: 3761103
    num_examples: 16097
  - name: validation
    num_bytes: 2998735
    num_examples: 3226
  download_size: 3676086479
  dataset_size: 18809719
---

# Dataset Card for WIDER FACE

## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
  - [Dataset Summary](#dataset-summary)
  - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
  - [Languages](#languages)
- [Dataset Structure](#dataset-structure)
  - [Data Instances](#data-instances)
  - [Data Fields](#data-fields)
  - [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
  - [Curation Rationale](#curation-rationale)
  - [Source Data](#source-data)
  - [Annotations](#annotations)
  - [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
  - [Social Impact of Dataset](#social-impact-of-dataset)
  - [Discussion of Biases](#discussion-of-biases)
  - [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
  - [Dataset Curators](#dataset-curators)
  - [Licensing Information](#licensing-information)
  - [Citation Information](#citation-information)
  - [Contributions](#contributions)

## Dataset Description

- **Homepage:** http://shuoyang1213.me/WIDERFACE/index.html
- **Repository:**
- **Paper:** [WIDER FACE: A Face Detection Benchmark](https://arxiv.org/abs/1511.06523)
- **Leaderboard:** http://shuoyang1213.me/WIDERFACE/WiderFace_Results.html
- **Point of Contact:** shuoyang.1213@gmail.com

### Dataset Summary

WIDER FACE dataset is a face detection benchmark dataset, of which images are
selected from the publicly available WIDER dataset. We choose 32,203 images and
label 393,703 faces with a high degree of variability in scale, pose and
occlusion as depicted in the sample images. WIDER FACE dataset is organized
based on 61 event classes. For each event class, we randomly select 40%/10%/50%
data as training, validation and testing sets. We adopt the same evaluation
metric employed in the PASCAL VOC dataset. Similar to MALF and Caltech datasets,
we do not release bounding box ground truth for the test images. Users are
required to submit final prediction files, which we shall proceed to evaluate.

### Supported Tasks and Leaderboards

- `face-detection`: The dataset can be used to train a model for Face Detection. More information on evaluating the model's performance can be found [here](http://shuoyang1213.me/WIDERFACE/WiderFace_Results.html).

### Languages

English

## Dataset Structure

### Data Instances

A data point comprises an image and its face annotations.

```
{
  'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=1024x755 at 0x19FA12186D8>, 'faces': {
    'bbox': [
      [178.0, 238.0, 55.0, 73.0],
      [248.0, 235.0, 59.0, 73.0],
      [363.0, 157.0, 59.0, 73.0],
      [468.0, 153.0, 53.0, 72.0],
      [629.0, 110.0, 56.0, 81.0],
      [745.0, 138.0, 55.0, 77.0]
    ], 
    'blur': [2, 2, 2, 2, 2, 2],
    'expression': [0, 0, 0, 0, 0, 0],
    'illumination': [0, 0, 0, 0, 0, 0],
    'occlusion': [1, 2, 1, 2, 1, 2],
    'pose': [0, 0, 0, 0, 0, 0],
    'invalid': [False, False, False, False, False, False]
  }
}
```

### Data Fields

- `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`
- `faces`: a dictionary of face attributes for the faces present on the image
  - `bbox`: the bounding box of each face (in the [coco](https://albumentations.ai/docs/getting_started/bounding_boxes_augmentation/#coco) format)
  - `blur`: the blur level of each face, with possible values including `clear` (0), `normal` (1) and `heavy`
  - `expression`: the facial expression of each face, with possible values including `typical` (0) and `exaggerate` (1)
  - `illumination`: the lightning condition of each face, with possible values including `normal` (0) and `exaggerate` (1)
  - `occlusion`: the level of occlusion of each face, with possible values including `no` (0), `partial` (1) and `heavy` (2)
  - `pose`: the pose of each face, with possible values including `typical` (0) and `atypical` (1)
  - `invalid`: whether the image is valid or invalid.

### Data Splits

The data is split into training, validation and testing set. WIDER FACE dataset is organized
based on 61 event classes. For each event class, 40%/10%/50%
data is randomly selected as training, validation and testing sets. The training set contains 12880 images, the validation set 3226 images and test set 16097 images.

## Dataset Creation

### Curation Rationale

The curators state that the current face detection datasets typically contain a few thousand faces, with limited variations in pose, scale, facial expression, occlusion, and background clutters,
making it difficult to assess for real world performance. They argue that the limitations of datasets have partially contributed to the failure of some algorithms in coping
with heavy occlusion, small scale, and atypical pose.

### Source Data

#### Initial Data Collection and Normalization

WIDER FACE dataset is a subset of the WIDER dataset.
The images in WIDER were collected in the following three steps: 1) Event categories
were defined and chosen following the Large Scale Ontology for Multimedia (LSCOM) [22], which provides around 1000 concepts relevant to video event analysis. 2) Images
are retrieved using search engines like Google and Bing. For
each category, 1000-3000 images were collected. 3) The
data were cleaned by manually examining all the images
and filtering out images without human face. Then, similar
images in each event category were removed to ensure large
diversity in face appearance. A total of 32203 images are
eventually included in the WIDER FACE dataset.

#### Who are the source language producers?

The images are selected from publicly available WIDER dataset.

### Annotations

#### Annotation process

The curators label the bounding boxes for all
the recognizable faces in the WIDER FACE dataset. The
bounding box is required to tightly contain the forehead,
chin, and cheek.. If a face is occluded, they still label it with a bounding box but with an estimation on the scale of occlusion. Similar to the PASCAL VOC dataset [6], they assign an ’Ignore’ flag to the face
which is very difficult to be recognized due to low resolution and small scale (10 pixels or less). After annotating
the face bounding boxes, they further annotate the following
attributes: pose (typical, atypical) and occlusion level (partial, heavy). Each annotation is labeled by one annotator
and cross-checked by two different people.

#### Who are the annotators?

Shuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang.

### Personal and Sensitive Information

[More Information Needed]

## Considerations for Using the Data

### Social Impact of Dataset

[More Information Needed]

### Discussion of Biases

[More Information Needed]

### Other Known Limitations

[More Information Needed]

## Additional Information

### Dataset Curators

Shuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang

### Licensing Information

[Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0)](https://creativecommons.org/licenses/by-nc-nd/4.0/).

### Citation Information

```
@inproceedings{yang2016wider,
    Author = {Yang, Shuo and Luo, Ping and Loy, Chen Change and Tang, Xiaoou},
    Booktitle = {IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
    Title = {WIDER FACE: A Face Detection Benchmark},
    Year = {2016}}
```

### Contributions

Thanks to [@mariosasko](https://github.com/mariosasko) for adding this dataset.