Datasets:
IGNF
/

ArXiv:
License:
AGarioud commited on
Commit
67ffafe
1 Parent(s): d01abd8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -14
README.md CHANGED
@@ -10,8 +10,7 @@ size_categories:
10
  # Datset Card for FLAIR land-cover semantic segmentation
11
 
12
  ## Context & Data
13
- <hr style='margin-top:-1em' />
14
-
15
  The hereby FLAIR (#2) dataset is sampled countrywide and is composed of over 20 billion annotated pixels of very high resolution aerial imagery at 0.2 m spatial resolution, acquired over three years and different months (spatio-temporal domains).
16
  Aerial imagery patches consist of 5 channels (RVB-Near Infrared-Elevation) and have corresponding annotation (with 19 semantic classes or 13 for the baselines).
17
  Furthermore, to integrate broader spatial context and temporal information, high resolution Sentinel-2 satellite 1-year time series with 10 spectral band are also provided.
@@ -168,9 +167,9 @@ The dataset covers 50 spatial domains, encompassing 916 areas spanning 817 km².
168
 
169
  <br><br>
170
 
171
- ## Dataset Structure
172
- <hr style='margin-top:-1em' />
173
 
 
 
174
  The FLAIR dataset consists of 77 762 patches. Each patch includes a high-resolution aerial image (512x512) at 0.2 m, a yearly satellite image time series (40x40 by default by wider areas are provided) with a spatial resolution of 10 m
175
  and associated cloud and snow masks, and pixel-precise elevation and land cover annotations at 0.2 m resolution (512x512).
176
 
@@ -249,8 +248,7 @@ Official domain split: <br/>
249
  <br><br>
250
 
251
  ## Baseline code
252
- <hr style='margin-top:-1em' />
253
-
254
  We propose the U-T&T model, a two-branch architecture that combines spatial and temporal information from very high-resolution aerial images and high-resolution satellite images into a single output. The U-Net architecture is employed for the spatial/texture branch, using a ResNet34 backbone model pre-trained on ImageNet. For the spatio-temporal branch,
255
  the U-TAE architecture incorporates a Temporal self-Attention Encoder (TAE) to explore the spatial and temporal characteristics of the Sentinel-2 time series data,
256
  applying attention masks at different resolutions during decoding. This model allows for the fusion of learned information from both sources,
@@ -274,8 +272,7 @@ domains_test : ["D015_2020","D022_2021","D026_2020","D036_2020","D061_2020","D06
274
 
275
 
276
  ## Reference
277
- <hr style='margin-top:-1em' />
278
-
279
  Please include a citation to the following article if you use the FLAIR dataset:
280
 
281
  ```
@@ -289,20 +286,17 @@ Please include a citation to the following article if you use the FLAIR dataset:
289
  ```
290
 
291
  ## Acknowledgment
292
- <hr style='margin-top:-1em' />
293
-
294
  This work was performed using HPC/AI resources from GENCI-IDRIS (Grant 2022-A0131013803). This work was supported by the project "Copernicus / FPCUP” of the European Union, by the French Space Agency (CNES) and by Connect by CNES.<br>
295
 
296
 
297
  ## Contact
298
- <hr style='margin-top:-1em' />
299
-
300
  If you have any questions, issues or feedback, you can contact us at: ai-challenge@ign.fr
301
 
302
 
303
  ## Dataset license
304
- <hr style='margin-top:-1em' />
305
-
306
  The "OPEN LICENCE 2.0/LICENCE OUVERTE" is a license created by the French government specifically for the purpose of facilitating the dissemination of open data by public administration.<br/>
307
  This licence is governed by French law.<br/>
308
  This licence has been designed to be compatible with any free licence that at least requires an acknowledgement of authorship, and specifically with the previous version of this licence as well as with the following licences: United Kingdom’s “Open Government Licence” (OGL), Creative Commons’ “Creative Commons Attribution” (CC-BY) and Open Knowledge Foundation’s “Open Data Commons Attribution” (ODC-BY).
 
10
  # Datset Card for FLAIR land-cover semantic segmentation
11
 
12
  ## Context & Data
13
+ <hr style='margin-top:-1em; margin-bottom:0' />
 
14
  The hereby FLAIR (#2) dataset is sampled countrywide and is composed of over 20 billion annotated pixels of very high resolution aerial imagery at 0.2 m spatial resolution, acquired over three years and different months (spatio-temporal domains).
15
  Aerial imagery patches consist of 5 channels (RVB-Near Infrared-Elevation) and have corresponding annotation (with 19 semantic classes or 13 for the baselines).
16
  Furthermore, to integrate broader spatial context and temporal information, high resolution Sentinel-2 satellite 1-year time series with 10 spectral band are also provided.
 
167
 
168
  <br><br>
169
 
 
 
170
 
171
+ ## Dataset Structure
172
+ <hr style='margin-top:-1em; margin-bottom:0' />
173
  The FLAIR dataset consists of 77 762 patches. Each patch includes a high-resolution aerial image (512x512) at 0.2 m, a yearly satellite image time series (40x40 by default by wider areas are provided) with a spatial resolution of 10 m
174
  and associated cloud and snow masks, and pixel-precise elevation and land cover annotations at 0.2 m resolution (512x512).
175
 
 
248
  <br><br>
249
 
250
  ## Baseline code
251
+ <hr style='margin-top:-1em; margin-bottom:0' />
 
252
  We propose the U-T&T model, a two-branch architecture that combines spatial and temporal information from very high-resolution aerial images and high-resolution satellite images into a single output. The U-Net architecture is employed for the spatial/texture branch, using a ResNet34 backbone model pre-trained on ImageNet. For the spatio-temporal branch,
253
  the U-TAE architecture incorporates a Temporal self-Attention Encoder (TAE) to explore the spatial and temporal characteristics of the Sentinel-2 time series data,
254
  applying attention masks at different resolutions during decoding. This model allows for the fusion of learned information from both sources,
 
272
 
273
 
274
  ## Reference
275
+ <hr style='margin-top:-1em; margin-bottom:0' />
 
276
  Please include a citation to the following article if you use the FLAIR dataset:
277
 
278
  ```
 
286
  ```
287
 
288
  ## Acknowledgment
289
+ <hr style='margin-top:-1em; margin-bottom:0' />
 
290
  This work was performed using HPC/AI resources from GENCI-IDRIS (Grant 2022-A0131013803). This work was supported by the project "Copernicus / FPCUP” of the European Union, by the French Space Agency (CNES) and by Connect by CNES.<br>
291
 
292
 
293
  ## Contact
294
+ <hr style='margin-top:-1em; margin-bottom:0' />
 
295
  If you have any questions, issues or feedback, you can contact us at: ai-challenge@ign.fr
296
 
297
 
298
  ## Dataset license
299
+ <hr style='margin-top:-1em; margin-bottom:0' />
 
300
  The "OPEN LICENCE 2.0/LICENCE OUVERTE" is a license created by the French government specifically for the purpose of facilitating the dissemination of open data by public administration.<br/>
301
  This licence is governed by French law.<br/>
302
  This licence has been designed to be compatible with any free licence that at least requires an acknowledgement of authorship, and specifically with the previous version of this licence as well as with the following licences: United Kingdom’s “Open Government Licence” (OGL), Creative Commons’ “Creative Commons Attribution” (CC-BY) and Open Knowledge Foundation’s “Open Data Commons Attribution” (ODC-BY).