File size: 17,316 Bytes
d465419
 
9eb9e72
 
 
 
 
 
 
 
 
 
 
 
 
d465419
9eb9e72
 
 
 
a33e081
 
9eb9e72
a33e081
 
 
9eb9e72
a33e081
 
 
9eb9e72
a33e081
 
9eb9e72
a33e081
 
9eb9e72
a33e081
9eb9e72
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a6d9ce4
a33e081
 
 
 
f520d6d
9eb9e72
 
 
 
 
 
 
 
 
a6d9ce4
9eb9e72
 
 
 
 
 
 
 
 
 
 
 
a6d9ce4
 
e959f97
 
17701a9
9eb9e72
 
 
 
e959f97
 
9eb9e72
 
 
 
17701a9
 
 
9eb9e72
 
 
 
 
6b2f019
e02573f
 
 
 
 
6b2f019
9eb9e72
6b2f019
 
 
 
9eb9e72
caa29b1
9eb9e72
6b2f019
 
 
 
9eb9e72
2fc2f63
beddae8
d957f2a
 
 
 
 
9eb9e72
 
 
 
 
f76e35a
7ab3e64
 
 
f76e35a
1de129d
 
 
 
 
 
 
 
 
 
 
 
 
9eb9e72
 
 
b6e575f
e02573f
9eb9e72
4bc6cac
 
 
 
 
 
 
 
9eb9e72
 
 
 
 
 
 
 
 
a5b25f6
 
 
9eb9e72
a5b25f6
 
9eb9e72
 
 
493b7fd
 
 
 
 
 
 
9eb9e72
5eaaaf0
 
 
 
 
 
 
e05a294
 
ed16d52
9eb9e72
a33e081
 
 
 
9eb9e72
a33e081
ed16d52
a33e081
af59084
a33e081
 
 
 
 
ed16d52
a33e081
ed16d52
a33e081
 
ed16d52
a33e081
771a72c
f0b1f9f
771a72c
 
 
f0b1f9f
771a72c
 
 
f0b1f9f
771a72c
 
 
 
f0b1f9f
771a72c
9eb9e72
771a72c
 
9eb9e72
f520d6d
ed16d52
771a72c
 
ed16d52
771a72c
 
ed16d52
771a72c
 
 
9eb9e72
771a72c
 
 
9eb9e72
771a72c
9eb9e72
771a72c
 
 
9eb9e72
771a72c
9eb9e72
771a72c
 
9eb9e72
 
 
 
 
 
 
 
 
 
6328a2e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
---
license: apache-2.0
datasets:
- openclimatefix/era5
language:
- es
- en
metrics:
- mse
library_name: transformers
pipeline_tag: image-to-image
tags:
- climate
- transformers
- super-resolution
---


# Europe Reanalysis Super Resolution

The aim of the project is to create a Machine learning (ML) model that can generate high-resolution regional reanalysis data (similar to the one produced by CERRA) by 
downscaling global reanalysis data from ERA5. 

This will be accomplished by using state-of-the-art Deep Learning (DL) techniques like U-Net, conditional GAN, and diffusion models (among others). Additionally, 
an ingestion module will be implemented to assess the possible benefit of using CERRA pseudo-observations as extra predictors. Once the model is designed and trained, 
a detailed validation framework takes the place. 

It combines classical deterministic error metrics with in-depth validations, including time series, maps, spatio-temporal correlations, and computer vision metrics, 
disaggregated by months, seasons, and geographical regions, to evaluate the effectiveness of the model in reducing errors and representing physical processes. 
This level of granularity allows for a more comprehensive and accurate assessment, which is critical for ensuring that the model is effective in practice. 

Moreover, tools for interpretability of DL models can be used to understand the inner workings and decision-making processes of these complex structures by analyzing 
the activations of different neurons and the importance of different features in the input data.

This work is funded by [Code for Earth 2023](https://codeforearth.ecmwf.int/) initiative. The model **ConvSwin2SR** is released in Apache 2.0, making it usable without 
restrictions anywhere.

# Table of Contents

- [Model Card for Europe Reanalysis Super Resolution](#model-card-for--model_id-)
- [Table of Contents](#table-of-contents)
- [Model Details](#model-details)
  - [Model Description](#model-description)
- [Uses](#uses)
  - [Direct Use](#direct-use)
  - [Out-of-Scope Use](#out-of-scope-use)
- [Bias, Risks, and Limitations](#bias-risks-and-limitations)
- [Training Details](#training-details)
  - [Training Data](#training-data)
  - [Training Procedure](#training-procedure)
    - [Preprocessing](#preprocessing)
    - [Speeds, Sizes, Times](#speeds-sizes-times)
- [Evaluation](#evaluation)
  - [Testing Data, Factors & Metrics](#testing-data-factors--metrics)
    - [Testing Data](#testing-data)
    - [Factors](#factors)
    - [Metrics](#metrics)
  - [Results](#results)
- [Technical Specifications](#technical-specifications-optional)
  - [Model Architecture](#model-architecture)
    - [Components](#components)
    - [Configuration details](#configuration-details)
    - [Loss function](#loss-function)
  - [Computing Infrastructure](#computing-infrastructure)
    - [Hardware](#hardware)
    - [Software](#software)
- [Authors](#authors)

# Model Details

## Model Description

<!-- Provide a longer summary of what this model is/does. -->
We present the ConvSwin2SR tranformer, a vision model for down-scaling (from 0.25º to 0.05º) regional reanalysis grids in the mediterranean area.

- **Developed by:** A team of Predictia Intelligent Data Solutions S.L. & Instituto de Fisica de Cantabria (IFCA)
- **Model type:** Vision model
- **Language(s) (NLP):** en, es
- **License:** Apache-2.0
- **Resources for more information:** More information needed
    - [GitHub Repo](https://github.com/ECMWFCode4Earth/DeepR)


# Uses


## Direct Use

The primary use of the ConvSwin2SR transformer is to enhance the spatial resolution in the Mediterranean area of global reanalysis grids using a regional reanalysis grid 
as groundtruth. This enhancement is crucial for more precise climate studies, which can aid in better decision-making for various stakeholders including policymakers, 
researchers, and weather-dependent industries like agriculture, energy, and transportation.


## Out-of-Scope Use

The model is specifically designed for downscaling ERA5 reanalysis grids to the CERRA regional reanalysis grid and may not perform well or provide accurate results 
for other types of geospatial data or geographical regions. 


# Bias, Risks, and Limitations

Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) 
and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes 
across protected classes; identity characteristics; and sensitive, social, and occupational groups.

# Training Details

## Training Data

The datasets that are mainly used in the project can be found in the following Copernicus Climate Data Store catalogue entries:

- [ERA5 hourly data on single levels from 1940 to present](https://cds.climate.copernicus.eu/cdsapp#!/dataset/reanalysis-era5-single-levels?tab=overview)
  
- [CERRA sub-daily regional reanalysis data for Europe on single levels from 1984 to present](https://cds.climate.copernicus.eu/cdsapp#!/dataset/reanalysis-cerra-single-levels?tab=overview)

1. Input low-resolution grids (ERA5):

  The input grids are structured as a 3D array with dimensions of (time, 60, 44), where 60 and 44 are the number of grid points along the longitude and latitude axes, 
  respectively. Geographically, these grids cover a longitude range from -8.35 to 6.6 and a latitude range from 46.45 to 35.50. 
  This implies that the data covers a region extending from a westernmost point at longitude -8.35 to an easternmost point at longitude 6.6, and from a 
  northernmost point at latitude 46.45 to a southernmost point at latitude 35.50.

2. Target high-resolution grids (CERRA):

  They are represented as a 3D array with larger dimensions of (time, 240, 160), indicating a finer grid resolution compared to the input grids. Here, 240 and 160 are 
  the number of grid points along the longitude and latitude axes, respectively. The geographical coverage for these high-resolution grids is defined by a longitude 
  range from -6.85 to 5.1 and a latitude range from 44.95 to 37. This region extends from a westernmost point at longitude -6.85 to an easternmost point at longitude 5.1,
  and from a northernmost point at latitude 44.95 to a southernmost point at latitude 37.

![spatial-coverages](spatial-coverages.png)

The dataset's temporal division is structured to optimize model training and subsequent per-epoch validation. 
The training duration spans 29 years, commencing in January 1985 and culminating in December 2013. 
Sequentially, the validation phase begins, covering the period from January 2014 to December 2017. This 4-year interval is solely dedicated to evaluating the model's 
aptitude on data it hasn't been exposed to during training. This separation ensures the model's robustness and its capability to make dependable predictions for the 
validation period.

## Training Procedure

### Preprocessing

The preprocessing of climate datasets ERA5 and CERRA, extracted from the Climate Data Store (CDS), is a critical step before their utilization in training models. 
This section defines the preprocessing steps undertaken to homogenize these datasets into a common format. 
The steps include unit standardization, coordinate system rectification, and grid interpolation. 
The rationale and methodologies employed in each step are discussed comprehensively, setting a robust foundation for the subsequent training procedure.

1. Unit Standardization: A preliminary step in the preprocessing pipeline involved the standardization of units across both datasets.
  This was imperative to ensure a uniform unit system, facilitating a seamless integration of the datasets in later stages.
  The units in both datasets were scrutinized and amended to adhere to a common unit system, thereby eliminating any discrepancies that could hinder the analysis.

2. Coordinate System Rectification: The coordinate system of the datasets was rectified to ensure a coherent representation of geographical information. 
  Specifically, the coordinates and dimensions were renamed to a standardized format with longitude (lon) and latitude (lat) as designated names. 
  The longitude values were adjusted to range from -180 to 180 instead of the initial 0 to 360 range, while latitude values were ordered in ascending order, 
  thereby aligning with conventional geographical coordinate systems.

3. Grid Interpolation: The ERA5 dataset is structured on a regular grid with a spatial resolution of 0.25º, whereas the CERRA dataset inhabits a curvilinear
   grid with a Lambert Conformal projection of higher spatial resolution (0.05º). To overcome this disparity, a grid interpolation procedure was initiated.
  This step was crucial to align the datasets onto a common regular grid (with different spatial resolution), thereby ensuring consistency in spatial representation.
  The interpolation transformed the CERRA dataset to match the regular grid structure of the ERA5 dataset, keeping its initial spatial resolution of 0.05º (5.5 km).

### Speeds, Sizes, Times

- Training time: The training duration for the ConvSwin2SR model is notably extensive, clocking in at 3,648 days to complete a total of 100 epochs with
  a batch size of 2 for a total number of batches equal to ~43000.

- Model size: The ConvSwin2SR model is a robust machine learning model boasting a total of 12,383,377 parameters.
  This size reflects a substantial capacity for learning and generalizing complex relationships within the data, enabling the model to
  effectively upscale lower-resolution reanalysis grids to higher-resolution versions.

- Inference speed: The ConvSwin2SR model demonstrates a commendable inference speed, particularly when handling a substantial batch of samples.
  Specifically, when tasked with downscaling 248 samples, which is synonymous with processing data for an entire month at 3-hour intervals,
  the model completes the operation in a mere 21 seconds. This level of efficiency is observed in a local computing environment outfitted with 16GB o
  f RAM and 4GB of GPU memory. 
 
# Evaluation

<!-- This section describes the evaluation protocols and provides the results. -->

## Testing Data, Factors & Metrics

### Testing Data

In terms of spatial dimensions, both the input grids from ERA5 and the target high-resolution grids from CERRA remain consistent throughout the training and testing phases.
This spatial consistency ensures that the model is evaluated under the same geographic conditions as it was trained, allowing for a direct comparison of its performance 
across different temporal segments.

The testing data samples correspond to the three-year period from 2018 to 2020, inclusive. This segment is crucial for assessing the model's real-world applicability and 
its performance on the most recent data points, ensuring its relevance and reliability in current and future scenarios.

## Results 

In our evaluation, the proposed model displayed a significant enhancement over the established baseline, which employs bicubic interpolation for the same task. 
Specifically, our model achieved a noteworthy 34.93% reduction in Mean Absolute Error (MAE), a metric indicative of the average magnitude of errors between 
predicted and actual values. Furthermore, there was a near 30% improvement in the Root Mean Square Error (RMSE), which measures the square root of the average 
of squared differences between predictions and actual values.

These metrics not only underscore the model's capability to predict with greater precision but also emphasize its reduced propensity for errors. 
In comparison to the bicubic interpolation baseline, our model's superior predictive accuracy is evident, positioning it as a more reliable tool for this task.

- Mean absolute error (MAE):
  
  ![mae](metric_global_map_diff_var-mae.png)

- Root mean squared error (RMSE):

  ![rmse](metric_global_map_diff_var-rmse.png)


# Technical Specifications

## Model Architecture

Our model's design is deeply rooted in the Swin2 architecture, specifically tailored for Super Resolution (SR) tasks.
We've harnessed the [transformers library](https://github.com/huggingface/transformers) to streamline and simplify the model's design.

![Model Architecture](architecture.png)

### Components

- **Transformers Component**: Central to our model is the [transformers.Swin2SRModel](https://huggingface.co/docs/transformers/model_doc/swin2sr#transformers.Swin2SRModel). This component amplifies the spatial resolution of its inputs by a factor of 8. Notably, Swin2SR exclusively supports upscaling ratios that are powers of 2. 
- **Convolutional Neural Network (CNN) Component**: Given that our actual upscale ratio is approximately 5 and the designated output shape is (160, 240),
  we've integrated a CNN. This serves as a preprocessing unit, transforming inputs into (20, 30) feature maps suitable for the Swin2SRModel.
  
The underlying objective of this network is to master the residuals stemming from bicubic interpolation.

### Configuration Details

For those inclined towards the intricacies of the model, the specific parameters governing its behavior are meticulously detailed in the 
[config.json](https://huggingface.co/predictia/convswin2sr_mediterranean/blob/main/config.json).

### Loss function

The Swin2 transformer optimizes its parameters using a composite loss function that aggregates multiple L1 loss terms to enhance its predictive 
accuracy across different resolutions and representations:

1. **Primary Predictions Loss**: 
   - This term computes the L1 loss between the primary model predictions and the reference values. It ensures that the transformer's outputs
   - closely match the ground truth across the primary spatial resolution.

2. **Downsampled Predictions Loss**: 
   - Recognizing the importance of accuracy across varying resolutions, this term calculates the L1 loss between the downsampled versions of the
   - predictions and the reference values. By incorporating this term, the model is incentivized to preserve critical information even when the data is represented
   - at a coarser scale.

3. **Blurred Predictions Loss**: 
   - To ensure the model's robustness against small perturbations and noise, this term evaluates the L1 loss between blurred versions of the
   - predictions and the references. This encourages the model to produce predictions that maintain accuracy even under slight modifications in the data representation.

By combining these loss terms, the Swin2 transformer is trained to produce accurate predictions across different resolutions and under various data transformations, 
ensuring its versatility and robustness in diverse scenarios.

## Computing Infrastructure

Leveraging GPUs in deep learning initiatives greatly amplifies the pace of model training and inference. This computational edge not only diminishes the total 
computational duration but also equips us to proficiently navigate complex tasks and extensive datasets.

Our profound gratitude extends to our collaborative partners, whose invaluable contribution and support have been cornerstones in the fruition of this project. 
Their substantial inputs have immensely propelled our research and developmental strides.

- **AI4EOSC**: Representing "Artificial Intelligence for the European Open Science Cloud," AI4EOSC functions under the aegis of the European Open Science Cloud (EOSC).
  Initiated by the European Union, EOSC endeavors to orchestrate a cohesive platform for research data and services. AI4EOSC, a distinct arm within EOSC, concentrates
  on embedding and leveraging artificial intelligence (AI) techniques within the open science domain.

- **European Weather Cloud**: Serving as a cloud-centric hub, this platform catalyzes collective efforts in meteorological application design and operations
  throughout Europe. Its offerings are manifold, ranging from disseminating weather forecast data to proffering computational prowess, expert counsel, and
  consistent support.

### Hardware Specifications

Our endeavor harnesses the capabilities of two virtual machines (VMs), each embedded with a dedicated GPU. One VM is bolstered with a 16GB GPU, while its counterpart 
is equipped with an even potent 20GB GPU. This strategic hardware alignment proficiently caters to diverse computational needs, spanning data orchestration to model 
fine-tuning and evaluation, ensuring the seamless flow and success of our project.

### Software Resources

For enthusiasts and researchers inclined towards a deeper probe, our model's training and evaluation code is transparently accessible. 
Navigate to our GitHub Repository [ECMWFCode4Earth/DeepR](https://github.com/ECMWFCode4Earth/DeepR) under the ECWMF Code 4 Earth consortium.


### Authors

<!-- This section provides another layer of transparency and accountability. Whose views is this model card representing? How many voices were included in its construction? Etc. -->

- Mario Santa Cruz. Predictia Intelligent Data Solutions S.L.

- Antonio Pérez. Predictia Intelligent Data Solutions S.L.

- Javier Díez. Instituto de Física de Cantabria (IFCA)