Update README.md
Browse files
README.md
CHANGED
@@ -6,31 +6,43 @@ license: cc-by-4.0
|
|
6 |
|
7 |
<!-- Provide a quick summary of what the model is/does. -->
|
8 |
|
9 |
-
|
10 |
|
11 |
## Model Details
|
|
|
|
|
|
|
|
|
|
|
12 |
|
13 |
### Model Description
|
14 |
|
15 |
-
|
16 |
|
|
|
17 |
|
|
|
18 |
|
19 |
-
|
20 |
-
- **Funded by [optional]:** [More Information Needed]
|
21 |
-
- **Shared by [optional]:** [More Information Needed]
|
22 |
-
- **Model type:** [More Information Needed]
|
23 |
-
- **Language(s) (NLP):** [More Information Needed]
|
24 |
-
- **License:** [More Information Needed]
|
25 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
26 |
|
27 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
28 |
|
29 |
-
|
30 |
|
31 |
-
- **Repository:**
|
32 |
-
- **Paper [optional]:**
|
33 |
-
- **Demo [optional]:**
|
34 |
|
35 |
---
|
36 |
Table 1: Linear probing results on six classification tasks. All models are trained
|
@@ -80,4 +92,14 @@ this domain.
|
|
80 |
|
81 |
## Uses
|
82 |
|
83 |
-
Please refer to the Github repo [DOFA](https://github.com/zhu-xlab/DOFA) for more details.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
|
7 |
<!-- Provide a quick summary of what the model is/does. -->
|
8 |
|
9 |
+
What is DOFA: DOFA is a unified multimodal foundation model for different data modalities in remote sensing and Earth observation.
|
10 |
|
11 |
## Model Details
|
12 |
+
Differences with existing foundation models: DOFA is pre-trained using five different data modalities in remote sensing and Earth observation. It can handle images with any number of input channels.
|
13 |
+
|
14 |
+
DOFA is inspired by neuroplasticity Neuroplasticity is an important brain mechanism for adjusting to new experiences or environmental shifts. Inspired by this concept, we design DOFA to emulate this mechanism for processing multimodal EO data.
|
15 |
+
|
16 |
+
For more details, please take a look at the paper [Neural Plasticity-Inspired Foundation Model for Observing the Earth Crossing Modalities](https://arxiv.org/abs/2403.15356).
|
17 |
|
18 |
### Model Description
|
19 |
|
20 |
+
**Why develop DOFA**
|
21 |
|
22 |
+
The learned multimodal representation may not effectively capture such an intersensor relationship.
|
23 |
|
24 |
+
The performance of foundation models will degrade when downstream tasks require the utilization of data from unseen sensors with varying numbers of spectral bands and spatial resolutions or different wavelength regimes.
|
25 |
|
26 |
+
The development of individual, customized foundation models requires considerably more computing resources and human efforts.
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
|
28 |
+
The increasing number of specialized foundation models makes it difficult to select the most appropriate one for a specific downstream task.
|
29 |
+
|
30 |
+
DOFA supports input images with any number of channels using our pre-trained foundation models.
|
31 |
+
The examples in the Github repo [DOFA](https://github.com/zhu-xlab/DOFA) show how to use DOFA for Sentinel-1 (SAR), Sentinel-2, NAIP RGB.
|
32 |
+
We will add example usage for Gaofen Multispectral, and Hyperspectral data soon.
|
33 |
+
|
34 |
+
---
|
35 |
+
|
36 |
+
- **Developed by:** Techinical University of Munich, [Chair of Data Science in Earth Observation](https://www.asg.ed.tum.de/en/sipeo/home/)
|
37 |
+
- **Funded by:** Ekapex, ML4Earth
|
38 |
+
- **Model type:** Multimodal Foundation Model for Remote Sensing and Earth Observation
|
39 |
+
- **License:** CC-BY-4.0
|
40 |
|
41 |
+
### Model Sources [optional]
|
42 |
|
43 |
+
- **Repository:** https://github.com/zhu-xlab/DOFA
|
44 |
+
- **Paper [optional]:** https://arxiv.org/abs/2403.15356
|
45 |
+
- **Demo [optional]:** https://github.com/ShadowXZT/DOFA-pytorch/blob/master/demo.ipynb
|
46 |
|
47 |
---
|
48 |
Table 1: Linear probing results on six classification tasks. All models are trained
|
|
|
92 |
|
93 |
## Uses
|
94 |
|
95 |
+
Please refer to the Github repo [DOFA](https://github.com/zhu-xlab/DOFA) for more details.
|
96 |
+
|
97 |
+
|
98 |
+
```
|
99 |
+
@article{xiong2024neural,
|
100 |
+
title={Neural Plasticity-Inspired Foundation Model for Observing the {Earth} Crossing Modalities},
|
101 |
+
author={Xiong, Zhitong and Wang, Yi and Zhang, Fahong and Stewart, Adam J and Hanna, Jo{\"e}lle and Borth, Damian and Papoutsis, Ioannis and Saux, Bertrand Le and Camps-Valls, Gustau and Zhu, Xiao Xiang},
|
102 |
+
journal={arXiv preprint arXiv:2403.15356},
|
103 |
+
year={2024}
|
104 |
+
}
|
105 |
+
```
|