File size: 8,256 Bytes
1c53fc3
 
4aa59b3
3f72d00
4aa59b3
 
 
1c53fc3
4aa59b3
3f72d00
 
4aa59b3
 
 
 
3f72d00
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4aa59b3
 
3f72d00
 
 
 
 
 
 
4aa59b3
 
 
 
 
 
 
3f72d00
 
 
 
 
4aa59b3
 
 
 
 
 
3f72d00
 
4aa59b3
 
3f72d00
 
 
 
 
 
 
4aa59b3
 
3f72d00
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4aa59b3
 
 
3f72d00
 
 
 
 
 
 
 
4aa59b3
 
3f72d00
 
4aa59b3
 
 
 
 
 
3f72d00
 
 
4aa59b3
 
 
 
 
 
 
 
 
 
 
 
 
 
3f72d00
4aa59b3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3f72d00
4aa59b3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
---
license: apache-2.0
tags:
- multi-modal
- 3D medical segmentation
size_categories:
- 1K<n<10K
---

![Data_visualization](M3D_Seg.jpg)

## Dataset Description
Large-scale General 3D Medical Image Segmentation Dataset (M3D-Seg)

### Dataset Introduction
3D medical image segmentation poses a significant challenge in medical image analysis. 
Currently, due to privacy and cost constraints, publicly available large-scale 3D medical images 
and their annotated data are scarce. To address this, we have collected 25 publicly available 
3D CT segmentation datasets, including CHAOS, HaN-Seg, AMOS22, AbdomenCT-1k, KiTS23, KiPA22, 
KiTS19, BTCV, Pancreas-CT, 3D-IRCADB, FLARE22, TotalSegmentator, CT-ORG, WORD, VerSe19, VerSe20,
SLIVER07, QUBIQ, MSD-Colon, MSD-HepaticVessel, MSD-Liver, MSD-lung, MSD-pancreas, MSD-spleen, LUNA16. 
These datasets are uniformly encoded from 0000 to 0024, totaling 5,772 3D images and 149,196 3D mask annotations.
The semantic labels corresponding to each mask can be represented in text. 
Within each sub-dataset folder, there are multiple data folders (containing image and mask files), 
and each sub-dataset independently utilizes its JSON file to split.

- **data_load_demo.py**: Provides an example code on reading images and masks from the dataset.
- **data_preocess.py**: Describes how to convert raw `nii.gz` or other format data into a more efficient `npy` format and preprocess them,
saving them in a unified format. This dataset has already been preprocessed, so there is no need to use data_preocess.py again.
If adding new datasets, please follow a unified processing approach.
- **dataset_info.json & dataset_info.txt**: Contain the names of each dataset and their label texts.
- **term_dictionary.json**: Provides multiple definitions or descriptions for each semantic label in the dataset,
generated by `ChatGPT` for each term. Researchers can convert category IDs in the dataset to label texts
using the information in dataset_info.txt and further convert them into text descriptions using term_dictionary.json as text inputs for segmentation models,
enabling tasks such as segmentation based on text prompts and referring segmentation.

This dataset supports not only traditional semantic segmentation tasks but also text-based segmentation tasks. 
For detailed methods, please refer to [SegVol](https://github.com/BAAI-DCAI/SegVol) and [M3D](https://github.com/BAAI-DCAI/M3D). 
As a general segmentation dataset, we provide a convenient, unified, and structured dataset organization 
that allows for the uniform integration of more public and private datasets in the same format as this dataset,
thereby constructing a larger-scale general 3D medical image segmentation dataset.

### Supported Tasks
This dataset not only supports traditional image-mask semantic segmentation tasks
but also represents data in the form of image-mask-text, where masks can be converted into box coordinates 
through bounding boxes. Based on this, the dataset can effectively support a series of image segmentation 
and positioning tasks, as follows:

- **3D Segmentation**: Semantic segmentation, text-based segmentation, referring segmentation, reasoning segmentation, etc.
- **3D Positioning**: Visual grounding/referring expression comprehension, referring expression generation.

## Dataset Format and Structure

### Data Format
<pre>
    M3D_Seg/
        0000/
            1/
                image.npy
                mask_(1, 512, 512, 96).npz
            2/
            ......
            0000.json
        0001/
        ......
</pre>

### Dataset Download
The total dataset size is approximately 224G.

#### Clone with HTTP
```bash
git clone https://huggingface.co/datasets/GoodBaiBai88/M3D-Seg
```

#### SDK Download
```bash
from datasets import load_dataset
dataset = load_dataset("GoodBaiBai88/M3D-Seg")
```
#### Manual Download
Manually download all files from the dataset files. It is recommended to use batch download tools for efficient downloading.
Please note the following:

- **Downloading in Parts and Merging**: Since dataset 0024 has a large volume,
the original compressed file has been split into two parts: `0024_1` and `0024_2`.
Make sure to download these two files separately and unzip them in the same directory to ensure data integrity.

- **Masks with Sparse Matrices**: To save storage space effectively,
foreground information in masks is stored in sparse matrix format and saved with the extension `.npz`.
The name of each mask file typically includes its shape information for identification and loading purposes.

- **Data Load Demo**: There is a script named data_load_demo.py, which serves as a reference for correctly
reading the sparse matrix format of masks and other related data.
Please refer to this script for specific loading procedures and required dependencies.



### Dataset Loading Method
#### 1. Direct Usage of Preprocessed Data
If you have already downloaded the preprocessed dataset, no additional data processing steps are required. 
You can directly jump to step 2 to build and load the dataset. 
Please note that the contents provided by this dataset have been transformed and numbered through data_process.py,
differing from the original `nii.gz` files. To understand the specific preprocessing process, 
refer to the data_process.py file for detailed information. If adding new datasets or modifying existing ones, 
please refer to data_process.py for data preprocessing and uniform formatting.


#### 2. Build Dataset 
To facilitate model training and evaluation using this dataset, we provide an example code for the Dataset class. 
Wrap the dataset in your project according to the following example:

```python

```


### Data Splitting
Each sub-dataset folder is splitted into `train` and `test` parts through a JSON file, 
facilitating model training and testing.



### Dataset Sources

| ID | Dataset  | Link |
| ------------- | ------------- | ------------- |
| 0000 |CHAOS|	https://chaos.grand-challenge.org/ |
| 0001 |HaN-Seg|	https://han-seg2023.grand-challenge.org/|
| 0002 |AMOS22|	https://amos22.grand-challenge.org/|
| 0003 |AbdomenCT-1k|	https://github.com/JunMa11/AbdomenCT-1K|
| 0004 |KiTS23|	https://kits-challenge.org/kits23/|
| 0005 |KiPA22|	https://kipa22.grand-challenge.org/|
| 0006 |KiTS19|	https://kits19.grand-challenge.org/|
| 0007 |BTCV|	https://www.synapse.org/#!Synapse:syn3193805/wiki/217753|
| 0008 |Pancreas-CT|	https://wiki.cancerimagingarchive.net/display/public/pancreas-ct|
| 0009 | 3D-IRCADB  | https://www.kaggle.com/datasets/nguyenhoainam27/3dircadb |
| 0010 |FLARE22|	https://flare22.grand-challenge.org/|
| 0011 |TotalSegmentator|	https://github.com/wasserth/TotalSegmentator|
| 0012 |CT-ORG|	https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=61080890|
| 0013 |WORD|	https://paperswithcode.com/dataset/word|
| 0014 |VerSe19|	https://osf.io/nqjyw/|
| 0015 |VerSe20|	https://osf.io/t98fz/|
| 0016 |SLIVER07|	https://sliver07.grand-challenge.org/|
| 0017 |QUBIQ|	https://qubiq.grand-challenge.org/|
| 0018 |MSD-Colon|	http://medicaldecathlon.com/|
| 0019 |MSD-HepaticVessel|	http://medicaldecathlon.com/|
| 0020 |MSD-Liver|	http://medicaldecathlon.com/|
| 0021 |MSD-lung|  	http://medicaldecathlon.com/|
| 0022 |MSD-pancreas|	http://medicaldecathlon.com/|
| 0023 |MSD-spleen|	http://medicaldecathlon.com/|
| 0024 |LUNA16|	https://luna16.grand-challenge.org/Data/|


## Dataset Copyright Information

All datasets in this dataset are publicly available. For detailed copyright information, please refer to the corresponding dataset links.

## Citation
If you use this dataset, please cite the following works:

```BibTeX
@misc{bai2024m3d,
      title={M3D: Advancing 3D Medical Image Analysis with Multi-Modal Large Language Models}, 
      author={Fan Bai and Yuxin Du and Tiejun Huang and Max Q. -H. Meng and Bo Zhao},
      year={2024},
      eprint={2404.00578},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}
@misc{du2024segvol,
      title={SegVol: Universal and Interactive Volumetric Medical Image Segmentation}, 
      author={Yuxin Du and Fan Bai and Tiejun Huang and Bo Zhao},
      year={2024},
      eprint={2311.13385},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}
```