M3D-Seg / README.md
GoodBaiBai88's picture
Update README.md
3f72d00 verified
|
raw
history blame
8.26 kB
---
license: apache-2.0
tags:
- multi-modal
- 3D medical segmentation
size_categories:
- 1K<n<10K
---
![Data_visualization](M3D_Seg.jpg)
## Dataset Description
Large-scale General 3D Medical Image Segmentation Dataset (M3D-Seg)
### Dataset Introduction
3D medical image segmentation poses a significant challenge in medical image analysis.
Currently, due to privacy and cost constraints, publicly available large-scale 3D medical images
and their annotated data are scarce. To address this, we have collected 25 publicly available
3D CT segmentation datasets, including CHAOS, HaN-Seg, AMOS22, AbdomenCT-1k, KiTS23, KiPA22,
KiTS19, BTCV, Pancreas-CT, 3D-IRCADB, FLARE22, TotalSegmentator, CT-ORG, WORD, VerSe19, VerSe20,
SLIVER07, QUBIQ, MSD-Colon, MSD-HepaticVessel, MSD-Liver, MSD-lung, MSD-pancreas, MSD-spleen, LUNA16.
These datasets are uniformly encoded from 0000 to 0024, totaling 5,772 3D images and 149,196 3D mask annotations.
The semantic labels corresponding to each mask can be represented in text.
Within each sub-dataset folder, there are multiple data folders (containing image and mask files),
and each sub-dataset independently utilizes its JSON file to split.
- **data_load_demo.py**: Provides an example code on reading images and masks from the dataset.
- **data_preocess.py**: Describes how to convert raw `nii.gz` or other format data into a more efficient `npy` format and preprocess them,
saving them in a unified format. This dataset has already been preprocessed, so there is no need to use data_preocess.py again.
If adding new datasets, please follow a unified processing approach.
- **dataset_info.json & dataset_info.txt**: Contain the names of each dataset and their label texts.
- **term_dictionary.json**: Provides multiple definitions or descriptions for each semantic label in the dataset,
generated by `ChatGPT` for each term. Researchers can convert category IDs in the dataset to label texts
using the information in dataset_info.txt and further convert them into text descriptions using term_dictionary.json as text inputs for segmentation models,
enabling tasks such as segmentation based on text prompts and referring segmentation.
This dataset supports not only traditional semantic segmentation tasks but also text-based segmentation tasks.
For detailed methods, please refer to [SegVol](https://github.com/BAAI-DCAI/SegVol) and [M3D](https://github.com/BAAI-DCAI/M3D).
As a general segmentation dataset, we provide a convenient, unified, and structured dataset organization
that allows for the uniform integration of more public and private datasets in the same format as this dataset,
thereby constructing a larger-scale general 3D medical image segmentation dataset.
### Supported Tasks
This dataset not only supports traditional image-mask semantic segmentation tasks
but also represents data in the form of image-mask-text, where masks can be converted into box coordinates
through bounding boxes. Based on this, the dataset can effectively support a series of image segmentation
and positioning tasks, as follows:
- **3D Segmentation**: Semantic segmentation, text-based segmentation, referring segmentation, reasoning segmentation, etc.
- **3D Positioning**: Visual grounding/referring expression comprehension, referring expression generation.
## Dataset Format and Structure
### Data Format
<pre>
M3D_Seg/
0000/
1/
image.npy
mask_(1, 512, 512, 96).npz
2/
......
0000.json
0001/
......
</pre>
### Dataset Download
The total dataset size is approximately 224G.
#### Clone with HTTP
```bash
git clone https://huggingface.co/datasets/GoodBaiBai88/M3D-Seg
```
#### SDK Download
```bash
from datasets import load_dataset
dataset = load_dataset("GoodBaiBai88/M3D-Seg")
```
#### Manual Download
Manually download all files from the dataset files. It is recommended to use batch download tools for efficient downloading.
Please note the following:
- **Downloading in Parts and Merging**: Since dataset 0024 has a large volume,
the original compressed file has been split into two parts: `0024_1` and `0024_2`.
Make sure to download these two files separately and unzip them in the same directory to ensure data integrity.
- **Masks with Sparse Matrices**: To save storage space effectively,
foreground information in masks is stored in sparse matrix format and saved with the extension `.npz`.
The name of each mask file typically includes its shape information for identification and loading purposes.
- **Data Load Demo**: There is a script named data_load_demo.py, which serves as a reference for correctly
reading the sparse matrix format of masks and other related data.
Please refer to this script for specific loading procedures and required dependencies.
### Dataset Loading Method
#### 1. Direct Usage of Preprocessed Data
If you have already downloaded the preprocessed dataset, no additional data processing steps are required.
You can directly jump to step 2 to build and load the dataset.
Please note that the contents provided by this dataset have been transformed and numbered through data_process.py,
differing from the original `nii.gz` files. To understand the specific preprocessing process,
refer to the data_process.py file for detailed information. If adding new datasets or modifying existing ones,
please refer to data_process.py for data preprocessing and uniform formatting.
#### 2. Build Dataset
To facilitate model training and evaluation using this dataset, we provide an example code for the Dataset class.
Wrap the dataset in your project according to the following example:
```python
```
### Data Splitting
Each sub-dataset folder is splitted into `train` and `test` parts through a JSON file,
facilitating model training and testing.
### Dataset Sources
| ID | Dataset | Link |
| ------------- | ------------- | ------------- |
| 0000 |CHAOS| https://chaos.grand-challenge.org/ |
| 0001 |HaN-Seg| https://han-seg2023.grand-challenge.org/|
| 0002 |AMOS22| https://amos22.grand-challenge.org/|
| 0003 |AbdomenCT-1k| https://github.com/JunMa11/AbdomenCT-1K|
| 0004 |KiTS23| https://kits-challenge.org/kits23/|
| 0005 |KiPA22| https://kipa22.grand-challenge.org/|
| 0006 |KiTS19| https://kits19.grand-challenge.org/|
| 0007 |BTCV| https://www.synapse.org/#!Synapse:syn3193805/wiki/217753|
| 0008 |Pancreas-CT| https://wiki.cancerimagingarchive.net/display/public/pancreas-ct|
| 0009 | 3D-IRCADB | https://www.kaggle.com/datasets/nguyenhoainam27/3dircadb |
| 0010 |FLARE22| https://flare22.grand-challenge.org/|
| 0011 |TotalSegmentator| https://github.com/wasserth/TotalSegmentator|
| 0012 |CT-ORG| https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=61080890|
| 0013 |WORD| https://paperswithcode.com/dataset/word|
| 0014 |VerSe19| https://osf.io/nqjyw/|
| 0015 |VerSe20| https://osf.io/t98fz/|
| 0016 |SLIVER07| https://sliver07.grand-challenge.org/|
| 0017 |QUBIQ| https://qubiq.grand-challenge.org/|
| 0018 |MSD-Colon| http://medicaldecathlon.com/|
| 0019 |MSD-HepaticVessel| http://medicaldecathlon.com/|
| 0020 |MSD-Liver| http://medicaldecathlon.com/|
| 0021 |MSD-lung| http://medicaldecathlon.com/|
| 0022 |MSD-pancreas| http://medicaldecathlon.com/|
| 0023 |MSD-spleen| http://medicaldecathlon.com/|
| 0024 |LUNA16| https://luna16.grand-challenge.org/Data/|
## Dataset Copyright Information
All datasets in this dataset are publicly available. For detailed copyright information, please refer to the corresponding dataset links.
## Citation
If you use this dataset, please cite the following works:
```BibTeX
@misc{bai2024m3d,
title={M3D: Advancing 3D Medical Image Analysis with Multi-Modal Large Language Models},
author={Fan Bai and Yuxin Du and Tiejun Huang and Max Q. -H. Meng and Bo Zhao},
year={2024},
eprint={2404.00578},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
@misc{du2024segvol,
title={SegVol: Universal and Interactive Volumetric Medical Image Segmentation},
author={Yuxin Du and Fan Bai and Tiejun Huang and Bo Zhao},
year={2024},
eprint={2311.13385},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
```