Skip to content

Latest commit

 

History

History
115 lines (91 loc) · 5.61 KB

README.md

File metadata and controls

115 lines (91 loc) · 5.61 KB

Upping the Game: How 2D U-Net Skip Connection Flip 3D Segmentation

NeurIPS 2024 (poster)

📖 Abstract

In the present study, we introduce an innovative structure for 3D medical image segmentation that effectively integrates 2D U-Net-derived skip connections into the architecture of 3D convolutional neural networks (3D CNNs). Conventional 3D segmentation techniques predominantly depend on isotropic 3D convolutions for the extraction of volumetric features, which frequently engenders inefficiencies due to the varying information density across the three orthogonal axes in medical imaging modalities such as computed tomography (CT) and magnetic resonance imaging (MRI). This disparity leads to a decline in axial-slice plane feature extraction efficiency, with slice plane features being comparatively underutilized relative to features in the time-axial. To address this issue, we introduce the U-shaped Connection (uC), utilizing simplified 2D U-Net in place of standard skip connections to augment the extraction of the axial-slice plane features while concurrently preserving the volumetric context afforded by 3D convolutions. Based on uC, we further present uC 3DU-Net, an enhanced 3D U-Net backbone that integrates the uC approach to facilitate optimal axial-slice plane feature utilization. Through rigorous experimental validation on five publicly accessible datasets—FLARE2021, OIMHS, FeTA2021, AbdomenCT-1K, and BTCV, the proposed method surpasses contemporary state-of-the-art models. Notably, this performance is achieved while reducing the number of parameters and computational complexity. This investigation underscores the efficacy of incorporating 2D convolutions within the framework of 3D CNNs to overcome the intrinsic limitations of volumetric segmentation, thereby potentially expanding the frontiers of medical image analysis.

🚀Methodology

We introduce uC3D U-Net, which integrates U-Shaped Connection (uC) into a 3D U-Net backbone, augmented with a dual feature integration (DFi) module.

🚀Installation

We run the code with cuda=11.8, python=3.9.0, pytorch=2.0.0+cu118, and torchvision=0.15.1+cu118. Please follow the instructions here to install cuda dependencies, and follow the instructions here to install both pytorch and torchvision dependencies.

Clone the repository locally

git clone https://github.com/IMOP-lab/U-Shaped-Connection.git --recurse

Environment setup

conda env create --file environment.yaml

🚀Getting Started

First download the dataset according to the dataset download link provided datasets_download.md.

Datasets format

The format of the dataset folder is as follows:

./datasets/
├── ABCT1K
├── FeTA
├── FLARE
├── OIMHS
├── BTCV
├── ...

The dataset subfolders are further divided into training sets, validation sets, and test sets in the following format:

path to the dataset/
├── imagesTr
├── labelsTr
├── imagesVal
├── labelsVal
├── imagesTs
├── original_labelTs
├── shapes.json

The original_labelTs folder includes all raw unlabeled data, shapes.json is each sample and its corresponding shape, for example: "train_000.nii.gz": [512, 512, 110], currently our framework only supports data in .nii.gz format.

Pre-trained model

We provide the model checkpoint at baidu netdisk.

Run the codes

You can use pre-trained models to inference by running:

bash pretrained_test.sh

And you can also train and test your own model by running:

bash run_gpu.sh

The model parameters, FLOPs, and inference time can be tested by running:

python cost.py

🚀Experimental results

Quantitative Results on OIMHS dataset

Method #Params FLOPs mIoU Dice VOE HD95 AdjRand
3D U-Net 4.81M 135.9G 86.02 92.05 13.98 6.77 91.34
Swin UNETR 62.2M 328.4G 86.73 92.53 13.27 5.09 91.85
3D UX-Net 53.0M 639.4G 87.43 92.90 12.57 4.41 92.27
SASAN 22.96M 282.92G 88.44 93.53 11.56 3.14 92.96
nnFormer 149.3M 240.2G 72.16 81.60 27.84 23.49 80.36
TransBTS 31.6M 110.4G 74.80 83.08 25.20 31.43 82.05
UNETR 92.8M 82.6G 80.52 88.11 19.48 30.07 87.21
uC 3DU-Net 21.7M 286.43G 89.48 94.13 10.52 2.98 93.62

Qualitative Results on FLARE2021 dataset

Qualitative Results on backbones

🎫License

This project is licensed under the MIT license.

🎫 Acknowledgment

We sincerely appreciate the outstanding contributions of monai and 3DUX-Net projects, which have been helpful in the successful implementation of this project.

🎫 Contributors

The project is implemented with the help of the following contributors:

Xingru Huang, Yihao Guo, Jian Huang, Tianyun Zhang, Hong He, Shaowei Jiang, Yaoqi Sun.