Need help with FewShot_GAN-Unet3D?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

280 Stars 74 Forks MIT License 40 Commits 16 Opened issues


Tensorflow implementation of our paper: Few-shot 3D Multi-modal Medical Image Segmentation using Generative Adversarial Learning

Services available


Need anything else?

Contributors list

# 117,761
32 commits
# 389,544
1 commit

Few-shot 3D Medical Image Segmentation using Generative Adversarial Learning

This repository contains the tensorflow and pytorch implementation of the model we proposed in our paper of the same name: Few-shot 3D Medical Image Segmentation using Generative Adversarial Learning

The code is available in both tensorflow and pytorch. To run the project kindly refer to the individual readme file.


iSEG 2017 dataset was chosen to substantiate our proposed method. It contains the 3D multi-modal brain MRI data of 10 labeled training subjects and 13 unlabeled testing subjects. We split the 10 labeled training data into training, validation and testing images for both the models.(Eg- 2,1 and 7) Rest of the 13 unlabeled testing images are only used for training the GAN based model.

MRBrainS13 dataset was also used to test the robustness of our proposed model. It contains 3 modalities (T1-weighted, T1-weighted inversion recovery and FLAIR). Original Dataset consists of 5 labeled training and 6 unlabeled testing subjects. We split the 5 labeled training data into 1,1,3 for training, validation and testing. The 6 testing subjects are used as unlabeled data for the GAN based models.

Proposed model architecture

The following shows the model architecture of the proposed model. (Read our paper for further details)

Some results from our paper

  • Visual comparison of the segmentation by each model, for two test subjects of the iSEG-2017 dataset, when training with different numbers of labeled examples.

  • Segmentation of Subject 10 of the iSEG-2017 dataset predicted by different GAN-based models, when trained with 2 labeled images. The red box highlights a region in the ground truth where all these models give noticeable differences.

  • More such results can be found in the paper.


You can mail us at: [email protected], [email protected] If you use this code for your research, please consider citing the original paper:

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.