• Corpus ID: 254069545

Generalisable 3D Fabric Architecture for Streamlined Universal Multi-Dataset Medical Image Segmentation

  title={Generalisable 3D Fabric Architecture for Streamlined Universal Multi-Dataset Medical Image Segmentation},
  author={Siyu Liu and Weiqun Dai and Craig B. Engstrom and Jurgen Fripp and Stuart Crozier and Jason A Dowling and Shekhar S. Chandra},
Data scarcity is common in deep learning models for medical image segmentation. Previous works proposed multi-dataset learning, either simultaneously or via transfer learning to expand training sets. However, medical image datasets have diverse-sized images and features, and developing a model simultaneously for multiple datasets is challenging. This work proposes Fabric Image Representation Encoding Network (FIRENet), a universal 3D architecture for simultaneous multidataset segmentation and… 

Figures and Tables from this paper



Med3D: Transfer Learning for 3D Medical Image Analysis

A heterogeneous 3D network called Med3D is designed to co-train multi-domain 3DSeg-8 so as to make a series of pre-trained models which can accelerate the training convergence speed of target 3D medical tasks and improve accuracy ranging from 3% to 20%.

3D U2-Net: A 3D Universal U-Net for Multi-Domain Medical Image Segmentation

Experimental results show that this universal architecture that handles multiple medical segmentation tasks and is extendable for new tasks, regardless of different organs and imaging modalities is capable of competing with traditional models in terms of segmentation accuracy.

3D deeply supervised network for automated segmentation of volumetric medical images

CAN3D: Fast 3D Medical Image Segmentation via Compact Context Aggregation

C2FNAS: Coarse-to-Fine Neural Architecture Search for 3D Medical Image Segmentation

A coarse-to-fine neural architecture search (C2FNAS) is proposed to automatically search a 3D segmentation network from scratch without inconsistency on network size or input size and achieves state-of-the-art performance with the network searched using one dataset.

V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation

This work proposes an approach to 3D image segmentation based on a volumetric, fully convolutional, neural network, trained end-to-end on MRI volumes depicting prostate, and learns to predict segmentation for the whole volume at once.

UNETR: Transformers for 3D Medical Image Segmentation

This work reformulates the task of volumetric (3D) medical image segmentation as a sequence-to-sequence prediction problem and introduces a novel architecture, dubbed as UNEt TRansformers (UNETR), that utilizes a transformer as the encoder to learn sequence representations of the input volume and effectively capture the global multi-scale information.

nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation.

Biomedical imaging is a driver of scientific discovery and a core component of medical care and is being stimulated by the field of deep learning. While semantic segmentation algorithms enable image

DS-TransUNet: Dual Swin Transformer U-Net for Medical Image Segmentation

A novel deep medical image segmentation framework called dual swin transformer U-Net (DS-TransUNet), which aims to incorporate the hierarchical swin transformers into both the encoder and the decoder of the standard U-shaped architecture and adopts a well-established dual-scale encoding mechanism.