2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017
DOI: 10.1109/cvpr.2017.398
|View full text |Cite
|
Sign up to set email alerts
|

Joint Sequence Learning and Cross-Modality Convolution for 3D Biomedical Segmentation

Abstract: Deep learning models such as convolutional neural network have been widely used in 3D biomedical segmentation and achieve state-of-the-art performance. However, most of them often adapt a single modality or stack multiple modalities as different input channels. To better leverage the multimodalities, we propose a deep encoder-decoder structure with cross-modality convolution layers to incorporate different modalities of MRI data. In addition, we exploit convolutional LSTM to model a sequence of 2D slices, and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
100
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 156 publications
(105 citation statements)
references
References 22 publications
(41 reference statements)
0
100
0
Order By: Relevance
“…Two-dimensional methods [6,12] that are based on inplane segmentation does not utilize contextual information along the slice direction. Three-dimensional convolutional neural networks (CNN) [4,13] have been proposed to capture image dependencies between consecutive slices. However, by trading-off between the convolution kernel size and the number of pooling layers, these methods can only capture limited local receptive field and short-range dependencies.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Two-dimensional methods [6,12] that are based on inplane segmentation does not utilize contextual information along the slice direction. Three-dimensional convolutional neural networks (CNN) [4,13] have been proposed to capture image dependencies between consecutive slices. However, by trading-off between the convolution kernel size and the number of pooling layers, these methods can only capture limited local receptive field and short-range dependencies.…”
Section: Introductionmentioning
confidence: 99%
“…The contributions of our RSANet are three folds. First, unlike methods [3,13] using RNN or LSTM to capture the slice-wise dependencies, where RNN and LSTM have inherent drawbacks [2,10,14] of capturing long-range dependencies, we propose a novel slice-wise attention module, called SA Block (see Fig. 2) to compute the response at a slice as a weighted sum of the features from all slices along the same direction.…”
Section: Introductionmentioning
confidence: 99%
“…In medical images, combining multi-modality images such as CT, T1 MRI, T2 MRI, etc. to realize multi-organ segmentation [17] and lesion segmentation [18,19] is widely adopted due to distinct responses of different modalities datasets for different tissues. According to the review [20] of deep learning for medical image segmentation using multi-modality fusion, multi-modal segmentation network architectures can be categorized into input-level fusion network, layer-level fusion network and decision fusion network.…”
Section: Multi-modal Fusionmentioning
confidence: 99%
“…In decision-level fusion segmentation networks, multiple pathways are set to process separately multimodal images and the final features [25] or results [26] are combined for decision making. The layer-level fusion network [17,18,23,24] fuses multi-source features in mediate layers to obtain complementary and interdependent features.…”
Section: Multi-modal Fusionmentioning
confidence: 99%
See 1 more Smart Citation