2023
DOI: 10.1016/j.jksuci.2023.02.012
|View full text |Cite
|
Sign up to set email alerts
|

IB-TransUNet: Combining Information Bottleneck and Transformer for Medical Image Segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(7 citation statements)
references
References 14 publications
0
1
0
Order By: Relevance
“…A deep supervised transformer full-resolution residual network was presented in [ 37 ]. Its feature fusion is better and suppresses irrelevant features, while the deep supervision mechanism reduces the gradient vanishing problem.…”
Section: Organsmentioning
confidence: 99%
“…A deep supervised transformer full-resolution residual network was presented in [ 37 ]. Its feature fusion is better and suppresses irrelevant features, while the deep supervision mechanism reduces the gradient vanishing problem.…”
Section: Organsmentioning
confidence: 99%
“…The BGRD-TransUNet model is an improvement upon the TransUNet model. Currently, some models based on Tran-sUNet, such as DA-TransUNet [28] and IB-TransUNet [10], introduce bottleneck modules between the CNN and Transformer to optimize the feature maps fed into the Transformer. However, the CNN components of these models still utilize the original ResNet50 [13] structure of TransUNet without modification.…”
Section: A Overall Structurementioning
confidence: 99%
“…However, due to the inherent limitations of CNN operations, U-Net often struggles to capture global contextual information, [9]. This limitation arises because the size of the convolutional kernel in CNN determines its local perception ability, restricting a single convolutional kernel to focus on a limited local region, [10]. While CNN models can gradually expand the receptive field of convolutional kernels by stacking numerous convolutional layers (which allows higher-level convolutional kernels to focus on global features), lower and intermediate-level convolutional kernels remain constrained and can only focus on local regions.…”
Section: Introductionmentioning
confidence: 99%
“…These transformers divide images into patches and apply the attention mechanism between each set of patches. Yet, since there is no natural way of reducing or expanding the dimensionality of the data, the ViT has been used to dimensionality reduction tasks only to a limited degree [33][34][35][36] . Importantly, current ViTs have not been integrated with increasing numbers of channels in convolutional layers to represent increasingly complicated features.…”
Section: Transformer-based Dimensionality Reductionmentioning
confidence: 99%