2019
DOI: 10.3390/a12030051
|View full text |Cite
|
Sign up to set email alerts
|

Optimized Deep Convolutional Neural Networks for Identification of Macular Diseases from Optical Coherence Tomography Images

Abstract: Finetuning pre-trained deep neural networks (DNN) delicately designed for large-scale natural images may not be suitable for medical images due to the intrinsic difference between the datasets. We propose a strategy to modify DNNs, which improves their performance on retinal optical coherence tomography (OCT) images. Deep features of pre-trained DNN are high-level features of natural images. These features harm the training of transfer learning. Our strategy is to remove some deep convolutional layers of the s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
18
0
1

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 76 publications
(20 citation statements)
references
References 30 publications
1
18
0
1
Order By: Relevance
“…Fig 6(A)–6(E) demonstrate how factorization of 5×5, 7×7 convolutions into smaller convolutions 3×3, or asymmetric convolutions 1×7 and 7×1, is conducted during the experiment. This reduces the number of deep network parameters [ 46 ]. Compared against the new classification layers added, the original classification layers replaced include the global average pooling layer and the last dense layer with 1000 outputs.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Fig 6(A)–6(E) demonstrate how factorization of 5×5, 7×7 convolutions into smaller convolutions 3×3, or asymmetric convolutions 1×7 and 7×1, is conducted during the experiment. This reduces the number of deep network parameters [ 46 ]. Compared against the new classification layers added, the original classification layers replaced include the global average pooling layer and the last dense layer with 1000 outputs.…”
Section: Methodsmentioning
confidence: 99%
“…In the case of a convolution block, its input dimensions are smaller than its output dimensions due to the convolutional layer’s availability at a shortcut ( Fig 8(C) ). In both blocks, 1×1 convolution is implemented at the beginning, and the end of the network through a technique called a bottleneck design [ 46 ]. This technique decreases the number of parameters without degrading the performance of the network.…”
Section: Methodsmentioning
confidence: 99%
“…The identity block has not convolution layer in shortcut, and the output possesses the same dimension as the input dimension. [ 28 ] As shown in Figure 3 both these blocks consist of two 3 × 3 convolution layers, followed by ReLU and Batch normalization layers.…”
Section: Proposed Methodsmentioning
confidence: 99%
“…Once the three layers are added to the implemented CNN architectures, we use "Adam" as the adaptive optimizer of the learning rate. Figure 5 displays the summary of DenseNet121 architecture with the three added layers, and Figure 6 shows the detailed DenseNet121 architecture [39]. We have built the eight CNN architectures that we have already mentioned in Int J Artif Intell ISSN: 2252-8938…”
Section: Network Building and Learningmentioning
confidence: 99%