2016
DOI: 10.1109/tcsvt.2015.2477937
|View full text |Cite
|
Sign up to set email alerts
|

Multi-loss Regularized Deep Neural Network

Abstract: A proper strategy to alleviate overfitting is critical for deep neural network (DNN). In this work, we introduce the cross-loss-function regularization for boosting the generalization capability of the DNN, which results in the Multi-Loss regularized Deep Neural Network (ML-DNN) framework. For a particular learning task, e.g., image classification, only a single loss function is used for all previous DNNs, and the intuition behind the multiloss framework is that the extra loss functions with different theoreti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
46
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
5
3
1

Relationship

2
7

Authors

Journals

citations
Cited by 79 publications
(46 citation statements)
references
References 21 publications
0
46
0
Order By: Relevance
“…Additional works where state-of-the-art results were obtained without data augmentation include those by McFonnell and Vladusich [35], who reported a test error rate of 0.37% using a fast-learning shallow convolutional neural network, Mairal et al [36], who achieved 0.39% using convolutional kernel networks, Xu et al [37], who explored multi-loss regularization in CNNs obtaining an error rate of 0.42%, and Srivastava et al [38] used so-called convolutional "highway" networks (inspired by LSTM recurrent networks) to achieve an error rate of 0.45%.…”
Section: State Of the Artmentioning
confidence: 99%
“…Additional works where state-of-the-art results were obtained without data augmentation include those by McFonnell and Vladusich [35], who reported a test error rate of 0.37% using a fast-learning shallow convolutional neural network, Mairal et al [36], who achieved 0.39% using convolutional kernel networks, Xu et al [37], who explored multi-loss regularization in CNNs obtaining an error rate of 0.42%, and Srivastava et al [38] used so-called convolutional "highway" networks (inspired by LSTM recurrent networks) to achieve an error rate of 0.45%.…”
Section: State Of the Artmentioning
confidence: 99%
“… Multi-loss function Different loss functions lead the networks to reach different local minima. The analysis of different losses also showed they have their own strengths and limitations (Janocha & Czarnecki, 2017;Rosasco, De Vito, Caponnetto, Piana, & Verri, 2004;C. Xu et al, 2016), while the question arises as how to combine these advantages in a unified system without the heavy computational load required for independent runs of different loss functions and combining them.…”
Section:  Ce Loss and Csd Lossmentioning
confidence: 99%
“…From another point of view, different loss functions have complementary advantages and limitations (Janocha & Czarnecki, 2017;C. Xu et al, 2016).…”
Section: Introductionmentioning
confidence: 99%
“…Ren et al [25] proposed a Region Proposal Network (RPN) that shared full-image convolutional features with the detection network, thus enabling nearly cost-free region proposals. In [26], a Multi-Loss regularized Deep Neural Network (ML-DNN) framework was proposed, which exploited multiple loss functions with different theoretical motivations to mitigate overfitting during semantic concept learning. He et al [27] proposed a residual learning framework to alleviate the training of neural networks.…”
Section: Related Workmentioning
confidence: 99%