2021
DOI: 10.1109/tpami.2021.3058891
|View full text |Cite
|
Sign up to set email alerts
|

Deep Polynomial Neural Networks

Abstract: Deep Convolutional Neural Networks (DCNNs) are currently the method of choice both for generative, as well as for discriminative learning in computer vision and machine learning. The success of DCNNs can be attributed to the careful selection of their building blocks (e.g., residual blocks, rectifiers, sophisticated normalization schemes, to mention but a few). In this paper, we propose Π-Nets, a new class of DCNNs. Π-Nets are polynomial neural networks, i.e., the output is a high-order polynomial of the input… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
21
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 28 publications
(21 citation statements)
references
References 64 publications
0
21
0
Order By: Relevance
“…Then, depending on the interactions between the layers we want to forge, we can share the corresponding factor matrices. That results (see [164], [169] for detailed derivation) in a simple recursive relationship, that can be expressed as:…”
Section: F Tensor Structures In Polynomial Network and Attention Mechanismsmentioning
confidence: 99%
“…Then, depending on the interactions between the layers we want to forge, we can share the corresponding factor matrices. That results (see [164], [169] for detailed derivation) in a simple recursive relationship, that can be expressed as:…”
Section: F Tensor Structures In Polynomial Network and Attention Mechanismsmentioning
confidence: 99%
“…Recently, higher-order units were revisited [22]- [25]. In [22], a quadratic convolutional filter of the complexity O(n 2 ) was proposed to replace the linear filter, while in the work by Chrysos et al [23] the higher-order units as described by Eq. ( 3) were embedded into a deep network to reduce the complexity of the individual unit via tensor decomposition and factor sharing.…”
Section: Related Workmentioning
confidence: 99%
“…Such a network achieved cutting-edge performance on several tasks. Compared to [23], our group proposed a simplified quadratic neuron with O(3n) parameters and argued that more complicated neurons are not necessary based on algebraic fundamental theorem [26]. Interestingly, when only the first and second-order terms are kept, and the rank is set to two in tensor decomposition, the network of [23] becomes a special case of our quadratic model.…”
Section: Related Workmentioning
confidence: 99%
“…In the machine learning domain, Simon revealed an interesting correspondence between the multiplicative neuron and the additive neuron, according to the identify [ 47 ], and remarked that the multiplicative neuron network can be expressed in the form of an additive neuron network with a different nonlinearity [ 51 ]. At another venue, polynomial neural networks have been in progress, and their advantage over additive networks has been investigated [ 52 , 53 , 54 ]. Relations between polynomial regression and classical neural networks were discussed and polynomial activation functions were shown on the basis of the Taylor theorem [ 54 ].…”
Section: Introductionmentioning
confidence: 99%