The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2018 International Joint Conference on Neural Networks (IJCNN) 2018
DOI: 10.1109/ijcnn.2018.8489651
|View full text |Cite
|
Sign up to set email alerts
|

Deep Quaternion Networks

Abstract: The field of deep learning has seen significant advancement in recent years. However, much of the existing work has been focused on real-valued numbers. Recent work has shown that a deep learning system using the complex numbers can be deeper for a fixed parameter budget compared to its real-valued counterpart. In this work, we explore the benefits of generalizing one step further into the hyper-complex numbers, quaternions specifically, and provide the architecture components needed to build deep quaternion n… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
140
2

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 147 publications
(157 citation statements)
references
References 16 publications
2
140
2
Order By: Relevance
“…Similar to [65] and [67], the diagonal of γ is initialized to 1/ 8 , the off diagonal terms of γ and all components of β are initialized to 0.…”
Section: Octonion Batch Normalization Modulementioning
confidence: 99%
See 1 more Smart Citation
“…Similar to [65] and [67], the diagonal of γ is initialized to 1/ 8 , the off diagonal terms of γ and all components of β are initialized to 0.…”
Section: Octonion Batch Normalization Modulementioning
confidence: 99%
“…Since the images in datasets of CIFAR-10 and CIFAR-100 are real-valued, however, the input of the proposed deep octonion networks needs to be octonion matrix, which we should to be obtained first. The octonion has one real part and seven imaginary parts, we put the original N training real images into the real part, and similar to [65] and [67], To speed up the training, the following layer is an AveragePooling2D layer, which is then followed by a fully connected layer called Dense to classify the input. The deep octonion network model sets the number of residual blocks in the three stages to 10, 9, and 9, respectively, and the number of convolution filters is set to 32, 64, and 128.…”
Section: Octonion Input Constructionmentioning
confidence: 99%
“…The goal of learning effective representations lives at the heart of deep learning research. While most neural architectures for NLP have mainly explored the usage of real-valued representations (Vaswani et al, 2017;Bahdanau et al, 2014;Parikh et al, 2016), there have also been emerging interest in complex (Danihelka et al, 2016;Arjovsky et al, 2016;Gaudet and Maida, 2017) and hypercomplex representations (Parcollet et al, 2018b,a;Gaudet and Maida, 2017).…”
Section: Related Workmentioning
confidence: 99%
“…Notably, progress on Quaternion and hypercomplex representations for deep learning is still in its infancy and consequently, most works on this topic are very recent. Gaudet and Maida proposed deep Quaternion networks for image classification, introducing basic tools such as Quaternion batch normalization or Quaternion initialization (Gaudet and Maida, 2017). In a similar vein, Quaternion RNNs and CNNs were proposed for speech recognition (Parcollet et al, 2018a,b).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation