2015 IEEE Winter Conference on Applications of Computer Vision 2015
DOI: 10.1109/wacv.2015.75
|View full text |Cite
|
Sign up to set email alerts
|

A Multi-modal Sparse Coding Classifier Using Dictionaries with Different Number of Atoms

Abstract: Most of classification methods including the ones based on sparse representation (SRC), look at every training sample and its extracted modalities as a single point in a high dimensional space and a collection of these points build the training space used to train the classifier. In a multimodality classification problem, there might be lots of redundancies associated with different modalities of the training data which degrade the performance of the classifier. This paper considers the problem of multi-modali… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 15 publications
0
2
0
Order By: Relevance
“…In particular, many researchers use deep Convolutional Neural Network (CNN) to conduct cross-modal matching [84,85,86] or SBIR which is essentially to learn some feature subspaces to match multi-modal data. Moreover, the convolutional sparse coding technology can also learn subspace satisfying certain qualities [87,88,89], which illustrates the convolutional idea and subspace learning can be reasonably combined. Therefore, it is natural to also utilize cross-modal subspace learning concepts to improve CNN for SBIR, and potentially incorporating saliency information [90,91] to improve partlevel examination in the same network.…”
Section: Discussion and Future Workmentioning
confidence: 93%
“…In particular, many researchers use deep Convolutional Neural Network (CNN) to conduct cross-modal matching [84,85,86] or SBIR which is essentially to learn some feature subspaces to match multi-modal data. Moreover, the convolutional sparse coding technology can also learn subspace satisfying certain qualities [87,88,89], which illustrates the convolutional idea and subspace learning can be reasonably combined. Therefore, it is natural to also utilize cross-modal subspace learning concepts to improve CNN for SBIR, and potentially incorporating saliency information [90,91] to improve partlevel examination in the same network.…”
Section: Discussion and Future Workmentioning
confidence: 93%
“…Originated to explaining neuronal activation of human visual system Olshausen and Field [1997], sparse coding model has been widely used in machine learning applications to model different uni-modal data, such as image, text and audio Mairal et al [2010], Yang et al [2009], Yogatama et al [2015, Arora et al [2018], Whitaker andAnderson [2016], Grosse et al [2012]. Also, there is a line of research to develop sparse representations for multiple modalities simultaneously Yuan et al [2012], Shafiee et al [2015], Gwon et al [2016].…”
Section: Data Distributionmentioning
confidence: 99%