Proceedings of the 15th Conference of the European Chapter of The Association for Computational Linguistics: Volume 1 2017
DOI: 10.18653/v1/e17-1104
|View full text |Cite
|
Sign up to set email alerts
|

Very Deep Convolutional Networks for Text Classification

Abstract: The dominant approach for many NLP tasks are recurrent neural networks, in particular LSTMs, and convolutional neural networks. However, these architectures are rather shallow in comparison to the deep convolutional networks which have pushed the state-of-the-art in computer vision. We present a new architecture (VD-CNN) for text processing which operates directly at the character level and uses only small convolutions and pooling operations. We are able to show that the performance of this model increases wit… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

5
581
0
8

Year Published

2017
2017
2023
2023

Publication Types

Select...
3
2
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 776 publications
(594 citation statements)
references
References 18 publications
5
581
0
8
Order By: Relevance
“…Fast Fourier Transform (FFT) provides an alternative approach to calculate the correlation coefficient with a high computational speed as compared to Equation (6) [65,66]. The correlation coefficient between A and B is computed by locating the maximum value of the following equation:…”
Section: Correlation Coefficientmentioning
confidence: 99%
See 1 more Smart Citation
“…Fast Fourier Transform (FFT) provides an alternative approach to calculate the correlation coefficient with a high computational speed as compared to Equation (6) [65,66]. The correlation coefficient between A and B is computed by locating the maximum value of the following equation:…”
Section: Correlation Coefficientmentioning
confidence: 99%
“…Deep convolutional neural networks (CNNs) recently have shown remarkable success in a variety of areas such as computer vision [1][2][3] and natural language processing [4][5][6]. CNNs are biologically inspired by the structure of mammals' visual cortexes as presented in Hubel and Wiesel's model [7].…”
Section: Introductionmentioning
confidence: 99%
“…37 Deep learning model architecture are designed based on the learning task, number of 38 the parameters and size of the dataset. Well-known deep learning models, e.g., ResNet 39 and VGGNet, from computer vision [2] have been reused to build advanced systems for 40 text processing such as Very Deep Convolution Network (VDCNN) c [5] operating at 41 character level directly. Text modeling and sentence classification have been also tackled 42 with a small number of convolution layers such as one layer, two layers and six 43 layers [6][7][8].…”
mentioning
confidence: 99%
“…The character level CNN model of with six convolutional layers was outperformed by a bag-of-words model for three out of four data sets for topic classification. The 29 layer model of Conneau et al (2016) improves the results; however, only for two out of the four data sets, the CNN performs better than the bag-of-words model. Wang et al (2015) developed a CNN with one convolutional layer and a layer that extracts several representations of the texts by applying multiple windows with various width over the pre-trained word embeddings.…”
Section: Topic and Question Classificationmentioning
confidence: 90%
“…For one of the data sets, a bag-of-words model outperformed the character CNN, while for the other three data sets, the CNN showed better performance. Conneau et al (2016) showed that the performance is further improved with a CNN containing 29 convolutional layers. This CNN also performs better than the bag-of-word model for the data set for which the bag-of-words model outperformed the smaller CNN.…”
Section: Sentiment Classificationmentioning
confidence: 99%