2018
DOI: 10.14419/ijet.v7i3.8.15210
|View full text |Cite
|
Sign up to set email alerts
|

A Review of Different Text Categorization Techniques

Abstract: In this paper, we focus on a major internet problem which is a huge amount of uncategorized text. We review existing techniques used for feature selection and categorization. After reviewing the existing literature, it was found that there exist some gaps in existing algorithms, one of which is a requirement of the labeled dataset for the training of the classifier.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
5
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(5 citation statements)
references
References 5 publications
(5 reference statements)
0
5
0
Order By: Relevance
“…And second type of text classification, is called the "multi-label classification" it is considered as a multi-label if there are two or more classes assigned in a document. [8] And today, neural networks and deep learning models create major advances in the natural language processing (NLP) field such as text categorization. HD Wehle (2017) defined deep learning as a form of machine learning that can be either utilized by a supervised or unsupervised learning or both.…”
Section: Neural Network and Deep Learning For Text Categorizationmentioning
confidence: 99%
See 1 more Smart Citation
“…And second type of text classification, is called the "multi-label classification" it is considered as a multi-label if there are two or more classes assigned in a document. [8] And today, neural networks and deep learning models create major advances in the natural language processing (NLP) field such as text categorization. HD Wehle (2017) defined deep learning as a form of machine learning that can be either utilized by a supervised or unsupervised learning or both.…”
Section: Neural Network and Deep Learning For Text Categorizationmentioning
confidence: 99%
“…HD Wehle (2017) defined deep learning as a form of machine learning that can be either utilized by a supervised or unsupervised learning or both. [8] Recently, the success of deep learning models in the image classification have attracted considerable attentions to used it in the text classification problem. [9] In 2014 Yoon Kim, used the Convolutional Neural Network (CNN) to classify sentences.…”
Section: Neural Network and Deep Learning For Text Categorizationmentioning
confidence: 99%
“…Features refined through selection and extraction are fed into classifiers for training and prediction. Traditionally, the most popular classifiers include Naive Bayes, K Nearest Neighbour, Decision Tree, Random Forest, and Support Vector Machine (Aggarwal et al, 2018). Lately, deep-learning-based classifiers have achieved impressive results in TC as they are able to model complex non-linear relationships within data (Kowsari et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…A number of review studies have already been carried out. For instance, Aggarwal et al (2018) and Kowsari et al (2019) presented a general overview of TC algorithms; Manikandan and Sivakumar (2018) and Kadhim (2019) conducted surveys on machine-learningbased techniques for TC; Altinel and Ganiz (2018) reviewed the history and development of semantic approaches to TC; Shah and Patel (2016) compared existing methods for feature selection and extraction. However, to our knowledge, no research has been conducted to systematically review TC research with large-scale bibliographic data from a bibliometric perspective.…”
Section: Introductionmentioning
confidence: 99%
“…Data classification is one of the most important tasks for different applications, such as text categorization, tone recognition, image classification, microarray gene expression, and protein structure prediction ( Choi et al, 2017 ; Johnson and Zhang, 2017 ; Malhotra et al, 2017 ; Aggarwal et al, 2018 ; Fang et al, 2018 ; Mikołajczyk and Grochowski, 2018 ; Kerkeni et al, 2019 ; Saritas and Yasar, 2019 ; Yildirim et al, 2019 ; Chandrasekar et al, 2020 ). Many types of information (e.g., language, music, and gene) can be represented as sequential data that often contains related information separated by many time steps, and these long-term dependencies are difficult to model as we must retain information from the whole sequence with greater complexity of the model ( Trinh et al, 2018 ; Liu et al, 2019 ; Shewalkar, 2019 ; Yu et al, 2019 ; Zhao et al, 2020 ).…”
Section: Introductionmentioning
confidence: 99%