Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2008
DOI: 10.1145/1401890.1401936
|View full text |Cite
|
Sign up to set email alerts
|

Fast logistic regression for text categorization with variable-length n-grams

Abstract: A common representation used in text categorization is the bag of words model (aka. unigram model). Learning with this particular representation involves typically some preprocessing, e.g. stopwords-removal, stemming. This results in one explicit tokenization of the corpus. In this work, we introduce a logistic regression approach where learning involves automatic tokenization. This allows us to weaken the a-priori required knowledge about the corpus and results in a tokenization with variable-length (word or … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
50
0

Year Published

2012
2012
2022
2022

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 70 publications
(50 citation statements)
references
References 35 publications
0
50
0
Order By: Relevance
“…In this work we study a sequence classification method, the Sequence Learner (SEQL), introduced in [10], [11]. Due to its greedy optimization approach, SEQL can quickly capture the distinct patterns of sequence data in very high-dimensional spaces.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…In this work we study a sequence classification method, the Sequence Learner (SEQL), introduced in [10], [11]. Due to its greedy optimization approach, SEQL can quickly capture the distinct patterns of sequence data in very high-dimensional spaces.…”
Section: Related Workmentioning
confidence: 99%
“…Sequence Learner SEQL learns discriminative subsequences from training data by exploiting the all-subsequence space using a coordinate gradient descent approach [10], [11]. The key idea is to exploit the structure of the subsequence space in order to efficiently optimize a classification loss function, such as the binomial log-likelihood loss of Logistic Regression or squared hinge loss of Support Vector Machines.…”
Section: Classification With Sequence Learnermentioning
confidence: 99%
See 2 more Smart Citations
“…Studies in [76], [77] designed linear classifiers to train explicit mappings of sequence data, where features correspond to subsequences. Using the relation between subsequences, they are able to design efficient training methods for very high dimensional mappings.…”
Section: A Training and Testing Explicit Data Mappings Via Linear CLmentioning
confidence: 99%