2017
DOI: 10.1007/978-3-319-71246-8_3
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Sequence Regression by Learning Linear Models in All-Subsequence Space

Abstract: We present a new approach for learning a sequence regression function, i.e., a mapping from sequential observations to a numeric score. Our learning algorithm employs coordinate gradient descent and Gauss-Southwell optimization in the feature space of all subsequences. We give a tight upper bound for the coordinate wise gradients of squared error loss that enables efficient Gauss-Southwell selection. The proposed bound is built by separating the positive and the negative gradients of the loss function and expl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 15 publications
0
3
0
Order By: Relevance
“…Gsponer et al used a different approach, where a linear model is trained on a set of features extracted from sequences of the dataset [20]. Those features correspond to the search space of all possible subsequences.…”
Section: Heuristic Methodsmentioning
confidence: 99%
“…Gsponer et al used a different approach, where a linear model is trained on a set of features extracted from sequences of the dataset [20]. Those features correspond to the search space of all possible subsequences.…”
Section: Heuristic Methodsmentioning
confidence: 99%
“…In this work we adopt SEQL a linear sequence classifier algorithm, as we want to learn a model that is interpretable but still achieves high accuracy. The main idea behind SEQL is to use a greedy coordinate gradient descent with the Gauss-Southwell rule [13] which allows to avoid the explicit generation of the feature vectors [5]. A key step of this approach is the efficient search for the current best k-mer, in the sense of maximum absolute gradient value, followed by an update of the corresponding weight value β.…”
Section: Seqlmentioning
confidence: 99%
“…The main problem of using the k-mer frequency feature is dealing with a largedimension feature space when aiming to obtain high accuracy [14]. To solve this problem, Kusuma [15] introduced spaced k-mers, inspired by PatternHunter [16], to reduce the feature space dimension and improving accuracy.…”
Section: Introductionmentioning
confidence: 99%