2016
DOI: 10.48550/arxiv.1604.00077
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Neural Attention Models for Sequence Classification: Analysis and Application to Key Term Extraction and Dialogue Act Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(3 citation statements)
references
References 14 publications
0
3
0
Order By: Relevance
“…Instead of producing heatmaps [28] or highlighting the words with different colours [29], we directly enlarge and darken the words in the visualization proportional to the attention weights to try reduce the cognitive load of the reader.…”
Section: Analysis Of How Human Interpret Attention From Supervised De...mentioning
confidence: 99%
“…Instead of producing heatmaps [28] or highlighting the words with different colours [29], we directly enlarge and darken the words in the visualization proportional to the attention weights to try reduce the cognitive load of the reader.…”
Section: Analysis Of How Human Interpret Attention From Supervised De...mentioning
confidence: 99%
“…This kind of attention mechanism has been integrated with LSTMs in natural language processing (NLP). Shen and Lee [19] applied an attention-mechanism LSTM for key term extraction and dialogue act detection. They showed that the attention mechanism enables the LSTM to ignore noisy or irrelevant passages in long input sequences, as well as to identify important portions of a given sequence to improve sequence labeling accuracy.…”
Section: Attention Mechanismsmentioning
confidence: 99%
“…We compare the performance of the proposed HA-TCN model to that of the classical SVM operating on the handcrafted features described above, and to those of recurrent architectures based on LSTMs and Bidirectional LSTMs (Bi-LSTM), and of recurrent architectures with attention mechanisms like attention-based LSTMs [19] and attentionbased Bi-LSTMs [26], as well as that of the traditional TCN [3]. For deep learning models, each raw handgrip time series is truncated or padded to a total length of 750 time steps, and its amplitude normalized between [0, 1].…”
Section: Dataset and Experimental Setupmentioning
confidence: 99%