2019
DOI: 10.1609/aaai.v33i01.33016932
|View full text |Cite
|
Sign up to set email alerts
|

COALA: A Neural Coverage-Based Approach for Long Answer Selection with Small Data

Abstract: Current neural network based community question answering (cQA) systems fall short of (1) properly handling long answers which are common in cQA; (2) performing under small data conditions, where a large amount of training data is unavailable—i.e., for some domains in English and even more so for a huge number of datasets in other languages; and (3) benefiting from syntactic information in the model—e.g., to differentiate between identical lexemes with different syntactic roles. In this paper, we propose COALA… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
35
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 18 publications
(35 citation statements)
references
References 24 publications
0
35
0
Order By: Relevance
“…To test whether our strategy to train models with title-body pairs is also suitable for answer selection, we use the data and code of Rücklé et al (2019a) and train two different types of models with WS-TB on their five datasets that are based on StackExchange Apple, Aviation, Academia, Cooking, and Travel. We train (1) a siamese BiLSTM, which learns question and answer representations; and (2) their neural relevance matching model COALA.…”
Section: Answer Selectionmentioning
confidence: 99%
See 1 more Smart Citation
“…To test whether our strategy to train models with title-body pairs is also suitable for answer selection, we use the data and code of Rücklé et al (2019a) and train two different types of models with WS-TB on their five datasets that are based on StackExchange Apple, Aviation, Academia, Cooking, and Travel. We train (1) a siamese BiLSTM, which learns question and answer representations; and (2) their neural relevance matching model COALA.…”
Section: Answer Selectionmentioning
confidence: 99%
“…https://github.com/huggingface/ pytorch-transformers same as for BiLSTM (e.g., loss calculation). We train the models until they do not improve for at least 20 epochs, and we restore the weights of the epoch that obtained the best development score.For all other datasets (AskUbuntu-Lei and Answer Selection datasets) we add BERT to the experimental software ofRücklé et al (2019a). We do not include it in the software ofLei et al (2016) because it is tightly coupled to the Theano framework, which is not actively maintained.…”
mentioning
confidence: 99%
“…Output matrix η nm ∈ R |Q |× |P | contains the relevance scores of all pairs between n-grams in query and m-grams in answer. From η nm , we conduct a row-wise max-pooling to obtain A nm , relaxing the length constraint in interactions (Rücklé et al, 2019).…”
Section: Matching Modulementioning
confidence: 99%
“…InsuranceQA WikiPassageQA Accuracy MAP MRR P@5 P@10 nDCG R@5 R@10 R@20 IR Based (Rücklé et al, 2019) . COALA syntax-aware (Rücklé and Gurevych, 2017) is a variant of COALA using dependency parse trees (Schuster and Manning, 2016).…”
Section: Modelmentioning
confidence: 99%
See 1 more Smart Citation