Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1515
|View full text |Cite
|
Sign up to set email alerts
|

A strong baseline for question relevancy ranking

Abstract: The best systems at the SemEval-16 and SemEval-17 community question answering shared tasks-a task that amounts to question relevancy ranking-involve complex pipelines and manual feature engineering. Despite this, many of these still fail at beating the IR baseline, i.e., the rankings provided by Google's search engine. We present a strong baseline for question relevancy ranking by training a simple multi-task feed forward network on a bag of 14 distance measures for the input question pair. This baseline mode… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
4

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 12 publications
0
4
0
Order By: Relevance
“…In general, the effectiveness of these MTL baselines on the QC task is limited because there are only a small amount of QD pairs available for training. Both our method and its ablated variant outperform the Python SQL MAP nDCG MAP nDCG MTL-MLP (Gonzalez et al, 2018) MTL baselines. This shows that it may be more effective to use a data scarce task to regularize the adversarial learning of a relatively data rich task, than using those scarce data in MTL.…”
Section: Results and Analysesmentioning
confidence: 90%
See 2 more Smart Citations
“…In general, the effectiveness of these MTL baselines on the QC task is limited because there are only a small amount of QD pairs available for training. Both our method and its ablated variant outperform the Python SQL MAP nDCG MAP nDCG MTL-MLP (Gonzalez et al, 2018) MTL baselines. This shows that it may be more effective to use a data scarce task to regularize the adversarial learning of a relatively data rich task, than using those scarce data in MTL.…”
Section: Results and Analysesmentioning
confidence: 90%
“…The results are reported in Table 4. The MTL-MLP model is originally proposed to improve question-question relevance prediction by using question-comment relevance prediction as a secondary task (Gonzalez et al, 2018). It does not perform as well as MTL-DCS, which basically uses hard parameter sharing between the two tasks and does not require additional similarity feature definitions.…”
Section: Results and Analysesmentioning
confidence: 99%
See 1 more Smart Citation
“…We approximate a point p ∈ S by specifying some error margin > 0 so that dist(p, q) ≤ (1 + )(dist(p * , q)), where p * is the real nearest neighbor. Because we use approximate search, we rerank the retrieved utterances using a feed-forward ranking model, introduced in Gonzalez et al (2018). Their ranking model is a multi-task model, which relies on simple textual similarity measures combined in a multi-layered perceptron architecture.…”
Section: Exemplar-hredmentioning
confidence: 99%