Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.11
|View full text |Cite
|
Sign up to set email alerts
|

Self-Supervised Knowledge Triplet Learning for Zero-Shot Question Answering

Abstract: The aim of all Question Answering (QA) systems is to generalize to unseen questions. Current supervised methods are reliant on expensive data annotation. Moreover, such annotations can introduce unintended annotator bias, making systems focus more on the bias than the actual task. This work proposes Knowledge Triplet Learning (KTL), a self-supervised task over knowledge graphs. We propose heuristics to create synthetic graphs for commonsense and scientific knowledge. We propose using KTL to perform zero-shot q… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
40
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 32 publications
(45 citation statements)
references
References 51 publications
0
40
0
Order By: Relevance
“…In Table 9, we observe that self-talk significantly hurts performance for this dataset. On the other hand, contrastive explanations are found to be useful and approach the zero-shot performance of the stateof-the-art, which uses ConceptNet (Banerjee and Baral, 2020b). These results demonstrate that the set of contrastive prompts are generalizable to other commonsense reasoning datasets, and that while our contrastive prompts are limited to contrasting two answer choices at a time, the framework can be easily extended to tasks with multiple foils.…”
Section: Generalizability Of Promptsmentioning
confidence: 67%
See 1 more Smart Citation
“…In Table 9, we observe that self-talk significantly hurts performance for this dataset. On the other hand, contrastive explanations are found to be useful and approach the zero-shot performance of the stateof-the-art, which uses ConceptNet (Banerjee and Baral, 2020b). These results demonstrate that the set of contrastive prompts are generalizable to other commonsense reasoning datasets, and that while our contrastive prompts are limited to contrasting two answer choices at a time, the framework can be easily extended to tasks with multiple foils.…”
Section: Generalizability Of Promptsmentioning
confidence: 67%
“…Table 9: Zero-shot test performance on Common-senseQA for baselines as well as contrastive models which ensemble fact/foil pairs by voting (V) and maximum margin (MM). The best reported unsupervised performance (Banerjee and Baral, 2020b) uses Con-ceptNet, which was used to construct the dataset.…”
Section: Discussionmentioning
confidence: 99%
“…There are different ways to frame the task of answering questions by machine reading, including sentence retrieval (Momtazi and Klakow, 2015), multi-hop reasoning (Khot et al, 2020), and reasoning about multiple paragraphs or documents at the same time (Dua et al, 2019;Cao et al, 2019). Recent work has considered the development of reasoning-based QA systems (Weber et al, 2019) as well as the integration of external (Banerjee and Baral, 2020) and commonsense knowledge tured knowledge graphs. A prominent example is the series of Question Answering over Linked Data (QALD) evaluation campaigns (Usbeck et al, 2018), now in its 9th edition and going back to 2011.…”
Section: Related Workmentioning
confidence: 99%
“…There are different ways to frame the task of answering questions by machine reading, including sentence retrieval (Momtazi and Klakow, 2015), multi-hop reasoning (Khot et al, 2020), and reasoning about multiple paragraphs or documents at the same time (Dua et al, 2019;. Recent work has considered the development of reasoning-based QA systems (Weber et al, 2019) as well as the integration of external (Banerjee and Baral, 2020) and commonsense knowledge (Dua et al, 2019) focuses on complex reasoning tasks in form of both, open and closed domain, questions. The task combines the challenge of extractive QA with testing a models ability to perform limited numerical reasoning, e.g.…”
Section: Related Workmentioning
confidence: 99%