Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.47
|View full text |Cite
|
Sign up to set email alerts
|

SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer Matching Retrieval

Abstract: We introduce SPARTA, a novel neural retrieval method that shows great promise in performance, generalization, and interpretability for open-domain question answering. Unlike many neural ranking methods that use dense vector nearest neighbor search, SPARTA learns a sparse representation that can be efficiently implemented as an Inverted Index. The resulting representation enables scalable neural retrieval that does not require expensive approximate vector search and leads to better performance than its dense co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 27 publications
(16 citation statements)
references
References 45 publications
0
13
0
Order By: Relevance
“…Moreover, recent works addressed ranking with sequence-to-sequence transformers based approach as the Mono-T5 model [31] for re-ranking documents returned by a BM25 ranker. Using a weak initial ranker such as BM25 may be the bottleneck of reaching higher performances, some approaches are thus reconsidering dense retrieval [14,15,7,44]. All these models are data-dependent, relying on word/topic/query distribution in the training dataset and their application to new domains is not always straightforward [28,32].…”
Section: Related Workmentioning
confidence: 99%
“…Moreover, recent works addressed ranking with sequence-to-sequence transformers based approach as the Mono-T5 model [31] for re-ranking documents returned by a BM25 ranker. Using a weak initial ranker such as BM25 may be the bottleneck of reaching higher performances, some approaches are thus reconsidering dense retrieval [14,15,7,44]. All these models are data-dependent, relying on word/topic/query distribution in the training dataset and their application to new domains is not always straightforward [28,32].…”
Section: Related Workmentioning
confidence: 99%
“…Along with the success of deep learning that offers remarkable semantic representation, various deep retrieval models have been developed in the past few years, greatly enhancing retrieval effectiveness and thus lifting final QA performance. According to the different ways of encoding the question and document as well as of scoring their similarity, dense retrievers in existing OpenQA systems can be roughly divided into three types: Representation-based Retriever [16], [29], [36], [72], Interaction-based Retriever [15], [31], and Representation-interaction Retriever [17], [81], as illustrated in Fig. 5.…”
Section: Dense Retrievermentioning
confidence: 99%
“…Representation-interaction Retriever: In order to achieve both high accuracy and efficiency, some recent systems [17], [81] combine representation-based and interaction-based methods. For instance, ColBERT-QA [17] develops its retriever based on ColBERT [82], which extends the dual-encoder architecture by performing a simple token-level interaction step over the question and document representations to calculate the similarity score.…”
Section: Retriever-only Denspimentioning
confidence: 99%
See 1 more Smart Citation
“…There are two popular approaches in conversational chatbot modeling namely Transformer network-based models such as [13]- [15] and recurrent neural network (RNN)-based sequence to sequence learning (Seq2Seq) models such as [16]- [22]. The Transformer network is based on the feed-forward network [11], wherein sentences are processed as a whole rather than word by word by utilizing a self-attention mechanism, which can be highly parallelized.…”
Section: Introductionmentioning
confidence: 99%