Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.241
|View full text |Cite
|
Sign up to set email alerts
|

Database reasoning over text

Abstract: Neural models have shown impressive performance gains in answering queries from natural language text. However, existing works are unable to support database queries, such as "List/Count all female athletes who were born in 20th century", which require reasoning over sets of relevant facts with operations such as join, filtering and aggregation. We show that while state-of-the-art transformer models perform very well for small databases, they exhibit limitations in processing noisy data, numerical operations, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2
2

Relationship

1
8

Authors

Journals

citations
Cited by 16 publications
(8 citation statements)
references
References 26 publications
0
3
0
Order By: Relevance
“…These models have been evaluated primarily via perplexity and non-NLP benchmarks (Tay et al, 2020b). These metrics may not be ideal (Sun et al, 2021) and may not reflect performance on complex NLP tasks (Arutiunian et al, 2020;Thorne et al, 2021). We argue these metrics have not been sufficient for the development of efficient Transformers and their practical application on long texts, and that existing benchmarks are insufficient guides for architecture selection.…”
Section: Introductionmentioning
confidence: 90%
“…These models have been evaluated primarily via perplexity and non-NLP benchmarks (Tay et al, 2020b). These metrics may not be ideal (Sun et al, 2021) and may not reflect performance on complex NLP tasks (Arutiunian et al, 2020;Thorne et al, 2021). We argue these metrics have not been sufficient for the development of efficient Transformers and their practical application on long texts, and that existing benchmarks are insufficient guides for architecture selection.…”
Section: Introductionmentioning
confidence: 90%
“…Answering Database Queries There has been substantial effort put into converting queries expressed in natural language into SQL queries for databases with known structure [1,16,38], and there have also been advancements in adapting this approach for semistructured data and knowledge bases [4,20]. Recently, Thorne et al [28,29] proposed NeuralDB as a way to perform database queries over a collection of textual documents without the need to translate data or queries into a predefined database schema but using parallel neural techniques instead. Their approach is very effective but it: (i) requires preprocessing and analysis for the aggregation operator; (ii) is limited to simple queries and (iii) is capable of handling data just in textual format.…”
Section: Related Workmentioning
confidence: 99%
“…In this perspective paper, we propose to study, design, and build MMNDBs by combining the capabilities of large multimodal models, multi-media information retrieval, and database query processing, as shown in Figure 1. We have been inspired by the work on neural databases [26,28,29] that have garnered interest in the NLP, database, and IR communities. However, we differentiate from that work as we position ourselves as an evolution of the field of MMIR by means of modern and, more recently proposed, multimodal AI technologies.…”
Section: Introductionmentioning
confidence: 99%
“…While the dataset is bilingual, it uses crowdsourced questions and is not designed for compositionality analysis. Recently, Thorne et al (2021) proposed WIKINLDB, a Wikidata-based English KBQA dataset, focusing on scalability rather than compositionality. Other related datasets include QALM (Kaffee et al, 2019), a dataset for multilingual question answering over a set of different popular knowledge graphs, intended to help determine the multilinguality of those knowledge graphs.…”
Section: Related Workmentioning
confidence: 99%