2021
DOI: 10.48550/arxiv.2105.07065
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Visual analogy: Deep learning versus compositional models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 15 publications
0
4
0
Order By: Relevance
“…4 The failure of large language models in acquiring functional linguistic competence 4.1 LLMs are great at pretending to think Large text corpora contain a wealth of non-linguistic information, from mathematical and scientific facts (e.g., "two plus seven is nine") to factual knowledge (e.g., "the capital of Texas is Austin") to harmful stereotypes (e.g., "women belong in the kitchen"). This is not particularly surprising since even simple patterns of co-occurrence between words capture rich conceptual knowledge, including object properties [e.g., Grand et al, 2022, Huebner and Willits, 2018, Unger and Fisher, 2021, Utsumi, 2020, van Paridon et al, 2021, abstract analogies [Ichien et al, 2021], social biases [e.g., Bolukbasi et al, 2016, Caliskan et al, 2017, Lewis and Lupyan, 2020, and expert knowledge in specialized domains [e.g., Tshitoyan et al, 2019]. Moreover, statistical regularities extracted from language and from visual scenes exhibit a substantial degree of correspondence [Roads andLove, 2020, Sorscher et al, 2021], indicating that linguistic information can capture at least some aspects of experiential input [e.g., Abdou et al, 2021, Patel and.…”
Section: Interim Conclusionmentioning
confidence: 99%
“…4 The failure of large language models in acquiring functional linguistic competence 4.1 LLMs are great at pretending to think Large text corpora contain a wealth of non-linguistic information, from mathematical and scientific facts (e.g., "two plus seven is nine") to factual knowledge (e.g., "the capital of Texas is Austin") to harmful stereotypes (e.g., "women belong in the kitchen"). This is not particularly surprising since even simple patterns of co-occurrence between words capture rich conceptual knowledge, including object properties [e.g., Grand et al, 2022, Huebner and Willits, 2018, Unger and Fisher, 2021, Utsumi, 2020, van Paridon et al, 2021, abstract analogies [Ichien et al, 2021], social biases [e.g., Bolukbasi et al, 2016, Caliskan et al, 2017, Lewis and Lupyan, 2020, and expert knowledge in specialized domains [e.g., Tshitoyan et al, 2019]. Moreover, statistical regularities extracted from language and from visual scenes exhibit a substantial degree of correspondence [Roads andLove, 2020, Sorscher et al, 2021], indicating that linguistic information can capture at least some aspects of experiential input [e.g., Abdou et al, 2021, Patel and.…”
Section: Interim Conclusionmentioning
confidence: 99%
“…A new set of VAPs was also constructed in [168]. In contrast to V-PROM, matrices from [168] focus on renderings of realistic cars from the ShapeNet dataset [167], as shown in Fig. 18b.…”
Section: Avr-like Tasks For Representation Learningmentioning
confidence: 99%
“…The images vary in texture, shading and viewpoint. Using this dataset, the authors presented a case where a general segmentation model performed better than task-specific architectures trained solely on the automotive VAPs [168].…”
Section: Avr-like Tasks For Representation Learningmentioning
confidence: 99%
See 1 more Smart Citation