2020
DOI: 10.1101/2020.10.26.354258
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Neural Network Models of Object Recognition Exhibit Human-Like Limitations when Performing Visual Search Tasks

Abstract: What limits our ability to find an object we are looking for? There are two competing models: one explains attentional limitations during visual search in terms of a serial processing computation, the other attributes limitations to noisy parallel processing. Both models predict human visual search behavior when applied to the simplified stimuli often used in experiments, but it remains unclear how to extend them to account for search of complex natural scenes. Models exist of natural scene search, but they do… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 89 publications
0
2
0
Order By: Relevance
“…Avoiding affirming the consequent can also be subserved by the clarification that ANNs "may simply rely on brute-force memorization and interpolation to learn how to generate the appropriate linguistic outputs in light of prior contexts" (Goldstein et al, 2021) -something which does appear to be true in certain contexts (Zhang, Bengio, Hardt, Recht, & Vinyals, 2016). Others solve this by carefully couching their findings as explicitly correlational where applicable, handin-hand with conceptually analysing the capacity under study (e.g., Lindsay & Miller, 2018;Nicholson & Prinz, 2020). Either way, leaving our syllogisms ambiguous -this includes not explaining that it is Q → P and not P → Q -leads to, or at least does not protect from affirming the consequent.…”
Section: Impediments To Inferencementioning
confidence: 99%
“…Avoiding affirming the consequent can also be subserved by the clarification that ANNs "may simply rely on brute-force memorization and interpolation to learn how to generate the appropriate linguistic outputs in light of prior contexts" (Goldstein et al, 2021) -something which does appear to be true in certain contexts (Zhang, Bengio, Hardt, Recht, & Vinyals, 2016). Others solve this by carefully couching their findings as explicitly correlational where applicable, handin-hand with conceptually analysing the capacity under study (e.g., Lindsay & Miller, 2018;Nicholson & Prinz, 2020). Either way, leaving our syllogisms ambiguous -this includes not explaining that it is Q → P and not P → Q -leads to, or at least does not protect from affirming the consequent.…”
Section: Impediments To Inferencementioning
confidence: 99%
“…Examples include automatically recognizing a person in selfie-images in social media platforms [108], diagnosing diseases from scanned medical images in health care [2], autonomous driving in self-driving cars [110], facial recognition [39] or fingerprint recognition [17] in security industry, automatically adding item details from product images in retail industry [4], and searching images instead of text in visual search engine [142] [87].…”
Section: Problem Statementmentioning
confidence: 99%