2022
DOI: 10.48550/arxiv.2206.14754
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Distilling Model Failures as Directions in Latent Space

Abstract: Existing methods for isolating hard subpopulations and spurious correlations in datasets often require human intervention. This can make these methods labor-intensive and dataset-specific. To address these shortcomings, we present a scalable method for automatically distilling a model's failure modes. Specifically, we harness linear classifiers to identify consistent error patterns, and, in turn, induce a natural representation of these failure modes as directions within the feature space. We demonstrate that … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(8 citation statements)
references
References 11 publications
0
8
0
Order By: Relevance
“…Instead, we propose to describe the visual biases with language. Some recent works use vision-language models to analyze the model failures by detecting outliers in the visual embedding (Eyuboglu et al, 2022;Jain et al, 2022a). In contrast, we directly generate descriptive captions from images instead of embeddings, and could find multiple and fine-grained biases.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Instead, we propose to describe the visual biases with language. Some recent works use vision-language models to analyze the model failures by detecting outliers in the visual embedding (Eyuboglu et al, 2022;Jain et al, 2022a). In contrast, we directly generate descriptive captions from images instead of embeddings, and could find multiple and fine-grained biases.…”
Section: Related Workmentioning
confidence: 99%
“…A few recent works use vision-language models to discover biases (they call slices or model failures). Concretely, they define biased groups as the outliers in the embedding space of the visual encoder, estimated by a Gaussian mixture model (Eyuboglu et al, 2022) or support vector machine (Jain et al, 2022a). In contrast, we directly generate captions from images, which may contain more detailed information than the encoder embeddings.…”
Section: G Additional Related Workmentioning
confidence: 99%
“…Many recent works aim to understand model systematic errors by finding subsets of inputs with similar characteristics where the model performs significantly worse. This is referred to as slice discovery (Chung et al, 2019;Singla et al, 2021;d'Eon et al, 2022;Eyuboglu et al, 2022;Jain et al, 2022a). However, these algorithms fail to address the most fundamental challenge for slice discovery -the lack of data.…”
Section: Rectified Model Misbehaviorsmentioning
confidence: 99%
“…However, certain factors such as lighting and contrast may not be captured by the object detector, which can limit the effectiveness of our approach. To mitigate this limitation, our method can be used in conjunction with previous work that utilizes system/content metadata or discovered visual features for general failure analysis (Nushi et al, 2018;Singla et al, 2021;Chung et al, 2019;Jain et al, 2022;Eyuboglu et al, 2022). For instance, one can enrich the test data with additional meta-data, such as contrast, blur, lighting, and camera angle, and apply our method as well as previous approaches to understand if the model's performance drops for some of these conditions.…”
Section: B2 Spurious Correlation Detectionmentioning
confidence: 99%