2024
DOI: 10.31234/osf.io/vf3se
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Pseudo Factor Analysis of Language Embedding Similarity Matrices: New Ways to Model Latent Constructs

Nigel Guenole,
E. Damiano D'Urso,
Andrew Samo
et al.

Abstract: This article builds on recent work using Large Language Models (LLMs) in psychometrics and, in particular, the use of sentence transformer models to generate pseudo-discrimination parameters. Pseudo-discrimination parameters are discrimination estimates that correlate with empirical discrimination parameters without needing empirical data collection. While earlier work looked at pseudo-discrimination on an item-by-construct basis, we introduce and evaluate the use of pseudo-factor analysis. Pseudo-factor analy… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 21 publications
(28 reference statements)
0
1
0
Order By: Relevance
“…Pre-transformer era attempts to use semantic features of items to predict associations between measurement scales using latent semantic analysis have demonstrated moderate utility (Arnulf et al, 2014;Larsen & Bong, 2016;Rosenbusch et al, 2020). As the ability of computerised language models to capture meaning has grown, researchers have sought to directly quantify relationships between adjectives from textual data (Cutler & Condon, 2022), to assign items to constructs (Fyffe et al, 2024;Guenole et al, 2024), to directly predict item responses (Abdurahman et al, 2024;Argyle et al, 2023) and quantify openended answers to questions (Kjell et al, 2019(Kjell et al, , 2024. used large language models (LLMs) to map survey items to vector space and predict empirical item correlations.…”
Section: Introductionmentioning
confidence: 99%
“…Pre-transformer era attempts to use semantic features of items to predict associations between measurement scales using latent semantic analysis have demonstrated moderate utility (Arnulf et al, 2014;Larsen & Bong, 2016;Rosenbusch et al, 2020). As the ability of computerised language models to capture meaning has grown, researchers have sought to directly quantify relationships between adjectives from textual data (Cutler & Condon, 2022), to assign items to constructs (Fyffe et al, 2024;Guenole et al, 2024), to directly predict item responses (Abdurahman et al, 2024;Argyle et al, 2023) and quantify openended answers to questions (Kjell et al, 2019(Kjell et al, , 2024. used large language models (LLMs) to map survey items to vector space and predict empirical item correlations.…”
Section: Introductionmentioning
confidence: 99%