2020
DOI: 10.5281/zenodo.4143612
|View full text |Cite
|
Sign up to set email alerts
|

Who's Learning? Using Demographics in EDM Research

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 14 publications
(15 citation statements)
references
References 190 publications
1
14
0
Order By: Relevance
“…Bias of this nature can also arise even when demographics are not explicitly encoded in models, for instance when a proxy for a demographic variable is unintentionally included as a predictor. This form of bias is surprisingly common in education -in a recent survey of the role of demographics in the Educational Data Mining (EDM) community, Paquette et al (2020) found that roughly half of papers including demographics in analyses used at least one demographic attribute as a predictive feature within the model, without incorporating demographics into model testing or validation.…”
Section: Origins Of Bias and Harm In The Machine Learning Pipelinementioning
confidence: 99%
See 3 more Smart Citations
“…Bias of this nature can also arise even when demographics are not explicitly encoded in models, for instance when a proxy for a demographic variable is unintentionally included as a predictor. This form of bias is surprisingly common in education -in a recent survey of the role of demographics in the Educational Data Mining (EDM) community, Paquette et al (2020) found that roughly half of papers including demographics in analyses used at least one demographic attribute as a predictive feature within the model, without incorporating demographics into model testing or validation.…”
Section: Origins Of Bias and Harm In The Machine Learning Pipelinementioning
confidence: 99%
“…Evaluation bias occurs when the test sets used to evaluate a model do not represent the eventual population where the model will be applied. As reviewed in Paquette et al (2020), many models in educational data mining are developed on non-representative populations, and many papers do not even report what populations the models were tested on, making detection of evaluation bias quite difficult. Finally, deployment bias involves a model being used in inappropriate ways --being designed for one purpose and then used for a different purpose, such as using a model designed to identify student disengagement for formative purposes to assign participation grades.…”
Section: Origins Of Bias and Harm In The Machine Learning Pipelinementioning
confidence: 99%
See 2 more Smart Citations
“…On one hand, the observed historical gaps capture systematic inequities in the educational environment of different student groups, which may well apply to future students from the same groups and therefore contribute to similar gaps. In this sense, explicitly using demographic data can result in more accurate predictions and improve the efficiency of downstream interventions and actions based on those algorithmic decisions [33]. From an ethics and equity perspective, however, including demographic variables may lead to discriminatory results as predictive models would systematically assign differential predicted values across student groups according to the records of their historical counterparts.…”
Section: Introductionmentioning
confidence: 99%