Proceedings of the 50th ACM Technical Symposium on Computer Science Education 2019
DOI: 10.1145/3287324.3287370
|View full text |Cite
|
Sign up to set email alerts
|

An Item Response Theory Evaluation of a Language-Independent CS1 Knowledge Assessment

Abstract: Tests serve an important role in computing education, measuring achievement and differentiating between learners with varying knowledge. But tests may have flaws that confuse learners or may be too difficult or easy, making test scores less valid and reliable. We analyzed the Second Computer Science 1 (SCS1) concept inventory, a widely used assessment of introductory computer science (CS1) knowledge, for such flaws. The prior validation study of the SCS1 used Classical Test Theory and was unable to determine w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
33
4

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 32 publications
(37 citation statements)
references
References 27 publications
(28 reference statements)
0
33
4
Order By: Relevance
“…FCS1 was developed as a concept inventory for an introductory undergraduate programming course with the goal of assessing all of the core concepts covered in the course (Tew & Guzdial, 2017). The SCS1 assessment (Parker, Guzdial & Engelman, 2016) was a revision and revalidation of the FCS1, though even through this refinement process other researchers continue to question the high difficulty level of some of the items (Xie, Davidson, Li, & Ko, 2019). At the K-12 level, there has been work on developing assessments of student CS conceptual understanding.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…FCS1 was developed as a concept inventory for an introductory undergraduate programming course with the goal of assessing all of the core concepts covered in the course (Tew & Guzdial, 2017). The SCS1 assessment (Parker, Guzdial & Engelman, 2016) was a revision and revalidation of the FCS1, though even through this refinement process other researchers continue to question the high difficulty level of some of the items (Xie, Davidson, Li, & Ko, 2019). At the K-12 level, there has been work on developing assessments of student CS conceptual understanding.…”
Section: Related Workmentioning
confidence: 99%
“…For item development around block-based programming code, prior work on high-school and 4 / 24 middle-school assessments were used to guide our work (Du Boulay, 1986;Shneiderman, & Mayer, 1979, Weintrop & Wilensky, 2015Xie et al, 2019). For each concept, we designed items with varying levels of difficulty and question types: comprehension of a code snippet, debugging a partially wrong code snippet, and developing/completing a partially built code snippet.…”
Section: Item Developmentmentioning
confidence: 99%
“…Many studies have investigated students' performance using the FCS1/SCS1 (e.g., [394,256,270,363]). One main advantage of using a validated instrument is to produce comparable and generalizable results, providing meaningful information about the investigated context.…”
Section: 30mentioning
confidence: 99%
“…We found similar student performance on the SCS1 and similar metrics of the SCS1's internal reliability as earlier studies. We now seek to provide a more nuanced analysis to investigate if our data suggest difficulties in the same questions by conducting an Item Response Theory (IRT) analysis using previous research, mainly Xie et al [394] and Parker et al [270], as benchmarks for methods and results.…”
Section: Sei W7mentioning
confidence: 99%
“…We were guided in our selection of statistical methods by our external evaluator, another evaluator she referred us to, and by recent work in CS education on other CIs. We referred to Chasteen et al [6] as a general guide for how to statistically validate a CI, and then consulted recent work by Xie et al [33] as an example of validating a computer science CI. Guided by both of these works, we applied both Classical Test Theory (CTT) and Item Response Theory (IRT).…”
Section: Step 5: Statistical Validationmentioning
confidence: 99%