2019
DOI: 10.3389/fpsyg.2019.00884
|View full text |Cite
|
Sign up to set email alerts
|

Automatic Generation of Number Series Reasoning Items of High Difficulty

Abstract: Number series reasoning items have been frequently used in educational assessment. This study reviewed the previous literature investigating features relating to item difficulty and developed a new automatic generator for number series reasoning items. Ninety-two items were generated and administered to 466 students. The results showed that the test achieved acceptable reliability. Items requiring two arithmetic operations were of particularly high difficulty. All stimulus features implemented in the automatic… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 35 publications
(43 reference statements)
0
4
0
Order By: Relevance
“…Additionally, in the literature, it is seen that there are studies in the field of medicine and mathematics in template-based automatic item generation (Colvin, 2014;Lai et al, 2016;Singley & Bennett, 2002;Sun et al, 2019). AIG should be used in psychological testing areas where cognitive models are involved, and individuals' reasoning skills should also be measured (Hommel et al, 2022;Sun et al, 2019;Yang et al, 2021) and cognitive ability items can be developed in the reasoning areas (Freund et al, 2008;Poinstingl, 2009). In the previous study, AIG was used to generate items in the field of Turkish literature.…”
Section: Discussionmentioning
confidence: 99%
“…Additionally, in the literature, it is seen that there are studies in the field of medicine and mathematics in template-based automatic item generation (Colvin, 2014;Lai et al, 2016;Singley & Bennett, 2002;Sun et al, 2019). AIG should be used in psychological testing areas where cognitive models are involved, and individuals' reasoning skills should also be measured (Hommel et al, 2022;Sun et al, 2019;Yang et al, 2021) and cognitive ability items can be developed in the reasoning areas (Freund et al, 2008;Poinstingl, 2009). In the previous study, AIG was used to generate items in the field of Turkish literature.…”
Section: Discussionmentioning
confidence: 99%
“…Whereas when the value of the βj is less than 0.20, the test item is described as extremely difficult and should be reviewed in subsequent tests. The optimal test item difficulty factor is 0.50, and it insures maximum discrimination between high and low ability [52][53][54]. To maximize item discrimination, the desired difficulty levels are slightly higher than halfway between the probability of answering correctly by chance (1.00 divided by the number of alternatives for the item) and the ideal score for the item (1.00) [55][56][57][58].…”
Section: A Item Difficultymentioning
confidence: 99%
“…Moreover, further consideration should be given to the item which was responded to better by those who generally performed poorly on the test than those who performed better on the test as a whole. The test item may be confusing in some way to top-performing respondents [52,53,58,59]. It is recommended to directly delete the item VI.…”
Section: B Item Discriminationmentioning
confidence: 99%
“…The difficulty of mathematics items is important in assessing the quality of examinations and the value of educational outcomes [1]. It is generally determined by several features, including the knowledge required, the depth of thinking, the problem-solving ability, and the time constraints [2][3][4]. Understanding item difficulty has practical implications for intelligent educational applications such as knowledge tracking [5,6], automatic test item generation [7][8][9], intelligent paper generation [10] and personalized recommendations [11,12].…”
Section: Introductionmentioning
confidence: 99%