2011
DOI: 10.1177/0265532211413445
|View full text |Cite
|
Sign up to set email alerts
|

Re-fitting for a different purpose: A case study of item writer practices in adapting source texts for a test of academic reading

Abstract: The important yet under-researched role of item writers in the selection and adaptation of texts for high-stakes reading tests is investigated through a case study involving a group of trained item writers working on the International English Language Testing System (IELTS). In the first phase of the study, participants were invited to reflect in writing, and then audio-recorded in a semantic-differential-based joint discussion, on the processes they employed to generate test material. The group were next obse… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
16
0

Year Published

2013
2013
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 26 publications
(25 citation statements)
references
References 13 publications
0
16
0
Order By: Relevance
“…However, for testing reading, it might not be necessary to reach the minimal reading level for all students; it might be sufficient to assess their reading proficiency and to allow for shortfalls in some situations. Perhaps due to this concern, doubt has been cast on the use of 95% vocabulary coverage as the minimal target in reading tests (see Green & Hawkey, ; Webb & Paribakht, ). For example, in Green and Hawkey (), the BNC 2,000 word families and the Academic Word List (Coxhead, ) were not expected to reach 90% vocabulary coverage of the four reading texts for IELTS (Nation & Waring, ).…”
Section: Implications and Conclusionmentioning
confidence: 99%
See 1 more Smart Citation
“…However, for testing reading, it might not be necessary to reach the minimal reading level for all students; it might be sufficient to assess their reading proficiency and to allow for shortfalls in some situations. Perhaps due to this concern, doubt has been cast on the use of 95% vocabulary coverage as the minimal target in reading tests (see Green & Hawkey, ; Webb & Paribakht, ). For example, in Green and Hawkey (), the BNC 2,000 word families and the Academic Word List (Coxhead, ) were not expected to reach 90% vocabulary coverage of the four reading texts for IELTS (Nation & Waring, ).…”
Section: Implications and Conclusionmentioning
confidence: 99%
“…Perhaps due to this concern, doubt has been cast on the use of 95% vocabulary coverage as the minimal target in reading tests (see Green & Hawkey, ; Webb & Paribakht, ). For example, in Green and Hawkey (), the BNC 2,000 word families and the Academic Word List (Coxhead, ) were not expected to reach 90% vocabulary coverage of the four reading texts for IELTS (Nation & Waring, ). Likewise, in examining the vocabulary coverage of 38 reading texts from CanTEST, Webb and Paribakht () found that 3,000 word families were sufficient to reach 95% vocabulary coverage of a text, but 14,000 word families were required to reach 95% vocabulary coverage of another text.…”
Section: Implications and Conclusionmentioning
confidence: 99%
“…As mentioned earlier, these measures reflect five discrete subconstructs, including clausal coordination, clausal subordination, nonfinite elements, phrasal coordination, and noun phrase complexity, and three composite features including overall sentence complexity, overall T‐unit complexity, and elaboration at the clause level. These results suggest that expert teachers and material developers attended to different subconstructs or dimensions of syntactic complexity in selecting and adapting teaching materials (Green & Hawkey, ). All eight syntactic complexity measures exhibited a clear upward trend across the 12 grade levels, indicating that overall, texts across the 12 grade levels become increasingly more complex in all dimensions of syntactic complexity.…”
Section: Discussionmentioning
confidence: 97%
“…The importance of including syntactic complexity as an explicit and distinct component in text complexity models has been well argued for (Frantz et al., ), and tools for evaluating text complexity do indeed all incorporate measures for this construct (Graesser et al., ; Sheehan et al., ). However, to date, little, if any, systematic effort has been made to establish benchmarks for syntactic complexity for text adaptation purposes and to make feedback regarding syntactic complexity accessible to teachers during the text adaptation process, despite the importance of this construct in this process (Crossley et al., ; Green & Hawkey, ). There is thus a clear need for more research in this area.…”
Section: Text Adaptation and Benchmarksmentioning
confidence: 99%
See 1 more Smart Citation