Understanding the sources of variance in speaking assessment is important in Japan where society's high demand for English speaking skills is growing. Three challenges threaten fair assessment of speaking. First, in Japanese university speaking courses, teachers are typically the only raters, but teachers' knowledge of their students may unfairly influence scoring. Second, in many speaking classes, teachers are often native speakers of English who may not be appropriately trained. Third, conducting practical speaking assessment is a major concern in classroom. Semi-direct testing using digital recorders could solve such issue and achieve reliability and efficiency in speaking assessment (Qian, 2009).
This study examined these sources of score variation--the effects of rater characteristics and scoring methods on speaking assessment by university instructors in Japan. Rater characteristics included scorers' knowledge of examinees and to a much lesser extent their native language differences. Scoring methods were live and delayed scoring with digital recorders. These factors were believed to be potential threats to fairness.
Participants were the researcher and the instructors (N = 9) as well as students (N = 169) in nine classes in four universities. In oral skills classes in Japanese universities, students were given one speaking task scored by instructors live and recorded. Overall raters' knowledge of examinees, despite some perceived influence of knowledge of examinees, was not a significant source of variance. For scoring methods, despite some raters' preference for delayed scoring, there were no statistically significant findings between live and delayed scoring. The two groups of raters (7 English and 1 Japanese), not including the researcher, focused on different speech characteristics. These results suggested that variations in speaking scores were not derived from differences in rater characteristics and scoring methods.
Implications include the following: instructors being the only rater depending on the purpose of assessment and circumstances, wider choices of speaking assessment methods to facilitate fairness and practicality, and possibly more involvement of Japanese instructors in teaching and assessing of speaking skills.
|Commitee:||Kang, Okim, McClure, John, McGroarty, Mary|
|School:||Northern Arizona University|
|School Location:||United States -- Arizona|
|Source:||DAI-A 74/09(E), Dissertation Abstracts International|
|Subjects:||Educational tests & measurements, English as a Second Language, Foreign Language|
|Keywords:||English language learners, Rater characteristics, Scoring methods, Speaking assessments|
Copyright in each Dissertation and Thesis is retained by the author. All Rights Reserved
The supplemental file or files you are about to download were provided to ProQuest by the author as part of a
dissertation or thesis. The supplemental files are provided "AS IS" without warranty. ProQuest is not responsible for the
content, format or impact on the supplemental file(s) on our system. in some cases, the file type may be unknown or
may be a .exe file. We recommend caution as you open such files.
Copyright of the original materials contained in the supplemental file is retained by the author and your access to the
supplemental files is subject to the ProQuest Terms and Conditions of use.
Depending on the size of the file(s) you are downloading, the system may take some time to download them. Please be