School of Languages and Linguistics - Research Publications

Permanent URI for this collection

Search Results

Now showing 1 - 4 of 4
  • Item
    Thumbnail Image
    Blending technologies in ESL courses: A reflexive enquiry
    Gruba, P ; Clark, C ; Ng, K ; Wells, M (The University of Auckland, 2009-12-01)
  • Item
    Thumbnail Image
    Perceptions of technologies in the assessment of foreign languages
    GRUBA, P ; CHERUBIN, L ; LAY-CHENCHABI, K ; MERA, H ; CARDENAS CLAROS, MS (RMIT University, 2009)
  • Item
    Thumbnail Image
    Doing normal: Discursive constructions of youth in transformative educational programs
    ELLWOOD, C. ; LAWS, C. (Cambridge Scholars Publishing, 2009)
  • Item
    Thumbnail Image
    Diagnostic assessment of writing: A comparison of two rating scales
    Knoch, U (SAGE PUBLICATIONS LTD, 2009-04)
    Alderson (2005) suggests that diagnostic tests should identify strengths and weaknesses in learners' use of language and focus on specific elements rather than global abilities. However, rating scales used in performance assessment have been repeatedly criticized for being imprecise and therefore often resulting in holistic marking by raters (Weigle, 2002). The aim of this study is to compare two rating scales for writing in an EAP context; one `a priori' developed scale with less specific descriptors of the kind commonly used in proficiency tests and one empirically developed scale with detailed level descriptors. The validation process involved 10 trained raters applying both sets of descriptors to the rating of 100 writing scripts yielded from a large-scale diagnostic assessment administered to both native and non-native speakers of English at a large university. A quantitative comparison of rater behaviour was undertaken using FACETS. Questionnaires and interviews were administered to elicit the raters' perceptions of the efficacy of the two types of scales. The results indicate that rater reliability was substantially higher and that raters were able to better distinguish between different aspects of writing when the more detailed descriptors were used. Rater feedback also showed a preference for the more detailed scale. The findings are discussed in terms of their implications for rater training and rating scale development.