• Skip to primary navigation
  • Skip to main content

site logo
The Electronic Journal for English as a Second Language
search
  • Home
  • About TESL-EJ
  • Vols. 1-15 (1994-2012)
    • Volume 1
      • Volume 1, Number 1
      • Volume 1, Number 2
      • Volume 1, Number 3
      • Volume 1, Number 4
    • Volume 2
      • Volume 2, Number 1 — March 1996
      • Volume 2, Number 2 — September 1996
      • Volume 2, Number 3 — January 1997
      • Volume 2, Number 4 — June 1997
    • Volume 3
      • Volume 3, Number 1 — November 1997
      • Volume 3, Number 2 — March 1998
      • Volume 3, Number 3 — September 1998
      • Volume 3, Number 4 — January 1999
    • Volume 4
      • Volume 4, Number 1 — July 1999
      • Volume 4, Number 2 — November 1999
      • Volume 4, Number 3 — May 2000
      • Volume 4, Number 4 — December 2000
    • Volume 5
      • Volume 5, Number 1 — April 2001
      • Volume 5, Number 2 — September 2001
      • Volume 5, Number 3 — December 2001
      • Volume 5, Number 4 — March 2002
    • Volume 6
      • Volume 6, Number 1 — June 2002
      • Volume 6, Number 2 — September 2002
      • Volume 6, Number 3 — December 2002
      • Volume 6, Number 4 — March 2003
    • Volume 7
      • Volume 7, Number 1 — June 2003
      • Volume 7, Number 2 — September 2003
      • Volume 7, Number 3 — December 2003
      • Volume 7, Number 4 — March 2004
    • Volume 8
      • Volume 8, Number 1 — June 2004
      • Volume 8, Number 2 — September 2004
      • Volume 8, Number 3 — December 2004
      • Volume 8, Number 4 — March 2005
    • Volume 9
      • Volume 9, Number 1 — June 2005
      • Volume 9, Number 2 — September 2005
      • Volume 9, Number 3 — December 2005
      • Volume 9, Number 4 — March 2006
    • Volume 10
      • Volume 10, Number 1 — June 2006
      • Volume 10, Number 2 — September 2006
      • Volume 10, Number 3 — December 2006
      • Volume 10, Number 4 — March 2007
    • Volume 11
      • Volume 11, Number 1 — June 2007
      • Volume 11, Number 2 — September 2007
      • Volume 11, Number 3 — December 2007
      • Volume 11, Number 4 — March 2008
    • Volume 12
      • Volume 12, Number 1 — June 2008
      • Volume 12, Number 2 — September 2008
      • Volume 12, Number 3 — December 2008
      • Volume 12, Number 4 — March 2009
    • Volume 13
      • Volume 13, Number 1 — June 2009
      • Volume 13, Number 2 — September 2009
      • Volume 13, Number 3 — December 2009
      • Volume 13, Number 4 — March 2010
    • Volume 14
      • Volume 14, Number 1 — June 2010
      • Volume 14, Number 2 – September 2010
      • Volume 14, Number 3 – December 2010
      • Volume 14, Number 4 – March 2011
    • Volume 15
      • Volume 15, Number 1 — June 2011
      • Volume 15, Number 2 — September 2011
      • Volume 15, Number 3 — December 2011
      • Volume 15, Number 4 — March 2012
  • Vols. 16-Current
    • Volume 16
      • Volume 16, Number 1 — June 2012
      • Volume 16, Number 2 — September 2012
      • Volume 16, Number 3 — December 2012
      • Volume 16, Number 4 – March 2013
    • Volume 17
      • Volume 17, Number 1 – May 2013
      • Volume 17, Number 2 – August 2013
      • Volume 17, Number 3 – November 2013
      • Volume 17, Number 4 – February 2014
    • Volume 18
      • Volume 18, Number 1 – May 2014
      • Volume 18, Number 2 – August 2014
      • Volume 18, Number 3 – November 2014
      • Volume 18, Number 4 – February 2015
    • Volume 19
      • Volume 19, Number 1 – May 2015
      • Volume 19, Number 2 – August 2015
      • Volume 19, Number 3 – November 2015
      • Volume 19, Number 4 – February 2016
    • Volume 20
      • Volume 20, Number 1 – May 2016
      • Volume 20, Number 2 – August 2016
      • Volume 20, Number 3 – November 2016
      • Volume 20, Number 4 – February 2017
    • Volume 21
      • Volume 21, Number 1 – May 2017
      • Volume 21, Number 2 – August 2017
      • Volume 21, Number 3 – November 2017
      • Volume 21, Number 4 – February 2018
    • Volume 22
      • Volume 22, Number 1 – May 2018
      • Volume 22, Number 2 – August 2018
      • Volume 22, Number 3 – November 2018
      • Volume 22, Number 4 – February 2019
    • Volume 23
      • Volume 23, Number 1 – May 2019
      • Volume 23, Number 2 – August 2019
      • Volume 23, Number 3 – November 2019
      • Volume 23, Number 4 – February 2020
    • Volume 24
      • Volume 24, Number 1 – May 2020
      • Volume 24, Number 2 – August 2020
      • Volume 24, Number 3 – November 2020
      • Volume 24, Number 4 – February 2021
    • Volume 25
      • Volume 25, Number 1 – May 2021
      • Volume 25, Number 2 – August 2021
      • Volume 25, Number 3 – November 2021
      • Volume 25, Number 4 – February 2022
    • Volume 26
      • Volume 26, Number 1 – May 2022
      • Volume 26, Number 2 – August 2022
      • Volume 26, Number 3 – November 2022
  • Books
  • How to Submit
    • Submission Procedures
    • Ethical Standards for Authors and Reviewers
    • TESL-EJ Style Sheet for Authors
    • TESL-EJ Tips for Authors
    • Book Review Policy
    • Media Review Policy
    • APA Style Guide
  • TESL-EJ Editorial Board

Constructing Analytic Rubrics for Assessing Open-Ended Tasks in the Language Classroom

February 2021 – Volume 24, Number 4

Mary Lou Vercellotti
Ball State University
<mlvercellottatmarkbsu.edu>

Dawn E. McCormick
University of Pittsburgh
<mccormicatmarkpitt.edu>

Abstract

With the increased emphasis in communicative language learning and task-based language teaching, classroom assessment increasingly includes performance-based assessments (e.g., essays, speeches, projects, presentations), which require careful planning and can be challenging to implement and assess. The ability to design and use well-designed assessments for language performance tasks is a critical skill for classroom instructors. A rubric is a tool to more objectively measure the quality of the language performance when assessing language use in open-ended tasks. Analytic rubrics, with multiple categories and descriptions reflecting various levels of performance, help the instructor evaluate the effectiveness of instruction, document evidence of learner progress, and give feedback to learners. This article synthesizes theoretical and empirical scholarship in order to describe how to construct well-designed analytic rubrics for classroom language assessment. Four main steps are described: 1) establishing categories, 2) describing levels of performance, 3) reviewing the components of the rubric before implementation, and 4) evaluating the effectiveness of the rubric after implementation.

Keywords: rubrics, scoring scales, assessment, grading, performance-based, task-based

Classroom Assessment

Classroom assessment documents evidence of learner progress, gives feedback to learners, and allows teachers to evaluate the effectiveness of instruction. Instructors should be able to design and implement “varied and valid assessments…to support student learning” (TESOL International Association, 2018, p. 60). Classroom instructors may choose a variety of assessments, from objective tests, such as form-focused multiple choice tests, to open-ended performance-based tasks, with many tasks in between these two sides of the assessment continuum. Assessing what a learner knows about a language can be relatively straightforward. For instance, the learner supplies answers on an objective test (e.g., vocabulary matching quiz) whose questions have expected answers, and the instructor can score the learners’ work reliably with an answer key. For classroom assessment with specific form-focused objectives, other assessment tools, such as detailed grading criteria, checklists, and tallies may be useful. Learner knowledge about the language, though, does not reflect what the learner can do with that knowledge (Van Gorp & Deygers, 2014), and the results of objective tests do not reflect the learner’s language skills, such as the ability to speak or write the language. With increased emphasis on communicative competence and task-based language teaching, language instructors have begun to assign more open-ended language tasks as pedagogical activities and as assessments (Purpura, 2016; Van Gorp & Deygers, 2014). Given the variety of assessment types and assessment tools, a crucial issue in assessment is the alignment between the content to be assessed and the assessment procedure. In a performance-based (or task-based) approach to language assessment, the learner’s performance in open-ended tasks is seen as evidence of the learner’s knowledge, skills, and ability (Purpura, 2016). With an open-ended language production task (e.g., essays, speeches, projects, presentations), the learner has the opportunity to choose the vocabulary and grammatical structures in response to the prompt or assignment. For instance, one learner may choose to use more complex grammar to express an idea while another may choose to create simpler but accurate structures. These open-ended performance-based assessments are vital in classroom language assessment because they more directly measure productive skills (e.g., speaking, writing). Performance-based assessments, however, require careful planning and can be challenging to implement. Instructors may not yet have the requisite knowledge to assess these language performances. Because of these challenges and others, instructors may borrow an existing assessment from another context (e.g., another course, another proficiency level), but using a rubric designed for another context is a potential misuse of the assessment tool (Purpura, 2016). Accordingly, it is important for instructors to create well-designed assessments for their classroom. In fact, designing valid assessments is a key teaching principle (TESOL International Association, 2018).

This article describes how to construct an analytic rubric for classroom language assessment, informed by theoretical and empirical scholarship. After introducing rubrics as a tool for assessing student performance of open-ended tasks, we then describe how to construct an analytic rubric by identifying categories and describing levels of performance for each category. The third and fourth sections provide recommendations for pre-use and post-use reviews, critical steps in the assessment process. Last, we provide a checklist that summarizes steps in constructing analytic rubrics for open-ended tasks and provides practical suggestions.

Performance-based Assessments and Rubrics

Classroom instructors choose from a range of assessments, with consideration of the link between course objectives and the assessments themselves. Open-ended tasks focus on course-level communication objectives, rather than on specific grammar or vocabulary of any particular lesson, thus requiring an assessment tool that measures the quality of the language produced during these tasks. Analytic rubrics are one tool for open-ended tasks in an instructor’s assessment tool kit, thus instructors must determine whether a rubric is the right tool for the assessment. The concepts discussed below provide guidance to determine if a rubric is an appropriate assessment tool.

Open-ended tasks often are evaluated through the rater’s impressions or judgements (Green & Hawkey, 2012), which are inherently subjective, even when learners are to be evaluated by how well they meet stated assessment objectives. The ratings could differ because different raters can have different impressions of the performance (Green & Hawkey, 2012), but even a single rater may have difficulty scoring consistently. For instance, scoring may become more conservative (or more generous) as fatigue increases (Van Moere, 2014), or a rater might start assessing quite strictly and lower expectations after viewing multiple performances. Raters can also be influenced by the previous language performance during the grading process (Upshur & Turner, 1995). After scoring a really strong essay, for instance, the rater may score the next essay lower than if it had followed a weaker essay. Having clearly ranked descriptions of language performance will help increase the consistency when assessing, and so rubrics have been increasingly used to assess the quality of the performance in open-ended tasks, such as speaking and writing.

The term “rubric” has sometimes been used to refer to any grading criteria. In this article, the term rubric refers more narrowly to a specific assessment tool with descriptions of various levels of the quality of the performance. Goldberg (2014) has defined a rubric as “a scoring guide that outlines features of work at different levels of performance” (p. 1). Holistic rubrics and analytic rubrics are two types of common tools for language performance. With holistic rubrics, the descriptions of the levels of performance include multiple traits, resulting in a single score. Holistic rubrics can be more practical when only an overall score is needed (e.g., a quick placement test, a proficiency test) (Brown & Abeywickrama, 2019) as a measure of learning that has happened (Brookhart, 2018). Since holistic rubrics provide less specific information, they are less useful for classroom assessment. Analytic rubrics, on the other hand, are more useful during the learning process (Brookhart, 2018). Analytic rubrics list multiple traits or categories separately for the rater to rank (e.g., good, better, best) based on descriptions of levels of performance. Analytic rubrics strengthen the reliability of the assessment for language produced during such tasks (Green & Hawkey, 2012). Since the instructor can give separate ratings for each category, the learner receives specific feedback about what they can do and what skills need more improvement (Brown, 2018). Additionally, the results of the assessment with an analytic rubric provides the instructor information about the strengths and weaknesses of the learners.

Table 1 shows an author-created analytic rubric for an individual informational speech in a language class for beginners to serve as an example. It is designed to assess three language skills: vocabulary, grammar, and pronunciation at three possible levels: meets standard, approaching standard, and needs improvement. In Table 1’s format, the lowest performance is at the bottom and each higher row has descriptions of higher level of performance.

Table 1. Analytic Rubric for Informational Speech- City; High Beginner Learners

Criterion Vocabulary Grammar Pronunciation
Meets Standard Speaker used specific, descriptive words to create a visual picture. Speaker created simple sentences using present tense. Speaker was clear, comprehensible and listener could understand meaning.
Approaching Standard Speaker used appropriate but general descriptive words. Speaker created simple and incomplete sentences using present tense. Speaker was comprehensible but listener had to focus to understand.
Needs Improvement Speaker used incorrect (or missing) descriptive words. Speaker created incomplete sentence with base form verbs. Speaker was generally not comprehensible even with listener effort.

In summary, language instructors often assign communicative language tasks, but these open-ended assignments are challenging to assess because the resulting language performances are all different. Analytic rubrics are an appropriate choice for instructors when assessing open-ended extended language performances because they provide a structure for consistent assessment and feedback.

In the following sections, we review how to construct an analytic rubric for teacher-implemented classroom assessment. Author-created rubrics are provided to model the concepts and steps presented. These are illustrative and not intended to be applied to specific classroom contexts. The imagined context for the following rubric is an adult, mixed-gender, integrated-skills English as an additional language (EAL) class in an English-speaking country. The students’ levels range from high-intermediate to low advanced [i] with a minimum of 52 on the TOEFL iBT or 5.5 on the IELTS to meet the entrance requirement. The students identify their motivation for language learning as academic, professional, and/or personal. The learning goals focus on speaking and listening, and reading, writing, grammar, and pronunciation skills are developed explicitly when they support speaking and listening. We focus on a rubric for an individual speaking activity. Students engage in language practice that facilitates transfer to language use in everyday, work, and academic settings. Although a specific class situates the example rubric, the concepts and steps are applicable across multiple teaching contexts.

Constructing Analytic Rubrics

Analytic rubrics have two main parts: categories to evaluate and ranked descriptions of performance (Brookhart, 2018), usually set up like a table, with the categories either listed in the columns (shown in Table 1) or the categories set up in the rows (shown in Table 2). Both the categories in the analytic rubric and levels of student performance are designed for a specific context, a specific course with a specific student population (Moskal & Leydens, 2000). The following two sections describe how to identify categories and determine performance descriptors for specific classroom assessments.

Categories

Each category (i.e., evaluation criteria) in the analytic rubric assesses a different aspect of the learner’s skill (e.g., Popham, 1997; Van Moere, 2014). In other words, the categories of an analytic rubric are independent. For classroom assessments, each learning objective to be assessed can be represented as a category in the rubric (Brookhart, 2018). The rubric in Table 2, for instance, has categories focused on linguistic skills (i.e., vocabulary, grammar structures, pronunciation, fluency). These categories reflect the skills required for effective communication. The rubric also includes a category on content to represent a learning objective of presenting factual information with supporting statements. The inclusion of discourse-level categories, such as content and fluency, reflects the understanding that effective communication is more than adherence to sentence-level forms that is captured by the other categories. In Table 2’s format, the categories are along the leftmost column.

The categories in the rubric must focus on observable and measurable skills of the performance and be directly linked to the assessment’s learning objectives. For instance, perhaps the students are asked to complete the task, “Present a ≤ 4-minute prepared speech,” in order to meet objectives that include:

  1. Use topic-appropriate vocabulary
  2. Use a variety of grammatical structures, including compound and complex sentences, and transitions that support meaning
  3. Produce segmentals and suprasegmentals that are comprehensible to listeners
  4. Use appropriate pacing and pausing
  5. Present original or given information with relevant supporting ideas

The five categories in Table 2 relate to the students’ language learning goals of the unit and reflect the course learning objectives, and the instructor can hear the evidence (i.e., observable) and evaluate performance (i.e., measurable) during the speech.

A rubric with many categories may become impractical because it is difficult to assess all of the language skills listed (Schreiber et al., 2012). Generally, five categories may be the maximum a rater can attend to (Green, 2014; Popham, 1997). No assessment can assess every skill; a rubric’s categories must be carefully selected from the learning objectives for that particular assessment.

Table 2. Analytic Rubric Categories: Formal Informational Speech

Since productive language skills tap language knowledge, analytic rubrics to assess writing and speaking may have the same linguistic categories of vocabulary and grammar. Despite these similarities, the grammar patterns and vocabulary use differ in written text and in speech because written language can be revised (Vasylets et al., 2017; Vercellotti, 2018).

In some rubrics for assessing speaking, instructors may include nonverbal behaviors (e.g., eye-contact, gestures) to evaluate that aspect of oral communication. Raters have been found to attend to nonverbal behavior in dialogues, even when that category is not included in the rubric (May, 2009). Nevertheless, research with native English speakers has shown that nonverbal behaviors do not correlate with speech grades (Schreiber et al., 2012), which indicates that those skills do not seem to be integral to effective monologic speaking. Accordingly, nonverbal communication could be included when the course (or unit) objectives specifically include development of such skills, particularly in interactive communication, but a category for nonverbal behavior during speaking performances may not be necessary to assess speaking skill.

Levels of Performance

After identifying each category in the analytic rubric, the next step is to create descriptions for different levels of performance in each category. The description of each level of performance should convey the criteria for that category to help the rater identify which description most aligns with the learner’s performance (Goldberg, 2014). In other words, the descriptions should help the rater place the performance at a particular level for each category. It is often helpful to first describe the expected performance at the “Meets Standard” level because the expectations of the stated standard is most evident. After describing the expectation for the standard, a description of the level of performance in that category at one level higher or one level lower can be developed. The differences between each level of performance should be distinct (Moskal, 2002). Labeling each level can help focus the rating process around the standard (Suskie, 2009). Such labels are also useful to anchor the descriptions around the standards when drafting the ranked descriptions (e.g., Exceeds Standard, Meets Standard, Approaching Standard). The example rubric in Table 3 includes descriptive labels as well as the corresponding letter grade (used in schools in the United States). It is important to note that the rubric’s top level of performance should not expect learners to have a native-like performance; learners should not be expected to have a perfect performance to receive full points.

Importantly, although the labels for the levels of performance may be evaluative terms (such as “excellent” or “exceeds standards), the descriptions differentiating the levels of performance should not use evaluative terms. Only “descriptive language helps students envision where they are in their learning and where they should go next” (Brookhart, 2018, p. 2) and “provide a clear description of the learning goal” (Brookhart, 2018, p 5). In addition, the descriptions should predominantly focus on what the learner demonstrates rather than what is absent. This recommendation does not limit the descriptions to “can do” statements. Some negative descriptions might be used, but without describing what is missing. For instance, the phrase “word choice may sometimes be too specific or vague” (rather than “did not use a variety of words”) focuses on what the learner produced.

Since rubrics are designed to evaluate the quality of the language performance, the descriptions must focus on distinguishing quality, rather than measuring quantity. Descriptions which differentiate levels of performance primarily with quantifiers (e.g., few, some, many) should be avoided also because it is difficult to consistently measure “few” of any aspect of a language performance. For instance, rather than a focus on the number of errors, the descriptions for a category for grammatical accuracy in a language classroom can include phrases about whether the errors interfere with understanding the speaker’s message. Likewise, descriptions should not attempt to differentiate levels of performance with relativistic or comparison terms (e.g., exceptional, stronger) because such terms may encourage scoring in comparison to others rather than to the stated objective. Further, since the categories in the rubric focus on the language skills (e.g., grammar, vocabulary) that contribute to a successful performance, the categories should not include requirements of the assessment itself, such as a word count or time fulfillment. In fact, student learning can be hindered by rubrics that focus on the assignment requirements (or directions) rather than describing the quality of the work (Brookhart, 2018).  Failure to meet the requirements of the assessment can be penalized, ideally outside of the rubric’s categories because rubrics are a tool to assess the quality of the language performance.

It is best practice to have only “as many scale points as can be well defined and that adequately cover the range” of performances (Perlman, 2002, p. 8). Generally, rubrics have between three and five levels of performance (Brookhart, 2018; Suskie, 2009). With too many ranked levels, the task of matching a learner’s performance to the most similar level becomes difficult and more time-consuming. Additionally, the number of categories in the rubric may influence the number of levels. For instance, an analytic rubric with a higher number of categories may, in turn, have fewer levels of performance, in order to keep the rubric a manageable size. The example rubric in Table 3 has four levels of performance arranged with the lowest performance in the column next to the categories and each higher level of performance builds up to the highest performance in the rightmost column.

Table 3. Analytic Rubric: Formal Informational Speech

Category In Progress
(below D)
3 points
Approaching Standard (C-)
4 points
Meets Standard
(B+)
4.5 points
Exceeds Expectations 
(A)
5 points
Vocabulary Words often unsuitable for task; errors interfere with meaning Word choice is limited or reduces effective expression of meaning Variety of words; word choice may sometimes be too specific or vague but meaning is clear Variety of words;
word choice supports meaning & appropriate for task
Grammar Structures Limited variety of structures & transitions; errors interfere with meaning Variety of useful structures & transition words;
some choices
reduce effective expression of meaning
Variety of structures; useful transition words; grammar choices may be generic but appropriate for meaning Variety of complex structures & effective transition words; grammar choices support meaning
Pronunciation Substitution of sounds, suprasegmental patterns interfere with meaning; difficult to understand by all audiences Sounds substitutions and suprasegmentals occasionally slow down comprehension but overall meaning is understandable to people familiar w/ENL speech Minor, predictable sound substitutions which are understandable to people unfamiliar with ENL speech Segmentals and suprasegmental patterns are understandable to all audiences
Fluency Pacing, pausing, and/or fillers interferes with comprehen-sibility Pacing, pausing and/or fillers slows down comprehensibil-ity Appropriate pacing and pausing Pacing and pausing enhance message
Content Topic not clear; information may not be factual or relevant to topic Speech is focused on a single unstated topic; information was factual but incomplete, general, or less relevant Topic was explicitly stated; information was factual and relevant to the topic and generally supported Topic was explicitly stated; information was factual and relevant to the topic; specific details enhance message

The descriptions of the levels of performance could be created based on previously scored learner performances, for example, by sorting the graded performances and identifying the features that differentiate the quality at each level for each category. Often, however, the instructor must create the descriptions at each performance level based only on the course expectations and previous experience with the teaching context.

Points and Weighting

When designing classroom assessment, considerations for assessing student work and assigning a score to the assessment must be addressed. In Table 1, we presented a rubric that described student performance without assigning a point value, but instructors and students may benefit from assigning points to performance because assigning points allows the instructor to grade the performance and the learner has concrete feedback of the distance between their performance and the expected standards of the activity. In other words, the levels in a rubric serve to assign a score and provide feedback to the learner (Brown, 2018; Goldberg, 2014).  In this section, we explain setting the point value of each performance level in relation to other levels and the weight of each category in relation to other categories. At the end of this section, we explore the question of using a range of scores for each performance level.

When assigning point values, the difference between point values should represent the difference in quality between the levels. For instance, in Table 2, the point values in the “Meets Standard (B+)” level is only slightly lower than the “Exceeds Expectation (A)” level because the quality of the performances described is only slightly less successful. It is also important to consider how the resulting grade percentage and/or letter grade align with the descriptions of performance. For instance, if the second-best level of performance equates to a “B” performance, the numbers assigned at that level should equal a percentage in the B range. Note also that according to the rubric in Table 2, meeting the standard is not sufficient to earn full points, which may not be appropriate in some teaching contexts. The point values also make sense mathematically and do not penalize students unintentionally. For instance, assigning point values of 1, 2, 3, and 4 when there are four levels of performance may be incongruous for grading, unless the descriptions reflect vastly different performances which should earn 25%, 50%, 75%, and 100%.

Additionally, analytic rubrics allow the instructor to weight the categories (Green, 2014; Popham, 1997), which means that a specific category can be worth more points in the analytic rubric, based on the course focus, unit focus, or learner level. For instance, the rubric shown in Table 4 lists five total categories with the categories of fluency and content worth half the value of the other categories, indicating the relative importance of those categories. Ballard (2019) found that the arrangement of the categories may inadvertently influence raters’ perception of which categories are most important, in that the leftmost column may be perceived as most important while the rightmost column may be perceived as the least important. This finding suggests that weighted categories should be placed left-most or topmost to align with rater’s expectations, to avoid causing a conflict in expectations.  Finally, when weighting categories, it is more efficient to list the point values in the individual cells, which can be circled, rather than having a multiplication notation (e.g., x2 for a category) because the notation requires another calculation for the rater and is often confusing to (or overlooked by) learners.

Table 4. Weighted Analytic Rubric: Formal Informational Speech

Category In Progress
(below D)
Approaching standard
(C-)
Meets Standard
(B+)
Exceeds Expectations 
(A)
Vocabulary Words often unsuitable for task; errors interfere with meaning
(6)
Word choice is limited or hinders effective expression of meaning
(8)
Variety of words; word choice may sometimes be too specific or vague but meaning is clear
(9)
Variety of words;
word choice supports meaning & appropriate for task
(10)
Grammar Structures Limited variety of structures & transitions; errors interfere with meaning
(6)
Variety of useful structures & transition words;
some choices hinder effective expression of meaning
(8)
Variety of structures; useful transition words; grammar choices may be generic but appropriate for meaning
(9)
Variety of complex structures & effective transition words; grammar choices support meaning
(10)
Pronunciation Substitution of sounds, suprasegmental patterns interfere with meaning Difficult to understand by all audiences
(6)
Sounds substitutions and suprasegmentals occasionally hinder but overall meaning is understandable to people familiar w/ENL speech
(8)
Minor, predictable sounds substitutions and suprasegmentals which do not interfere with meaning; understandable to people unfamiliar with ENL speech
(9)
Segmentals and suprasegmental patterns are understandable to all audiences
(10)
Fluency Pacing, pausing, and/or fillers interferes with
comprehensibility
(3)
Pacing, pausing and/or fillers hinders comprehensibility
(4)
Appropriate pacing and pausing
(4.5)
Pacing and pausing enhance message
(5)
Content Topic not clear; information may not be factual or irrelevant to topic
(3)
Speech is focused on a single unstated topic Information was factual but incomplete, general, or less relevant
(4)
Topic was explicitly stated; information was factual and relevant to the topic and generally supported
(4.5)
Topic was explicitly stated; information was factual and relevant to the topic; specific details enhance message
(5)

Some rubrics include a range of points within each level of performance (e.g., 18-19-20). This option adds flexibility to the scoring, but it counters one of the main purposes of the analytic rubric, namely to describe the features of each level of performance for the rater and for the learner. As Perlman (2002, p. 8) has stated “each point on the scale should be clearly labeled and defined”. Therefore, a range of scores within a level lowers the reliability because the rubric does not include guidelines for awarding the points within each level; thus, the instructor may not consistently score student work as an 18 vs. 19 vs. 20 when each value is connected to the same description of performance. Additionally, such rubrics are less helpful for the student to know how to improve or for the instructor to know what to teach. Accordingly, a range of available points within levels of performance should be used sparingly in classroom assessment. An alternative solution is to describe the levels of performances for specific point values (e.g., 20, 18, 16, and 14) and leave scores in-between (e.g., 19, 17, 15) unspecified. The rubric can note that the scores in-between the described levels of performance would represent a performance slightly stronger than the lower score but not quite fulfilling the description of the higher score. (See Appendix B for an example.) This formatting allows performances to be placed in a single point in the scoring according to how it maps onto the descriptions or in between the stated descriptions. This option may be the best balance between reliability and practicality (e.g., time, effort) when assessing tasks where the performances will vary greatly and when differentiating slight differences in quality is desired.

Pre-use Review

As with all materials development, rubrics require a careful review and an iterative revision process, particularly the descriptions of the levels of performance. This revision process, labeled as pre-use review, involves 1) ensuring that the expectations in the descriptions match the teaching context, 2) checking for consistency in the performance descriptors across levels, 3) considering the practicality of the rubric, and 4) improving the beneficial consequences of the rubric on student assessment and language development. These pre-use review components are described below.

First, the descriptions of level of performance should describe expectations appropriate to the context. Recall that descriptors must focus on the quality of performance, allow the rater to identify the appropriate level of the student performance, and provide the student with insight into their current and potential learning. The descriptions can sometimes unintentionally muddle the underlying construct represented in the learning objective that the category is meant to evaluate. For instance, Isaacs (2014) warned about including allusions to accentedness because learners of a language will commonly have an accent, but the accent itself does not prevent listeners from comprehending the speaker. In fact, the descriptions should not expect native-speaker-like performance (Green, 2014) because in most contexts learners should not be expected to have native-like performance, for instance in pronunciation (Isaacs, 2014; Purpura, 2016).

Second, the descriptions across the levels of performance should display parallelism, including consistency in language and syntax (Goldberg, 2014) so that the learners’ work can be consistently assessed across the levels. Further, the descriptions should be concise without “dysfunctional detail,” which includes long descriptors that try to describe every possible nuance within each level of performance (Popham, 1997, p. 74).

Third, the resulting rubric should be practical. Well-crafted descriptions of levels of performance will make the grading process more efficient (Ambrose et al., 2010) when the levels are easily compared. Practicality is further improved when the descriptions help the rater focus on the specific features which differentiate levels.  Font enhancements, such as bolding, may be useful, when used judiciously (Goldberg, 2014). Then, extraneous and redundant wording perhaps can be eliminated. Depending on the context, phrases (rather than sentences) may be sufficient and efficient to describe the performances. When fewer words are used in the descriptions, the grading can be quicker because the instructor has less to compare to identify the most appropriate level. It is also important to review if the rubric’s format is easy to use. Analytic rubrics can be visually crowded, so spacing and formatting should be carefully considered. At the same time, a one-page rubric is likely easier to use than having categories or descriptions of levels of performance across pages.

Fourth, a well-constructed rubric also improves the beneficial consequences of the assessment because classroom assessments should be in service to learning. An analytic rubric can also clarify the instructor’s expectations, serving as a blueprint for success for the learners, a benefit of the assessment process known as positive washback (Green, 2014). Accordingly, the categories and the descriptions of the levels of performance should be understandable to all users, the instructors who will assess and the learners who will be assessed (Moskal, 2002; Pui et al., 2020), so that the grading process is transparent and the scores provide feedback to the learner (Brown, 2018; Goldberg, 2014). While reviewing the rubric for transparency, think about how to explain the rubric’s components and scoring procedure to the students. In addition to giving a score for each of the categories, instructors may give individual feedback on the language performance. The rubric formatting can facilitate feedback by including a place for comments, either for each category, or general comments, perhaps below the rubric.

Post-use Evaluation

After using any assessment tool, an evaluation and reflection of the assessment process is necessary so that the instructor can reflect on the teaching implications (Ambrose et al., 2010) and the opportunities for improving the assessment. Despite careful design, no rubric will be perfect. Through a review process, both the rubric and the assessment process can be improved. The following content presents specific, feasible suggestions for reviewing a rubric after using it for classroom assessment.

The post-use review should include an investigation of the resulting scores. The class mean (average) and the range (the lowest and highest scores) for the assessment should be calculated to consider whether the scores are meaningful and useful for understanding the learners’ skills. It is also useful to review the mean and the range of scores for each section of the assessment. If the learners’ performances were organized in rank order, lowest to highest score, they should reveal a corresponding increase in quality. Any discrepancies can be reviewed to improve the assessment’s scoring. Additionally, any two or more performances which were given the same score in the stated scoring system should have equivalent quality. If the post-use review suggests that they are not, the scoring system may be revised to reflect the difference in quality. Of course, any changes should align with the purpose of the assessment and the assessment’s stated learning objectives.

Both the categories and the descriptions of the levels of performance of analytic rubrics should be carefully reviewed after using the rubric. A key principle of an analytic rubric is that the categories are distinct with each giving valuable information about the learner’s skills (e.g., Youn, 2015). The independence of the rubric categories can be reviewed by checking if any two categories pattern together. For instance, speakers who are rated as “meets expectations” in category X have also been rated “meets expectations” in category Y; and speakers rated “insufficient” in X have also been rated as “insufficient” in Y. If this is the case, those categories may be confounded or intertwined based on how the descriptions are written. With confounded categories, learners who are deficient (or superior) in one category will not be able to score higher (or lower) in another category; a rubric with confounded categories is incongruent with a main advantage of analytic rubrics, which is assessing various skills separately. To remedy this issue, the descriptions can be revised so that the categories are independent, ensuring that the skill level in one category does not limit the speaker’s opportunity to demonstrate skill in another category. Another possibility is that the categories are inherently connected. For instance, productive vocabulary use and collocation use both increase with proficiency as part of vocabulary development (Bonk, 2011). If the two categories are inherently related, those categories could be combined in the rubric, or one category could be eliminated, knowing that the other category has been shown to effectively measure the construct.

Research has shown that raters sometimes evaluate language performances using their own ideas of what is relevant (Fulcher, 2003), perhaps implicit, unstated expectations (Moskal & Leydens, 2000). Consequently, it is important to honestly consider whether the scoring followed the descriptions in the rubric. Consider possible sources for any inconsistency and how the assessment could help the rater evaluate more consistently. Further, the best learners’ performances can be reviewed for shared features; such a review may suggest features that better differentiate quality of performances. If it was difficult to place student work into one of two levels of performance in a certain category, those descriptions of level of performance can be clarified or more finely separated to capture the differences among the learners’ performance in that particular context.

A post-use review of an analytic rubric should also look carefully at the number of students placed at each level of performance. If the majority of learners have not met the stated level of performance, further instruction/practice is needed so that the learners are meeting the objective or the stated expectations in the rubric may need to be adjusted to the teaching context. Conversely, if most learner performances earned the same score in one category in the rubric, that category fails to differentiate between levels of performance. That result may, in fact, be appropriate, such as when all learners have met the stated standard. With a standards-based assessment, each learner’s work is assessed to the stated standard (not to the other learners), which means potentially every student can pass or even get full points (Green, 2014).

Alternatively, if all students appear to perform at the same level, this result may indicate that the descriptions of levels of performance for that category are too broad. If so, an adjustment should be made to better reflect the context, specifically differences in proficiency for those learners. Typically, the full range in the rubric should be used, sorting learners into each of the rubric’s levels of performance because the classroom assessment is designed for those specific learners. If, for instance, a post-use evaluation reveals that no performance was rated as “inadequate” for any category, the instructor should consider deleting that level when none of the performances of your particular learners fit the description for that level. On the other hand, an unused lowest level of performance can be used to shield learners from negative affect. Even though the ratings will accurately describe the learner’s performance, the rubric format can allow every learner to be above the “worst” description.  Likewise, the same inquiry can be considered within each category, where superfluous descriptions can be deleted. Generally, when the rubric can be simplified by deleting unnecessary descriptions, practicality is strengthened. The practicality can also be evaluated with a review of how easy it was to grade the student work with the rubric. If the rubric (or a section of the rubric) was difficult to use, consider how the assessment could be improved. Often, simply adjusting the order of the evaluation categories (to match the order during the assessment process) or adjusting the physical layout of the rubric improves practicality.

Summary


An analytic rubric is a tool to more objectively and reliably measure the quality of the language performance; it is a tool to assess language production in open-ended tasks where the learner has freedom to decide what vocabulary and grammar constructions to use. Despite the recognition of the importance of assessing productive language skills, the assessment of language performance has several challenges. No rubric is perfect; therefore, revision before and after using it is always necessary. The steps listed below provide a summary of the process for constructing strong analytic rubrics for classroom use. (Note: A more detailed list in provided in Appendix A):

  • Step 1: Categories: Identify categories that reflect separate skills of the stated learning objectives.
  • Step 2: Levels of performance: Describe the expected levels of performance in each category appropriate for the context.
  • Step 3: Pre-Use Review: Review the rubric for validity (e.g., categories are aligned with the assessment’s stated learning objectives), reliability (e.g., performances can be consistently scored with the descriptions), practicality (e.g., rubric is easy to use), and the beneficial consequences of using the rubric.
  • Step 4: Post-Use Evaluation: Check that the scores are meaningful and based on the descriptions, the categories are independent, the descriptions are level-appropriate, and the rubric is easy to use.

The effort to carefully design a rubric for a language assessment, while challenging, improves grading consistency and transparency during the grading. Analytic rubrics are a powerful tool for open-ended tasks to help the instructor evaluate their own teaching, document evidence of learner progress, and give feedback to learners.

About the Authors

Mary Lou Vercellotti is an Associate Professor of English at Ball State University where she teaches courses in linguistics and TESOL, including an assessment course. Dr. Vercellotti earned her PhD in Applied Linguistics at the University of Pittsburgh. Her research interests include investigating language development of adult instructed learners of English, often using speech data.

Dawn E. McCormick is a Senior Lecturer and the Director of the English Language Institute in the Department of Linguistics at the University of Pittsburgh. She is interested in IEP program administration, professional development for language teachers, and student self-assessment and self-correction.

Acknowledgements

The authors thank the participants of the “Creating Rubrics for Assessing Language Performance in the TESOL Classroom” workshop at the 2019 TESOL International Pre-conference Institute. The authors also thank Taylor Lutz, Virginia Robson, and M. Christine O’Neill, and the journal’s reviewers for giving feedback on a draft of this article.

To cite this article:

Vercellotti, M. L. & McCormick. D. E.(2021). Constructing analytic rubrics for assessing open-ended tasks in the language classroom. Teaching English as a Second Language Electronic Journal (TESL-EJ), 24(4). https://tesl-ej.org/pdf/ej96/a2.pdf

References

Ambrose, S. A., Bridges, M. W., DiPietro, M., Lovett, M. C., & Norman, M. K. (2010). How learning works: Seven research-based principles for smart teaching. John Wiley & Sons, Inc.

Ballard, L. (2019) Analytic rubric format:  How category position affects raters’ mental rubric. In S. Papageorgiou & K. M. Bailey (Eds.) Global perspectives on language assessment:  Research, theory, and practice (pp. 3-17). Taylor & Francis Group.

Bonk, W. J. (2001). Testing ESL learners’ knowledge of collocations. In T. Hudson & J. D. Brown (Eds.), A focus on language test development: Expanding the language proficiency construct across a variety of tests (pp. 114-132). University of Hawaiʻi Press.

Brookhart, S. M. (2018). Appropriate criteria: Key to effective rubrics. Frontiers in Education 3(22). doi: 10.3389/feduc.2018.00022.

Brown, J. D. (2018). Rubrics. In B. B. Frey (Ed.), The SAGE Encyclopedia of Educational Research, Measurement, and Evaluation (pp. 1436 – 1440). SAGE Publications.

Brown, H. D., & Abeywickrama, P. (2019). Language assessment: Principles and classroom practices. Pearson Longman.

Fulcher, G. (2003). Testing second language speaking. Pearson Longman.

Goldberg, G. L. (2014). Revising an Engineering Design Rubric: A case study illustrating principles and practices to ensure technical quality of rubrics. Practical Assessment, Research & Evaluation, 19(8). https://doi.org/10.7275/vq7m-e490

Green, A. (2014). Exploring language assessment and testing. Routledge.

Green, A., & Hawkey, R. (2012). Marking assessments: Rating scales and rubrics. In C. Coombe, P. Davidson, B. O’Sullivan & S. Stoynoff (Eds.), The Cambridge guide to second language assessment (pp. 299-306). Cambridge University Press.

Isaacs, T. (2014). Assessing pronunciation. In A. J. Kunnan (Ed.), The companion to language assessment (Vol. 1, pp. 140-155). John Wiley & Sons, Inc. https://doi.org/10.1002/9781118411360.wbcla012

May, L. (2009). Co-constructed interaction in a paired speaking test: The rater’s perspective. Language Testing, 26(3), 397-421. https://doi.org/10.1177/0265532209104668

Moskal, B. M. (2002). Recommendations for developing classroom performance assessments and scoring rubrics. Practical Assessment, Research, & Evaluation, 8(14). https://doi.org/10.7275/jz85-rj16

Moskal, B. M. & Leydens, J. A. (2000). Scoring rubric development: Validity and Reliability. Practical Assessment, Research & Evaluation, 7(10).

Perlman, C. (2002). An introduction to performance assessment scoring rubrics. Understanding scoring rubrics: A guide for teachers (pp. 5-13). ERIC Clearinghouse on Assessment and Evaluation.

Popham, W. J. (1997). What’s wrong–and what’s right–with rubrics. Educational Leadership 55(2), 72–75.

Pui, P., Yuen, B., & Goh, H. (2020). Using a criterion-referenced rubric to enhance student learning: A case study in a critical thinking and writing module. Higher Education Research & Development. https://doi.org/10.1080/07294360.2020.1795811

Purpura, J. E. (2016). Second and foreign language assessment. Modern Language Journal, 100(S1), 190-208. https://doi.org/10.1111/modl.12308

Schreiber, L. M., Paul, G. D., & Shibley, L. R. (2012). The development and test of a public speaking competence rubric. Communication Education, 61(3), 205-233. https://doi.org/10.1080/03634523.2012.670709

Suskie, L. (2009). Assessing student learning: A common sense guide. Jossey-Bass.

TESOL International Association. (2018). The 6 principles for exemplary teaching of English learners, grades K-12. TESOL Press.

Upshur, J. A., & Turner, C. E. (1995). Constructing rating scales for second language tests. ELT Journal, 49(1), 3-12.

Van Gorp, K., & Deygers, B. (2014). Task-based language assessment. In A. J. Kunnan (Ed.), The companion to language assessment (Vol. 2, pp. 578-593). John Wiley & Sons, Inc.

Van Moere, A. (2014). Raters and ratings. In A. J. Kunnan (Ed.), The companion to language assessment (Vol. 3, pp. 1358-1374). John Wiley & Sons, Inc.

Vasylets, O., Gilabert, R., & Manchón, R. M. (2017). The effects of mode and task complexity on second language production. Language Learning, 67(2), 394-430. https://doi.org/10.1111/lang.12228

Vercellotti, M. L. (2018). Finding variation: Assessing the development of syntactic complexity in ESL speech. International Journal of Applied Linguistics, 29(2), 1-15. https://doi.org/10.1111/ijal.12225

Youn, S. J. (2015). Validity argument for assessing L2 pragmatics in interaction using mixed methods. Language Testing, 32(2), 199-225. https://doi.org/10.1177/0265532214557113

Appendix A

Constructing Analytic Rubrics for Assessing Open-Ended Tasks in the Language Classroom: Checklist

Choosing Analytic Rubrics

  • Rubrics are tools to assess open-ended performance-based tasks.
  • Rubrics are used when assessing longer language samples, especially monologic tasks.
  • Analytic rubrics results provide information to instructors and students.

Categories

  • Map each category to a skill listed in the learning objectives of the course or assignment.
  • Check that categories reflect observable and measurable skills.
  • Check that each category is independent.
  • Limit the rubric to no more than five categories.

Levels of Performance

  • Describe the expected level of performance (e.g., meets standard) for each category first.
  • Describe the performance one level higher (e.g., exceeds standard) and/or one level lower (e.g., approaching standard) after the expected performance is set.
  • Use descriptive language (rather than evaluative terms) to help learners understand what they should do to reach the next higher level.
  • Describe quality rather than quantity.
  • Consider the number of levels of performance in relation to the number of categories for ease of use.
  • Check that point values make sense and that category weights are representative of course and activity foci.

Pre-use Review

  • Review the descriptions for context appropriate expectations.
  • Check for consistency among descriptions across levels.
  • Evaluate if descriptions can be shortened while maintaining meaning.
  • Evaluate formatting or text enhancement for ease of use.
  • Consider how well the rubric’s format facilitates students’ understanding of the feedback.

Post-use Evaluation

  • Consider whether the assigned scores adequately represent the quality of the performances.
  • Confirm the independence of each category in the rubric.
  • Evaluate whether each category was assessed consistently, following the descriptions at each level.
  • Evaluate how useful the levels of performance are for this context and these learners.
  • Evaluate how easy the rubric was to use. [back to article]

Appendix B

Table 5. Analytic Rubric: Travel Advertisement – Low-Advanced Learners

20
Strong student example
19* 18
Great work
17* 16
Good work
15* 14
Not yet meeting expectations
Content factual information claims well-supported relevant and specific details factual information statements generally supported
relevant but perhaps vague details
factual but incomplete information claims minimally supported some details less relevant some incorrect or irrelevant information confusing or misleading to audience in parts
Word Choices variety of words appropriate for audience word choice supports meaning variety of words appropriate for audience word choice may be too specific or general but meaning is clear words may be repetitive word choice may reduce understanding of meaning in isolated parts words often unsuitable for task and/or audience
errors interfere with meaning
Grammar variety of structures grammar choices support meaning variety of structures grammar choices may be generic but appropriate for meaning some variety of structures but some overused
some choices reduce effective expression of meaning
limited variety of structures errors interfere with meaning
Accuracy isolated errors
do NOT interfere with meaning
repeated, predictable errors
do NOT interfere with meaning
multiple, repeated errors
distract from understanding meaning
multiple, repeated errors
slows down understanding of meaning
*Fulfills the expectations of the lower level but not quite fulfills the expectations of the higher level

[back to article]

 

[i] Students’ proficiency levels are approximately B1 and B2 on the CERF scale

Copyright rests with authors. Please cite TESL-EJ appropriately.
Editor’s Note: The HTML version contains no page numbers. Please use the PDF version of this article for citations.

© 1994–2023 TESL-EJ, ISSN 1072-4303
Copyright of articles rests with the authors.