Category Archives: Assessment

Using a Rubric to Assess Presentational Writing Tasks

As my long-time followers will have noticed, my understandings related to assessment have continued to evolve over the past several years.  Between this blog and my TPT store, I have shared dozens of different rubrics, each of which reflected my knowledge and beliefs at that time.  In my latest presentational writing rubric, I’ve tried to address certain challenges inherent in using rubrics in the world language classroom. Here are a few that come to mind:

  1. Rubrics must be written in student-friendly language so that they can be used to provide feedback to students.
  2. Rubrics must be concise so that teachers can assess student work efficiently.
  3. Rubrics must ensure that we are addressing our national standards by incorporating the ACTFL Proficiency Guidelines and ACTFL Performance Descriptors.
  4. For most teachers, rubrics must lend themselves to producing a numerical or letter grade.

An additional obstacle in developing rubrics for world language students is that in measuring progress toward proficiency, it is not enough to use a well-worded rubric.  We must also assign performance tasks that are appropriate to the targeted proficiency level of our students. (Click here for information about proficiency level expectations.) I find that the simplest way to do so is by customizing relevant NCSSFL-ACTFL Can-Do Statements when creating assessments. 

Example: I can present personal information about my life and activities in a message to a keypal , using simple sentences most of the time. 

In this example, the italicized text is taken directly from one of the Novice High Presentational Can-Do Statements and the bold print text describes the performance task.  The teacher might, of course, specify additional details or length requirements.

Having created an appropriate task, we can use the rubric to select the proficiency level that most closely aligns with each student’s performance.  In this new rubric, I have clarified the two major considerations that must be taken into account when placing students’ writing along the proficiency continuum. The first is to define the extent to which the student’s writing is comprehensible.  Therefore, first  bold print sentence in each column specifies who can understand the student’s text and how well. 

It is important to note that while comprehensibility is important, it is not enough to determine a student’s level of performance. In fact, our Novice Low-Mid students might be the most comprehensible of all! Because they rely exclusively on memoried words and phrases, these students can often produce error-free lists or other appropriate novice tasks.  Therein lies the importance of the second bold print sentence, which specifies the text type, amount of detail and organizational  features of the student’s performance.  When using this rubric, the student’s level is determined by the highest level for which the student meets all of the criteria.

Most of us will have one more challenge, which is to assign a letter or percentage grade to the performance. Doing so may depend on your school culture, personal philosophy and other factors.  Personally, I feel most comfortable assigning a 9/10 (or 90%) to students who meet the targeted proficiency level.  For standards-based grading, I would use this level as the Level 3 or “Meets the standard.”  The next highest level would be assigned a 10/10 or Level 4 “Exceeds the standard.”  I would assign a score of 8/10 for the proficiency level that is one below the target, a 7/10 to the level that is two below the target, and so on. Here’s a table that reflects this scoring method.

I’d love to hear how you’ve addressed the challenges in using rubrics to assess written tasks. Please share by clicking on “Leave a reply” at the top of this post!

5 Tips for Encouraging Students to Speak in Class

One of the most frequent frustrations that I hear from the teachers I work with is that their students are not willing to speak the target language in class.  While there are many reasons that students hesitate to speak, I have found the following practices considerably improved my students’ willingness to participate in the frequent formative interpersonal speaking assessments that are included in my curriculum.

  1. Provide frequent opportunities for low stakes speaking tasks. As those that use my materials know, I include frequent interpersonal speaking activities in my units. As a result, the majority of my students quickly became comfortable speaking with a partner.  
  2. Make it clear to students that the purpose of the oral interpersonal tasks is to speak.  Most of the tasks include a written component upon which the conversation is based.  However, I make it clear to students that I don’t ever grade the written aspect of the activity.  This clarification removes the student’s temptation to avoid speaking by covertly showing their papers to their partners, speaking English, etc.
  3. End the activity as soon as the first 2-3 pairs have finished (or provide enrichment for early finishers). Since the goal of these tasks is to provide feedback on interpersonal communication, it is not necessary for every student to complete the entire task. Setting a time limit both encourages the students to remain on task and avoids having several students without a meaningful task to complete while their classmates finish their work.
  4. Circulate among the students as they complete the task. In addition to encouraging the students to stay on-task, this practice allows the teacher to provide individualized oral feedback to students. This coaching guides students to make progress toward proficiency as well as normalizes the teacher as an observer during interpersonal tasks.
  5. Provide written feedback. Depending on their own philosophy and school culture, teachers may or may not regularly provide grades on formative tasks.  Because I had a category in my gradebook for formative tasks, I felt it was important to record occasional scores for these interpersonal tasks. However, I found that I could not possibly grade every student during a 15-minute task.  Instead, I simply chose 5-7 students to formally assess during each activity.  By the end of a week, I had a grade and written feedback for every student.  Click here for a generic interpersonal rubric can be used with students of various proficiency levels to quickly provide feedback and/or a formative grade.  

If you’d like to provide additional opportunities for your students to engage in interpersonal speaking tasks, you might consider incorporating one of these spring-themed mini-units:

  • Click here for a 6-day mini-unit on secular Easter traditions for French 1 students.
  • Click here for a 90-minute lesson on Saint Patrick’s Day for French 1 or 2 students.
  • Click here for a 4-day mini-unit on Saint Patrick’s Day for French 2 or 3 students.
  • Click here for a 90 minute lesson on Ramaddan and Aïd-al-fitre for French 2 students.
  • Click here for a 7-day mini-unit on rainy day activities for French 1 or 2 students.

Please share your own tips for encouraging students to speak the target language in class by clicking on Leave a reply at the top of this page!

Grading for Proficiency, 2020

Although I have written about my grading policies in the past, my ideas have continued to evolve since these earlier posts.  Prompted by a recent conversation on the French Teachers in the US Facebook page, I thought I would share my most recent policies for grading in a proficiency-based classroom.

Step 1: Gradebook

My preferred gradebook configuration is to have five different categories: Reading, Listening, Speaking, Writing and Miscellaneous .  I weight each of these categories evenly at 20%.  Other people use the modes (Interpretive, Interpersonal and Presentational) as their categories and I think that’s great.  I preferred to use the skills rather than modes, because my students and I got more specific feedback this way.  I found that a student’s interpersonal speaking and presentational speaking skills were much more congruous than their presentational speaking and presentational writing skills were.  If I used Presentational as a category, I would not immediately be clear whether a struggling student needed support for writing or speaking. Note: If I were in a school with standards-based grading, I would eliminate the Miscellaneous category and move the other categories to 25%. 

Step 2: Formative Assessments

In my classroom the majority of the each class period was spent on communicative activities that I assessed as follows:

Interpretive Reading: When the students read something in class and completed some type of comprehension-based task, I either gave whole class feedback (providing the correct answers) or collected their work for written feedback. This feedback was usually limited to a score (I scored out of 10 rather than percentages) based on accuracy. 

Interpretive Listening: I assigned a lot of Edpuzzles (but only because FluentKey wasn’t yet available!) I used the computer-generated score as a basis for a score out of 10. Because I didn’t expect my students to get all of my questions correct (I designed some questions to encourage my high flyers to stretch), I usually gave a 10/10 for 90% or higher, 9/10 for 80% or higher, etc.  

Interpersonal Speaking: I circulated among my students as they worked on interpersonal tasks. Sometimes I filled out a rubric with specific feedback and recorded a score (out of 10) but most times I jotted down notes on a class roster.  Occasionally I used these notes to determine a grade. I did not generally assign any interpersonal writing tasks. 

Presentational Writing: I often assigned short written tasks based on the interpretive and interpersonal tasks from the lesson and gave basic feedback and a score out of 10.  

Presentational Speaking: I seldom assigned a presentational speaking task as a formative assessment.  When I did so, I provided feedback and a score out of 10.

In addition to these communicative activities, I occasionally gave a quiz of some kind.  For example, if the students had done a pair activity that involved describing pictures to each other, I might describe 5 pictures and have the students identify them. If they had done a reading activity with a partner, I might ask some oral true/false questions about the reading.  

At the end of a 4-6 week unit, I had a LOT of scores.  They ALL went in the Miscellaneous category. A better name for this category would have been Formative Assessment, but I was leaving myself a little wiggle room for the rare instances that I needed to include something that I didn’t consider to be a formative assessment but that I was required to assign by my department.  

Step 3: Summative Assessments

At the end of each unit, I assigned an Integrated Performance Assessment.  Each task in the assessment resulted in a separate grade.  So for most IPAs I had a grade for each performance category: Interpretive Reading, Interpretive Listening, Interpersonal Speaking and Presentational Writing.  If I included a Presentational Speaking task, that would be an additional grade in my Speaking category.  

The grades for the IPAs were the only grades that went into my performance categories.  In other words, 80% of my grades (all but the formative assessments) came from IPAs.  

When I graded my IPAs, I used the following rubrics.

Interpretive Reading: I used the rubric from the ACTFL publication, Implementing Integrated Performance Assessment by Bonnie Adair-Hauck, Eileen W. Glisan and Francis J. Troyan. I assigned 10/10 for students whose work fell in the Accomplished Comprehension category, and 8/10 or 9/10 for Strong Comprehension, a 7/10 for Minimal Comprehension and a 5/10 or 6/10 for Limited Comprehension.  I did not assign scores of less that 5/10.  (See this article for an explanation of this policy.)  

Interpretive Listening: Because I did not use the IPA Interpretive Template for my listening assessments, I developed individual rubrics for my listening tasks.  I usually used a graphic organizer or comprehension questions for listening assessment and a rubric something like this: 

10: Identifies the main ideas and supports each one with relevant details

9: Identifies the main ideas and supports most of them with relevant details

8: Identifies some main ideas and/or several details

7: Identifies a main idea and a couple of details.

6: Identifies a few details.

5: Is unable to identify a main idea or provide any relevant details. 

Interpersonal Speaking: I used this proficiency-based rubric from the Ohio Department of Education. 

Presentational Speaking and/or Presentational Writing: I used this proficiency-based rubric from the Ohio Department of Education. 

Note: The Interpersonal and Presentational rubrics do not provide a numeric or letter grade.  I preferred to just check the relevant boxes to provide feedback to my students.  In order to determine a numerical grade for my gradebook, I used this conversion chart:

*In each case (9) represents the ACTFL expectation.

Standards-Based Grading

Many schools are adopting standards-based grading practices.  I have created standards-based rubrics for each mode of communication and they are available here: https://www.teacherspayteachers.com/Product/Standards-Based-Rubrics-for-World-Languages-4743987

I know that what worked for me may not work in other environments and I look forward to hearing from other teachers who are willing to share the grading practices they have developed!

Making Learning Visible

Thursday evening on #langchat we had a great discussion about what had worked well for us in 2017.  When I shared that I had begun having my students graph their progress toward proficiency, several people expressed interest. So I thought I’d type up a quick post about what we’ve been doing and what I’ve learned from the process.

I began by passing out this document when I returned the students’ first IPA of the year. Although I am clearly evaluating classroom performance in my assessments, it is important to me that my students see their progress in terms of language proficiency, so I use the proficiency rubrics from the Ohio Department of Education for presentational writing and interpersonal speaking. As the directions on the graph document indicate, I had the students place a dot in the appropriate square for each IPA and then draw a line connecting their dots.  I had them use the same graph, but two different colored lines for presentational writing and interpersonal speaking.  Here are a few of the graphs from my French 2 classes:

    

    

There are a few reasons why I consider the simple task of graphing proficiency/performance progress to be one of my successes in 2017.  For one, I LOVED the conversations that I heard among my students as they completed their graphs each time.  It is so much more meaningful to hear “I moved up to Intermediate Low 1” than “I got a B.”  These graphs are also a great visual for my students.  As we transition toward teaching for proficiency (and away from discrete point assessments) some students question “what” they’re learning.  These graphs help students to see their progress in a concrete way.

These graphs also inform my instructional and assessment practices.  Because nearly every student performed lower on the 3rd IPA, I had to take a long, hard look at this assessment (from this post).  My conclusion is that the prompts that I used for the writing and speaking tasks did not encourage the students to stretch in a way that would demonstrate their highest proficiency/performance.  Few students ask the variety of questions that would enable them to demonstrate Intermediate interpersonal communication and most did not include the compound sentences, creativity and cultural competence described in the Intermediate descriptors for the presentational mode. I clearly need to either change these prompts and provide additional direction and targeted practice the next time I teach this unit. I also wonder whether some of the results may indicate a lack of evaluator reliability on my part.  As with most rubrics, there is a bit of subjectivity and I hope to increase my consistency as I continue to use these rubrics.

Image Credit: http://www.publicdomainfiles.com/show_file.php?id=13489790414892

In a Nutshell: 5 Steps to Designing a Thematic Unit

As a result of several recent questions by members of my PLN who are beginning their journey to a more proficiency-based methodology, I have created this outline of the steps I take when creating a thematic unit. While I am planning a series of posts with more detailed information about each step, I’ve included basic information about the process I use, as well as an agenda (updated 8/4/18) with resources for an Intermediate Low unit on vacations, in this post.

Step 1: Determine what I want the students to be able to do at the end of the unit and write a Can-Do statement for each mode of communication. Because ACTFL has not yet released their new version of the Can-Do Statements, I based these Can-Do’s on the current benchmarks. These statements are based on the Intermediate Low descriptors, which is my targeted performance level for these students.

  • Interpersonal Communication: I can participate in conversations about vacations using simple sentences.
  • Presentational Speaking: I can present information on a vacation using a series of simple sentences.
  • Presentational Writing: I can write briefly about a vacation using a series of simple sentences.
  • Interpretive Listening: I can understand the main idea in short, simple messages and presentations about vacations.
  • Interpretive Reading: I can understand the main idea of short and simple texts about vacations.

Step 2: Create the Integrated Performance Assessment. For an in-depth explanation of how I design IPA’s, please refer to this previous post. In short, I 1) Select an authentic written and/or recorded resource, 2) Create a comprehension guide based on the ACTFL IPA template, 3) Create an interpersonal task based on the authentic text and 4) Create a presentational writing and/or speaking task based on the authentic resource and interpersonal task. 

Step 3:  Identify the structures, vocabulary and skills the students need in order to demonstrate the targeted proficiency level on the IPA.  In this unit, I determined that the students would need to learn/acquire the following language, structures and content.

  • Vocabulary related to the topic of vacations.  This would include terms for vacation activities, lodging, transportation, etc. While these students will be familiar with some leisure activities that are part of a typical vacation, a greater variety of vocabulary will allow for more detailed performances.
  • The ability to use past tenses to describe vacations they have taken. While these students used some past tenses in French 2, they will need lots of exposure and practice to be able to use these structures, albeit with expected errors, on these performances. Because the descriptor, “I can usually talk about events and experiences in various time frames” is part of the Intermediate High benchmark,   it will be some time before I will expect these students to easily use these structures. However, by providing opportunities for students to use past tenses in a variety of contexts in this unit I am preparing them to eventually reach this level of proficiency.
  • Cultural background on French products, practices and perspectives.  Because I assess my students’ cultural competence as part of each mode of communication, it is important that they have adequate preparation in determining these aspects of culture throughout the unit.

Step 4. Create a series of lessons that will allow the students to demonstrate the targeted proficiency level on the IPA. Having determined the students’ needs in terms of vocabulary, structures and content, I create individual lessons designed to fill these gaps. These lessons will provide the students with multiple exposures to the targeted vocabulary and structures as well as learning activities that will allow the students to practice/receive feedback on their use of these structures. Here is a simple explanation of the steps that I usually take in designing each individual lesson for a thematic unit. 

A. Determine an organizational structure for the lessons. Based on the theme of a given unit, there are many ways to break the topic into smaller subtopics to provide an integrated structure for individual lessons.  In general, I find it works best to begin with lessons that will provide general information on the topic before focusing on more specific details. So in this case, I began with lessons focusing on general vacation practices and then added tasks related to specifics such as beach destinations, vacation activities, traveling with friends, camping vacations and packing for vacation. Because I curate authentic resources on Pinterest boards for each unit that I teach, I often begin the process of creating subtopics by looking at the resources I already have, and grouping them according to subtopic. This saves a considerable amount of time compared to choosing subtopics and then finding appropriate resources. (Of course, I end up searching for additional resources after I have a skeleton of the unit design.)

B. Create a hook for the lesson.  I choose an authentic written or recorded text to present at the beginning of each lesson.  Presenting simple texts such as infographics or short videos allows me to provide comprehensible input as I talk about the information in the text and ask personalized questions incorporating the vocabulary, structure and content of the text. Click here for a transcript of a sample discussion during the hook portion of the first lesson in this unit based on this infographic.

C. Design an interpretive activity for the lesson. I choose an authentic resource that the students will read or listen to and create a corresponding learning activity/formative assessment that will allow the students to interact with this text.  While I will go into greater detail about this aspect of lesson design in a future post, you will find several different examples in this and other units in this blog. In my opinion, this is the most important part of each lesson, as it provides the basis of the interpersonal and presentational activities that follow.  In addition, because I don’t use a textbook in my classroom, the authentic resources used in the hook and interpretive activities provide the vocabulary and some structures that the students will use in their performance assessments. Note: You will notice that most of the authentic resources used for the interpretive activities in this unit are written texts. In order to ensure that my students have adequate opportunities to interpret recorded texts, I’ve included several video-based formative assessments (using Edpuzzle) that the students will complete in class or at home throughout the unit.

D. Construct an interpersonal activity based on the content, vocabulary and/or structures in the authentic resource. The interpersonal activity provides students with an opportunity to use the vocabulary and structures that were introduced in the authentic resource to create their own meaning.  In addition, as they negotiate meaning on these tasks they are practicing the skills they will use on the IPA with additional scaffolding. Based on the authentic resource and the targeted proficiency level, I incorporate a variety of different types of interpersonal activities.  At the novice level, I often focus on vocabulary-building activities such as those described in this post or even this one. As students reach the Intermediate level and are able to create more with the language, I often integrate interpersonal and interpretive activities by having the students co-create graphic organizers (such as in the 1st and 2nd lesson in this unit) or discuss responses on target language interpretive assessments.

E. Devise a presentational writing and/or speaking formative assessment. These activities provide the students with scaffolded opportunities to synthesize the vocabulary and structures introduced in the lesson to create a written or oral product. The scaffolding provided in these formative assessments, as well as the individualized feedback I will give on many of these tasks, will provide the support the students need to demonstrate growth in proficiency on the IPA. Note: While I have included an idea for a written or spoken presentational task for each lesson, it is unlikely that time will permit me to actually assign all of these tasks.  Instead, I will choose from among those tasks as time allows.

Step 5: Administer and assess the IPA. Because the format of the IPA mimics the organizational structure of the lessons in the unit, the students should feel confident in their ability to be successful on this assessment.

Stay tuned for additional posts on each step of the lesson design and let me know if you have any questions!

Image Credit: http://maxpixel.freegreatpicture.com/Peanut-Shell-Nutshell-Peanut-Shell-Nuts-Nut-390081

Closing the Feedback Loop: An Action Plan

 As my understanding about how languages are acquired continues to evolve, so does my vision of my role in the classroom.  When I began teaching, I considered my prime responsibility to be that of providing vocabulary lists and explanations of grammatical rules followed by opportunities to practice them. A lot has changed over the past few years! I now see my primary role as that of creating contexts for my students to communicate using language suitable to their proficiency level and then providing feedback on their use of that language.  Specifically, I provide language input via culturally-rich authentic resources (as well as my own language use) and create activities that require the students to interpret this language and use the vocabulary and structures they acquire to communicate with others. Of course, my work isn’t finished when these learning opportunities have been created!  These students need feedback on their language use.  They need to know whether their interpretation of a text is accurate and whether their own oral and written communication is comprehensible.  More importantly, they need to know what they can do to increase their proficiency in the language.  

In an ideal world, this means that the students would engage in communicative activities, I would provide immediate feedback on this communication, and the students would use this feedback to set goals enabling them to communicate more proficiently in the future.  However, in the imperfect world of my classroom, this process has not been working they way it should. My feedback has not been timely enough and I have not provided adequate opportunities for the students to use this feedback in a way that would inform their subsequent communication.  As a result, my feedback process looked like the image linked to this post rather than the loop it should have been. The feedback my students received from me often seemed to be a dead end–clearly I need to do much better at closing my feedback loop!

After careful reflection, I’ve come up with the following action plan for the upcoming school year.

Interpretive Communication

My students read a lot of authentic materials in class, but I often fail to provide timely feedback on the accuracy of their interpretation for several reasons.  First of all, I’ve been using the ACTFL IPA template to create comprehension guides for many of these texts.  While I think it’s important that instruction mirror assessment, the use of English for these formative assessments (which I support) would impede my ability to stay in the target language. Furthermore, I worried that my students will be less likely to focus on interpretive tasks if they know that I would be providing the answers at the end of the class.  As a result, I collected way too many papers, spent way too much time grading and recording them (and cajoling absent students to complete them) and wasted valuable class time passing them back to students who looked at the grade and threw them away. I plan to address these obstacles this year by 1) creating formative comprehension tasks that don’t require English, 2) letting go of the idea that grades can be used to control student behavior and 3) providing whole-class feedback directly after the formative interpretive task.  As a result of these changes, I will spend less time grading and my students will receive immediate feedback on their interpretive communication.

Interpersonal and Presentational Communication

While whole-class feedback can be effective on interpretive tasks that often have right or wrong answers, students need specific, individualized feedback to improve their performance on this mode.  While I am able to provide some feedback as I circulate among the students during these activities, I think I could provide more global feedback if each student had an opportunity to receive feedback on the entirety of their performance.  Therefore, my plan is to provide each student an opportunity to be formatively assessed on the same prompt they will have on the IPA, although with a different partner in order to maintain spontaneity on the summative task.  I will then use this document to provide feedback, an opportunity for goal-setting and a means of self-reflection for the students. As the document shows, the students will check the level of proficiency that their formative performance demonstrated (see note below).  They will then check which steps they need to take to improve on their performance on the IPA, based on the feedback given on the rubric on the back of the page. I will then assess their performance on the IPA using the rubric on the second (identical) rubric.  After the IPA, the students will complete the reflection portion of the document which I will then file until the next round of IPA’s. (I might end up making the process digital, rather than paper and pencil.)  I am hoping that the requiring the students to choose action steps, simplifying the rubrics and providing an opportunity for reflection will help close the feedback loop on interpersonal assessments.

I will follow this same process for the presentational task of the IPA.  Using either the presentational speaking or presentational writing feedback form, the students will again record their formative proficiency level, create an action plan and then reflect on whether they were able to achieve their proficiency goal.  

Note about the rubrics

One of my favorite aspects of the Ohio Department of Education rubrics that I had been using is the fact that they break down each proficiency level into 3 different sublevels.  This has allowed me to track small changes, which helps my students see their progress and me to use proficiency-based grading.  However, this specificity makes the rubrics very wordy.  While this would not be especially problematic if I were using them as they were intended–to document proficiency growth from the beginning to the end of an academic year– I found that my students did not have the patience to read through the lengthily descriptors.  Therefore, I created the simplified versions I have included in the documents.  However, in order to document smaller increments of growth, I will add the following sublevels to their proficiency level.

Sublevel 1 Meets all relevant criteria for previous level and at least 70% of the relevant criteria for the targeted level.
Sublevel 2 Meets all relevant criteria for the targeted level.
Sublevel 3 Meets all relevant criteria for the targeted level and at least 30% of the relevant criteria for the targeted level.

While I may adjust the percentages, I think these sublevels will enable the students to see growth and allow me to continue to assign grades based on proficiency levels.

I’d love to hear suggestions on what procedures you’ve developed to create a successful feedback loop!

Using Rubrics to Assess Interpretive Reading

rubricLast night’s #langchat was hopping!  One of the most lively discussions had to do with the topic of using rubrics to assess students’ communication in the interpretive mode.  So, at the request of @MmeBlouwolff, I’m sharing a few thoughts about how I use rubrics to assess reading in my classes.

Like many of my colleagues, I did not understand how I could use a rubric to assess reading comprehension when I first began using IPA’s.  It was not until I saw the ACTFL Interpretive template, that I realized I didn’t have to assess comprehension with discrete point measures.  After adopting the question types suggested by this guide, the switch to a more holistic grading system made perfect sense. A student’s comprehension is not adequately assessed by the number of questions they answered correctly, any more than their presentational writing can be evaluated by counting spelling errors. Furthermore, our current understanding of the interpretive mode of communication does not limit us to evaluating our students’ literal comprehension of a text.  Instead, we are encouraged to assess inferential strategies such as guessing meaning from context, making inferences, identifying the author’s perspective and making cultural connections.  Using a rubric to measure student growth on these skills allows me to show my students what they can do, as well as how they can improve their interpretive strategies.

Here’s a look at a sample of student work and how I used a rubric to assess both the student’s literal and interpretive comprehension. Please note that although I relied heavily on ACTFL’s Interpretive IPA Rubric, I changed the format to make it more similar to the Ohio proficiency rubrics that I use for the interpersonal and presentational modes.  In addition, I modified some of the wording to reflect my own practices and added a numerical score to each column.

As the completed rubric shows, I ask my students to assess themselves by circling the box which best reflects their own understanding of their performance on each section.  In addition to providing an opportunity for self-assessment, this step ensures that the students have a clear understanding of the expectations for the assessment and encourages goal-setting for future performances. This process also provides me with important information about the students’ metacognition. In this case, the student seemed to feel very confident about his/her responses to the Guessing Meaning from Context section, in spite of the fact that he only guessed one word correctly.

After collecting the assessments and student-marked rubrics, it’s my turn to assess the students.  The use of a rubric streamlines this process considerably, as I can quickly ascertain where each student’s performance falls without the laborious task of tallying each error.  I simply check the appropriate box on the rubric, and then project a key when I return the papers so that each student receives specific feedback on the correct responses for each item.  

When it comes to determining a score on the assessment, as a general rule I assign the score for which the student has met all, or nearly all of the descriptors. I do consider, however, how the class does as a whole when assigning numerical grades.  I am frequently unrealistic in my expectations for the Guessing Meaning from Context, for example, and as a result I do not weigh this category very heavily when assigning a final score.  In the case of this student’s work, I assigned a grade of 9.5/10 as s/he met many of the descriptors for Accomplished and demonstrated greater comprehension than the majority of his/her classmates.

While the use of rubrics for interpretive communication might not work for everyone, I have found that holistic grading provides better opportunities for self-assessment, encourages students by providing feedback on what they can do and saves me time on grading.  

As always, I look forward to your feedback, questions and suggestions!

Image credit: https://commons.wikimedia.org/wiki/File:Rubric.jpg

Taking the plunge into proficiency-based grading

grade-28199_960_720A couple of years ago when I decided to drastically change what I taught (cultural content instead of vocabulary and structures) and how I taught it (by using authentic resources instead of textbook exercises), I took a close look at my assessment practices.  While I embraced the concept of IPA’s, I struggled a bit on how to assign a grade to these assessments.  In the beginning I used my own holistic rubrics and later adopted the Ohio Department of Education’s Performance Scoring Guides for World Languages. Being a rule follower, I chose the Performance rubrics because that’s what ODE’s website said that teachers should use for IPA’s.  Although I knew that some teachers were linking students’ grades to their proficiency level this practice didn’t fit with my understanding of proficiency, which I’ve been taught can only be measured by an OPI.  Because I understood that my classroom assessments were clearly performances (measurements of what my students had learned as a result of my instruction), I used the Performance rubrics.  While these are great rubrics, as I continue to adapt my instruction, I find that I will need to make some changes to my assessment practices in order to meet my goals for this year.  Specifically, I want my students to be more involved in their own learning. Rather than passively waiting for me to assign a numerical score to all of their performances, I want my students to understand their proficiency level, set their own proficiency goals, understand how to meet those goals, and self-assess their progress in reaching these goals. Because the descriptors in ODE’s Performance Rubrics do not reflect different proficiency levels (There is only one scoring guide for each skill/mode.), my students were not able to determine their current level of proficiency based on my completing this rubric.  Furthermore, they were not able to determine exactly what they needed to do to improve their proficiency (or grade). In the absence of clear descriptors for each level of proficiency, the students were faced with trying to hit a moving target.  As my performance assessments required increasingly greater levels of proficiency, a similar score on a string of assessments did not allow the students to see the progress that they were making.

In order to remedy this situation, I’ve decided to use ODE’s Proficiency Scoring Guides this year. Based on my current understanding of the common language of world language educators, I will be able to describe my students’ performances as exhibiting characteristics of a proficiency level, without implying that I am able to assign a specific proficiency level to an individual student.  But most importantly, because these rubrics contain separate descriptors for each proficiency level, they will enable my students to define their performances as exemplifying a targeted proficiency level.  Not only will my feedback allow them to identify their current level of performance, they will know exactly what they need to do to achieve the next level.  I especially love that these rubrics include three levels for each proficiency level (NH1, NH2, NH3, for example).  As a result, I hope to be able to measure each increment of progress in my students’ path to proficiency.

For many of us, of course, it is not enough to only identify a student’s proficiency level, we must also assign a numerical (or letter) grade for each performance.  After reading many outstanding teachers’ methodology for doing so, I’ve determined the following guidelines for implementing my proficiency-based grading system.

  1. Students who reach  ACTFL Proficiency Target will earn an 85% (B).  Because it seems unfair and unrealistic for the students to reach an end of course target first semester, I have (somewhat arbitrarily) determined that the first semester goal will be two sublevels below the end of the course target.  For example, since Novice High 2 is the targeted proficiency level for the end of French 2, Novice Mid 3 is the targeted level for first semester.   This table shows what score a student will earn for each proficiency level. (The numerical scores reflect my preferred maximum score of 10 rather than 100 [a percentage].)
  2. In order to more easily implement this system, I have prepared a first semester and a second semester rubric for each course. As indicated on the rubrics, the language is taken directly from the ODE scoring guides for each skill/mode. I simply chose which 5 columns I felt would be the most likely to cover the range of levels for a particular course and typed them on a single page, with an additional column for comments. I also took the liberty of creating a separate rubric for each Presentational skill and removed the comments about pronunciation from the Writing rubric in order to streamline the feedback process. I can easily use a lower level rubric (changing the scores accordingly) for those students who are unable to meet the lowest level on the rubric for his/her course.  Note: I have not included a 2nd semester rubric for French 4, as the ODE rubrics stopt Intermediate Mid 3. I’ll use my own judgment in assigning a score for any students who exceed this level.
  3. Because ODE does not have an Interpretive rubric (They provide only a link to the ACTFL IPA Interpretive Rubric), I will use the ACTFL rubric for interpretive reading tasks at each level. Because it is the task, rather than level of performance which demonstrates a student’s proficiency in interpretive assessments, the same rubric is appropriate for all levels. I will assign the following numerical scores to each level on the rubric: Limited Comprehension (7), Minimal Comprehension (8), Strong Comprehension (9) and Accomplished Comprehension (10).  A student who does not meet the descriptors for Limited Comprehension will earn a 6.

I’m sure that I’ll make modifications to these guidelines as I implement proficiency-based grading, so if you’re assessing according to proficiency, I’d love to know how it’s working in your classes!

Resources for Planning and a Food Unit for Intermediate Low French Students

undefined

As regular readers may have noticed, I ended up taking a hiatus from blogging this spring. It all started when I welcomed an awesome student teacher to my classroom who was so well-skilled in proficiency-based instructional methods that I didn’t need to create any new lessons for several weeks. Then I decided to relocate closer to family, creating a whirlwind of life changes which including finding a new position, selling a house, buying a new house, moving and setting up a new household. To say it’s been hectic would be an understatement! Anyway, with everything going on, needless to say, I had to put aside my blogging for a few months! However, now that I’m settled into my new home I’m anxious to share some of the materials I’ve been working on for my new students.

Creating units for students that I’ve never met, in a school with a different curriculum and culture than the one I left has been a bit of a challenge. Although I don’t know much about the proficiency level or personal interests of my new students, I can’t wait until August to begin preparing instructional materials for my new kiddos.

Besides, reading Chapter 1 of The Keys to Planning for Learning for #langbook has me thinking about all of the ways I can improve my planning and I’m excited to start implementing some of the ideas that are reinforced in this book.

I decided to start with my French 3 curriculum, since I will have three different French classes this year–half of my school day. In addition to reading The Keys to Planning for Learning, I completed the self-assessment survey provided by the TELL Project before developing this unit. As a result of this self-assessment, I realized I needed to be more intentional in developing daily objectives for my lessons. Although I had previously created Can Do Statements for each unit, I hadn’t provided my students with a clear objective for each lesson. I have therefore included daily performance objectives in addition to the Essential Questions and Can Do Statements for this unit.

Because the first theme in my new French 3 curriculum, “Nourriture,” is so broad, I have broken it down into three topics–breakfast, school lunch, and Francophone specialties. This Google Slide Presentation (updated link 7/3/2018) contains the unit plan as well as links to the materials I’ve created/borrowed for each of the 19 lessons in the unit. I am hoping that this format will improve transitions, encourage the students to work more independently and allow absent students to complete work from home. It will also facilitate sharing this work as I can continue to make edits/correct errors without having to reload word documents to this blog. While I’ve previously shared some of these materials, many others are new, including several Edpuzzle video quizzes that will serve as formative assessments in the 1:1 learning environment of my new school.

While I have not included assessments in the presentation, you can click here for the breakfast IPA and here for the school lunch IPA. As the agenda shows, the students will prepare a presentation, rather than a full IPA as a summative assessment on the Francophone specialty topic.

As always, I welcome feedback on these materials!

Image Credit: http://commons.wikimedia.org/wiki/File:Italian_cooking_icon.svg?uselang=fr

4 Suggestions for Assessing Interpersonal Communication with Novice Learners

600px-Two-people-talking-logoAs I’ve evolved in my teaching practice, I’ve made significant changes to how I assess oral communication. Here are a few suggestions that have helped me improve my assessment of my Novice students’ interpersonal communication, resulting in increased proficiency among these early language learners.

Suggestion #1: Just do it! It seems that many of my colleagues are hesitant to assess their Novice students’ interpersonal communication. It is their belief that that because these students are entirely dependent on memorized language, no true interpersonal communication can occur.  Fortunately, we’ve agreed to disagree on this point! In my opinion, a Novice speaker’s reliance on practiced or memorized language does not preclude her from true communication on an unrehearsed task.  While the topics that these learners discuss will be limited to those which have been practiced, they will still be able to demonstrate interpersonal communication when given an appropriate communicative task.  In fact, the NCSSFL-ACTFL Can-Do Statements list the following for Novice Mid Interpersonal Communication:

  • I can greet and leave people in a polite way.
  • I can introduce myself and others.
  • I can answer a variety of simple questions.
  • I can make some simple statements in a conversation
  • I can ask some simple questions
  • I can communicate basic information about myself and people I know
  • I can communicate basic information about my everyday life

Clearly we can expect students to be able to demonstrate an ability to communicate about such basic topics as likes/dislikes, leisure activities, family members, school subjects and supplies, eating habits, etc. While appropriate questions, answers, and rejoinders may be practiced in advance with Novice learners, we can ensure that they are demonstrating actual communication in this mode by creating prompts that prevent memorization of a script.  For example, by pairing a student with a classmate with whom he hasn’t yet practiced, we ensure that he is unable to memorize the exact questions he will ask and the answers or rejoinders that he will give.  Consider a unit on likes and dislikes, often one of the first topics in a Level 1 curriculum. We might give an assessment prompt such as, “You are choosing a roommate for choir camp and you want to make sure you end up with someone that likes the same things you do. Discuss your likes and dislikes in order to find out what you have in common.”  While these students will have practiced expressing their preferences (“I like…”) , asking questions (“Do you like…?), and replying to a partner (“Me, too.” “Not me.” “Me neither.”), they will not know in advance which questions they will be asked or which responses their partner will provide.  As a result, each student must be prepared to ask a variety of questions, in order to avoid repeating those asked by his partner. Likewise, working with a new partner will require a student to comprehend his interlocutor’s response (rather than simply memorizing a script) in order to choose the appropriate rejoinder. (“Me neither” is not an appropriate response to a partner who has stated that she likes something, for example.) Furthermore, even Novice learners can make some adjustments in order to clarify meaning for an interlocutor who has demonstrated a lack of comprehension.  Requests for repetition are often all that’s needed in order to understand a message, whether because the original speaker is able to correct an error that impeded comprehension or because the repetition enable the interlocutor to establish additional meaning. Clearly, even a task as simple as this one does require the negotiation of meaning which typifies interpersonal communication.

Suggestion #2:  Stay out of it! I rely almost entirely on student-to-student interaction for my interpersonal assessments, even for Novice learners, for the following reasons:

  1. In my experience, allowing students talk to each other great increases the quality of the interaction. I have followed the suggestion of Colleen Lee (@CoLeeSensei) who said in a #langchat discussion, “I teach my [students] that your partner not understanding you is YOUR responsibility to clear up!” Nothing is more magical than hearing a level 1 student encourage a classmate by suggesting possible language chunks that would provide the necessary clarification to allow communication to occur. In an early unit this year, for example, one student negotiated meaning by asking C’est ta mère ou ta sœur? when her partner used the incorrect vocabulary word when describing his family photos. This clarification gave valuable feedback to the speaker who had made the error, as well as allowed his interlocutor to stretch beyond the rehearsed statements and questions she had anticipated using during this assessment. As a result of this negotiation, both students are likely to make progress toward proficiency that wouldn’t have been likely had the conversation occurred between a teacher and student.
  2. I have found that being able to talk to a peer, rather than the teacher, greatly reduces the students’ affective filter. A conversation between a teacher and student, regardless of the prompt, is a conversation between an expert/evaluator and a student, which creates a certain level of anxiety in many learners. When a student’s focus is on communicating with a peer, however, she is often able to disregard the presence of the teacher (who is most likely busily taking notes in order to provide feedback to the speakers). As a result of this decrease in anxiety, the quality of the communication is considerably greater than it would have been if the student was speaking to the teacher.
  3. In my opinion, the authenticity of the communication is significantly reduced when one of the speaker’s primary motivation is assessment, rather than comprehension. In a teacher-student interpersonal assessment, the student’s goal is most likely to avoid errors, and the teacher’s is to note them as part of the feedback process. As a result, the communicative content of the conversation often loses its significance.
  4. Lastly, assessing pairs of students saves valuable class time. While I can generally assess all 30 of my students in one class period when listening to two speakers at a time, I would not be able to do so if I were assessing each one individually.

Suggestion #3: Use a great rubric. I love the one from the Ohio Department of Education (http://education.ohio.gov/getattachment/Topics/Ohios-Learning-Standards/Foreign-Language/World-Languages-Model-Curriculum/World-Languages-Model-Curriculum-Framework/Instructional-Strategies/Scoring-Guidelines-for-World-Languages/3-Interpersonal-Rubric_unit_august_2015.pdf.aspx ) because it includes an interculturality component, as well as great descriptors related to the quality of the interaction.  The wording in the comprehensibility section makes it clear that some errors are to be expected, even for those students rated as Strong. Knowing that the content and quality of the interaction are as important as accuracy encourages students to make more risks during interpersonal tasks. This risk-taking leads allows the learners to demonstrate greater proficiency than they would if their only goal was to avoid grammatical errors.

Suggestion #4: Don’t forget to incorporate culture. As I discussed in an earlier post, I am experimenting with using role plays in order to incorporate more culture into my novice interpersonal communication assessments.  My previous prompts, in which I asked students to discuss their own personal preferences and experiences, often failed to produce adequate evidence of interculturality.  On the other hand, I was pleased with the results I had during a recent holiday unit when I assigned a role to each member of the conversation pair. In this assessment, I asked one student to play the role of someone who had traveled to France for the holidays and the other to play the role of someone who had traveled to Canada.  When these students discussed the pictures they had “taken” (a Google Slides presentation I prepared), they were able to demonstrate their cultural competence in a comprehensible way, in spite of their Novice proficiency level.

I’ll look forward to hearing what has worked for you when assessing impersonal communication with Novice students!