I use a standardsbased assessment system in my classroom. Assessments are marked based on evidence of conceptual understanding using a pretty extensive rubric. Most of my students are new to this idea, and it’s a struggle for them to understand. The students who struggle the most are often the ones who have had some success in traditional math classes, where grades are based more on the ability to perform algorithms, procedures, and calculations fluently. Assessing for understanding requires that they not only are able to perform procedures correctly (which is of course still important), but also show evidence that they understand the underlying concepts. This is incredibly frustrating for students who are able to learn procedures without understanding why they work. It’s also tricky for me, as a teacher, to illustrate why this is so important. On a recent precalculus assessment over vectors, some students produced interesting responses that, sometimes within the same student’s test, made me think a great deal about this difference, and provided me with some great fodder for explaining it. Context: The learning target for this assessment is NVM.A and B: Represent and model with vector quantities and perform operations on vectors. I knew the students needed some scaffolding for the process of adding vectors. It is complicated, with many opportunities for errors. It’s also a great example of the kind of problem where you can do the calculations absolutely correctly without having any understanding of what you’re doing: just give me some formulas for r and theta, and I’m golden, right? Well, not really, especially when you have to figure out the angle at the end. There might be a bunch of “rules” to teach students about when to subtract from 180, add to 180, ditto for 360, but I don’t use ‘em, or know ‘em: in my opinion, you really have to understand what you're doing to reach valid solutions on these problems. I don't know a better way to show this understanding than visual models. The week before the test, we went through this problem together as a class. Students worked it out, and put it in their notebooks for reference. My assessments are opennotes, so the intention of this exercise was to give students a thorough walkthrough of one vector addition problem to use to solve problems on the assessment. The point I made repeatedly while working this through with three sections of students was this:
Student A’s response to the boating problem made me really hopeful: precise use of notation, clear reasoning, good calculations, and a visual model that, while not very accurate, at least shows that A has a reasonable idea of where the boat’s going. Then I saw the angle at the end, and said to myself, “Dangit! A is just blindly following the procedure we did in class from his notebook. What a bummer!” Student C used COLORS!!!, which they know makes me biased. But hey! C screwed up the calculation for the angle at the end. WRONG, RIGHT? Well, yeah, until C did this supersweet confirmation to check if the answer made sense. C gets it, just made a calculation error. Proficient conceptual understanding, needs work on procedures and showing reasoning. Here’s student D, who’s a little further from the goal than student B, but using some correct procedures. What’s missing? Well, there are quite a few things missing, but most of all, it’s sensemaking. Getting negative x and y values for a vector in the first quadrant should be a red flag for a student who understands what s/he is doing. D’s calculator is in radian mode. These are just a few examples of how interesting assessing responses can be when you look for understanding and reasoning rather than right or wrong. I’d love to hear opinions from anyone who would like to discuss.
But back to the students. I’m only giving them written feedback on this assessment, no grade (we take two assessments over every topic, and this is just the first). I’m hoping that sharing these responses with the students next week and having them do a little assessment or comparison of their own work will help make the point clear: using the right algorithms, even if you do it well, isn’t enough to prove that you understand the concepts.
0 Comments

Jon LindLet's see if I can keep up with a blog! Archives
November 2018
Categories
All
