As a teacher, I struggled with the fuzzy logic of grading every term. I was invested in all those points I totaled and calculated, in categories I devised and weighted on assessments I wrote. I considered their relative value, their worth as a measure of learning, their objectivity and subjectivity. Did I grade that first paper, the one I graded just after dinner, when I was fresh, full, and in a good mood, on the same relative scale as that last paper, when I was exhausted, and just wanted to get to bed? Did the midterm test comprehension or rote memorization? I agonized over these details as if they were my final and unequivocal communication of educational truth.
I realized that the current system of points-based grading is highly subjective. As Alfie Kohn has written, “what grades offer is spurious precision—a subjective rating masquerading as an objective evaluation.” A few years ago, I told my students about a study I’d read that showed judges rule more favorably after breaks, so from then on, students left snacks in my office and reminded me to take breaks when they knew I would be grading their work. If the purpose of grading is to objectively evaluate student learning and achievement, surely my work breaks and snacking habits should prove irrelevant in their calculation.
Teachers are trapped in a Catch-22. We are asked to assess our students precisely (many grading programs track scores to the hundredths place) and with the appearance of objectivity while using an inherently subjective process. Teachers are then asked to present their calculations on official documents and defend those numbers at parent-teacher conferences as if they are objective measures of student learning. For all the effort, time, and best intentions teachers invest in those reams of grade reports, we are lying to ourselves and to our students’ parents, cheating our students out of clear and accurate feedback on their academic process, and contributing to the greater illusion that grades are an accurate reflection of skill mastery.
Teachers have struggled for years with the calculation and purpose of grades. The evolution of the grading system we use today reflects that search for a valid system of evaluation and assessment. In 1913, I. E. Finkelstein sought to find answers to a few basic questions about grading in his book The Marking System in Theory and Practice:
What should the mark really represent? Should the mark be based upon ability or performance, or even upon zeal and enthusiasm? What is the best set of symbols to represent ability or achievement?
At the heart of his book is the question of what a grade ought to represent. In the early days of American education, teachers used all sorts of distinctions in order to evaluate and differentiate students for the convenience of the teacher and the institution. As former Harvard University president Charles William Eliot explained in his book Harvard Memories, 18th-century Harvard students were arranged “in an order determined by the occupational standing of their parents.” As colleges moved toward a more academically relevant measure of distinction, Yale was the first institution to use a system of evaluating achievement, first with a series of descriptive adjectives, and later with a numerical scale of 1 to 4, which probably led to the 4.0 scale we use today. In 1877, Harvard began using academic “divisions” and a system of “classes” to rank students. Finally, in 1897, Mount Holyoke College adopted the familiar system of A-D and F for grading students.