By John D. Mays, Ph.D. | May 16, 2021
Mastery learning has been the subject of hundreds of research studies since Benjamin Bloom’s ground-breaking 1968 paper.[1] The concept actually goes back way before Bloom, but it was Bloom’s prestige that finally put the subject permanently before the attention of educational researchers.
Fundamental Principles in Bloom’s Model
Bloom’s 1968 mastery-learning proposal includes these fundamental ideas:
Summary of Mastery-Learning Literature
The research literature on mastery learning is extensive. Since 1976, five metastudies of this literature have been conducted: Block and Burns (1976); Guskey and Gates (1986); Slavin (1987); Guskey and Pigott (1988); and Kulik, Kulik, and Bangert-Drowns (1990). All these except Slavin (1987) report strong, positive research findings in various areas associated with mastery learning, including student achievement, student retention, student affective variables, and teacher affective variables. Kulik, Kulik, and Bangert-Drowns (1990) include a detailed critique of the procedures and findings reported by Slavin (1987), discussed below.
In addition to the metastudies, several other significant analyses and discussions of the literature have been published, including:
Nearly all the writers mentioned so far hold that the extensive body of research literature supports the use of mastery-learning principles, Slavin (1987) being the single exception. In his 1987 metastudy, Slavin found positive results from mastery learning on criterion-referenced tests (i.e., teacher-written tests that measure student performance on learning objectives associated with the specific curriculum in use) but found no significant effect from mastery learning on student performance on norm-referenced tests (i.e., standardized tests).[6] Here I will mention two issues associated with Slavin’s analysis. The first, noted by Kulik, Kulik, and Bangert-Drowns (1990), is that the selection criteria used by Slavin are problematic and that he unreasonably excludes from his analysis many of the studies that would lead to different results.[7]The second concerns Slavin’s focus on norm-referenced tests. Kulik, Kulik, and Bangert-Drowns (1990) as well as Zimmerman and DiBenedetto (2008) argue that student performance on criterion-referenced tests is exactly what we should be concerned about, since criterion-referenced tests measure how well students have learned the specific content of what the teachers are teaching them.[8]
I agree with the position expressed by Kulik, Kulik, and Bangert-Drowns (1990) and Zimmerman and DiBenedetto (2008). As a classroom teacher, my concern is exactly with how well students learn what I teach them. Of course, those teaching in public schools that mandate the administration of norm-referenced tests must be concerned with those results as well. However, unless one is specifically teaching to the test, I would not expect the mastery-learning methods used in short-term (one year or less) research studies to have much influence on performance on norm-referenced tests for the simple reason that such tests are designed to measure student development over the long term. I would expect a positive influence on performance on norm-referenced tests if mastery-learning methods were used over the course of years, and if the tests were designed to test what the schools are teaching, but this is not what the research studies have examined. In his short 1990 paper, Slavin concedes that “the findings of positive effects of mastery learning on experimenter-made measures can be interpreted as supporting the view that this technique can help focus teachers on a given set of objectives. Educators who value this have a good rationale for using group-based mastery learning.”[9]
The findings of the metastudies are well-summarized in Davis and Sorrell (1995). They report the finding from Block (1971) “that students with minimal prior knowledge of material have higher achievement through mastery learning than with traditional methods of instruction.”[10] Other points noted in this paper are:
Davis and Sorrell conclude their study with a comment that is consistent with claims made in my own publications: “School systems must recognize that traditional methods of teaching and learning are unsuccessful for many students.”[15]
The Bloom-Guskey Mastery-Learning Model
Early studies of mastery learning varied in the details of their applications of various mastery-learning models, making it difficult to summarize the models in terms of a set of common features. But over the years, the common features coalesced into a well-defined mastery-learning model, articulated primarily through the writings of Thomas Guskey, an educational researcher and leading proponent of mastery learning. Guskey (2005) notes that Bloom’s model is all about reducing variability in student achievement by increasing variability in instructional approaches and learning time.[16]He also notes that from a researcher’s perspective, achievement gaps are matters of variation.
Guskey (2005) describes Bloom’s model with a list of essential elements and a flow chart similar to the one shown below.[17] The flow chart illustrates the basic idea. A class is taught the material from a unit of instruction, using whatever high-quality instructional practices the teacher chooses. The students are then administered a formative assessment that measures students’ mastery of the material. (Formative assessments are used for learning instead of the usual assessments of learning.[18]) Those who score below a mastery level of typically 80% or 90% on this assessment are given additional instruction consisting of tasks, known as correctives, specifically targeting the objectives for which mastery was not demonstrated. This additional instruction lasts for a couple of days and is not mere reteaching. Instead, instruction and correctives are individualized for each student to address students’ different “learning styles, learning modalities, [and] types of intelligence.”[19] After the additional instruction and correctives, the students take a second formative assessment, this time targeting the specific skills addressed in the correctives. In this retest, problems and questions are similar but not identical to those on the first assessment. Those who demonstrate mastery are ready for the next unit; those who do not enter into a second set of correctives.
During the class time needed for correctives and retesting, students who scored at or above the mastery level on the first formative assessment are engaged in enrichment activities designed to broaden or deepen their knowledge of and experience with the material. The nature of the enrichment activities can vary widely, and student interest is one of the factors that guides the selection of particular activities.
Two comments about this process might be mentioned here, neither of which are mentioned in Guskey (2005). The first is that some students may take up to three formative assessments before moving to the next unit.[20] Second, two different approaches to grading are mentioned in the mastery-learning literature. Guskey (2010) says that
Mastery learning teachers make a point of recognizing those students who do well on the initial formative assessments. But they also acknowledge that students who do well on the second assessment have learned just as much and deserve the same grades as those who scored well on their first try.[21]
A second approach is suggested by Bloom:
Implicit in this way of defining the outcomes and preparing evaluation instruments is a distinction between the teaching-learning process and the evaluation process. At some point in time, the results of teaching and learning can be reflected in the evaluation of the students. But these are separate processes. That is, teaching and learning are intended to prepare the student in an area of learning, while evaluation (summative) is intended to appraise the extent to which the student has developed in the desired ways.[22]
In other words, the scores on the formative assessments do not figure into the student’s final grade. The final grade is established by a summative assessment given later. As I note in the comparison section below, the manner in which student grades are established in my own mastery-learning model is distinct from both of these approaches. The justification for this grading policy is described below.
Guskey (2005) summarizes Bloom’s mastery-learning model as including two essential elements: 1) a feedback, corrective, and enrichment process, and 2) “instructional alignment.” The first of these is described above. Instructional alignment entails four components: a) learning objectives for each unit, b) instruction targeted at the objectives, c) formative assessments targeting the objectives, and d) feedback and correctives for objectives not mastered.[23]
The FWTM Model
The FWTM model is described fully in From Wonder to Mastery, cited at the beginning of this paper.[24] The goal of the FWTM model is to eliminate the Cram-Pass-Forget Cycle, the normal operating mode in most classrooms in America. When teachers and students operate this way, students cram for tests, pass them, and forget most of what they crammed in about three weeks. In the FWTM model, Cram-Pass-Forget behavior is replaced with a Learn-Master-Retain pattern:
The mastery component of the FWTM model may be implemented in any courses in any subject area. But the full FWTM framework applies to science instruction and includes these three major elements:
The power of the FWTM model resides in the unified orchestration of science classes around wonder, integration, and mastery. However, for the purpose of comparing the FWTM model to the Bloom model, we must focus the present discussion on the mastery or mechanical element of the FWTM model, since the Bloom model entails only the mechanics of teaching, testing, correctives, and retesting.
Implementation of the mastery element in the FWTM model requires specific attention in three areas:
Comparing Mastery-Learning Models
Both the Bloom model and the FWTM model result in significant gains in student achievement, significant increases in positive student affects, and significant increases in positive teacher affects. Both models also establish mastery as the expectation for students. However, the two models differ in key respects and, in my view, the differences result in some benefits that favor the FWTM model over the Bloom model. The purpose of this section is to make these distinctions and benefits clear. In the tables below, I compare the two models in several specific categories. For this comparison, I use the weekly quiz regimen described in From Wonder to Masteryand recommended for grades 7–9.
Some additional discussion is warranted on some of the differences between Bloom’s model and the FWTM model.
The first issue relates to the way student grades are established. As noted above, Bloom held that formative assessments should not count in the students’ grades, which should be based instead on a summative assessment after the learning process is complete. I certainly understand the principle involved here, that learning and assessment of learning are separate processes. (This is why one should assign little or no grade credit for homework—homework is a necessary learning activity but not a legitimate assessment of whether learning has occurred.) However, there is an inevitable decrease in motivation for some students when they know that if they don’t score high enough on a quiz the grade won’t count. I have implemented mastery-learning programs in both math and science that have run for many years. While I was developing the FWTM model for use in science classes, I developed a separate model for mathematics that we applied in middle and high school to all math classes and all students from pre-algebra through algebra 2. The model we used was similar to Bloom’s and required students to take a retest if they didn’t score at least 83 on the chapter test. Early on, we discovered a fundamental property of student psychology: many students are not motivated to apply themselves diligently to prepare for a test if they know the test grade doesn’t count. If the score doesn’t count, there is no penalty for not being prepared for the test. This results in some students not taking the steps they need to take in order to be prepared for the test; they are content simply to retest over and over until they finally hit the magic 83. This experience persuaded us that it is essential that scores on all quizzes and tests must be included in the students’ grades.
Second, a closely related psychological flaw resides in the use of required retesting itself. In our mathematics mastery program, we found that over time an increasing number of students discovered that even if all the scores were counted in the grade they could game the system by essentially not studying or working at all, other than showing up at the required after-school tutorial sessions. Thus, the teacher became the one doing all the work writing retests, scoring retests, and tutoring students while the students sat more or less passively listening in the belief that sooner or later they would have seen a problem worked enough times that the student finally scores the 83 by some accident. Needless to say, the load on teachers when some students require five or six retests is enormous. There is no remedy for this problem other than forcing students to accept a failing grade if they don’t score at the mastery standard within a specific number of test attempts, and such a policy would be politically very unpopular.
Third, without a requirement for cumulative assessments, the Cram-Pass-Forget Cycle—the enemy of education—still occurs. Since the Bloom model does not require cumulative assessments, the higher achievement does not translate well into students taking what they have learned with them into future years and courses. Research does show improved retention, but the studies measure this in terms of only a few weeks (four months in a few cases). Thus, the model addresses the level of initial achievement, not the tendency to forget things after a few weeks. The cumulative nature of the assessments is at the very core of the success of the FWTM model and results in students remembering things for months or years into the future.
Fourth, Guskey remarks that for a student to be engaged in additional tutoring and practice on older material while simultaneously moving on to new material is an “impossible situation” that should be avoided.[27] However, I have used the FWTM mastery-learning model in the classroom for a decade myself and I have not found this to be at all the case. When students study correctly in the first place, they do not fall behind and do not need much in the way of tutoring. I found that each year a few students did fall behind in their practice of older material and did have to work hard to get caught back up. But never once did any student or parent complain that a mastery standard requiring students to get tutoring on their own time when needed was an unreasonable challenge.
Benefits of the FWTM Model
Structuring mastery learning around cumulative assessments instead of retests results in several benefits that make the FWTM model successful and preferable. I conclude by itemizing a few of these.
References
Block, James H. and Burns, Robert B. (1976), Mastery Learning, Review of Research in Education, 4, 3–49
Bloom, Benjamin. (1968), Learning for Mastery, Evaluation Comment, 1, 2, 1–11
Davis, Denese and Sorrell, Jackie. (1995), Mastery Learning in Public Schools, Educational Psychology Interactive, December, accessed http://www.edpsycinteractive.org/files/mastlear
Guskey, Thomas and Gates, Sally L. (1986), Synthesis of Research on the Effects of Mastery Learning in Elementary and Secondary Classrooms, Educational Leadership, 43, 8, 73–80
Guskey, Thomas and Pigott, Therese D. (1988), Research on Group-Based Mastery Learning Programs: A Meta-Analysis, The Journal of Educational Research, 81, 4, 197–216
Guskey, Thomas. (2005), Formative Classroom Assessment and Benjamin S. Bloom: Theory, Research, and Implications, paper presented at the 2005 Annual Meeting of the American Educational Research Association
—— (2010), Lessons of Mastery Learning, Educational Leadership, 68, 2, 52–57
—— (2011), Response-to-Intervention and Mastery Learning: Tracing Roots and Seeking Common Ground, The Clearing House, 84, 249–255
Karpicke, Jeffrey D. and Roediger III, Henry L. (2007), Repeated retrieval during learning is the key to long-term retention, Journal of Memory and Language, 57, 2, 151–162
Kulik, Chen-Lin C.; Kulik, James A.; and Bangert-Drowns, Robert L. (1990), Effectiveness of Mastery Learning Programs: A Meta-Analysis, Review of Educational Research, 60, 2, 265–299
Mays, John D. (2021), From Wonder to Mastery: A Transformative Model for Science Education, Centripetal Press/Classical Academic Press
Motamedi, Vahid and Sumrall, William J. (2000), Mastery Learning and Contemporary Issues in Education, Action in Teacher Education (Association of Teacher Educators), 22, 1, 32–42
Roediger III, Henry L. and Butler, Andrew C. (2011), The critical role of retrieval practice in long-term retention, Trends in Cognitive Sciences, 15, 1, 20–27
Rowland, C.A. (2014), The effect of testing versus restudy on retention: A meta-analytic review of the testing effect, Psychological Bulletin, 140, 6, 1432–1463
Slavin, Robert. (1987), Mastery Learning Reconsidered, Review of Educational Research, 57, 2, 175–213
—— (1990), Mastery Learning Re-Reconsidered, Review of Educational Research, 60, 2, 300–302
Zimmerman, Barry J. and DeBenedetto, Maria K. (2008), Mastery Learning and Assessment: Implications for Students and Teachers in an Era of High-Stakes Testing, Psychology in the Schools, 45, 3, 206–216
[1] Bloom (1968).
[2] We have Benjamin Bloom to thank for the elimination, at the secondary level at least, of grading according to a normal curve.
[3] Bloom (1968), p. 2.
[4] Davis and Sorrell (1995), p. 4.
[5] Note that while mastery learning adapts instruction to variability in aptitude by accommodating the different rates at which students learn, the spread in aptitude for secondary mathematics is still so wide that multiple tracks or pathways are essential.
[6] Kulik, Kulik, and Bangert-Drowns (1990) comment that “the demonstrably superior performance of LFM [Learning For Mastery] students on locally developed tests is not offset by demonstrably poorer performance on standardized tests. LFM students perform as well as conventionally instructed students on standardized tests, and they perform at a clearly higher level on the locally developed tests used in most LFM evaluations.” p. 290–291.
[7] ibid., p. 288–290.
[8] ibid., p. 291; Zimmerman and DiBenedetto (2008), p. 207–208.
[9] Slavin (1990), p. 301.
[10] Davis and Sorrell (1995), p. 1.
[11] ibid., p. 2.
[12] ibid., p. 3. This is not surprising: low-aptitude students have a wider range of improvement available to them.
[13] ibid., p. 4.
[14] ibid., p. 6.
[15] ibid., p. 6.
[16] Guskey (2005), p. 1.
[17] I have modified Guskey’s diagram to show the possibility of more than one pass through the correctives. This possibility is not mentioned in Guskey (2005) but is included in Zimmerman and DiBenedetto (2008).
[18] Guskey (2010), p. 55.
[19] Guskey (2005), p. 6.
[20] Zimmerman and DiBenedetto (2008), p. 208.
[21] Guskey (2010), p. 56.
[22] Bloom (1968), p. 8.
[23] Guskey (2005), p. 5, 7.
[24] Mays (2021).
[25] The practice effect: “[R]epeated recall of previously recalled items enhanced retention by more than 100% relative to dropping those items from further testing. Repeated retrieval is the key to long-term retention,” Karpicke and Roediger (2007), p. 151. See also Rowland (2014), p. 1432 and Roediger and Butler (2011), p. 20.
[26] These are the major enabling strategies. There are many more minor enabling strategies described in From Wonder to Mastery.
[27] Guskey (2010), p. 56.
John D. Mays
Centripetal Press/Novare Science/Classical Academic Press
Much of this article originally appeared as an appendix in the author’s book From Wonder to Mastery: A Transformative Model for Science Education © Classical Academic Press®, 2021. All rights reserved. Reprinted by permission from Classical Academic Press.