Professional Documents
Culture Documents
Cognitive Computer Tutors
Cognitive Computer Tutors
Albert Corbett
Abstract. Individual human tutoring is the most effective and most expensive
form of instruction. Students working with individual human tutors reach
achievement levels as much as two standard deviations higher than students in
conventional instruction (that is, 50% of tutored students score higher than 98%
of the comparison group). Two early 20th-century innovations attempted to offer
benefits of individualized instruction on a broader basis: (1) mechanized
individualized feedback (via teaching machines and computers) and (2) mastery
learning (individualized pacing of instruction). On average each of these
innovations yields about a half standard deviation achievement effect. More
recently, cognitive computer tutors have implemented these innovations in the
context of a cognitive model of problem solving. This paper examines the
achievement effect size of these two types of student-adapted instruction in a
cognitive programming tutor. Results suggest that cognitive tutors have closed
the gap with and arguably surpass human tutors.
1 Introduction
M. Bauer, P.J. Gmytrasiewicz, and J. Vassileva (Eds.): UM 2001, LNAI 2109, pp. 137–147, 2001.
© Springer-Verlag Berlin Heidelberg 2001
138 Albert Corbett
+1 +2
Bloom [1] reports four dissertation studies comparing human tutoring to conventional
classroom instruction. Learning time was held constant and the average achievement
level in the human tutoring condition typically fell +2 standard deviations (SD) above
the mean of the comparison group. Bloom characterizes the tutors in this study as
“good” tutors, and this effect size should be thought of as an upper bound on what
human tutors can achieve. Cohen, Kulik and Kulik [2] conducted a meta-analysis of
achievement gains in 52 studies of human tutors in elementary and secondary schools
and report a substantially smaller average effect size of +0.40 SD. One study yielded
an effect size of 2.3, in Bloom’s range, but only 16 of the 52 studies yielded an effect
size of +0.50 or higher. These are primarily studies of peer (student-student) tutoring
and paraprofessional tutoring, rather than professional tutors, so the tutors perhaps did
not meet Bloom’s standard of “good”.
The central idea of mastery learning is that virtually all students can achieve expertise
in a domain if two conditions are met: (1) the domain is analyzed into a hierarchy of
component skills and (2) students master prerequisite skills before moving on to
higher level skills. Early interest in mastery learning arose in the 1920s, but it only
achieved prominence in the 1960s. Two forms of mastery learning were developed.
Bloom [9] proposed “learning for mastery” in which a group of students move
through a curriculum together and individual students get extra assistance as needed.
In Keller’s [10] “personalized system of instruction”, students move through the
curriculum at their own pace. A meta-analysis of mastery learning [11] found an
average achievement effect of +0.59 SD for group-based learning for mastery and
+0.48 SD for personalized mastery studies. Mastery learning increases instructional
140 Albert Corbett
time, but only slightly. Median instructional time was 4% longer in mastery learning
than in the conventional conditions.
Like all cognitive tutors, APT is constructed around a cognitive model of the
knowledge students are acquiring. APT uses this knowledge for two purposes. First,
the cognitive model enables the tutor to trace the student's solution path, providing
immediate step-by-step feedback on problem solving actions and advice on steps that
achieve problem solving goals, in a process we call model tracing. Second, the tutor
tracks the student's growing problem solving knowledge across problems in a process
we call knowledge tracing, which in turn is used to implement cognitive mastery
learning. In cognitive mastery learning, the student completes an individualized
sequence of problems in each curriculum section until “mastering” each of the
cognitive problem solving rules introduced in the section.
Table 1. Posttest accuracy (probability correct) in the Non-Tutor and Model Tracing
conditions and effect size of Model Tracing
Table 2. Mean learning time (minutes) for the Non-Tutor and Model Tracing
conditions and the ratio of learning times (Model Tracing/Non-Tutor)
A second study [14] compared students working through the APT Lisp Tutor with
just model tracing to students working with both model tracing and cognitive mastery
(knowledge tracing). Students in the model tracing comparison group completed a
fixed set of programming problems. Students in the cognitive mastery condition
completed this fixed set plus additional problems as individually needed in each
curriculum section to master the cognitive programming rules introduced in the
section
Cognitive Computer Tutors: Solving the Two-Sigma Problem 143
Procedure. Twenty-two college students worked through the early sections in a Lisp
text, completed corresponding programming problems, then completed an on-line
programming test. Ten students were assigned to a “model tracing only” condition
while twelve students were assigned to cognitive mastery condition. Students in both
conditions completed a fixed set of 30 programming problems across the curriculum.
In both conditions APT continuously updated its estimates that the student had
learned the cognitive programming rules needed to solve the problems and displayed
these estimates in a “skillmeter” (depicted in the upper right corner of Fig. 2). In the
model tracing condition, these knowledge tracing estimates did not affect the
curriculum and students completed just the fixed set of 30 problems. In the cognitive
mastery condition, each student completed these 30 problems and completed an
individually tailored sequence of problems until the probability that each of the rules
had been learned reached a mastery criterion of 0.95.
Table 3. Posttest accuracy (probability correct) in the Model Tracing and Cognitive
Mastery conditions and effect size of Cognitive Mastery
Table 4. Mean learning time (minutes) for the Model Tracing and Cognitive Mastery
conditions and the ratio of learning times (Cognitive Mastery/Model Tracing) for
curriculum sections 2 through 5.
The knowledge tracing model used to estimate student knowledge can also be used to
predict student problem solving performance. Studies show that the knowledge
tracing model overestimates students’ test performance by about 10% and does so
because some students are learning suboptimal rules ([15], [16]). As a result, two
model-driven forms of augmented support were developed for the early part of the
APT Lisp Tutor curriculum. One enhancement employs animated feedback to make
key data structure relationships salient. The second employs subgoal scaffolding to
support students in developing simple programming plans. A recent study [17]
compared a cognitive mastery condition in which students work to mastery with this
augmented support to the standard cognitive mastery condition (like that in the
previous study) in which students work to mastery without augmented support.
Procedure. Thirty-nine college students worked through the first five sections of the
APT Lisp Tutor curriculum with both model tracing and knowledge tracing. Twenty-
one students worked to reach cognitive mastery with augmented feedback and plan
scaffolding. The other eighteen students worked to cognitive mastery without the
additional model-based support.
Table 6. Mean learning time (minutes) for the Standard Cognitive Mastery and
AugmentedCognitive Mastery conditions and the ratio of learning times (Augmented
Cognitive Mastery/Standard Cognitive Mastery)
model tracing. Scaffolded cognitive mastery learning yields a +0.42 effect size
compared to standard cognitive mastery. Each innovation adds substantially to the
impact on student achievement. What is the aggregate effect size the three
instructional interventions? We cannot simply add the three effect sizes, because the
unit (standard deviation of the comparison group) decreases across studies. But we
can derive the aggregate effect size, since the comparison group in each successive
study is the same as the experimental group in the previous study. Model tracing
serves as the experimental treatment in experiment 1, then serves as the comparison
group for standard cognitive mastery in experiment 2. Tables 1 and 2 reveal that test
performance in the model tracing condition is almost identical across the studies, both
mean (.67 vs. .68) and standard deviation (.23 vs. .19). Similarly, standard cognitive
mastery serves as the experimental treatment in experiment 2 and as the comparison
group for scaffolded mastery in experiment 3. Again, Tables 2 and 3 show that test
performance in this condition is almost identical across the two studies, both mean
(.85 vs. .83) and standard deviation (.14 vs. .12). As a result, we can reasonably
estimate the aggregate effect size of model tracing, cognitive mastery and scaffolded
cognitive mastery by computing the total difference between the scaffolded cognitive
mastery mean in experiment 3 and the non-tutor mean in experiment 1 and dividing
by the standard deviation of the non-tutor baseline in experiment 1:
5 Discussion
Acknowledgment
This research was supported by NSF grant number 9720359 to CIRCLE: Center for
Interdisciplinary Research in Constructive Learning Environments.
References
1. Bloom, B.S.: The 2 Sigma Problem: The search for methods of group instruction as
effective as one-to-one tutoring. Educational Researcher 13 (1984) 3-15
2. Cohen, P.A., Kulik, J.A., Kulik, C.C.: Educational outcomes of tutoring: A meta-analysis
of findings. American Educational Research Journal 19 (1984) 237-248
3. Pressey, S.L.: A simple apparatus which gives tests and scores – and teaches. School and
Society 23 (1926) 373-376
4. Kulik, J.A.: Meta-analytic studies of findings on computer-based instruction. In E. Baker
& H. O’Neil (Eds.) Technology assessment in education and training. Lawrence Erlbaum,
Mahwah, NJ (1994) 9-33
5. Kulik, J.A., Bangert, R.L., Williams, G.W.: Effects of computer-based teaching on
secondary school students. Journal of Educational Psychology 75 (1983) 19-26
6. Kulik, C.C., Kulik, J.A.: Effectiveness of computer-based instruction: An updated
analysis. Computers in Human Behavior 7 (1991) 75-94
7. Liao, Y.: Effects of computer-assisted instruction on cognitive outcomes: A meta-
analysis. Journ al of Research on Computing in Education 24 (1992) 367-380
8. Niemiec, R., Walberg, H.J.: Comparative effectiveness of computer-assisted instruction:
A synthesis of reviews. Journal of Educational Computing Research 3 (1987) 19-37
9. Bloom, B.S.: Learning for mastery. In Evaluation Comment, 1. UCLA Center for the
Study of Evaluation of Instructional Programs, Los Angeles, CA (1968)
10. Keller, F.S.: “Good-bye teacher…”. Journal of Applied Behavioral Analysis 1 (1968) 79-
89
11. Kulik, C.C, Kulik, J.A., Bangert -Drowns, R.L.: Effectiveness of mastery learning
programs: A meta-analysis. Review of Educational Research 60 (1990) 265-299
12. Anderson, J.R., Corbett, A.T., Koedinger, K.R., Pelletier, R.: Cognitive tutors: Lessons
learned. Journal of the Learning Sciences 4 (1995) 167-207
13. Corbett, A.T., Anderson, J.R.: Locus of feedback control in computer-based tutoring:
Impact on learning rate, achievement and attitudes. Proceedings of ACTM CHI’2001
Conference on Human Factors in Computing Systems (in press)
14. Corbett, A.T., Anderson, J.R.: Knowledge decomposition and subgoal reification in the
ACT Programming Tutor. Artificial Intelligence and Education, 1995: The Proceedings
of AI-ED 95. AACE., Charlottesville, VA (1995) 469-476
15. Corbett, A.T., Knapp, S.: Plan scaffolding: Impact on the process and product of learning.
In C. Frasson, G. Gauthier, & A. Lesgold (Eds.) Intelligent tutoring systems: Third
international conference, ITS ’96. Springer, New York (1996) 120-129
16. Corbett, A.T., Bhatnagar, A.: Student modeling in the ACT Programming Tutor:
Adjusting a procedural learning model with declarative knowledge. User Modeling:
Cognitive Computer Tutors: Solving the Two-Sigma Problem 147
Proceedings of the Sixth International Conference, UM97. Springer, New York, (1997)
243-254
17. Corbett, A.T., Trask, H.: Instructional interventions in computer-based tutoring:
Differential impact on learning time and accuracy. Proceedings of ACTM CHI’2000
Conference on Human Factors in Computing Systems. Springer, New York (2000) 97-104
18. Anderson, J.R., Gluck, K.: What role do cognitive architectures play in intelligent tutoring
systems. In D. Klahr & S. Carver (Eds.) Cognition and instruction: 25 years of progress.
Lawrence Erlbaum, Mahwah, NJ (in press)
19. Aleven, V., Koedinger, K. R.: Toward a tutorial dialog system that helps students to
explain solution steps. Building Dialogue Systems for Tutorial Applications: AAAI Fall
Symposium 2000, (2000)