Professional Documents
Culture Documents
1 Kulgemeyer 2018
1 Kulgemeyer 2018
RESEARCH ARTICLE
1
Physics Education Department, Institute
Abstract
of Science Education, University of
Bremen, Bremen, Germany In recent years, many studies have researched the impact of
2
Physics Education Department, Institute teachers’ professional knowledge on teaching quality. Those
of Physics IA, RWTH Aachen University, findings are still ambiguous providing unclear evidence for
Aachen, Germany supporting teacher education programs that aim at develop-
Correspondence ing professional knowledge. In this study, we followed a
Christoph Kulgemeyer, Physics new approach and used a “performance test” to take a closer
Education Department, Institute of look at the impact of professional knowledge on teaching
Science Education, University of quality. We simulated one particular teaching situation in a
Bremen, Otto-Hahn-Allee 1, Bremen
controlled, standardized setting in which student teachers
28359, Germany.
Email: Kulgemeyer@physik.uni-bremen.de
(N 5 109) enrolled in physics teacher training courses at five
German universities had to explain given phenomena to
Funding information high-school students. These high-school students were
Bundesministerium f€ur Bildung und For- trained to behave in a standardized way and to ask standar-
schung, Grant/Award Number: dized questions. Videos of these situations were analyzed
01PK11001B
using an established model of explaining physics. The valid-
ity, reliability, and objectivity of these performance tests for
explaining physics were examined in previous studies. In
this article, we report on the analysis and interpretation of
the results of our study concerning the impact of physics
content knowledge (CK), pedagogical content knowledge
(PCK), and two groups of beliefs (specific aspects of (i)
self-efficacy and (ii) teaching and learning) that mediate the
effect of CK and PCK on student teachers’ explaining per-
formance. Using path analysis, we can show that student
teachers’ PCK mediates the influence of their CK on
explaining performance in that CK only has a positive influ-
ence if PCK has also increased as well. Our findings stress
the key role of PCK. For one particular teaching situation,
we can show the positive influence of student teachers’ CK
and their PCK they acquired in academic teacher education
on their teaching quality.
KEYWORDS
explanation, instructional quality, pedagogical content knowledge, physics
education, teaching quality
1 | THEORETICAL BACKGROUND
2010; Cauet, Liepertz, Kirschner, Borowski, & Fischer, 2015; Hill, Rowan, & Ball, 2005). In these
studies, the three areas are sometimes described as including the seven categories. Our views accord
with those of these studies and we focus on the unclear evidence about the relationship between these
three areas and teaching quality. In this sense, CK is the knowledge of the content selected for teach-
ing, PCK is the domain-specific knowledge of how to teach this content and PK is the knowledge of
how to act in teaching situations in general. The focus on the three areas is a limitation of our study
and we cannot make claims about research on professional knowledge in general but just about these
aspects.
correlation between PCK and learning outcome (r 5 0.270, p < 0.05) and also stated that Finnish stu-
dents had the greatest achievement of knowledge while their teachers had the lowest PCK compared to
Switzerland and Germany.
All of these tests for professional knowledge were developed with a focus on curricular validity;
and therefore, the test results were a good representation of the content of the science-specific aspects
of teacher education (PCK and CK). Two assumptions emerged—either the content of teacher educa-
tion is not relevant for both teaching quality and students’ learning (as the facets of the content respec-
tively covered by the test instruments) or the measurement using written tests is not sufficiently valid.
Aufschnaiter and Bl€omeke (2010), for example, argued that written tests, especially those with multiple
choice items are not appropriate for measuring relevant knowledge, namely procedural knowledge but
only for measuring declarative knowledge focused on facts.
In our study, we wanted to contribute to researching the unclear relations between CK, PCK, and
teaching quality. We did not focus on student outcomes.
1.4 | Beliefs about self-efficacy, beliefs about teaching and learning, and their
impact on teaching quality
Prior research suggests that not only knowledge but also beliefs might determine teaching quality (e.g.,
Fang, 1996; Stipek, Givvin, Salmon, & MacGyvers, 2001), most importantly self-efficacy beliefs (e.g.,
Retelsdorf, Butler, Streblow, & Schiefele, 2010). Ross (1998), for instance, reported in a review that
teachers with high self-efficacy use more activating instructional methods and their behaviors are more
adaptive and supportive. Being adaptive and supportive is crucial for teaching quality in general and
for explaining situations in particular. In fact, adaptation is the most important strategy for effective
explanation (Wittwer & Renkl, 2008). In this regard, self-efficacy is important for the present study.
However, self-efficacy has been measured with very different instruments and notions; sometimes,
researchers followed the original distinction of Bandura (1977) between self-efficacy and outcome
expectancy (e.g., Enochs & Riggs, 1990) and sometimes they assumed that both cannot be distin-
guished clearly (e.g., Williams, 2010). In our study, we focused on self-efficacy, not outcome expect-
ancy because our focus was on behaviors leading to teaching quality, not beliefs on how likely it is
that these behaviors would result in student outcomes. In this study, we followed the definition of
Schiefele and Schaffner (2015, p. 161): “Teacher self-efficacy represents teachers’ belief that they are
able to perform those teaching behaviors that bring about student learning even when students are diffi-
cult or unmotivated.” We only focused on a small aspect of self-efficacy that is important for explain-
ing situations. Tschannen-Moran and Woolfolk Hoy (2001) differentiated between instructional
strategies, classroom management, and student engagement as essential dimensions of self-efficacy.
Concerning the present study on explaining situations, self-efficacy regarding instructional strategies
might be the most important part, which was our focus.
Besides self-efficacy, beliefs about teaching and learning are also likely to influence teaching qual-
ity (Staub & Stern, 2002). Beliefs about teaching and learning have been described as including very
different dimensions, for example, about the nature of knowledge, the way learning works, the role of
teachers (and students), effective instructional methods, and learning approaches (Chan & Elliott,
2004). Concerning their impact on teaching quality, one aspect seems to be of a distinguished impor-
tance: teachers’ beliefs toward learning facilitation and knowledge construction. Teachers with a
student-centered view are more likely to teach effectively than are teachers who believe that teaching
means only a transmission of information (Prosser & Trigwell, 1999; Rienties, Lygo-Baker, &
Brouwer, 2013). In particular, teachers with a constructivist view (instead of a transmissive view) on
teaching and learning have a higher teaching quality regarding cognitive activation (Dubberke, Kunter,
KULGEMEYER AND RIESE | 5
McElvany, Brunner, & Baumert, 2008). As we will discuss below, the difference between a constructi-
vist view and a transmissive one is of high importance for effective teacher explanations. Teacher
explanations, in general, are often confused with situations that build up on a simple transmission of
knowledge. Our research, therefore, focused on this aspect of beliefs on teaching and learning.
greatly between one lesson and the other, and that might have a massive influence (e.g., whether it is
the last lesson of the day or the first, or whether there was an examination administrated in the previous
lesson). It is very likely that there are many other factors. For a rather low number of lessons video-
taped in a study, these influencing factors are crucial for the outcome of the data analysis. If the occur-
rence of these factors is assumed to be random, it is likely that for a large number of lessons
videotaped in the study, these factors may cancel each other out and the “real effect” of teachers’ pro-
fessional knowledge can be revealed. Videotaping a large number of lessons and then analyzing them
appropriately, however, is nearly impossible even for very large research projects.
Another reason why it is hard to investigate a relationship between teachers’ professional knowl-
edge and teaching performance is that teachers have to make decisions to solve various problems in a
very short time during instruction (Weinert, 1996). Beginning teachers report that they perceive their
teaching as “action under pressure” (Wahl, 1991). Depending on variables like their experience, skills,
or personality this pressure leads to decisions that are not entirely based on their professional knowl-
edge. It sometimes leads to instruction that simulates the instruction as they experienced it as a student
(Brouwer, 2010). A common written test for professional knowledge is less useful for testing knowl-
edge required for authentic instruction compared to a test that demands decision-making under pressure
as in real instruction. Bl€omeke et al. (2014) proposed to use video-based testing (so-called “video-
vignettes”) to simulate an authentic teaching situation but still admitted that testing with video-
vignettes is much closer to testing cognition than to testing performance.
The literature on measuring knowledge and performance suggested one point for testing purposes
we regard as essential for our study: It is important to look for tests for professional performance
beyond written tests and videotaped lessons. We argued that the so-called “performance tests” (as out-
lined in the following) are a good alternative and maybe even a better choice when it comes to
validity.
When applied to teacher education, this would mean that a test for knowledge is a paper-and-pencil
test for declarative or procedural knowledge (which would include dealing with authentic classroom
situations or so-called vignettes [cf. Aufschnaiter & Bl€omeke, 2010]). A test for action would mean an
observation of teachers’ action in the classroom, which is basically the aim of the video-studies
described above. These ways of testing are common in teacher education. Uncommon is a test for per-
formance, which means an observation of teachers’ performance in simulated, standardized teaching
situations. That could mean that students have to be trained to behave in a specific manner and to chal-
lenge each test teacher with the same realistic problems. Those performance tests are much closer to
real teaching but have the advantages of standardized testing. All the contextual factors described
above that make it hard to research teachers’ skills can be controlled in such a setting. Indeed, the inde-
pendent action in the classroom still is something different. For example, the independent action inte-
grates a diversity of different situations, whereas performance tests can only simulate a particular
teaching situation. In a way, performance tests are a step closer regarding authenticity, but they are not
authentic.
The differentiation between tests for knowledge and performance is important for teacher educa-
tion. Terhart (2012) described a simplified “functional chain” of teacher education. Teacher education
(e.g., at a university) improves teachers’ knowledge (both declarative and procedural). Knowledge is
assumed to allow them to take more appropriate actions in teaching which in turn results in better stu-
dents’ learning. All these assumed causal relationships are not corroborated yet, as described above.
The study reported in this article aimed to contribute to the first two “links” of this chain. We used
knowledge tests that were developed with a focus on curricular validity to ensure that teacher education
potentially affected the development of the knowledge. Many test instruments have been developed in
this way, and it is well known that CK and PCK develop during science teacher education (e.g., Riese
& Reinhold, 2012). Our study adds a new perspective to these studies. We researched whether or not
teachers use this knowledge successfully for enacting specific elements of high-quality instruction. Pro-
fessional knowledge was measured by using paper-and-pencil tests, which were developed with a focus
on curricular validity. The teaching quality was measured by using standardized performance tests. We
decided to choose explaining situations for these performance tests as a start. For science teaching, we
can think of many standard situations, which someone working in the profession of a science teacher
has to face frequently. Explaining science is certainly one of such teaching situations.
fulfil the terms of logic. This “covering law model” (Hempel & Oppenheim, 1948) is similar to
Toulmin’s (1958) argumentation patterns. However, explaining in the classroom is more than just pre-
senting a logical connection between a phenomenon and an underlying principle. Treagust and Harri-
son (1999) therefore distinguished between scientific explanations on the one hand (a scientific
explanation could follow the covering law model, even though there are alternatives, e.g., Kitcher,
1981) and science teaching explanations (that means an explanation of a scientific phenomenon) on
the other. Explaining in the classroom needs to meet the needs of the students (e.g., their prior knowl-
edge) and sometimes has a reversed structure when compared to scientific explanations. A general law,
for example, Ohm’s law, might be explained by giving different examples in which the law shows that
it can be applied to different phenomena (Kulgemeyer & Tomczyszyn, 2015). Kulgemeyer and
Schecker (2013) stressed that even a well-reasoned claim—that means a good argumentative structure
—might lead to acceptance or conviction, but not necessarily to understanding. Understanding (e.g.,
by achieving a conceptual change) is the main goal of explaining in the classroom (Gage, 1968).
Concerning the present study, we focused on science teaching explanations), that means explana-
tions of a scientific idea to a particular addressee. They are sometimes also named instructional explan-
ations (e.g., Wittwer & Renkl, 2008). We wanted to focus on teaching quality in explaining situations.
Thus, that aspect of explaining seems to be more relevant.
as a possible prerequisite for student’s understanding—or as Brown (2006) stated: “an attempt to pro-
vide understanding of a problem to others” (p. 196). High-quality explaining helps students to con-
struct their own knowledge. The model has the characteristic of a probabilistic approach to
communication—the actions of the explainer aim to increase the likelihood that an addressee can con-
struct meaning from the explained information. The addressee plays an active role during this process.
Once he/she decides to take part in the explaining process, he/she will show verbally or non-verbally
signs of interest, comprehension, or misunderstanding. An explainer has to perceive and interpret those
signs and decide how to vary the explaining in order to make it more comprehensible. In particular, in
a physics-related explaining situation, there are some physics-specific means that can be varied for this
purpose. These means are: (i) the level of mathematization, (ii) the representational forms (e.g., dia-
grams or realistic pictures), (iii) the contexts and examples (e.g., everyday contexts or physics textbook
examples), and (iv) the code of the verbal language used (e.g., informal language, technical language,
or language of education). For each statement, an explainer has to decide on how to make appropriate
use of these variables. It is at the core of explaining skills to decide on how to choose these four varia-
bles to be both addressee-oriented and subject-adequate and to adapt them according to the addressee’s
needs in order to increase the comprehensibility.
With respect to the three characteristics of teaching quality (emotional support, classroom organiza-
tion, instructional support), adapting the four variables means providing emotional support (as the
adaptation is guided by students’ perspectives) and instructional support (as the adaptation regards con-
cept development and the quality of feedback). Classroom organization is not a part of the model as it
focuses on dialogic explaining.
2 | RESEARCH QUESTIONS
Prior studies could not corroborate the claim that teaching quality benefits from professional knowl-
edge. As this is a core claim in teacher education, our research question addresses this desideratum
concerning one particular situation of teaching: explaining physics. Our research question in this study
is: How do CK and PCK highlighted in academic physics teacher education impact explaining per-
formance? Our study, therefore, has a strong limitation to the quality of teaching in dialogic explaining
situations, the explaining performance quality.
10 | KULGEMEYER AND RIESE
We expected both CK and PCK to be important for explaining performance. However, PCK might
be a mediator for CK (as in the COACTIV study mentioned above). Without PCK, CK might not be
useful for explaining. We tested this hypothesis, but also the possibility that CK might affect explain-
ing performance directly. We decided to focus on the impact of CK and PCK and to leave out PK.
Testing time is limited, and these two are more likely to be of use in explaining situations than PK.
We also tested two groups of important beliefs that are likely to mediate the effects of CK and
PCK on explaining performance quality: (i) beliefs about self-efficacy regarding the instructional strat-
egy of explaining and (ii) beliefs on teaching and learning regarding a constructivist or transmissive
view on explaining. We assumed according to prior research (described above) that someone with a
high self-efficacy explains more effectively, for example, because he or she has enough self-
confidence to use various examples and to explain more independently from the exact knowledge
found in the textbook. We also assumed according to prior research (as described above) that someone
who considers explaining from a constructivist point of view explains more effectively than someone
who regards explaining as simple transmission of knowledge. The two groups of beliefs are used as
control variables for the effect of CK and PCK on explaining performance quality. Testing time is lim-
ited here as well, so we operationalized the two groups of beliefs close to explaining situations and by
using just two very small aspects of them. That was a limitation of our study.
3 | METHODS
3.1 | Design
We administered the paper-and-pencil tests about CK, PCK, the two beliefs and demographics to the
participating student teachers. The tests had been split into two separate test booklets, each of which
was a 90-min test. After having tested the student teachers with the paper-and-pencil tests, we tested
them with the performance test on explaining physics (20 min). Various test administrators followed a
clear documentation about the process to ensure objectivity in conducting the paper-and-pencil tests.
Conducting the performance test was more challenging. Thus, the same test administrator conducted
all of the tests at all of the universities over a period of 3 years. Three different high-school students
who had been trained thoroughly participated in performance tests (as described below). Pilot studies
were conducted to ensure that all of these students acted similarly in all the performance tests. All in
all, the testing time for each participating student was 200 min.
3.2 | Sample
German teacher education starts from the first semester on with a specific track for future physics
teachers. To become a physics teacher in Germany a master’s degree (10 semesters) is mandatory. We
wanted to increase the likelihood to reach students with a broad range of CK, PCK, and explaining per-
formance quality. Thus, our sample consisted of student teachers enrolled in a physics teacher training
program from all stages of their academic teacher education in German universities. The median of the
number of semesters for which they attended teacher training courses at university was 6, and the range
was from 1 to 20. One hundred ninety-eight of the student teachers participated in the performance
test, but not all of them were present when the other tests were conducted. We collected the data from
109 of these 198 student teachers for all three of the tests for explaining performance quality, PCK,
and CK. Those 109 student teachers came from five different German universities. The final sample
consisted of 68 male and 41 female student teachers. Their age ranged from 17 to 40 years with a
median of 23 years.
KULGEMEYER AND RIESE | 11
3.3 | Instruments
3.3.1 | Demographics
We collected data about the age, semesters in a physics teacher training program, Credit Points in a
physics teacher training program, high-school exam grade, and weeks of internships in schools.
physics? My teacher always wants me to use the appropriate terms.”). Some simply attempt to use a tool
(“Could you sketch that for me?”). At all of the universities, the same students were used as trained novi-
ces. Three high-school students participated as explainees. Even though only one of them was part of
each setting in a pilot study, it turned out that we needed to change the explainees after each five to seven
DEAs because they got tired. Also, in a pilot study, we ensured that all of them behave similarly and the
results of the DEAs were as independent of the explainee as possible. Ten test persons explaining the
same topic to two different explainees had comparable results (Kulgemeyer & Tomczyszyn, 2015).
Dialogic situations frequently happen in science teaching, for example, when students are working
in groups or on their own, and the teacher has time to help. Still, we cannot be sure that explaining a
phenomenon to a group of students needs the same skills as classroom explaining. We focus on dia-
logic situations because this setting can also be realized in large-scale studies.
We analyzed the videos using an established instrument and an established process for researching
explaining performance quality. Both the instrument and the process had been researched thoroughly regard-
ing validity, reliability, and objectivity in a previous study (Kulgemeyer & Tomczyszyn, 2015). Kulgemeyer
and Tomczyszyn (2015) used the “model of explaining physics” described above (cf. Theoretical back-
ground) as the starting point. Using this model, initial categories have been formed; further categories have
been added following the process of qualitative content analysis (Mayring, 2000). The final instrument con-
sists of 12 categories for appropriate explaining and for non-appropriate explaining (categories are given in
Table 1). Kulgemeyer and Tomczyszyn (2015) describe the process of analysis as follows:
Validity
The PI predicted experts’ decision for the better explaining quality when a pair of videos was com-
pared (Cohen’s j 5 0.78). The PI correlated with the rating of the trained students on the explaining
quality (Pearson’s r 5 0.53, p < 0.001); they were asked to rate the explaining quality on a Likert Scale
after each of the 109 DEAs. Both the expert rating and the student rating are arguments for concurrent
validity, even though the ratings of the students are probably confounded by factors such as sympathy.
An argument for content validity is that the categories represent the four “variables” of the model for
explaining physics (e.g., “contexts and examples”). For construct validity, a nomological network with
both aspects of convergent and discriminant validity was analyzed. Interview studies were conducted
to ensure that the test teachers perceived the test situations as authentic.
Reliability
The PI reached a good Cronbach’s a 5 0.772. A one-dimensional Rasch-model could be used to
model the scale applying the usual criteria (0.8 < Infit MNSQ < 1.2, T < 2.0).
Objectivity
Objectivity was examined by measuring inter-rater reliability. In the first step for all categories, two
raters reached an accordance ranging between 73% and 97% depending on the category. In the second
step, a consensus between the two raters could be reached for all categories and videos.
KULGEMEYER AND RIESE | 13
Category Description
Explaining physics concepts Explainer avoids technical terms by describing the underlying concept
in everyday language with everyday terms.
“push” instead of “momentum”; “. . . and the water lifts it a bit” instead
of “buoyancy”
Connecting non-verbal elements Explainer connects non-verbal elements like diagrams, pictures or
demonstrations by highlighting similarities and differences.
Using items in general Explainer uses small everyday items (e.g., a paper snarl) to illustrate a
process.
Connecting items with the Explainer not only uses small everyday items but connects them to the
topic by showing analogy topic he/she wants to explain (“The paper snarl stands for the asteroid.
Look at it when I am moving it”)
Small demonstrations The explainer conducts small demonstrations with everyday items (e.g.,
“In which direction does the paper snarl move when I push it?”)
Answering inadequately (scientifically Explainer does not answer an addressee’s question or ignores the
wrong answers do not belong question.
into this category) A: “Weightlessness, is that the same as in outer space?” - E: “I have
never been in outer space, how should I know?”
Review The explainer stresses that something has already been explained and is
needed now (“You remember that we were talking about friction
before, that is exactly what happens here.”)
Encouragement The explainer praises the explainee for good answers and encourages to
deal with difficult parts of the explanation.
Diagnosing understanding The explainer diagnoses the success of the explanation by asking
questions or giving tasks (NOT just: “Did you understand that?”)
Request action from explainee The explainer requests the explainee to act. (“What do you think how it
moves? Could you sketch that for me?”)
FIGURE 2 Sample item to measure PCK (subscale: students’ misconceptions and how to deal with them)
Content knowledge
The model for physics teachers’ CK comprises three subscales: (i) physics knowledge from school
textbooks, (ii) deeper understanding of physics knowledge from a school textbook (e.g., knowledge of
different procedures of solution, knowledge of boundary conditions, or knowledge of coherences of
various physical phenomena), and (iii) physics knowledge from a university textbook. To create items
with different requirements and difficulties, a further dimension called complexity was added; that is
made up of: (i) facts, (ii) links between facts, and (iii) dealing with advanced physical concepts.
Based on this CK-model, a paper-and-pencil test with 143 multiple-choice items (single select) was
developed, from which 40 items (for the test of 60 min) were chosen for the main study (with regard
to statistical fit, validity, and a balanced distribution of item difficulties). For content validity, curricular
analyses—of the courses in all participating universities on the one hand and different textbooks
(school textbooks and university textbooks) on the other—were conducted. Furthermore, the matching
between the test items and the underlying model was checked by several physics student teachers. For
construct validity, a one-dimensional Rasch-model (with just one global CK scale: variance was 0.90;
EAP-reliability .83; 0.8 < MNSQ < 1.2; 21.9 < T < 1.9) was compared with a three-dimensional
Rasch-model (using those three subscales mentioned above: variance was between 0.98 and 1.43;
EAP-reliability between 0.78 and 0.83; 0.8 < MNSQ < 1.2; 21.9 < T < 1.9). Thereby, significantly
higher and better matching of the three-dimensional model could be observed (v2 test: p < 0.001). So,
it seems justified to accept the assumption of three independent subscales for measuring CK. Finally, a
nomological network with both aspects of convergent (e.g., physics grades) and discriminant validity
(e.g., mathematical skills) was examined with correlational analysis (cf. Riese et al., 2015).
CK 5 content knowledge; PCK 5 pedagogical content knowledge; SE 5 self-efficacy; BT 5 belief on teaching and learning
“explaining means transmission”; EXP 5 explaining performance; M 5 means (in percent); SD 5 standard deviation (in percent).
4 | FINDINGS
CK 5 content knowledge; PCK 5 pedagogical content knowledge; SE 5 self-efficacy; BT 5 belief on teaching and learning
“explaining means transmission”; EXP 5 explaining performance.
*p < 0.05; **p < 0.01; ***p < 0.001.
case, is reflected by the three measures “semesters in a physics teacher training program,” “Credit
Points in a physics teacher training program,” and “weeks of internships in schools.” For the explain-
ing performance test, we could only identify correlations to the “Credit Points in a physics teacher
training program.” Interestingly, the high-school exam grade seems to be a fair predictor of the per-
formance in the tests as it correlates with all measures except the measured aspect of self-efficacy.
In order to have a closer look at the explaining performance test, we use Table 4 to present data on
how many of the student teachers used each category for their explaining. For example, 78 out of the
109 student teachers paraphrased a technical term in everyday language at least once. Analysis of the
numbers showed that most of the student teachers (N 5 89) were able to avoid examples that are non-
fitting from a scientific point of view but just a few (N 5 12) attempted to request actions from their
explainee. In general, the interaction with the explainee was the most difficult main category.
T A BL E 4 Number of student teachers who used a particular category during their explaining attempts
Small demonstrations 19 90
Answering inadequately 89 20
Review 42 67
Summary 33 76
Encouragement 18 91
Diagnosing understanding 27 82
CK PCK BT SC EXP
CK – – – – –
PCK 0.564*** – – – –
BT 20.266* 20.237* – – –
CK 5 content knowledge; PCK 5 pedagogical content knowledge; SE 5 self-efficacy; BT 5 belief on teaching and learning
“explaining means transmission”; EXP 5 explaining performance.
*p < 0.05; **p < 0.01; ***p < 0.001.
T A BL E 6 Manifest correlations between explaining performance (EXP) (Pearson’s r), respectively with two
particular subscales of PCK and CK
PCK CK
StuCon Concepts SchKno DeepSch PhyUn
the paths from CK and PCK were not significant. That implies that the model in Figure 3 more accu-
rately described the empirical data in this study.
We want to highlight in Figure 3 that PCK had a positive impact on explaining performance, and
so did the self-efficacy “physics in explaining situations.” The belief “explaining means transmission,”
however, had a negative impact on explaining performance. PCK decreased this belief. CK had a small
impact on the self-efficacy in teaching situations but was not found to be significant. The model
explained 29% of the variance, which is large based on Cohen’s (1988) suggestions to interpret R2.
An important question for teacher education is what parts exactly of CK and PCK are important
for high-quality performance and in what situations of teaching. These parts could be highlighted in
teacher education. For explaining physics, our study provides first insights because we can differentiate
empirically between subscales of PCK and CK as described above in the Methods section. In Table 6
we want to point out four correlations between the empirically verified subscales of PCK and CK and
explaining performance in our study. Table 6 should be read as first results and we want to remind that
a correlation does not necessarily mean a causal relationship. We discuss these results in the following
section.
5 | DISCUSSION
Regarding the descriptive results, we interpret the results as follows: the correlations between both the
CK and the PCK test results and the three aspects of the progress in the studies reflected the focus on
curricular validity during the test development. Still, there was a limitation because of the focus on
mechanics. There was a medium correlation between the Credit Points achieved in a physics teacher
training program and the explaining performance (r 5 0.482, p < 0.01) but not between explaining
performance and the semesters in a physics teacher training program. The reason could be that the
Credit Points are a better measure to reflect the actual study progress: the number of semesters would
also increase for “passive” students, whereas Credit Points could only be achieved after completing a
course successfully. Interestingly, the practical experience (“internships in schools”) did not correlate
with the explaining performance as well. One might have expected that because practical experience
perhaps without reflection is unlikely to result in better results. However, our results are limited: it was
unclear what the student teachers really did during their internships or how much teaching experience
they really had. There was a small negative, but significant correlation between the measured aspect of
self-efficacy and the weeks of internships in schools (r 5 20.174, p < 0.05). That could mean that
practical experience or a more elaborate picture of the profession of a science teacher helps student
teachers to become aware of how difficult it is to be a good explainer. This hypothesis would be sup-
ported by the medium correlations to the semesters (r 5 20.340, p < 0.01) and Credit Points
20 | KULGEMEYER AND RIESE
(r 5 20.385, p < 0.01) in a physics teacher training program. However, we can just present the corre-
lations and cannot claim to identify causal relationships.
Analyzing the difficulty of the single categories of the explaining test, we have to highlight that the
explainer’s interaction with the explainee was the category where the least student teachers collected
points. For example, just 12 of them requested actions from their explainee in a way they had to partic-
ipate in the explaining process actively (e.g., “Could you draw the graph in the diagram?”). That shows
how difficult it is to understand an explanation as more than just a presentation.
Concerning our research question, we interpret the findings of our study as follows: Both CK and
PCK were important in improving the quality of explaining. PCK, however, mediated the path of CK to
explaining performance, and therefore, played the key role in transferring CK to explaining performance.
We could not find a direct path from CK to explaining performance. Beliefs were important in transfer-
ring knowledge to explaining performance. In this particular case, first, the belief about teaching and
learning “explaining means transmission” decreased explaining performance but was itself decreased by
PCK. That means that an appropriate constructivist view on explaining increased the explaining perform-
ance and was itself increased by PCK. Second, a high self-efficacy “physics in teaching situations”
increased explaining performance. Of course, our results are limited to the particular aspects of the
beliefs we measured: beliefs on self-efficacy regarding the instructional strategy of explaining and beliefs
on teaching and learning regarding a constructivist or transmissive view on explaining.
This result is of interest not only for the researchers on teacher explanations but also for researchers
on teaching quality in general. For one particular situation of instruction, our study showed that CK
and PCK highlighted in teacher training programs at universities was useful for high-quality perform-
ance. After all, both the test instruments for CK and PCK had been developed in previous studies
stressing above all curricular validity—they focus on the content of German university courses.
Besides, we could show for explaining situations that PCK, in particular, plays a key role. This result
provides a good argument for teacher training programs that not only focus on the development of CK
but also highlight PCK. On the other hand, beliefs seemed to play an essential role for explaining per-
formance as well. There was an additional but indirect path from PCK to explaining performance via
the beliefs on teaching and learning. Also, there was a direct path from the measured aspect of self-
efficacy to explaining performance. Analyzing the curricula of German teacher education, we came
across an interesting observation. The curricula highlighted cognitive variables such as knowledge, and
therefore, focused on PCK and CK, as well as pedagogical knowledge. Probably, most examinations
focus on cognitive variables as well—which very much underestimates the worth of beliefs on actual
performance. Knowledge might affect beliefs implicitly, as PCK affects the epistemological beliefs in
our path model. However, teachers’ knowledge is certainly not the only trait to develop during teacher
education because, for example, their appropriate attitudes towards teaching are also important. In our
case, that could mean that someone might have a lot of knowledge about students’ misconceptions and
how to diagnose them (an important part of our PCK test) but he or she still acts as if explaining meant
transmission. We, as teacher educators, should value beliefs in teacher education and work on appro-
priate ways to test them. Of course, many teacher training programs already do so.
We want to highlight the correlation between explaining performance and the knowledge about stu-
dents’ misconceptions and how to deal with them (r 5 0.34, p < 0.01) as a particular aspect of PCK. It
was the subscale of PCK having the largest significant correlation with explaining performance. Sadler,
Sonnert, Coyle, Cook-Smith, and Miller (2013) found that teachers, who were able to identify miscon-
ceptions in items of a physics assessment instrument that has been administered to their students, had
larger classroom gains than teachers, who only knew the correct answer. Our study showed similar
results. First of all, the teaching quality in explaining situations correlated with the teachers’ knowledge
about students’ misconceptions and how to deal with them. We only focused on explaining situations
KULGEMEYER AND RIESE | 21
but a higher teaching quality, in general, might explain larger classroom gains. Second, in our study,
PCK mediated the path of CK to the teaching quality in explaining situations. That might explain why
in the study of Sadler et al. (2013) teachers, who were able to identify misconceptions (a part of PCK),
had larger classroom gains compared to those teachers who only had the CK to solve the items cor-
rectly on their own. Future studies should further investigate this result. It also yields a second question
for further studies: Can PCK courses better prepare student teachers at universities for explaining situa-
tions by focusing more on the knowledge of student’s misconceptions and how to deal with them (e.g.,
conceptual change)?
Still significant was the correlation between explaining performance and the knowledge on PCK-
related theoretical concepts (e.g., the model of conceptual change). The two subscales integrate the
knowledge about how to diagnose prior knowledge, how to identify common misconceptions in stu-
dents’ utterances, and how to deal with them. It is likely that this kind of knowledge helps in explain-
ing physics.
Both the CK-subscales “Physics knowledge from school textbooks” (r 5 0.32, p < 0.01) and
“Deeper understanding of physics knowledge from a school textbook” (r 5 0.34, p < 0.01) correlated
significantly with the explaining performance. There was no such correlation between explaining per-
formance and the more abstract physics knowledge from university textbooks (r 5 0.17, p > 0.05).
That means the more “abstract” physics knowledge from a university textbook showed no correlation
with explaining performance, whereas both the subscales “Physics knowledge from school textbooks”
and “Deeper understanding of physics knowledge from a school textbook” correlated with explaining
performance having a medium effect size. Of course, the explained topics are from school physics and
maybe the knowledge of university physics was simply not needed. Also, both explaining and prepara-
tion were limited to 10 min—maybe the student teachers just wanted to focus on the more basic
aspects. On the other hand, the more abstract knowledge of university physics might be regarded as
including the more applied school-related knowledge. We believe that it might be possible that some
student teachers are not capable of transferring the more abstract physics knowledge to explaining sit-
uations, whereas the more applied knowledge that relates directly to school textbooks can be used. It
might be the case that it is not an automatic process that someone very capable of solving problems
from university physics is also able to use this knowledge to explain physics on a school level. The
use of “physics knowledge from university textbooks” for teaching physics at a school level might be
a topic for further studies.
We want to highlight the most important limitation of our study—our study was a correlative study
but not a study with an experimental design; and therefore, we cannot interpret these relations simply
as causal relationships. Further studies are needed. Our sample size was rather small after all and just
provides results we can treat as empirically supported hypotheses. We regard our study merely as a
starting point for research that focuses on performance tests. However, performance tests are limited to
one single situation (in our case explaining) and, therefore, their results do not give holistic information
on teaching quality. Regarding our tests, there may be biases: maybe some pairs of explainer and
explainee just worked because they liked one another and that the explainer’s skills were overestimated
due to that. Maybe the measurement was confounded by sympathy. Still, the setting was much more
standardized than videotaping whole lessons (see the section “difficulties in testing teachers’ knowl-
edge and performance” above). Further studies should develop performance tests for other “standard
situations” of science education, such as conducting an experiment as a demonstration. The relation-
ship between PCK, CK, and the performance quality might substantially differ, especially on a sub-
scale level. A goal for science education research should be to find out more about what kind of
knowledge and beliefs are actually useful for improving teaching quality. This could provide very use-
ful evidence on how to develop better teacher education. Maybe the most important outcome of our
22 | KULGEMEYER AND RIESE
study is that we could show the high potential of performance test instruments for an empirical orienta-
tion of teacher education programs at universities. Performance tests might help to find a way to a
more evidence-based teacher education.
O R CI D
Christoph Kulgemeyer http://orcid.org/0000-0001-6659-8170
R EFE RE NC ES
Abell, S. (2007). Research on science teachers’ knowledge. In S. Abell & N. Lederman (Eds.), Handbook of
research on science education (pp. 1105–1149). Mahwah, NJ: Lawrence Erlbaum Associates.
Abrahams, I., & Millar, R. (2008). Does practical work really work? A study of the effectiveness of practical work as
ateaching and learning methodin school science. International Journal of Science Education, 30(14), 1945–1969.
Aikenhead, G. S. (1988). An analysis of four ways of assessing student beliefs about STS topics. Journal of
Research in Science Teaching, 25, 607–629.
Anderson, J. (1976). Language, memory, and thought. Hillsdale, NJ: Lawrence Erlbaum Associates.
Aufschnaiter, C. V., & Bl€omeke, S. (2010). Professionelle Kompetenz von (angehenden) Lehrkräften erfassen –
Desiderata [Measuring teachers’ professional competence - desiderata]. Zeitschrift f€ur Didaktik der Naturwissen-
schaften, 16, 361–367.
Bandura, A. (1977). Self-efficacy: Toward a unifying theory of behavioral change. Psychological Review, 84, 191–215.
Barrows, H., & Abrahamson, S. (1964). The programmed patient: A technique for appraising student performance
in clinical neurology. Journal of Medical Education, 39, 802–805.
Baumert, J., & Kunter, M. (2006). Stichwort: Professionelle Kompetenz von Lehrkräften [Keyword: teachers’ pro-
fessional competence]. Zeitschrift f€ur Erziehungswissenschaft, 9, 469–520.
Baumert, J., Kunter, M., Blum, W., Brunner, M., Voss, T., Jordan, A., . . . Tsai, Y. M. (2010). Teachers’ mathemati-
cal knowledge, cognitive activation in the classroom, and student progress. American Educational Research
Journal, 47(1), 133–180.
Berger, R., & Hänze, M. (2015). Impact of expert teaching quality on novice academic performance in the jigsaw
cooperative learning method. International Journal of Science Education, 37(2), 294–320.
Berland, L. K., & McNeill, K. L. (2012). For whom is argument and explanation a necessary distinction? A
response to Osborne and Patterson. Science Education, 96(5), 808–813.
Bl€omeke, S., Gustafsson, J., & Shavelsson, R. (2015). Beyond dichotomies. Competence viewed as a continuum.
Zeitschrift F€ur Psychologie, 223(1), 3–13.
Bl€omeke, S., K€onig, J., Busse, A., Suhl, U., Benthien, J., D€ohrmann, M., & Kaiser, G. (2014). Von der Lehreraus-
bildung in den Beruf - Fachbezogenes Wissen als Voraussetzung f€ur Wahrnehmung, Interpretation und Handeln
im Unterricht [Content related knowledge as a prerequisite for teachers’ observance, analysis and action]. Zeits-
chrift f€ur Erziehungswissenschaft, 17, 509–542.
Brophy, J. (2000). Teaching. Brussels, Belgium: UNESCO International Academy of Education.
Brouwer, C. (2010). Determining long-term effects of teacher education. In P. Peterson, E. Baker, & B. McGraw
(Eds.), International encyclopedia of education (pp. 503–510). Amsterdam, The Netherlands: Elsevier.
Brown, G. (2006). Explaining. In O. Hargie (Ed.), The handbook of communication skills (pp. 195–228). East Sus-
sex, UK: Taylor & Francis.
Carroll, J. (1963). A model of school learning. Teachers College Record, 64(8), 723–733.
Cauet, E., Liepertz, S., Kirschner, S., Borowski, A., & Fischer, H. E. (2015). Does it matter what we measure? Domain-
specific professional knowledge of physics teachers. Revue Suisse des sciences de l’education, 37(3), 462–479.
Chan, K.-W., & Elliott, R. G. (2004). Relational analysis of personal epistemology and conceptions about teaching
and learning. Teaching and Teacher Education, 20(8), 817–831.
KULGEMEYER AND RIESE | 23
Clark, L. A., & Watson, D. (1995). Constructing validity: Basic issues in objective scale development. Psychologi-
cal Assessment, 7(3), 309–319.
Cleland, A., Abe, K., & Rethans, J. (2009). The use of simulated patients in medical education: AMEE Guide No
42. Medical Teacher, 31(6), 477–486.
Cohen, J. (1988). Statistical power analysis for the behavioral sciences. Hillsdale, NJ: Lawrence Erlbaum
Associates.
Delaney, S. (2012). A validation study of the use of mathematical knowledge for teaching measures in Ireland.
ZDM Mathematics Education, 44(3), 427–441.
Dorfner, T., F€ortsch, C., & Neuhaus, B. (2017). Die methodische und inhaltliche Ausrichtung quantitativer Video-
studien zur Unterrichtsqualität im mathematisch-naturwissenschaftlichen Unterricht [The methodical and content-
related orientation of quantitative video studies on instructional quality in mathematics and science education].
Zeitschrift f€ur Didaktik der Naturwissenschaften. https://doi.org/10.1007/s40573-017-0058-3.
€
Dubberke, T., Kunter, M., McElvany, N., Brunner, M., & Baumert, J. (2008). Lerntheoretische Uberzeugungen von
Mathematiklehrkräften: Einfl€usse auf die Unterrichtsgestaltung und den Lernerfolg von Sch€ulerinnen und
Sch€ulern [Beliefs about teaching and learning of mathematic teachers: impact on teaching quality and achieve-
ment]. Zeitschrift f€ur Pädagogische Psychologie, 22, 193–206.
Duit, R., Gropengießer, H., Kattman, U., Komorek, M., & Parchmann, I. (2012). The model of educational reconstruc-
tion – A framework for improving teaching and learning science. In D. Jorde & J. Dillon (Eds.), Science education
research and practice in Europe: Retrospective and prospective (pp. 13–37). Dordrecht, The Netherlands: Springer.
Eckel, J., Merod, R., Vogel, H., & Neuderth, S. (2014). Einsatz von Schauspielpatienten in den “Psych”-Fächern
des Medizinstudiums – Verwendungsm€oglichkeiten in der Psychotherapieausbildung? [Simulated patients n
“psych”-subjects of medicine – is there a use in educating psychotherapists?] Psychotherapie, Psychosomatik,
Medizinische Psychologie, 64, 5–11.
Enochs, L. G., & Riggs, I. M. (1990). Further development of an elementary science teaching efficacy belief instru-
ment: A preservice elementary scale. School Science and Mathematics, 90(8), 694–706.
Erg€oneç, J., Neumann, K., & Fischer, H. (2014). The impact of pedagogical content knowledge on cognitive activa-
tion and student learning. In H. E. Fischer, P. Labudde, K. Neumann, & J. Viiri (Eds.), Quality of instruction in
physics. Comparing Finland, Germany and Switzerland (pp. 145–160). M€unster, Germany: Waxmann.
Fan, X., Thompson, B., & Wang, L. (1999). Effects of sample size, estimation methods, and model specification on
structural equation modeling fit indexes. Structural Equation Modeling, 6, 56–83.
Fang, Z. (1996). A review of research on teacher beliefs and practices. Educational Research, 38(1), 47–65.
Fechner, S. (2009). Effects of context-oriented learning on student interest and achievement in chemistry education.
Berlin, Germany: Logos.
Fischer, H., Borowski, A., & Tepner, O. (2012). Professional knowledge of science teachers. In B. Fraser, K. Tobin,
& C. McRobbie (Eds.), Second international handbook of science education (pp. 435–448). Dordrecht, The
Netherlands: Springer.
Fischer, H., Neumann, K., Labudde, P., & Viiri, J. (2014a). Theoretical framework. In H. E. Fischer, P. Labudde,
K. Neumann, & J. Viiri (Eds.), Quality of instruction in physics. Comparing Finland, Germany and Switzerland
(pp. 13–30). M€unster, Germany: Waxmann.
Fischer, H., Neumann, K., Labudde, P., & Viiri, J. (Eds.). (2014b). Quality of instruction in physics. Comparing
Finland, Germany and Switzerland. M€unster, Germany: Waxmann.
Gage, N. (1968). The Microcriterion of Effectiveness in Explaining. In N. Gage (Ed.), Explorations of the Teacher’s
Effectiveness in Explaining, Technical Report No. 4, S. 1–8. Stanford, CA: Stanford Center for Research and
Developement in Teaching.
Geelan, D. (2012). Teacher explanations. In B. Fraser, K. Tobin, & C. McRobbie (Eds.), Second international hand-
book of science education (pp. 987–999). Dordrecht, The Netherlands: Springer.
Gramzow, Y., Riese, J., & Reinhold, P. (2013). Modellierung fachdidaktischen Wissens angehender Physiklehrkräfte
[Modelling prospective teachers’ knowledge of physics education]. Zeitschrift f€ur Didaktik der Naturwissenschaf-
ten, 19(1), 7–30.
24 | KULGEMEYER AND RIESE
Harden, R., Stevenson, M., & Wilson, W., G., W. (1975). Assessment of clinical competence using objective struc-
tured examination. British Medical Journal, 1, 447–451.
Hattie, J. A. C. (2009). Visible learning. A synthesis of over 800 meta-analyses relating to achievement. Abingdon,
UK: Routledge.
Helmke, A. (2006). Unterrichtsqualität: Erfassen, Bewerten, Verbessern [Teaching quality: Measurement and
improvement]. Seelze, Germany: Kallmeyersche Verlagsbuchhandlung.
Hempel, C., & Oppenheim, P. (1948). Studies in the logic of explanation. Philosophy of Science, 15(2), 135–175.
Hestenes, D., Wells, M., & Swackhamer, G. (1992). Force concept inventory. The Physics Teacher, 30(3), 141–166.
Hill, H. C., Rowan, B., & Ball, D. L. (2005). Effects of teachers’ mathematical knowledge for teaching on student
achievement. American Educational Research Journal, 42(2), 371–406.
Hodges, B., Regehr, G., Hanson, M., & McNaughton, N. (1998). Validation of an objective structured clinical
examination in psychiatry. Academic Medicine, 73(8), 910–912.
Hofstein, A., & Lunetta, V. (2004). The laboratory in science education: Foundations for the twenty-first century.
Science Education, 88(1), S. 28–54.
Hofstein, A., & Kind, P. M. (2012). Learning in and from science laboratories. In B. Fraser, K. Tobin, & C.
McRobbie (Eds.), Second International Handbook of Science Education (pp. 189–207). Dordrecht, The Nether-
lands: Springer.
Keller, M., Neumann, K., & Fischer, H. (2016). The impact of teachers’ pedagogical content knowledge and moti-
vation on students’ achievement and interest. Journal of Research in Science Teaching, 54(5), 568–614.
Kitcher, P. (1981). Explanatory unification. Philosophy of Science, 48(4), 507–531.
Kulgemeyer, C., & Schecker, H. (2012). Physikalische Kommunikationskompetenz - Empirische Validierung eines
normativen Modells [Physics communication competence – Empirical validation of a normative model]. Zeits-
chrift f€ur Didaktik der Naturwissenschaften, 18, 29–54.
Kulgemeyer, C., & Schecker, H. (2013). Students explaining science – Assessment of science communication com-
petence. Research in Science Education, 43, 2235–2256.
Kulgemeyer, C., & Tomczyszyn, E. (2015). Physik erklären – Messung der Erklärensfähigkeit angehender Physi-
klehrkräfte in einer simulierten Unterrichtssituation [Explaining physics – Measuring teacher trainees’ explaining
skills using a simulated teaching setting]. Zeitschrift f€ur Didaktik der Naturwissenschaften, 21(1), 111–126.
Lee, E., & Luft, J. A. (2008). Experienced secondary science teachers’ representation of pedagogical content knowl-
edge. International Journal of Science Education, 30(10), 1343–1363.
Magnusson, S., Krajcik, J., & Borko, H. (1999). Nature, sources, and development of pedagogical content knowl-
edge for science teaching. In J. Gess-Newsome & N. Lederman (Ed.), Examining pedagogical content knowledge
(pp. 95–132). Dordrecht, The Netherlands: Kluwer Academic Publishers.
Mayring, P. (2000). Qualitative Content Analysis. Forum: Qualitative Social Research, 1(2), 1–10.
McNaughton, N., Ravitz, P., Wadell, A., & Hodges, B. (2008). Psychiatric education and simulation: A review of
the literature. The Canadian Journal of Psychiatry, 53(2), 85–93.
Miller, G. (1990). The assessment of clinical skills/competence/performance. Academic Medicine, 65(9), S63–567.
Nau, J., Halfens, R., Needham, I., & Dassen, T. (2010). Student nurses’ de-escalation of patient aggression: A
pretest-posttest intervention study. International Journal of Nursing Studies, 47, 699–708.
Nunally, J. (1978). Psychometric theory. New York, NY: McGraw Hill.
O’Neill, G. P. (1988). Teaching effectiveness: A review of the research. Canadian Journal of Education, 13(1),
162–185.
Osborne, J. F., & Patterson, A. (2011). Scientific argument and explanation: A necessary distinction? Science Edu-
cation, 95(4), 627–638.
Park, R. S., Chibnall, J. T., Blaskiewicz, R. J., Furman, G. E., Powell, J. K., & Mohr, C. J. (2004). Construct valid-
ity of an objective structured clinical examination (OSCE) in psychiatry: Associations with the clinical skills
examination and other indicators. Academic Psychiatry, 28(2), 122–128.
KULGEMEYER AND RIESE | 25
Park, S., & Oliver, J. S. (2008). Revisiting the conceptualisation of pedagogical content knowledge (PCK): PCK as
a conceptual tool to understand teachers as professionals. Research in Science Education, 38(3), 261–284.
Peterson, P., Carpenter, T., & Fennema, E. (1989). Teachers’ knowledge of students’ knowledge in mathematics
problem solving: Correlating and case analysis. Journal of Educational Psychology, 81(4), 558–569.
Pianta, R. C., Hamre, B. K., & Mintz, S. (2012). Classroom assessment scoring system: Secondary manual. Charlot-
tesville, VA: Teachstone.
Prosser, M., & Trigwell, K. (1999). Understanding learning and teaching: The experience in higher education.
Buckingham, UK: SRHE and Open University Press.
Retelsdorf, J., Butler, R., Streblow, L., & Schiefele, U. (2010). Teachers’ goal orientations for teaching: Association
with instructional practices, interest in teaching, and burnout. Learning and Instruction, 20(1), 30–46. https://
doi.org/10.1016/j.learninstruc.2009.01.001
Rienties, B., Lygo-Baker, S., & Brouwer, N. (2013). The effects of online professional development on higher edu-
cation teachers’ beliefs and intentions towards learning facilitation and technology. Teaching and Teacher Edu-
cation, 29, 122–131.
Riese, J. (2009). Professionelles Wissen und professionelle Handlungskompetenz von (angehenden) Physiklehrkräf-
ten [Professional knowledge and competence of physics student teachers]. Berlin, Germany: Logos.
Riese, J., Kulgemeyer, C., Borowski, A., Fischer, H., Gigl, F., Gramzow, Y., . . . Zander, S. (2015). Modellierung
und Messung des Professionswissens in der Lehramtsausbildung Physik [Modeling and measuring professional
knowledge in physics teacher training]. Beiheft der Zeitschrift f€ur Pädagogik, 61, 55–79.
Riese, J., & Reinhold, P. (2012). Die professionelle Kompetenz angehender Physiklehrkräfte in verschiedenen Aus-
bildungsformen [Professional competence of student teachers enrolled in different teacher training programmes].
Zeitschrift F€ur Erziehungswissenschaft, 15(1), 111–143.
Rochelson, B., Baker, D., Mann, W., Monheit, A., & Stone, M. (1985). Use of male and female professional patient
teams in teaching physical examination of the genitalia. The Journal of Reproductive Medicine, 30(11), 864–866.
Roelle, J., Berthold, K., & Renkl, A. (2014). Two instructional aids to optimise processing and learning from
instructional explanations. Instructional Science, 42, 207–228.
Ross, J. A. (1998). The antecedents and consequences of teacher efficacy. In J. Brophy (Ed.), Advances in research
on teaching (pp. 49–73). Greenwich, CT: JAI Press.
Sadler, P. M., Sonnert, G., Coyle, H. P., Cook-Smith, N., & Miller, J. L. (2013). The Influence of Teachers’ Knowl-
edge on Student Learning in Middle School Physical Science Classrooms. American Educational Research Jour-
nal, 50(5), 1020–1049.
Schecker, H. (1986). Sch€ulerinteresse und Sch€ulervorstellungen zur Mechanik (S II) [Students’ interests and students
conceptions in upper-secondary mechanics]. Physica Didatica, 2/3, 21–33.
Schiefele, U., & Schaffner, E. (2015). Teacher interests, mastery goals, and self-efficacy as predictors of instruc-
tional practices and student motivation. Contemporary Educational Psychology, 42, 159–171. https://doi.org/
10.1016/j.cedpsych.2015.06.005
Scott, P., Asoko, H., & Driver, R. (1992). Teaching for conceptual change — A review of strategies. In R. Duit, F.
Goldberg, & H. Niedderer (Eds.), Research in physics learning: Theoretical issues and empirical studies (pp.
310–329). Kiel, Germany: IPN.
Seidel, T., & Shavelson, R. J. (2007). Teaching effectiveness research in the past decade: The role of theory and
research design in disentangling meta-analysis results. Review of Educational Research, 77(4), 454–499.
Sevian, H., & Gonsalves, L. (2008). Analysing how scientists explain their research: A rubric for measuring the
effectiveness of scientific explanations. International Journal of Science Education, 30(11), 1441–1467.
Shulman, L. (1986). Those who understand: Knowledge growth in teaching. Educational Researcher, 15(2), 4–14.
Shulman, L. (1987). Knowledge and teaching: Foundations of the new reform. Harvard Education Review, 57(1), 1–22.
Staub, F. C., & Stern, E. (2002). The nature of teachers’ pedagogical content belief matters for students’ achieve-
ment gains: Quasi-experimental evidence from elementary mathematics. Journal of Educational Psychology, 94,
344–355.
26 | KULGEMEYER AND RIESE
Stipek, D., Givvin, K., Salmon, J., & MacGyvers, V. (2001). Teachers’ beliefs and practices related to mathematics
instruction. Teaching and Teacher Education, 17, 213–226.
Terhart, E. (2012). Wie wirkt Lehrerbildung? Forschungsprobleme und Gestaltungsfragen [What is the impact of
teacher education?]. Zeitschrift f€ur Bildungsforschung, 2(1), 3–21.
Toulmin, S. (1958). The uses of argument. Cambridge, UK: Cambridge University Press.
Treagust, D., & Harrison, A. (1999). The genesis of effective science explanations for the classroom. In J. Loughran
(Ed.), Researching teaching: Methodologies and practices for understanding pedagogy (pp. 28–43). Abingdon,
UK: Routledge.
Tschannen-Moran, M., & Woolfolk Hoy, A. (2001). Teacher efficacy: Capturing an elusive construct. Teaching and
Teacher Education, 17, 783–805.
Van Driel, J. H., Verloop, N., & De Vos, W. (1998). Developing science teachers’ pedagogical content knowledge.
Journal of Research in Science Teaching, 35(6), 673–695.
Vinzi, V. E., Trinchera, L., & Amato, S. (2010). PLS path modeling: from foundations to recent developments and
open Iissues for model assessment and improvement. In V. E. Vinzi, W. Chin, J. Henseler, & H. Wang (Eds.),
Handbook of Partial Least Squares (pp. 47–82). Berlin, Germany: Springer.
Vogelsang, C. (2014). Validierung eines Instruments zur Erfassung der professionellen Handlungskompetenz von
Physiklehrkräften - Zusammenhangsanalysen zwischen Lehrerkompetenz und Lehrerperformanz [Validating an
instrument to measure physics teachers’ professional competence]. Berlin, Germany: Logos.
Wahl, D. (1991). Handeln unter Druck. Der weite Weg vom Wissen zum Handeln bei Lehrern, Hochschullehrern
und Erwachsenenbildnern [Acting under pressure. The long way from knowledge to Action for teachers and
educators]. Weinheim, Germany: Deutscher Studien Verlag.
Walters, K., Osborn, D., & Raven, P. (2005). The development, validity and reliability of a multimodality objective
structured clinical examination in psychiatry. Medical Education, 39, 292–298.
Weinert, F. (1996). Der gute Lehrer” die gute Lehrerin” im Spiegel der Wissenschaft: Was macht Lehrende wirksam
und was f€uhrt zu ihrer Wirksamkeit? [What makes teachers impactful and what leads to their impact?]. Beiträge
zur Lehrerbildung, 14, 141–151.
Weinert, F. (2001). Concept of competence: A conceptual clarification. In D. Rychen & L. Salganik (Eds.), Defining
and selecting key competencies (pp. 45–65). G€ottingen, Germany: Hogrefe 1 Huber.
Williams, D. (2010). Outcome expectancy and self-efficacy: Theoretical implications of an unresolved contradiction.
Personality and Social Psychology Review, 14(4), 417–425.
Wilson, H., & Mant, J. (2011a). What makes an exemplary teacher of science? The pupils’ perspective. School Sci-
ence Review, 93(342), 121–125.
Wilson, H., & Mant, J. (2011b). What makes an exemplary teacher of science? The teachers’ perspective. School
Science Review, 93(343), 115–119.
Wittwer, J., & Renkl, A. (2008). Why instructional explanations often do not work: A framework for understanding
the effectiveness of instructional explanations. Educational Psychologist, 43(1), 49–64.
S U PP OR TI NG I NFO R M AT IO N
Additional Supporting Information may be found online in the supporting information tab for this
article.
How to cite this article: Kulgemeyer C, Riese J. From professional knowledge to professional
performance: The impact of CK and PCK on teaching quality in explaining situations. J Res Sci
Teach. 2018;00:1–26. https://doi.org/10.1002/tea.21457