Professional Documents
Culture Documents
Philosophy of Science Slides
Philosophy of Science Slides
LESSON 1: INTRODUCTION
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
AGENDA
Introduction
• Why Philosophy of Science – what is it?
• Practical info: Lesson overview, curriculum
• Learning objectives & the exam
• Illustrating a topic and type of exam question
• Video: How to study efficiently (for this course)
2
WHY AM I TEACHING THE COURSE?
Education
• Master’s degree in Philosophy
• PhD in Business Administration at BSS Aarhus
3
PHILOSOPHY OF SCIENCE: A META-LEVEL COURSE
1
Business Organizational
Math Financial Accounting
Statistics Behavior
5
Electives, or Stay
Electives, or Stay abroad Electives, or Stay abroad
abroad
6
Commerci Philosophy of
Bachelor Project
al law Science
4
WHY STUDY PHILOSOPHY OF SCIENCE (POS)? I
(DANISH TITLE: SCIENTIFIC METHOD)
5
WHY STUDY PHILOSOPHY OF SCIENCE (POS)? II
(DANISH TITLE: SCIENTIFIC METHOD)
8
PRACTICAL INFO II: ABOUT LECTURES
• Videos 2020-21: https://www.youtube.com/@PhilosophyScience/videos
9
PRACTICAL INFO III: A CUSTOMIZED
CHATGPT-4 CHATBOT
10
PRACTICAL INFO IV: EXAMS
Exam WO+: On-site written exam with upload of the exam paper in WISEflow, use of the
form Internet NOT allowed during the exam, own PC required. The assignment is
completed in FLOWlock+. Flowlock+: ”a closed browser which
5 hours on-site exam, March. prevents you from accessing
anything but PDF-files saved locally
Partially based on an unknown case. During the lectures, the students will have had
on your own computer. The
the opportunity to reflect upon similar types of questions, in order to prepare for this
lockdown browser must be installed
exam.
before the exam..”
Re-exam: Same as ordinary exam
https://kursuskatalog.au.dk/en/course/123022/Philosophy-of-Science
See Brightspace for prior exam sets (assessment guides & good answers uploaded later)
11
ILLUSTRATING A TOPIC & THE EXAM
RESEARCH DESIGNS, LESSON 2-5
• In Phil of Science focus on more general decisions: Certain research designs* (don’t)
allow certain kinds of conclusions
• What does this mean? Consider this example from a prior exam
*Research design: overall framework to collect and analyze data (more in lesson 4)
EXAMPLE OF PRIOR EXAM QUESTION (SPRING 2021)
READ AND REFLECT (3-4 MIN)
Question 2 (17 %)
Please explain what you think are the main weaknesses of the following study, in terms of its research
design.
Imagine a group of researchers recently set out to study how Covid-19 has impacted employees'
perception and behaviour concerning engaging in remote, virtual work. The study is based on a
survey, where individuals responded to a range of pre-validated questions about their work practices;
e.g. how useful online meetings are, and to what degree their organisation was able to facilitate high
quality online meetings. Survey responses were collected via online polls. One poll was sent out
during a large online conference for individuals in the particular industry surveyed in August 2020
(1400 responses). Another poll was sent out in December 2020 during a different, large online
conference for individuals in the particular industry surveyed (1700 responses). One of the results was,
for example, that most respondents appreciated virtual meetings, because they were easier to
access. The group of researchers also notes that the responses show that during 2020 organisations
became better at facilitating online work for their employees.
13
SIMILAR POINT IN THIS EXAMPLE: SELECTION BIAS
14
SELECTION BIAS: HOW DATA COLLECTION SHAPES
SAMPLE AND THUS CONCLUSIONS TO DRAW
http://www.motherjones.com/kevin-drum/2010/09/counterintuitive-world
https://medium.com/@penguinpress/an-excerpt-from-how-not-to-be-wrong-by-jordan-
ellenberg-664e708cfc3d#.yterlkdg0
Even if you randomly selected what
returned planes to investigate – you
would still face selection bias
15
A CORE TOPIC IN POS: HOW RESEARCH
DESIGN LIMIT CONCLUSIONS
• We can’t know if X causes Y if we don’t have an appropriate
research design, and e.g. know where the data comes from
(selection bias)
• Interviewing 100 successful entrepreneurs doesn’t
necessarily tell you much about how to succeed
• Finding a strong correlation between employee benefits
and performance doesn’t necessarily tell you if employee
benefits are important
• Many claims in business are not valid – will show why/how
16
QUESTIONS ABOUT THE COURSE, EXAM AND
SETUP? (DISCUSS 1-2 MIN)
17
NEED HELP? OR SOMEONE TO TALK TO?
• Our student welfare counsellors are there for you to talk to.
• They have an overall knowledge of rules and options at the degree
programmes at Aarhus BSS. The counselling sessions can relate to all
aspects of your student life, for example:
• Delays in your degree programme
• Doubts about your degree programme
• Illness with you or your relatives
• How to resume your degree programme
• Personal problems that affect your life as a student
If they are not able to help you directly, they can assist you in getting the help
you need.
HOW DO I LEARN MOST EFFICIENTLY AS A STUDENT?
IN GENERAL, AND FOR THIS CLASS
19
VIDEO WHERE I PRESENT GENERAL AND POS SPECIFIC INSIGHT ON HOW
TO STUDY EFFICIENTLY HTTPS://YOUTU.BE/AE5BPQSFIUU (13 MIN)
20
CONSIDERATIONS ON HOW TO PREPARE POS EXAMS
• Use reading guides to read texts. Often read twice.
• Use lectures/slides as pointers to identify key elements and get overview of curriculum
• Create own overview, e.g. including key quotes of key texts
• Read prior exam questions + answer exam answers to test yourself.
• Don’t only review notes, but practice exam questions
• Flowlock+: Remember to save notes / documents on your own computer, not online!
21
WHAT IS SOCIAL SCIENCE & AND DOES IT WORK, ACTUALLY?
(RITCHIE 2020, WATTS 2011)
p. 14
SCIENCE: ”UNMATCHED DEGREE OF OBJECTIVITY”
• Irrespective of flaws
• Science is still ”our best way of moving towards the truth” (p. 13)
• Science represents an ”unmatched degree of objectivity (p. 14) – more in lesson 7-9
p. 13
BUSINESS ADMINISTRATION AS A SCIENCE, COMPARED
TO E.G. MEDICINE (VEVOX.APP & 143-136-950, 1 MIN)
27
TO BE FAIR, ALSO LARGE DIFFERENCES BETWEEN
E.G. MEDICINE AND BUSINESS ADMINISTRATION
• Ritchie has introduced some core questions & indicated answers - to be expanded
2021 SPRING EXAM QUESTION 6
• How can Ritchie 2020 contribute to answer the following question? (other relevant texts are
presented later)
“A famous scientist recently claimed that science provides insight into objective reality. Please rely
on the curriculum to provide counterarguments to this claim.”
2021 SPRING EXAM QUESTION 6
• How can Ritchie 2020 contribute to answer the following question? (other relevant texts are
presented later)
“A famous scientist recently claimed that science provides insight into objective reality. Please rely
on the curriculum to provide counterarguments to this claim.”
• Science not an idealized, impartial process that clearly leads to truth (about objective reality)
• Science might ‘move us towards the truth’
• But human subjectivity, flaws and biases lead to imperfect social consensus seeking (constructs)
can shape and constrain outcomes
• …
WATTS 2011: WHY EVERYTHING IS (NOT) OBVIOUS
• Highlights how social sciences (incl. business administration) differs from e.g. physics
• But not due to methods or research designs
Social sciences: Economics, political
science, psychology, business
• Physics administration etc.
• In social science:
• Investigates the familar social world we live in
• Everything is obvious – once you know the answer
• Aim is to find what is a myth, and what works
https://www.youtube.com/watch?v=D9XF0QOzWM0 34
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
An experimental device in Cavendish’s 18th century lab able to measure the weight
of the earth (!) and following Newton’s law of gravity infer weight of other planets in the
solar system
This did not seem obvious at all!
What can social scientists infer from a small wooden / metal box in our lab?
35
https://www.insidescience.org/news/measuring-earth-wire
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
36
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
https://twitter.com/random_walker/status/909486051135299584
37
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
https://www.youtube.com/watch?v=D9XF0QOzWM0 38
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
39
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
• Complex world
• Many independent actors interact and influence each other (cf. lesson 13 on complexity)
• In other sciences one can ‘isolate’ what one is interested in (gravity, DNA, planetary movements…)
• Self-reflexive
• Definition: Humans are capable of changing their perception of the world and their behavior and targets
• Self-fulfilling prophecies? (cf. lesson 13 on complexity)
40
AARHUS UNIVERSITY
SUMMING UP LESSON 1
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
41
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
Philosophy of Science
https://youtu.be/Ae5bPqsFiuU
TATION
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
43
Improving Students’ Learning With Effective Learning Techniques
by Dunlosky et all, 2013
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
44
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
45
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
8
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
47
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
49
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
50
PHILOSOPHY OF SCIENCE
Causation: One event is the result of the other event, i.e. one caused an
effect. Causal terms: leads to, brings about, impacts, reduces, increases, weakens etc.
4
CORRELATION ≠ CAUSATION II
SOMETIMES CORRELATION DUE TO CHANCE
www.Tylervigen.com
CORRELATION ≠ CAUSATION III
CORRELATION DUE TO A THIRD - CAUSAL – VARIABLE?
https://pedermisager.org/blog/why_does_correlation_not_equal_causation/
https://imgs.xkcd.com/comics/correlation_2x.png
WHY IS IT DIFFICULT TO DISTINGUISH
CORRELATION AND CAUSALITY?
Data might show a correlation, but is it causal?
• Relation due to pure chance/randomness
• Reverse causation
• Confounders (or omitted variable bias)
• Selection bias
• Actual causation
REVERSE CAUSATION
CORRELATION OFTEN IMPLIES CAUSATION – BUT WHAT KIND?
Isager 2023: “If we see correlation, we have good reason to favor the hypothesis “exercise causes
mood” over the hypothesis “exercise and mood are completely independent”.”
Isager 2023: “This is what we mean when we say that stress confounds the relationship between
exercise and mood, and why confounding is also known as third variable problems.”
SELECTION BIAS
Diet Obesity
soda
Diet Weight
plan loss
Another Exercise
https://www.ft.com/content/0a5a4404-7c8e-11e7-ab01-a13271d1ee9c
factor
ADDITIONAL EXAMPLE: SELECTION BIAS
• Organisation X surveys relationship between job
satisfaction and productivity among employees. Not
everyone answers survey.
• Maybe respondents are both more likely to be highly
satisfied and highly productive
• Hence, relation would not be due to causality, but
selection bias (who responded)
Satisfaction
High S / Low P High P / High S
Low S / Low P High P / Low S
Productivity
12
WHAT ISSUE IS AT STAKE HERE?
VEVOX.APP 143-136-950 (REFLECT AND DISCUSS 2 MIN)
An organization with 100 employees introduced a wellness program. After 3 months, all employees
completed a survey on their use of the program and productivity levels. It was found that the more
an employee used the wellness program, the more they reported an increase in job productivity.
What should researchers primarily be worried about, when interpreting these results?
1) Selection bias
2) Neither selection bias nor confounding variables, since all data comes from within the
organization.
3) Confounding variables
4) Both selection bias and confounding variables.
CORRELATION VS.
CAUSALITY – WHICH IS IT?
Isager 2023: “Put simply, correlation indicates that
something interesting might be going on, but we need
to investigate further (do more research, collect more
data, and utilize our expert knowledge) to figure out
exactly what is going on!”
Made up numbers!
CONTROL FOR CONFOUNDERS:
STATISTICAL APPROACH TO SOLVE PROBLEM II
• Controlling for confounders can increase likelihood of finding
causal relations
https://thecorrespondent.com/100/the-new-dot-com-bubble-is-here-
its-called-online-advertising/109970897300-fd465983
18
COUNTERFACTUAL: NEED TO COMPARE A
WORLD OF ADS WITH A WORLD OF NO ADS
Maybe statistical
analysis / controls
could have ensured
counterfactual?
19
ADDITIONAL EXAMPLES: HOW TO GUARANTEE ALL
HIDDEN (POSSIBLY CONFOUNDING) VARIABLES HAVE
BEEN IDENTIFIED?
Org Behavior
• Employee Productivity Programs
• Organizational Restructuring
• Financial Incentives
Marketing: Financial
• Marketing Campaign Effectiveness incentives
Etc.
A FEW MORE WORDS ABOUT CAUSALITY I
https://www.thenewatlantis.com/publications/correlation-causation-and-confusion p. 3
22
A FEW MORE WORDS ABOUT CAUSALITY III
HOW ARE CAUSAL EFFECTS ‘REAL’? MORE IN LESSON 7-9
https://www.thenewatlantis.com/publications/correlation-causation-and-confusion p. 4
23
A FEW MORE WORDS ABOUT CAUSALITY IV
DIFFERENT TYPES OF CAUSAL EFFECTS
http://gusevlab.org/projects/hsq/#h.a6jctcodj87b
CREATING THE COUNTERFACTUAL:
RANDOMIZED CONTROLLED TRIALS (RCT)
We can’t clone a person, but
we can compare how
‘similar’ people respond to
exercise (or not)
(Barrowman p. 14)
BUSINESS EXAMPLE: RUNNING EXPERIMENTS TO
DISTINGUISH CORRELATION FROM CAUSATION
"Booking had learned early on that • At Expedia the scientific method is
it could not trust intuition and ”absolutely critical”
assumptions. "We see evidence • Thousands of experiments run
every day that people are terrible at continuously
guessing. Our predictions of how • User experience, human resources,
customers will behave are wrong sales etc.
nine out of ten times,"…We follow • Thousands of employees involved in
what the majority wants. And if you experimentation
fail fast, you can try a lot of things". • New employees take a crash-course
in the scientific method (experiments)
(Thomke , Experimentation Works p. 160)
(Thomke, Experimentation Works)
See also this short post for further perspectives on why and how businesses should run
experiments: https://www.linkedin.com/posts/carstenbergenholtz_innovation-experiments-
business-activity-6813800441286053888-Noil
26
CORRELATION OFTEN EXCHANGED WITH CAUSALITY
IN MEDIA STORIES ABOUT THE SCIENCE
28
Extra chance to meet the cand.merc. coordinators!
Still in doubt about which cand.merc. track to choose?
You can grab a free cup of coffee and meet the coordinators – to
make sure that you choose the right cand.merc. track.
PHILOSOPHY OF SCIENCE
LESSON 2 PART 2
INTRODUCTION TO EXPERIMENTS
VIDEO: HTTPS://YOUTU.BE/EDCYXONPTXQ
Based on: Mainly BRM ch 3 (pp. 50-59) (Experimental design), & also Bergenholtz 2024a
31
Treatment: Independent
variable
Outcome: Dependent
variable
• If 1 person gets treatment, and 1 person is control differences in outcome could have many
reasons (everybody is different)
34
EXAMPLE OF LAB-EXPERIMENT
(SAME EXAMPLE AS IN BERGENHOLTZ 2024A)
• Randomization: Two sessions run pr. day, randomly decided if 9 or 10 AM session was higher / lower pay
• Treatment: “For each correctly solved task, 1 [5 in treatment] DKK will be added to your payment”.
• Controlled setting: All received the same information, same room and tools, and solved the same two
kinds of tasks - addition and anagram tasks
35
IS THIS A PROPER RCT? WHY, WHY NOT?
VEVOX.APP 143-136-950 (1 MIN)
36
Randomized Controlled Trial
Allocation to each treatment Everything else the same 1 (only 1) treatment tried out
(not systematically different) No other changes
37
3 KEY ELEMENTS OF A RCT
Compared to sampling
strategy – different
1) Randomized process in allocation of participants “randomization” term!
38
2020 SPRING EXAM EXAMPLE
READ, REFLECT, DISCUSS 4-5 MIN
Question 2 (16%)
The COBE Lab at BSS, Aarhus University, is a social science lab that allows social science researchers to run
randomised controlled trials, similar in type to the experiment described in Bergenholtz (2024a). Individuals can sign
up to participate in studies via an online system. Before the government closed down universities in Denmark due to
the Coronavirus, the COBE Lab experienced that more individuals than usual cancelled the studies, for which they
had already signed up. This was probably due to more people preferring to stay at home, rather than being around
other people. Now, consider a situation before universities were locked down. Imagine that a study at COBE Lab has
started, and participants have been randomised into different treatments, creating two different groups. As usual,
after the randomisation but still at the beginning of the study, participants are informed about what they are
supposed to do. Imagine that, in the treatment group, they are being told that they are about to be part of a study
which involves participants being physically fairly close to others. This is related to the treatment, albeit participants of
course do not know if they are in the treatment or control group. In this hypothetical scenario, some people therefore
decide to drop out from the study, since they do not feel comfortable being close to others, while the Coronavirus is
beginning to spread in society. It is a common practice in the COBE Lab that participants can end their participation
at any time. Participants in the other (control) group are not physically close to each other, which might be the reason
why no participants drop out of that part of the study.
Question: How did it impact the quality of this RCT, that some participants decided to leave the experiment because they
did not want to be physically close to others?
Exam example & question has been slightly adapted 39
LAB VS FIELD EXPERIMENTS
Lab-experiments: “experiments conducted in an artificial environment where conditions
are much easier to control than in field experiments” (BRM p. 54)
40
FIELD EXPERIMENT I: DOES MANAGEMENT MATTER
(DIFFICULT QUESTION TO STUDY IN THE LAB)
41
FIELD EXPERIMENT II: DOES MANAGEMENT MATTER
42
A FIELD STUDY: IS THIS A PROPER RCT? WHAT ARE
POTENTIAL WEAKNESSES (READ AND DISCUSS, 2 MIN)
https://twitter.com/Devin_G_Pope/status/1258492772442681346?s=20
https://www.nytimes.com/2016/06/13/insider/which-headlines-attract-most-readers.html
44
ONLY MEN WERE INVITED TO PARTICIPATE IN THIS STUDY. IS IT
STILL A RCT? (VEVOX.APP 143-136-950 , 1-2 MIN)
http://blog.ted.com/2013/04/10/what-motivates-us-at-work-7-
fascinating-studies-that-give-insights/
45
ABOUT RANDOMIZATION
(TYPICAL ERROR IN EXAMS!)
• Internal validity: Can we be certain that we have identified the effects of the
treatment (or lack of effect) (typically high in lab-experiments)
• External validity: Can findings from this particular study be transferred to other
settings? (typically low / challenging in lab-experiments)
External validity:
Internal validity: Going beyond
The study this study,
generalizing
47
RCT THEME TO BE CONTINUED IN LESSON 3
48
PHILOSOPHY OF SCIENCE
LESSON 3
EXPERIMENTS - CONTINUED
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
SUMMING UP LESSON 2
?
Mood
Exercise Mood
https://rpsychologist.com/correlation/
Diet
2
SUMMING UP LESSON 2 Field vs. lab
experiments
• Jeff Bezos: Our success at Amazon is a function of how many experiments we run
• Hal Varian, chief economist at Google: If you want to find out what happens if you
change something…you have to change it
3
RANDOMIZATION VS. RANDOMIZATION
Randomization in RCTs:
Purpose: The primary aim is to evenly distribute known and unknown confounding variables across the treatment and
control groups to isolate the effect of the treatment being tested.
Implementation: Randomization occurs after the sample has been selected, and participants are randomly assigned
to different intervention groups within the study.
Outcome: This ensures that the groups are comparable, and any differences in outcomes can be attributed to the
treatment rather than other factors.
In summary, randomization in RCTs is about creating internal validity and ensuring that the causal relationship
between the treatment and the outcome is not confounded by other variables. In contrast, randomization in sampling
is about creating external validity and ensuring that the sample accurately represents the population, allowing the
results to be generalized.
ChatGPT-4 generated 4
WHAT IS THE MAIN CHALLENGE WITH THIS
KIND OF CLAIM (LESS SLEEP MORE POST-COVID RISK)
5
DOES THIS CONSTITUTE A RCT?
VEVOX.APP 143-136-950 (2 MIN)
6
AGENDA
• Internal vs. external validity (and lab vs. field experiments)
• Quasi-experiments
• Limits of RCTs
7
VIDEO ON LAB VS FIELD EXPERIMENTS, INTERNAL
AND EXTERNAL VALIDITY AND OTHER TOPICS
https://www.youtube.com/watch?v=TY1xoPon42U
(different order and structure, same message)
8
INTERNAL AND EXTERNAL VALIDITY
CRITERIA FOR ASSESSING THE QUALITY OF A STUDY
• Validity: Can we be certain about the results of a scientific study?
• Internal validity: Can we be certain that we have identified the effects of the
treatment (or lack of effect) - typically high in RCTs in particular lab-experiments
• External validity: Can findings from this particular study be transferred to other
settings? (typically low in lab-experiments)
External validity:
Internal validity: Other studies,
The study generalizing
9
THREATS TO INTERNAL VALIDITY THAT A
PROPER RCT DISMANTLES (BERGENHOLTZ 2024A: P. 2)
• Selection bias
• No self-selection but randomization
• Confounding variables
• Since treatment and control groups are ‘equivalent’ & only one
difference (treatment) between groups – no other confounding
variables can explain outcome.
10
Premise: If no internal validity – not possible to generalize
11
THREATS TO EXTERNAL VALIDITY
CRITERIA TO PAY ATTENTION TO (BRM P. 54-55, SEE ALSO BERGENHOLTZ 2024A P2-3)
https://us.macmillan.com/books/9780374710453
INTERACTION OF SELECTION Treatment could be so complicated, that it
only works on a specialist chef – and can’t
AND TREATMENT II generalize to a non-specialist. Thus limiting
external validity
INTERACTION OF SELECTION AND TREATMENT III
15
INTERACTION OF SETTING AND TREATMENT:
DO PARACHUTESWORK?
• Results: Parachute use did not significantly reduce death or
major injury (0% for parachute v 0% for control; P>0.9). This
finding was consistent across multiple subgroups.
https://www.bmj.com/content/363/bmj.k5094
Setting: Can refer to the task used (as in a lab-experiment), or the overall setup/environment – as in this example
16
WHAT ARE THE LIMITS TO THE EXTERNAL VALIDITY OF
EXAMPLE PRESENTED IN BERGENHOLTZ 2024A
CONSIDER EACH OF THE THREATS/CRITERIA (3-4 MIN)
18
SOME NOTES ON HOW TO GENERALIZE
• Not just binary – a spectrum (a lot, medium, a little)
• A study on paying 1 vs 5 DKK pr task will probably generalize to 1.1 vs. 5.1 pr. task
• Effect of headache pill can probably generalize
• People are quite similar, chemically speaking
• Much more challenging with humans / organizations
Be nuanced: How much can it generalize, how far?
• Assess multiple criteria (sample, setting, history etc.)
• Can it generalize to much higher sums, to people with high-paying jobs, to people
from other ages, regions, countries…
• Generalizations stronger when based on theory
• In exam setting, you might not know theory – speculate on what you find limiting
characteristics
19
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
21
SURVEY VS FIELD EXPERIMENTS: PROS AND CONS
22
QUASI-EXPERIMENTS
VIDEO: https://www.youtube.com/watch?
v=62QVoC7QBG4
QUASI EXPERIMENTS ≠ RCT
• RCT: Randomly assigns subjects to treatment vs control
• Quasi: No actual randomization – but ”as if” random and no self-selection into treatment
24
Key question: Has the treatment been randomized
across groups?
https://twitter.com/jenniferdoleac/status/1092057392542552067
25
IS THIS A QUASI-RCT?
26
QUASI-EXPERIMENT: AS GOOD AS RANDOM
Quote from paper: “In Denmark, as in many other places across the globe, test time
is determined by the weekly class schedule and computer availability at
schools….We claim that this creates variation in test time that is as good as random
to the individual.”
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4790980/#si1
Students haven’t self-selected into classrooms / best students have (probably) not
been put into certain timeslots.
27
HOW TO ASSESS IF SOMETHING IS A QUASI-RCT?
1) Check for Random Assignment:
• RCT: Deliberate randomization by researchers.
• Quasi-RCT: Participants assigned without researcher-controlled randomization.
2) Confounding Variable Scan:
• RCT: Randomization aims to balance confounders.
• Quasi-RCT: Higher risk of confounder influence.
3) Study Context Analysis:
• RCT: Controlled settings with researcher oversight.
• Quasi-RCT: Real-world contexts, where full control isn't feasible.
28
ANOTHER EXAMPLE: QUASI-EXPERIMENTS
(FROM BERGENHOLTZ 2024B: P. 2)
29
THE PROMISE OF QUASI-EXPERIMENTS?!
https://noahpinion.substack.com/p/the-
econ-nobel-we-were-all-waiting
30
TERMINOLOGY: NATURAL EXPERIMENT ≈ QUASI RCT
NATURE OR RESEARCHER THAT HAS ‘DECIDED’ TREATMENT? (CF. BRM P. 58)
31
Data collected on the same group of students
pre- and post Covid-19 lockdown
32
Lab-experiment
Natural experiment
33
Randomized Controlled Trial
Allocation to each treatment Everything else the same 1 (only 1) treatment tried out
(not systematically different) No other changes
Quasi: Can we assume allocation Quasi: Can we assume only one difference?
was as good as random? Or other systematic differences (e.g. created by
Treatment related to randomization? those that allocated participants?)
34
ECOLOGICAL VALIDITY
(SUB-CATEGORY OF EXTERNAL VALIDITY)
BRM p 55
35
HOW STRONG IS THE EXTERNAL & ECOLOGICAL
VALIDITY? (VEVOX.APP 143-136-950, 1-2 MIN)
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4790980/
36
INTENTION TO TREAT(MENT): IS THIS AN ACTUAL RCT?
VEVOX.APP 143-136-950, 1-2 MIN
37
TO CONSIDER AT THE EXAM
If you are asked to ”assess” an ‘experiment’”
• Validity: Internal, external, maybe ecological
• Internal: Is it actually an RCT (the 3 elements) – or a quasi-RCT
• External: Don’t just state ”it’s low” or ”high”, but use (some of the) criteria mentioned in
chapter/slides. What’s the sample, what’s the task, in what time/history was it done…
• Not your subjective assessment: Curriculum based assessment of the quality of the RCT
• Not just a binary (good/bad) assessment: Characterize strengths and weaknesses
38
REFLECT: HOW WOULD YOU SETUP AN
EXPERIMENT TO STUDY THE RQ IN YOUR BP
39
LIMITS OF EXPERIMENTS
For example, can you think of topics that can not be adressed via RCTs, be it
field- or lab-experiments? And why?
41
LIMITATIONS OF RCT’S I: OFTEN NOT POSSIBLE
• Long vs. short term: Experiments are usually relatively short-term – how to study impact of 5 year
organizational change process?
• Sometimes RCTs are impossible – can’t manipulate impact of gender, or economic impact of a depression
• Ethics: Maybe the control group is unethical – usefulness of parachutes, or impact of a psychopath
manager, or impact of new teaching design BRM p. 51
• Speed (and ethics): Maybe no time to wait for an RCT? – usefulness of facemasks etc.
• Dynamic world: The world does not change one (discrete) variable at a time (more in Nelson
2016 and lesson 13) – cf. multipath causality, e..g macro-economics or organizational culture
42
LIMITATIONS OF RCTS II: DON’T PROVIDE ALL THE
ANSWERS
Even when we have done an experiment…
• RCTs provide insight into average effects – not focus on differences across subjects
• Why do people decide to do what they do? (more in lesson 6, on theory)
• RCT can’t stand alone. Need theory (and other research designs) to generalize.
43
LIMITATIONS OF RCT’S III: WHAT CAN WE LEARN
FROM THEM? (SELECTION BIAS)
Hypothetical example
• Let’s say a company for the first time introduces something new (wellness program) – and
performance improves
• might the improvement be due to the wellness program, or just that something new
happened?
Real example
• Tennessee (a small’ish state) successfully reduced class size (improved performance)
• when California tried the same they failed. they lacked qualified teachers
• Hence, can be limits to how much one can generalize from – even great – field-experiments
44
LIMITATIONS OF RCT’S III: STATUS AND VALUE
OF RCT DEPENDS ON FIELD
45
BUSINESS ADMINISTRATION AND RCT’S
• Strong business case for business experimentation. Why not more widespread?
• Not aligned with incentives, org culture, and leadership
• But also difficult: Expedia CEO: ”People need to be trained in the scientific method” -> all
new employees get a crash-course (Thomke, Experimentation Works, p. 148)
Therefore: Teach the scientific method / RCT to business school students – to prepare them for
business life and to ensure more businesses run good RCT’s in the future
46
RCT’S: SUMMING UP
• Why are experiments considered the ”the golden standard”?
• The R and C and T – vs. the counterfactual
• Assessing the quality of RCTs
• Internal vs. external and ecological validty
• (dis)advantages of different types of experiments
• Lab vs. field experiments & quasi-experiments)
• Limitations of RCTs
• Experiments: A method that provides insight into general philosophy of science issues
• What is good, certain science
• How to balance certain (high internval validity) with more uncertain but more
generalizable knowledge (high externval validity)
47
PHILOSOPHY OF SCIENCE
LESSON 4
RESEARCH DESIGNS
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
RCT: MORE EXERCISE, BETTER MOOD?
DISCUSS AND VOTE, VEVOX.APP 143-136-950 (2 MIN)
• Result: Treatment group reported (statistically significant) better moods compared to control
2
WHAT IS THE EXTERNAL
VALIDITY?
Well, what if exercise program (treatment) was:
• Run 5 kilometer in forrest – still believe general claim?
Keep threats to external validity in mind when you see a study claiming: RCT showed X
3
SUMMING UP LESSON 3
4
LESSON 4: AGENDA AND LEARNING GOALS
Learning goals
• Research designs (RD) Not just to be able to summarize different types of RDs, but
• Experiments also to be able to assess research critically, and question if a
• Cross-sectional given RD can support the conclusion.
• Longitudinal
• Case-study Notice: There are no 100% unambiguous rules for when a
RD is appropriate.
https://www.youtube.com/watch?v=U9GNoIQ0qOw
6
RESEARCH DESIGN VS RESEARCH METHODS
7
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
8
WHERE ARE WE IN THE RESEARCH PROCES?
Research Question
Research Design
Theory (more in
lesson 6-7)
Data collection & analysis
9
A GOOD RESEARCH DESIGN IS A DRIVING
FORCE OF GOOD, CREDIBLE RESEARCH
10
RESEARCH QUESTIONS AND RESEARCH DESIGN:
WHAT DO WE WANT TO KNOW?
RCT can eliminate confounders and establish
causality. Yet, not only interest we have
Examples
• How and why has the introduction of daily exercise programs shaped the mood
• How much do employees at different age levels exercise, and what are their moods?
• How does the mood level generally change and fluctuate over a long period (a year e.g.)?
In other words
• Maybe we want to understand what happens in specific contexts (case)
• Maybe we want to know if something is correlated (cross-sectional)
• Maybe we really want to understand how something develops over time (longitudinal)
11
DIFFERENT TYPES OF RESEARCH DESIGNS I
• Experiments (cf. lesson 2 & 3)
• Cross-sectional
• Case study
• Longitudinal
• Experiments
• Cross-sectional
• Longitudinal
• Case study
14
LONGITUDINAL: STUDY OVER TIME (BRM P. 62-63)
Characteristics
• Data collected on the same units (individual, firm, case etc.) at
least twice – capturing change over time
• Ex: How motivation measured in time point 1 is related to
motivation and productivity measured in time point 2
18
CASE STUDY DESIGN (BRM P 63-70)
Characteristics
• Bounded organization / location / person / event
• In-depth focus on the unique / context – less about averages and variation (≠ cross-sectional)
• Can both be comparative (multiple cases) or longitudinal (e.g. lengthy study at one site)
Typical methods
• Interviews, observations, but also surveys, databases
• Usually qualitative, can also contain quantitative elements
19
SINGLE CASE VS. (MANY) MULTIPLE CASES (BRM P. 67)
• Single case study: Opportunity to focus on the particular context -> rich, deep
insight into the particular case
• Multiple case study (e.g. 4 or 8 cases): Focuses less on individual unique context,
more on variation/similarity across cases (closer to cross-sectional aim)
”Thus, case studies can be associated with both theory generation and theory testing.” (BRM p. 67)
20
A FEW FURTHER NOTES ON RESEARCH DESIGNS I
(BASED ON TYPICAL QUESTIONS)
Longitudinal
• Collecting data on alcohol consumption over multiple weeks to obtain a more reliable measurement of
consumption (creating a better measurement) not longitudinal change, thus cross-sectional.
• Collecting data on alcohol consumption in week 1 and 15 to assess development over time (longitudinal)
•A construct
• A construct is a general expression, very similar to the ‘concept’ word. Often one would call motivation a concept, while the
way one has measured / speaks of motivation would be constructs; e.g., job satisfaction or intention to quit (cf. Nohria). Any
empirical project relies on the use of constructs.
ANOTHER OVERVIEW OF RESEARCH DESIGNS
23
RESEARCH QUESTION: SHOULD WE BE WORRIED
ABOUT VIDEOGAMES & KIDS BEING OR
BECOMING AGGRESSIVE?
Version 1
Let’s assume a longitudinal design In this study, we investigated the extent to which
found a positive link between violent adolescents who spend time playing violent video games
videogames and aggression. Could exhibit higher levels of aggressive behaviour when
we be certain about the causal nature compared with those who do not. A large sample of British
adolescent participants (n = 1004) aged either 14 or 15
of this link? years were interviewed. Over a period of 2 years young
people provided weekly reports of how many hours a week
Why, why not? they played what they consider violent video games.
Furthermore, at the beginning of the 2 year period, after 1
year, and after 2 years they provided information on to
what extent they display aggressive behavior (based on
Likert-scales in a survey). Results were statistically analysed,
to identify if a correlation between playing violent video
games and displaying violent behavior can be identified.
26
https://royalsocietypublishing.org/doi/10.1098/rsos.171474 (slightly modified)
VIDEO GAMES AND AGGRESIVENESS: VERSION 3
WHAT RESEARCH DESIGN IS THIS & WHICH OF THE THREE DESIGNS PROVIDES
THE MOST CREDIBLE RESULTS, AND WHY? (VEVOX.APP 143-136-950 2-3 MIN)
https://www.nature.com/articles/s41380-018-0031-7
SPEAKING OF VIOLENT VIDEO GAMES
AND AGGRESSIVE BEHAVIORS…
https://nyheder.tv2.dk/video/bklXaHMzNTZCaF9fWEp0ZzNYNUhBM0thSkhvaEplTkY=
ASSESSING ONE STUDY VS. MULTIPLE
REFLECT ON FOLLOWING (HYPOTHETICAL) COMPARISON (1-2 MIN)
B) One RCT relying on one particular game and a small’ish sample shows a
positive link between violent videogames and aggression.
29
CASE STUDY VS. CROSS-SECTIONAL VS
LONGITUDINAL DESIGNS
You have now received some limited information on the above outlined research design. Based on the
provided information one can infer that it is not a randomised controlled trial. What information would you
have liked to know about this case in order to identify if the setup constitutes a quasi-experiment or not?
Please be specific when explaining what information you need. 31
FROM AN OVERALL CATEGORIZATION OF TYPES OF
RESEARCH DESIGNS TO IDENTIFYING SPECIFIC
CHALLENGES AMJ 2011
32
AMJ 2011: ON RESEARCH DESIGN
HTTPS://WWW.JSTOR.ORG/STABLE/PDF/23045105.PDF
• Audience: Top-researchers
• Aim: How to optimize research designs or; key reasons for rejecting a
scientific paper at one of the world’s leading scientific journals
• Provides: Incomplete list of challenges
AMJ 2011: LIST OF KEY REASONS FOR
REJECTION
• Limits of cross-sectional data (when adressing change and causality)
• Omitted variable bias (don’t overlook the confounding / causal variable)
• Don’t use inappropriate samples (if interested in CEO decisions, why study students)
• Engage in careful construct ’creation’ (data should match theory)
• Avoid common method variance/bias (when independent & dependent variables stem
from the same source)
34
Internal validity challenge
AMJ 2011 p. 657: “The problem with cross-sectional data is that they are mismatched with
research questions that implicitly or explicitly deal with causality or change, strong tests of
which require either measurement of some variable more than once, or manipulation of one
variable that is subsequently linked to another.”
OMITTED VARIABLE BIAS Internal validity challenge
Examples
• Exercise and mood
• School closures and learning
• Sleep and job performance
•…
Thus: Include relevant control variables in statistical analysis (cf. AMJ 2011 p. 659) to account for
confounders and reduce omitted variable bias.
• Does your construct capture the phenomena you are interested in?
• Measuring exercise – by asking people how much they exercise
• Measuring motivation – by asking people about motivation or effort
• Measuring firm performance – short / long term, turnover, profits, growth…
• …
• Reason to use established surveys/measures: Already tested (for reliability, construct
validity)
• Potential systematic error – if the source is biased, bias influences both independent
and dependent variable in same, systematic way
• Imagine a student grading their own paper
• Imagine a manager rating both work habits and performance of employee – manager might like
a particular employee, skewing both ratings?
• Imagine an accountant in charge of assessing asset quality and financial health of company –
any correlation might be due to CMB
• Preferably different sources for independent and dependent variable – less bias
• Limits of cross-sectional designs (change & causality direction) - via longitudinal and RCT designs
• Omitted variable bias – via RCT’s, careful inclusion of confounding variables and case studies
• Appropriate samples – the young we are interested in, both aggressive and not
• Construct creation – studying perception, behavior, archival data
• Avoid common methods bias – unbiased assessment of behavior from multiple sources
41
OVERALL AIM: IMPROVE YOUR ABILITY TO ASSESS IF OTHERS’
CLAIMS ARE VALID / CREATE GOOD PROJECTS YOURSELF
https://www.nytimes.com/2018/10/21/smarter-living/tips-better-morning-routine.html
https://towardsdatascience.com/correlation-vs-causation-a-real-world-example-9e939c85581e
https://www.rebeccabarter.com/cv/talks/Cal_Day_Presentation.pdf
42
CORRELATION OFTEN EXCHANGED WITH CAUSALITY
IN MEDIA STORIES ABOUT THE SCIENCE
44
PHILOSOPHY OF SCIENCE
LESSON 5
STUDY CAFE 1
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
Agenda
2
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
Imagine a study that investigates all books reviewed on a book review site
(about 50.000 books). The study has data on the average review rating (scale
from 1 to 5) submitted for all books, in all of 2011 - thus, one average number
for each book. Then the study compares if the average review rating a book
got, is correlated with the likelihood of getting a book award from a specific
institution. The award was given in the summer of 2011. Results show that the
reivew books receive are correlated with the likelihood of getting an award
(the higher the review, the higher the rating). One might then conclude that
getting a good average review rating increases the likelihood of getting an
award.
Feedback on A)
4
B) ‘Book reviews’ and ‘awards’, version 2
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
(discuss 5 min)
Kovács, B. and A. J. Sharkey. 2014, "The Paradox of Publicity: How Awards Can Negatively Affect the Evaluation of Quality."
Administrative Science Quarterly 59(1): 1-33. (slightly modified abstract) 5
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
Feedback on B)
6
C) What are strengths / weaknesses of this study + how could the authors’
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
Case description
How should one take notes while studying, e.g. in a class room setting? It is examined whether students learn the most
effectively (e.g. get the best grades) when taking notes using a computer or taking notes using a pen and paper. The
simple message is that students learn far better when writing with a pen, in contrast to typing on a computer. Mueller and
Oppenheimer carried out two lab-based experiments in order to test the varying hypotheses. In total 327 (primarily
psychology) students (about 20-22 years old)were involved in the two experimental studies; students were allocated
randomly to one of the following settings: either they had to use a computer to take notes, or they had to rely on the use of
pen and paper. All students were given a 45 min lecture in organizational behavior, and then tested immediately after the
lecture and a week later. The 15 min test both consisted of remembering theories and concepts.
Intuitively it could seem plausible that the reason why not using a computer would be beneficial is that wireless internet or
other technological distractions didn’t disturb the student while trying to listen to and learn what was being taught.
Classroom behavior clearly shows how much time students spend on the internet, both doing useful fact-finding search
and wasting time on Facebook. However, the researchers emphasize that the experimental setting did not allow the use of
any internet, and the reason for why using a pen and paper was more effective, is – they argue- because it facilitates the
learning process. When taking notes on a computer, students tend to write long sentences, while when using a pen, the
students attempt to abbreviate and simplify what is being taught. Students using a pen simply wrote fewer words.
This could be an exam challenge – discuss in detail what would be your answer 7
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
8
Assessing quality of RCT research design: What to
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
Not your personal assessment: Curriculum based assessment of the quality of the RCT
9
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
Feedback on D) Supermarkets
12
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
13
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
14
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
15
2020 Spring exam: Question 1 (16%)
Please identify theUNIVERSITY
AARHUS research design that the research study presented in section i) relies on. Furthermore, please
assess and explain if the interpretation (presented in section ii) of the research study (presented in section i)
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
seems credible. In other words, does the research design applied in the study justify the interpretation?
i) Research study
Is there any link between individuals having played a musical instrument at any point in their life, and how
well they perform on cognitive (IQ) tests, compared to people that have not played a musical instrument? In
a study that researchers made, 103 healthy, elderly humans complete a range of questions and tests, in
order to capture their IQ. In addition, at the same time the researchers also asked the 103 individuals if they
had played any musical instruments during their lifetime, and for how many years they had done this. The
researchers also contacted the study participants’ family members, in order to check this information.
Furthermore, the researchers collected data on the current health as well as current physical activity, e.g. if
the 103 individuals currently engaged in any sports. This data makes it possible to run a statistical regression
analysis, where the researchers could contrast and compare the cognitive ability of the group of people that
had played musical instruments in their life, with the group of individuals that had not, or at least had played
less. In the analysis, the researchers could control for how healthy the individuals currently were.
17
B) ‘Book reviews’ and ‘awards’,
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
version 2
The study’s setup
• A given book (X) receives ratings from 2007
• Intervention: Award given to X book in August 2009
• The book (X) continues to receive ratings in 2009-2011: Higher or lower than
before?
1) Thus a) longitudinal study + b) weak form of quasi-experiment
• Not RCT because the books are not selected at random (like test time in
elementary schools): A particular (probably good) type of books get the
awards
• One can argue for quasi: Compare one and the other group
• E.g. compare effect of reward: reviews before award and after reward
• Books with awards compared to books without rewards
different kinds of research designs can investigate the link between getting
a book award and the quality of the reviews that the book receives. I
Cross-sectional
• One could have data on 1000 books, and their awards, and (quality of) reviews at one point in time
in 2010. This would enable us to correlate what reviews books that get a reward have. However, let's
say the award is handed out in the middle of the year. If the book with awards have high reviews, we
do not know if these reviews have emerged because the book got an award, or if books with good
reviews get an award - since we only have data on the books at one point in time.
• Another angle. Let's say we have data on the nr. and quality of reviews on the exact day a book gets
a reward (or not). We might be able to show that books with good reviews are more likely to get a
reward, but we would not know if there is any clear causal relation. Maybe good books leads to
good reviews and awards - but maybe something else can explain it. Maybe books by famous
authors are more likely to get good reviews and are also more likely to get rewards - even though
the books really aren't better than other books.
• The key advantage here (compared to, e.g., the case study approach) is that one can create a quite
large sample. Thus, if one had selected 1000 books in the year 2011 (instead of 2010), there is a
reasonable chance one might get the same effect - unless maybe if completely different kinds of
books are picked in each year. So, having data from multiple years or investigate the relation to
different kinds of awards will strengthen ones ability to generalize from the particular dataset to other
situations.
• One could also engage in cross-sectional survey designs where one sends out a survey to a lot of
people, asking them what they think of rewards. I won't go further into this here, but just to emphasize
that correlational research designs could look very differently, depending on ones angle. 19
AARHUS UNIVERSITY
A short note expanding on the book award example: Let's imagine how
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
different kinds of research designs can investigate the link between getting
a book award and the quality of the reviews that the book receives. II
Longitudinal
• In the longitudinal design we have multiple data points on the reviews and if they get an award at a
given point (we still investigate the same 1000 books as in the cross-sectional design). So, let's say
we have the average review received in each month of a given year, from 2009 to 2015. If a book
gets an award in August 2011, we can now track if the average quality of reviews changes over
time. If the review quality decreases it is likely that getting an award has attracted different kinds of
readers that might think differently of the book or just write different kinds of reviews. However, we
can't rule out that the book got an award because it got a lot of attention (let's say it was mentioned
in Oprah Winfrey's talkshow in early 2011) and that this is both the reason for it getting award later
on, as well as the change in review quality (omitted variable bias). A more sophisticated study would
then statistically control for "was the book mentioned on TV shows" in order to eliminate the impact
of this. This is of course challenging, not just because one needs to find the data, but also one needs
to define what TV shows to include, what it means to be mentioned etc. Maybe some books are
mentioned with a bad review on a TV show, and that should maybe be categorized differently? This
is related to Nelson (2016) and his point about heterogeneous variables and numerous influences.
Experiment
• In an experimental design one randomly assigns books with the award or not. So, out of the 1000
books, 50 randomly selected books get an award. We can then investigate if getting an award has
any identifiable impact on the quality of reviews. Since we know that the only systematic difference
between books is if they got this particular award or not we can be more certain about the causal
20
effect. Maybe some books were also mentioned on Oprah's show, but if we study many books, such
singular effects should be averaged out.
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
A short note expanding on the book award example: Let's imagine how
different kinds of research designs can investigate the link between getting
a book award and the quality of the reviews that the book receives. III
Case study
• One could also interview readers of books and ask them if a book getting an award a) influences
them to buy a given book, b and if it is also influences if they might a review and what review they
write. Maybe some readers state that they had really high hopes for this book because of the
award it received, but the high expectations were not fulfilled. One could thus try to inductively
infer, that when the individual reads a book that did not get an award, they don't have the same
high expectation and the review might become better. The same meal will probably get a
different review at a local, cheap cafe vs. a (more expensive) restaurant with multiple Michelin
stars. However, we need to keep in mind that we get individuals' self-perceptions. Maybe readers
don't really know why they buy books. Maybe they think that the award has one impact, but it
actually influenced their thinking differently. Nevertheless, a case study design can get to the
question of why people buy something, in a more in-depth way than a survey e.g. could, or the
data that the above 3 designs imply. Several long'ish interviews with different kinds of readers
could generate insight into how different kinds of readers seem to think differently about rewards
and reviews. While the smaller scale involves drawbacks, there are also key advantages
associated to this approach.
• Overall, different designs allow different kinds of conclusions, and different certainty in the 21
conclusion.
C) Is the Pen mightier than the Sword? A few notes
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
• External validity
• Which other types of students can this be transferred to, e.g. in terms of age, educational level, countries
etc. (cf. interaction between selection and treatment)
• Might depend on type of course: What if we had to take notes on equations / drawings, or words (cf.
interaction between setting and treatment). What if the test is different, e.g. analyzing a case.
• Might depend on time study is carried: Are we worse at taking long-hand notes in 2023, compared to
decades ago? (interaction between history and treatment).
• Maybe students worked hard because the study was run by their professors? Maybe difference is therefore
smaller because everyone worked harder. (Hawthorne effect)
• What about Ewriters, where you write with a pen on a screen, and get the advantage of both?
Note other researchers have tried to replicate the results and ‘failed’: The pen is not mightier than
22
the sword https://twitter.com/HeatherUrry/status/1289013799828090880?s=20
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
However
• Different kinds of customers in different supermarkets?
Who looks at the receipt?
• Maybe Netto customers go there daily, while you only go
to Bilka once a month + maybe buy a lot?
Selection bias and risk of confounding variables
23
AARHUS UNIVERSITY
In order to pass this question, the following should be established in the answer:
• Identify that the research design is cross-sectional
• Identify that a causal claim is made
• Internal validity is weak’ish: Explain why the causal claim is not warranted due to the
cross-sectional nature of the data, e.g. by referring to the risk of confounding variables
(omitted variable bias), or that the causality could be reversed i.e. high IQ could lead to
playing a musical instrument.
LESSON 6
WHAT IS A THEORY
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
Dear students CONCERNING CHATBOT ACCESS
As promised in the first lesson, you can get access to a ChatGPT-4 driven, customized chatbot. Let's call it Phil (as in Philosophy of Science). Roughly
speaking, we have created a platform, where any question you ask is answered by a ChatGPT-4 search tool that relies on parts of the curriculum to
answer your question.
This is a pilot project and - as far as I know - the first of its kind at our university. A few things for you to consider before using the platform:
•Purpose: The platform is designed to only answer questions related to this course. Phil costs money, and therefore search has to be limited to the
course. If we encounter usage that goes clearly beyond the topics of the course, we can (and will) shut down the individual's access.
•Validity: While I have tested Phil, I cannot guarantee that all answers are true - or identical to what I would have answered. However, I am very confident
that the vast majority of answers will be as good as any I would have generated and better than using any other available chatbot. I will continuously
monitor the questions posed and answers provided, in order to ensure the quality of the answers. If errors emerge, I will of course flag them.
•Privacy: Please note that this also means that the questions you ask are stored, linked to your user id. However, my colleague (Oana Vuculescu) will only
send me the questions and answers, not user information. I am honestly not interested in who asks which questions, just the nature of the questions and
quality of the answers. Btw., nothing is sent to OpenAI (the owner of ChatGPT).
•Content: Copyright limits how much of the curriculum I can upload to Phil. Main documents are prior lectures, various texts I have written and good
student exam answers. Phil is probably strongest in relation to Research Design and Theory (lesson 2 to 6), while it also does contain information about
the rest of the curriculum.
•Usage limit: We are not entirely certain how costly this is going to be, so due to a limited budget we are for now setting a limit to 30 searches pr. week.
We will revisit this limit.
•Usage, how to: Note that when you have received an answer to your question, you can click on the citation (below the answer) and get information on
where the chatbot retrieved the answer from.
•Plagiarism (important!): Different students can get a) identical (!) answers to similar questions and b) answers can be very similar to material in the text.
Hence, if you want to use answers from Phil in the exam, you will have to quote it, in order not to plagiarize from the curriculum or other students. More
in lesson 12.
I might adjust the settings and functionality as we go along. If you encounter any errors or chatbot 'behavior' that you find odd, please let me know at
cabe@mgmt.au.dk. Link: https://app-backend-ubqtrwmmqio3c.azurewebsites.net/ (you should have access since Microsoft Azure is connected to your
AU email account). Best regards, Carsten
2
ANOTHER RESEARCH
DESIGN EXAMPLE:
HARM OF SCREENTIME?
Journalistic causal conclusion: Letting young children sit in front of a screen for
too long can have significant negative consequences…
Actual correlational claim of study: Infants who had around two hours of
screen time per day had weakened cognitive abilities later in life.
The study’s longitudinal research design: 437 one-year-olds, same children
examined when 18 months old and again at 9 years old.
Confounding variable: "I could imagine that those who give their children a
screen at the age of one are parents who may demand less of their children or
are less interested in their children…"
https://ekstrabladet.dk/nyheder/samfund/nyt-studie-chokerer-saadan-paavirker-skaermtid-boernene/9634816
3
AGENDA
HA / BSc bachelor project course description:
• Empirical thesis: The purpose is to collect knowledge about a given topic by relating theory to
empirical observations, for example by using existing data from databases, questionnaires, interviews,
observations, simulations or experiments”
• Irrespective of the topic chosen, the use of relevant theory and literature is fundamental to the thesis.”
https://kursuskatalog.au.dk/da/course/123045/Bachelors-Project
However: What is a theory, what do we use it for, and how is the use of theory different
depending on your approach?
4
VIDEO FROM 2021 ON WHAT THEORY IS (NOT)
https://youtu.be/NF1059J7KBo
5
RESEARCH QUESTION – RESEARCH
DESIGN - THEORY
Theory
• Informs relevance of research question
• Guides choice of research design
• Shapes how to measure and analyze data
• Provides framework for interpreting results
• …
6
A THEORY IS NOT JUST A THEORY (NYT 2016)
OR: THEORY IN EVERYDAY LIFE VS IN RESEARCH
• “In everyday conversation, we tend to use the word “theory” to mean a hunch, an
idle speculation, or a crackpot notion. That’s not what “theory” means to
scientists.” (NYT 2016)
• Ancient Greek word for theorizing: Attentive observer sent to observe foreign
practices and “theorize” about them (Swedberg 2014: p. 19)
• “In science, the word theory isn’t applied lightly,” Kenneth R. Miller, a cell biologist
at Brown University, said. “It doesn’t mean a hunch or a guess. A theory is a system
of explanations that ties together a whole bunch of facts…” (NYT 2016)
https://www.nytimes.com/2016/04/09/science/in-science-its-never-just-a-theory.html?_r=0
7
THEORIES AS MAPS – NAVIGATION GUIDE
“To judge a map’s quality, we can see how well it guides us through its territory. In a
similar way, scientists test out new theories against evidence. Just as many maps have
proven to be unreliable, many theories have been cast aside.” (NYT 2016)
https://www.youtube.com/watch?v=4WVelCswXo4&t=9s
8
FIRST: WHAT DO RESEARCHERS DO? I
DESCRIPTIONS VS EXPLANATIONS VS PREDICTIONS
Descriptions vs explanations
"Empirical analyses in economics have diverse
How much, how
goals—all valuable. Some analyses advance many etc.
knowledge by uncovering new facts or providing
richer descriptions of old facts ... Other analyses
seek to understand the [theoretical] mechanisms
producing outcomes with an eye toward
interpretation and counterfactual policy
evaluation."
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5804814/ (authored by Remember: Need theory to make
Heckman, 2000 Nobel Prize winner in Economics)
descriptions – e.g. to define & measure GDP,
inflation, cost of living etc.
Yet a description is not a theory
9
WHAT DO RESEARCHERS DO? II
DESCRIPTIONS VS EXPLANATIONS VS PREDICTIONS
Predictions vs explanations
Predicting future events – using large datasets
Remember: Need theory to define/analyse variables
Yet: A prediction is not necessarily a theory
10
Ok, so research often (not always) involves aiming to explain something. But what does
that mean?
11
A THEORETICAL EXPLANATION
(BERGENHOLTZ 2023A SECTION 1.3)
12
ELEMENTARY SCHOOL QUASI-EXPERIMENT EX
WHAT IS THE ”ABSTRACTED” EXPLANATION? (DISCUSS 1 MIN)
“We identify one potential source of bias that influences children’s performance
on standardized tests and that is predictable based on psychological theory: the
time at which students take the test. Test time is determined by the weekly class
schedule and computer availability at the school. Using test data for all children
attending Danish public schools between school years 2009/10 and 2012/13,
we find that, for every hour later in the day, test scores decrease by 0.9% of an SD.
In addition, a 20- to 30-minute break improves average test scores. Time of day
affects students’ test performance because, over the course of a regular day,
students’ mental resources get taxed.”
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4790980/, (p. 2621, order of sentences slightly modified)
13
A THEORY IS NOT
• List of references
• Data
• Hypotheses
• List of variables
• Graphs
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4790980/
14
AN ABSTRACTED DESCRIPTION CAN GUIDE US AND PREDICT –
BUT NOT NECESSARILY INVOLVE THEORETICAL EXPLANATION
OF HOW OR WHY
15
WHAT EXPLAINS EXERCISE I (DISCUSS 1-2 MIN)
How many different theoretical explanations for why people exercise (so much / so
little) can you think of?
16
WHAT EXPLAINS EXERCISE II
Let’s say we investigate why people exercise (so much / so little)
• Individual explanations
• Intrinsic motivation (enjoyment)
• Extrinsic motivation (social status)
• Individual well-being (stress, sleep)
• …
• Structural explanations:
• Economic factors (affordability)
• Urban design (availability of parks/gyms)
• Cultural norms (social values)
• …
In a project one would usually not investigate all
• Biological explanations possible theoretical explanations
• Genetics
• Age, gender
• …
17
HOW THEORY CAN HELP
GENERALIZE
https://twitter.com/I_Am_NickBloom/status/1512446919875842051 18
THEORY AND GENERALIZATION IN PHYSICS,
MEDICINE ≠ BUSINESS ADMINISTRATION
• Theoretical explanations – usually – more universal in e.g. physics and often also medicine
• Atoms, cells etc. less different than humans & organizations
19
TWO DIFFERENT METAPHORS FOR WHAT
CONSTITUTES A THEORETICAL EXPLANATION
What does it mean to have ‘theoretically explained’ something?
An explanation – understanding
• Make sense of a situation
• Subjective meanings/experiences
https://www.nature.com/articles/s41390-018-0071-3
20
CAUSAL EXPLANATIONS I
21
CAUSAL EXPLANATIONS II
COUNTERFACTUAL
22
EXPLANATION AS UNDERSTANDING I
23
COMPARING THE TWO METAPHORS I
CAUSALITY VS UNDERSTANDING
Explanation as causality:
• For example: What do I think about the incentive change? What do I think was
changed, and how does this matter for me? Why does it matter what incentives
others get? (intentions, beliefs, behaviors)
• Not just about getting data that can validate quantitative data – but open up
rich(er) perspective and make sense of situation
24
EXPLANATION AS UNDERSTANDING II
ALSO CONTAINS CAUSAL ELEMENT?
• Understanding perspective can, but does not have to, make causal claims.
• Understanding is primary, causality secondary
• World is complex (more in lesson 13), simple causal relationships less likely
25
COMPARING THE METAPHORS II
HOW DO WE KNOW IF EXPLANATIONS ARE GOOD?
Causality: Did the answer ‘hit the bullseye’ – comparison with the world (validity concepts)
Understanding:
26
WHAT IS YOUR IDEA OF A THEORY - IN YOUR
BACHELOR PROJECT FOR EXAMPLE
(VEVOX. APP 143-136-950, 1-2 MIN)
27
INDUCTION VS DEDUCTION
(VEVOX. APP 143-136-950, REFLECT AND DISCUSS 1-2 MIN)
Imagine one has a dataset consisting of 1000s of book reviews; incl. data on age,
gender, educational level and personality of those writing the reviews. One now
examines the dataset in order to see if one can find any correlations between age
groups, educational level and various types of personality and the review rating
provided. Results are then produced based on this analysis.
28
HOW DO WE GET TO A THEORETICAL EXPLANATION?
THEORY BUILDING VS THEORY TESTING
30
THEORY TESTING
• “De Vaus (2001: p. 8) notes: “If the theory is true
certain things should follow in the real world”, in
other words certain observations should be implied
by the theory.” (Bergenholtz 2024c, section 3.1)
• Theory: General explanation – the relation between independent and dependent variable (e.g.
incentives and performance)
33
THEORY VS. HYPOTHESIS VS MODEL II
34
RELATION BETWEEN HYPOTHESIS AND THEORY
(VEVOX. APP 143-136-950, REFLECT AND DISCUSS 1-2 MIN)
35
RELATION BETWEEN HYPOTHESIS AND THEORY
(VEVOX. APP 143-136-950, REFLECT AND DISCUSS 1-2 MIN)
a) If the hypothesis turns out to get statistical support (e.g. p-value <
0.01) – is the theory then true?
Concerning b)
• Imagine two dices
• One is is slightly off: 20% chance of a 6
• The other is perfect: 16.6% chance of getting a 6
• We always (see e.g. Bergenholtz 2024a) test a particular version of a theory - based on
particular tasks, with a particular sample, particular size incentive, particular time, country etc.
• Full test of all combined options is incredibly large (millions) – see also lesson 13 on complexity
37
ILLUSTRATING DEDUCTIVE APPROACH IN
QUALITATIVE RESEARCH
40
THEORY BUILDING: INDUCTION III
41
WHEN TO ENGAGE IN THEORY BUILDING
(INDUCTION) VS. TESTING (DEDUCTION)?
• Depends on research question, but general rules of thumb:
• Induction
• When little is known / clear theory does not exist
• If there are tensions and disagreement in existing knowledge
• One stumbles upon new empirical insight
• If case study, and no clear theory to be tested
• Deduction
• Applied when testing hypotheses derived from existing theories.
• If theory can be extended to new untested fields, or contexts (samples)
• If large quantitative dataset is to be studied – and theoretical field is well
established so hypotheses developed in advance
• …
42
INDUCTION VS DEDUCTION: NOT ALWAYS EASY
TO SEPERATE TYPES OF REASONING
Abduction?
43
ABDUCTION: DIFFERENT TYPE OF INFERENCE
44
ABDUCTION: NOT JUST SEQUENTIALLY SWITCHING BETWEEN
INDUCTION AND DEDUCTION
45
ABDUCTION VS INDUCTION / DEDUCTION
• Some/much research characterized as deductive, is actually abductive (less so in
experiments) – according to Heckman
• Important not to oversell: don’t claim deduction if it is not
However:
• Don’t just (uncritically) adopt the term abduction.
• Few projects are 100% deductive / inductive – small changes ok’ish
• Do reflect on your process
• Maybe add a critical paragraph outlining the degree to which your process was in fact
inductive/deductive
46
THIS ALL SOUNDS VERY RATIONAL AND CUMULATIVE
IS THIS HOW THEORIES ACTUALLY DEVELOP? (Kuhn 1970 vs. Popper 1962)
47
SUMMARY OF LESSON 6
• What is a theory (not)?
• A theoretical explanation
• Different aims of theory: Causality vs. understanding
• Theory vs. models vs. hypotheses
• Theory building and theory testing or Different theoretical reasonings:
Inductive vs. deductive – and abductive approaches
48
PHILOSOPHY OF SCIENCE
LESSON 7
KUHN AND POPPER ON SCIENTIFIC
PROGRESS - PART 1
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
From lesson 6 slides: Plagiarism (important!): Different students can get a) identical (!) answers to similar
questions and b) answers can be very similar to material in the text. Hence, if you want to use answers from Phil
in the exam, you will have to quote it, in order not to plagiarize from the curriculum or other students. More in
lesson 12.
https://app-backend-ubqtrwmmqio3c.azurewebsites.net/
2
Why is a deductive test better than a
non-deductive test? Example:
• Thow a dice X number of times
• Some outcome might be surprising (#
of eyes shown on dice)
• If not pre-registered expectation
• 6 times bigger chance of finding
something interesting due to random
chance
3
AGENDA
From improving particular research designs and getting insights on how to develop
(good) theory more general questions about nature and truth of theories and how
science develops
4
PHILOSOPHY OF SCIENCE: HOW DO THEORIES
DEVELOP AND IMPROVE OVER TIME?
Not about one individual theory – but the philosophy of science (in general)
Insights on research design & deduction and induction can give impression of
• Science directly building on prior work
• Science continuously improving
Positivism
• We can observe the world as it is
Science as a very rational, cumulative process
• We can test theories & verify truth
5
THE (POSITIVISM) POSITION THAT
KUHN & POPPER ARGUE AGAINST
Positivism: Classic ‘textbook’ conception of science pre-Kuhn (1962/1970)
• “…positivists believed that disputes between rival scientific theories could be solved in a perfectly
objective way, comparing the theories directly to the observations…” (PoS 5, p.80)
• The world can be observed as it is / there are neutral facts (PoS p. 80)
• (true) Theory corresponds to / mirrors world (PoS p. 78)
• Theory development as a rational proces (PoS p. 78-79)
Linear Lego brick metaphor: Scientific development is rational and cumulative ≠ Kuhn
Implication I: If rational, cumulative science, we can verify theories and establish truth
Implication II: If theory development is rational, a better theory should always be accepted
Right?
Kuhn & Popper disagrees
6
The following introduction of Kuhn is also available in this
video from 2021: https://youtu.be/gD84WMLmQes
7
KUHN 1962: MOST CITED SOCIAL SCIENCE RESEARCH
(= MOST OFTEN IN REFERENCE LIST)
http://blogs.lse.ac.uk/impactofsocialsciences/2016/05/12/what-are-the-most-cited-publications-in-the-social-
sciences-according-to-google-scholar/ 8
HOW RATIONAL IS SCIENCE?
SOMETIMES A CLEARLY BETTER EXPLANATION IS NOT ACCEPTED
BY RESEARCHERS
Semmelweis had no
(good) theory though
http://www.jameslindlibrary.org/articles/how-dramatic-were-the-effects-of-
handwashing-on-maternal-mortality-observed-by-ignaz-semmelweis/
9
THE RELEVANT GERM THEORY WAS ACTUALLY
PROPOSED THOUSANDS OF YEARS AGO
https://twitter.com/page_eco/status/1356953614544064516?s=20
10
KUHN: SCIENTIFIC DEVELOPMENT – 2 TYPES
http://www.thwink.org/sustain/glossary/KuhnCycle.htm
11
THE ESSENCE OF KUHN’S MESSAGE: NORMAL
SCIENCE VS. PARADIGM SHIFTS
https://ftp.cs.ucla.edu/pub/stat_ser/r487.pdf
12
BUT WHY ISN’T SCIENCE JUST RATIONAL -
WHAT ARE THE BARRIERS?
13
PERCEPTION: WE, LITERALLY, DO NOT SEE THE SAME I
≠ positivism
14
PERCEPTION: WE, LITERALLY, DO NOT SEE THE SAME II
”If two people stand at the same place and gaze in the same direction, we must…conclude that
they receive closely similar …But people do not see stimuli…Instead they have sensations, and
we are under no compulsion to suppose that the sensations of two viewers are the same…much
neural processing takes place between the receipt of a stimulus and the awareness of a
sensation…the route from stimulus to sensation is in part condition by education.” (Kuhn p. 192-
193, my emphasis)
Although, one can’t make up any kind of interpretation: If you can’t distinguish a wolf from a
dog, you won’t survive (Kuhn p. 195, not direct quote)
https://upload.wikimedia.org/wikipedia/commons/a/ac/Alpha_particle_and_ele
ctrons_from_a_thorium_rod_in_a_cloud_chamber.jpg
17
SOCIAL SCIENCE EXAMPLES: WE DON’T DIRECTLY SEE THE
WORLD AS IT IS – BUT THROUGH A PARADIGMATIC LENSE
If this is right – how can we be certain our theories actually represent the world / are true?
18
VIDEO OF WHAT YOU SAY?
Data is not ”raw”, always interpreted
on a paradigmatic background
https://twitter.com/michaelgalanin/status/98
8087010270285825?s=11
19
KUHN’S TAKE ON SCIENTIFIC DEVELOPMENT
http://www.thwink.org/sustain/glossary/KuhnCycle.htm
KUHN’S PARADIGM SHIFTS ILLUSTRATED
http://www.youtube.com/watch?v=3cp6pEzx3uw
21
A SCIENTIFIC PARADGM (DEFINITION)
WHAT PARADIGMS ARE AND WHY THEY ARE IMPORTANT
Paradigms refer to the way “things are done” (norms), in a scientific community, and are
implicit.
For example: How we usually measure X variable, what questions we ask, what science gets
funded, who a science institution would hire, what methods are popular in a field, what qualifies as
a good answer etc.
“A paradigm governs, in the first instance, not a subject matter but rather a group of
practitioners” (Kuhn, p. 180)
• For example
• Within the same quantitative approach and paradigm, we can
directly compare two (similar) surveys on job satisfaction.
• Across paradigms, comparing a quantitative survey to an
ethnographic study on job satisfaction risks comparing volume to
weight – the measures don’t easily translate.
23
KUHN’S TAKE ON SCIENTIFIC DEVELOPMENT
Typical (normal)
science
Revolutions are –
very – rare! Paradigm: fundamental assumptions that
members of scientific community accept
(PoS p. 81)
http://www.thwink.org/sustain/glossary/KuhnCycle.htm
KUHN’S TAKE ON SCIENTIFIC DEVELOPMENT
(REFLECT, DISCUSS 2 MIN – QUESTIONS?)
http://www.thwink.org/sustain/glossary/KuhnCycle.htm
25
SUMMING UP: A SCIENTIFIC PARADIGM
Paradigms showcase a research agenda
• What is important Theories can more easily change -
• Which skills a researcher should have paradigms are embedded practices
• Which data are collected that are diffucult to change
• What is being taught
• Which research is being accepted
• ….
26
ILLUSTRATING SLOW CHANGS IN (ECONOMICS) SCIENCE
1994 landmark (Nobel prize winning) study claims: Increasing
minimum wage did not negatively impact employment
An example of a
fundamental change in
assumptions about the world
https://noahpinion.substack.com/p/why-15-minimum-wage-is-pretty-safe
27
DOES WATER RUN UPHILL?
OF COURSE MINIMUM WAGE WOULD NEGATIVELY IMPACT EMPLOYMENT…RIGHT?
“The inverse relationship between quantity demanded and price is the core proposition in economic
science, which embodies the presupposition that human choice behavior is sufficiently rational to allow
predictions to be made. Just as no physicist would claim that “water runs uphill,” no self-respecting
economist would claim that increases in the minimum wage increase employment. Such a claim, if seriously
advanced, becomes equivalent to a denial that there is even minimal scientific content in economics, and
that, in consequence, economists can do nothing but write as advocates for ideological interests.
Fortunately, only a handful of economists are willing to throw over the teaching of two centuries; we have
not yet become a bevy of camp-following whores.”
~James M. Buchanan, 1986 Nobel laureate in economics, writing in the Wall Street Journal on April 25, 1996
28
SLOWLY CHANGING BELIEF ON EFFECT OF
MINIMUM WAGE
https://noahpinion.substack.com/p/why-15-minimum-wage-is-pretty-safe
29
CARD’S (2021 NOBEL PRIZE WINNER) EXPERIENCE
A RATIONAL DISCUSSION?
“[E]conomists who objected to our work were upset by the thought that we were giving free rein to
people who wanted to set wages everywhere at any possible level…I've subsequently stayed away from
the minimum wage literature for a number of reasons. First, it cost me a lot of friends. People that I had
known for many years, for instance, some of the ones I met at my first job at the University of Chicago,
became very angry or disappointed. They thought that in publishing our work we were being traitors to
the cause of economics as a whole.”
https://noahpinion.substack.com/p/why-15-minimum-wage-is-pretty-safe
• Changes in fundamental (paradigmatic) assumptions and beliefs are difficult, stubborn and slow
• Science not a fully rational process / institution
• Yet, not necessarily a full-blown Kuhnian revolution arguably Kuhn partially right
30
SO, HOW DO PARADIGMS CHANGE AND IMPROVE
AFTER ALL, ACCORDING TO KUHN?
If science can’t improve and two paradigms just are
different pure relativism, anything goes
https://en.wikipedia.org/wiki/Planck%27s_principle
31
SCIENCE, ONE FUNERAL AT A TIME?
What research
design is this?
What to pay
attention to?
http://www.econ.upf.edu/docs/papers/downloads/1498.pdf ”…a new scientific truth does not triumph by convincing its opponents and
making them see the light, but rather because its opponents eventually die, and a
new generation grows up that is familiar with it.” (Max Planck) 32
NEW PARADIGMS ARE NOT MORE TRUE I
Kuhn p. 206: There is, I think, no theory-independent way to reconstruct phrases like ‘really
there’; the notion of a match between the ontology of a theory and its ”real” counterpart in
nature now seems to me illusive in principle” (my emphasis. A clear critique of positivism)
POS, p. 84: ”…Kuhn questioned whether the concept of objective truth actually makes sense at
all. The idea that there is a fixed set of facts about the world, independent of any particular
paradigm, was of dubious coherence, he believed. Kuhn suggested a radical alternative: That
facts about the world are paradigm-relative, and thus change when paradigms change. If this
suggestion is correct, then it makes no sense to ask whether a given theory corresponds to the
facts ‘as they really are’, nor therefore to ask whether it is objectively true.” (my emphasis)
33
NEW PARADIGMS ARE NOT MORE TRUE II
Kuhn p. 206: There is, I think, no theory-independent way to reconstruct phrases like ‘really there’;
the notion of a match between the ontology of a theory and its ”real” counterpart in nature
now seems to me illusive in principle” (my emphasis. A clear critique of positivism)
POS, p. 84: ”…Kuhn questioned whether the concept of objective truth actually makes sense at
all. The idea that there is a fixed set of facts about the world, independent of any particular
paradigm, was of dubious coherence, he believed. Kuhn suggested a radical alternative: That
facts about the world are paradigm-relative, and thus change when paradigms change. If this
suggestion is correct, then it makes no sense to ask whether a given theory corresponds to the
facts ‘as they really are’, nor therefore to ask whether it is objectively true.” (my emphasis)
35
WHICH QUOTE IS NOT BY KUHN?
(VEVOX. APP 143-136-950, 2 MIN)
1) Later scientific theories are better than earlier ones for solving puzzles in the often quite different
environments to which they are applied. That is not a relativist's position, and it displays the sense in
which I am a convinced believer in scientific progress.
2) 'Normal science' means research firmly based upon one or more past scientific achievements,
achievements that some particular scientific community acknowledges for a time as supplying the
foundation for its further practice.
3) A new paradigm in the end wins over the former, established paradigm, by developing new theories
and practices that manage to obtain a clearer and better understanding of reality as it is.
https://todayinsci.com/K/Kuhn_Thomas/KuhnThomas-Quotations.htm
QUESTION FROM 2018 RE-EXAM
Question 3 (15%)
Please explain why Kuhn would NOT agree with the following
statement: “A new paradigm in the end wins over the former
established paradigm, by developing new theories and practices that
manage to obtain a clearer and better understanding of reality as it is.”
Popper’s agenda
• Problem with induction and idea of observing the world as it is
• Asymmetry
• Verification vs. falsification of theories
• Realism and tentative truths (-> post-positivism, more in lesson 9)
41
PROBLEM WITH INDUCTION II: THE NEXT SWAN
MIGHT BE BLACK
ABH: p. 47
ABH: p. 48
42
WE PROPOSE A THEORY AND TEST IT I
Yet: We should not try to verify a theory. Too easy to find examples that (seem to) support.
Thus: Too easy to deductively create a hypothesis and find support for it
43
WE PROPOSE A THEORY AND TEST IT SEVERELY
– TO FALSIFY IT II
Make conjectures (informed guesses) – and then seek to refute/falsify it (!)
ABH: p. 48
44
WE MAKE RISKY PREDICTIONS – AND AIM TO
FALSIFY IT
A theory should not just state:
• More motivation better performance Easy to “confirm”
• Movements by heavy objects influence other objects
•…
We want more specific, risky predictions More risky and challenging theory
• Challenging, specific goals better performance
• A new marketing campaign targeting a smaller, niche audience will result in higher profit
margins than mass marketing efforts.
45
VIDEO: FALSIFICATION
https://www.youtube.com/watch?v=wf-sGqBsWv4 46
POPPER: IS SCIENCE BASED ON INDUCTION OR
DEDUCTION?
We often look at the world / data and try to find a pattern – inductive theory building
However, actual science (according to Popper) is testing the theory to see if theory is
wrong - deductively
47
WHAT DOES THIS MEAN FOR QUALITATIVE /
QUANTITATIVE STUDIES?
Quantitative Hypothesis Testing:
• Keep in mind that a hypothesis test is specific, and always an incomplete test of a theory
48
WHAT DOES THIS MEAN FOR QUALITATIVE /
QUANTITATIVE STUDIES?
Qualitative Case Studies:
• While qualitative case studies are not typically structured around statistical hypothesis testing,
Popper's ideas still hold relevance. The critical approach to developing theory from case
studies involves looking for disconfirming evidence as well as confirming evidence,
challenging the researcher to refine their theories. Popper would likely caution against over-
reliance on inductive reasoning from case studies and stress the importance of seeking
potential falsification even in qualitative research. (ChatGPT-4 generated)
49
IF THE THEORY IS REFUTED/WRONG, IT IS NOT
UNSCIENTIFIC – JUST NOT VALID
ABH: p. 49
50
WE CAN ONLY APPROXIMATE TRUTH
http://www.proprofs.com/quiz-school/user_upload/ckeditor/Trooth.jpg 51
POPPER – TO BE CONTINUED IN LESSON 8
52
PHILOSOPHY OF SCIENCE
LESSON 8
KUHN AND POPPER ON SCIENTIFIC
PROGRESS - PART 2
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
AGENDA: FIRST HALF OF LESSON
2
REVISITING KUHN I: HOW SCIENCE DEVELOPS
Positivism, pre-Kuhn: “…positivists believed that disputes between rival
scientific theories could be solved in a perfectly objective way, comparing
the theories directly to the observations…” (PoS 5, p.80)
REVISITING KUHN II: PARADIGM SHIFTS AND DIFFICULTY
OF COMPARING BETWEEN PARADIGMS
4
REVISITING KUHN III: EXAMPLE ON HOW COMPARISON
ACROSS ‘PARADIGMS’ CAN BE DIFFICULT
For example: If studying emotions
• Approaches can be so different, so very difficult to compare
• Observing (subjectively) real world situations vs. what happens (objectively) inside the brain
• What is the correct / objective theory?
incomparable – or, very difficult to compare. No common standard
5
CONSIDER THIS: IS IT A GOOD THING IF 100% OF
SCIENTISTS AGREE ON SOMETHING? (REFLECT 1-2 MIN)
6
AFTER KUHN: SCIENCE & POLITICS
Politicians (mis)use science for politics
7
REVISITING POPPER I: HOW TO DO SCIENCE
Given how induction & deduction (do not) work this is what science can do
8
REVISITING POPPER II: WHY ONE (OR MANY) TESTS
DON’T VERIFY A THEORY
Imagine a theory: Higher financial incentives increase performance in simple tasks
10
EXAMPLE: KUHN, POPPER AND THE CHALLENGE OF
NEW CLAIMS
Why wasn’t a better explanation accepted – quicker?
Popper:
• The medical community should have conducted
objective tests to falsify hypotheses
• A critical, rational approach could have resolved it
Kuhn
• Paradigms explains resistance to germ
theory/handwashing
• Not just objective evidence that led to scientific
change
• Survival of the fittest explanation…
11
2021 SPRING EXAM QUESTION (4-5 MIN)
Question 6: ” A famous scientist recently claimed that science provides insight into objective
reality. Please rely on the curriculum to provide counterarguments to this claim.”
• How can Kuhn & Popper contribute to answer the following question? (note: further relevant
texts are presented later)
• Please consider, in some detail, how you would structure an answer to the question
OUTLINE OF ANSWER TO 2021 Q6
Kuhn
• Focuses on socio-psychological factors
• We see a processed version of the world
• Therefore no theory-independent observations of the world -> embedded in paradigms
• Observations embedded in (somewhat) incommensurable paradigms
• Evolutionary ‘survival of the fittest’ perspective on improvement of theories – not closer to reality
Popper
• Focuses on nature of induction & deduction -> what science can (not) do
• Conjectures and Refutations: Create theories/hypotheses and aim to falsify – not to verify/prove
• Problem of induction – next data point might prove us wrong
• We can always think of a different way to deductively test theory
• We can approximate truth, but will always ‘only’ be ”science of the time”
Both agree
• No completely objective truths
• No direct access to the world in itself / objective reality
• Later theories often improve upon and replace earlier ones
13
SYNTHESIZING KUHN AND POPPER: IN RESEARCH
AND IN BUSINESS
14
AGENDA: SECOND HALF OF LESSON
15
Following presentation of Nelson 2016 also available
in this 2021 video: https://youtu.be/t-yLe1LllbA
WHAT ARE WE MEASURING, AND HOW CERTAIN CAN WE
BE ABOUT OUR CONCLUSIONS?
Different approaches to these questions in different fields
17
NELSON 2016: WHAT KIND OF SCIENCE IS
ECONOMICS / BUSINESS ADMINISTRATION
“The location of any planet at any time can be described completely in terms of numbers,
as can its motion at that time” (Nelson 2016: p. 1693)
“Thus while Newton’s calculations treat only the masses of a particular planet and the sun,
the effect of other masses and potential influences like galactic dust on the planet’s orbit
turns out to be very small” (Nelson 2016: p. 1694)
BUSINESS: A SCIENCE? II
EVEN SEVERAL NATURAL SCIENCES AREN’T LIKE
CLASSICAL PHYSICS
Meteorology (e.g. storms) ≠ astrophysics, Newtonian laws
Nevertheless, meteorology is a
pretty precise science.
“Every storm is different in some often significant way…. Measures of physical characteristics of
storms, like wind velocity, provide important indicators of how destructive that storm is likely to
be. However, the measurements here are not precise, and in their nature cannot be, and the
relevant causal connections are somewhat variable” (Nelson 2016: p. 1699)
BUSINESS: A SCIENCE? III
THE WORLD(S) WE STUDY
1) “First, while the subject matter they study falls into different classes, within each
class there is considerable heterogeneity, which cannot be reduced to a small
number of well defined variables, given the questions being asked of that subject
matter. And in many cases the classes themselves do not have sharply defined
boundaries“ (Nelson 2016: p. 1697)
2) “Second, the forces and conditions that influence the subjects of study in ways
that the science seeks to understand are numerous, highly variable, and often
cannot be separated sharply one from another. Indeed the circumstances
associated with any particular set of observations may need to be understood as,
in some sense, unique.” (p. 1697)
3) …the subject matter they study often changes over time…” (p. 1698) – see also
lesson 9
https://eprints.qut.edu.au/221162/1/95492.pdf
23
BUSINESS: A SCIENCE? IV
NELSON 2016 ON NATURE OF SOCIAL SCIENCE
CONSTRUCTS [MEASURES OF CONCEPTS)
“I want to suggest that, where the underlying conception of the phenomenon being studied has blurry
edges, a constructed number meant to measure an aspect of it may take on a definitional quality.
That is, the number, and the assumptions behind its construction, comes to be treated as what the
phenomenon is, rather than as a useful if limited measure of that phenomenon. I propose that this is a
real danger, and needs to be guarded against.” (Nelson 2016: p. 1695, my emphasis)
EXAMPLES OF DIFFERENCES IN MEASURES - THE
DIFFERENCES MATTER
http://jreiss.org/Philosophy_of_Economics/PhilEcon_Manuscript_Master.pdf
BIG 5 VS OTHER
PERSONALITY CONCEPTS:
NO SHARP BOUNDARIES ->
FUZZY OVERLAP
https://twitter.com/PsychoSchmitt/status/1482023538710167556
Heterogeneous variables without clear boundaries
27
HETEROGENEOUS VARIABLES INFLUENCED BY
NUMEROUS OTHER VARIABLES (NELSON POINT #1+2)
https://doi.org/10.1038/nature13977
HETEROGENEOUS VARIABLES INFLUENCED BY
NUMEROUS OTHER VARIABLES (NELSON POINT #1+2)
29
HOW MANY VARIABLES MIGHT IMPACT
ORGANIZATIONAL PERFORMANCE?
Market demand Employee morale Customer satisfaction
Economic conditions Organizational structure Supply chain efficiency
Industry competition Innovation capacity Technological adoption
Company culture Financial health Regulatory environment
Management effectiveness Brand reputation Globalization effects
31
METEOROLOGY VS. SOCIAL SCIENCES
PEOPLE ARE NOT JUST HETEROGENEOUS - CAN ALSO DECIDE TO
CHANGE BEHAVIOR OVER TIME (NELSON POINT #3, MORE IN LESSON 13)
https://twitter.com/kjhealy/status/1040220704527011840
https://economics.stackexchange.com/questions/5706/ https://www.youtube.com/watch?v=GdujVdjAKUg
what-do-supply-demand-curves-really-look-like
Pfleiderer, Paul. 2014. Chameleons: The Misuse of Theoretical Models in Finance and Economics. Stanford University.
NELSON ON LIMITS OF EXPERIMENTS (P. 1699)
How homogeneous and
predictable is the world?
37
BUSINESS: A SCIENCE? III
THE WORLD(S) WE STUDY
1) “First, while the subject matter they study falls into different classes, within each
class there is considerable heterogeneity, which cannot be reduced to a small
number of well defined variables, given the questions being asked of that subject
matter. And in many cases the classes themselves do not have sharply defined
boundaries.“ (Nelson 2016: p. 1697)
2) “Second, the forces and conditions that influence the subjects of study in ways
that the science seeks to understand are numerous, highly variable, and often
cannot be separated sharply one from another. Indeed the circumstances
associated with any particular set of observations may need to be understood as,
in some sense, unique.” (p. 1697)
3) …the subject matter they study often changes over time…” (p. 1698) – see also
lesson 9
How easy is it to measure / make sense of what you are interested in?
CONSIDER THIS: IS IT A GOOD THING IF 100% OF
SCIENTISTS AGREE ON SOMETHING? (REFLECT 1-2 MIN)
43
IMPLICATIONS OF NELSON’S CHARACTERISTIC OF
BUSINESS ADMINISTRATION / ECONOMICS
• Physics’ quantitative ”law-focus” should not be the goal of business administration
• When phenomena and theoretical relations are heterogeneous, we have to
acknowledge limitations & be transparent in our categorization of our data/variables
• Both quantitative and qualitative work important – selection of method should be based
on specific problem field
• The more homogeneous the field, the easier it is to quantify average relations
…it is a mark of the trained mind never to expect more precision in the treatment of any subject than the nature of that
subject permits; for demanding logical demonstrations from a teacher of rhetoric is clearly about as reasonable as
accepting mere plausibility from a mathematician
(Aristotle, 1976: p. 65)
WHY READ KUHN, POPPER AND NELSON?
They have shaped scientists (and societies) ideas about what science is
• How rational and objective science is
• What certainty / kind of truth can we expect
• How are some sciences different than others
• Established counterpoints to positivism & a naive conception of science
However: Science still the best option we have, to learn about the world
• Might not be true, but best approximation
• Different (paradigmatic) approaches are difficult – but possible – to compare
• Fittest theories survive
• Science is (somewhat) self-correcting & explanations are – continuously - improving
45
PHILOSOPHY OF SCIENCE
LESSON 9
PHILOSOPHICAL ASSUMPTIONS IN SCIENCE
2
ANOTHER REMINDER: REMEMBER FLOWLOCK+ SETUP
NOTES: PDF + OWN COMPUTER
More info on how this works and how you can try out the setup to follow
3
SUMMING UP: NELSON 2016
4
AGENDA
5
TERMINOLOGY: ONTOLOGY & EPISTEMOLOGY
Easier to think about in physics compared to social sciences, cf. Nelson 2016?
One could argue that we in physics ”uncover” reality and the mathematical,
‘100%’ correct laws that explain behavior of atoms, planets etc.
Social sciences?
Overview
Guba 1990, p. 18
https://journals.sagepub.com/doi/full/10.1111/1467-9256.00173
9
”YES, PRIME MINISTER” I
https://www.youtube.com/watch?v=G0ZZJXw4MTA
10
”YES, PRIME MINISTER”: IS THE PROBLEM ONTOLOGICAL
OR EPISTEMOLOGICAL, OR BOTH?
https://x.com/keiranpedley/status/1759970566776197307?s=20
12
TYCHO BRAHE’S MEASUREMENTS
OF THE SOLAR SYSTEM –
ONTOLOGICAL DISTINCTION
• Measurements of planets ≠ measurements of
attitudes, unemployment, stock markets etc.
• If we struggle to measure something (e.g. unemployment, innovation) and can’t perfectly predict
future (≈ effect size of e.g. 0.2), more difficult to assume ‘real-ness’ and ‘real knowledge’?
Therefore:
• Issue with our ability to measure and understand the world (epistemology) or are these concepts
just not independently real (ontology), cf. Prime Minister ex.?
Epistemology We can observe the world as Objectivity the ideal, theories Theories / constructs are constructed,
it is. Have access to theories / / constructs are (inter)subjective knowledge, findings
constructs, which represent approximations, can’t see the are creation of interaction between
and correspond to the world = world as it is ≈ Popper’s inquiry / inquired into
objective knowledge position
Methodology Uncover truth, fly on the wall, RCTs are good, but valuable We can’t go outside our theories,
(or investigator) see ‘through’ a microscope, to add ‘naturalistic settings’, don’t aim to see unfiltered world, re-
RCTs the best many sources reduces construct the world, intersubjective
distortions, aim to remove approach, interesting world is in the
filter, see ‘with’ a microscope minds of those we research
RICKY GERVAIS ON (POSITIVISTIC) SCIENCE
16
GUBA: POSITIVISM, SOME RELEVANT QUOTES
P. 19: ”The basic belief system of positivism is rooted in a realist ontology,
that is, the belief that there exists a reality out there, driven by immutable
natural laws. The business of science is to discover the ”true” nature” -
ontology
P. 19: "If there is a real world operating according to natural laws, then the
inquirer must behave in ways that put questions directly to nature and
allow nature to answer back directly” - epistemology
p. 19: ”But how can that be done, given the possibility of inquirer bias, on
the one hand, and nature's propensity to confound, on the other? The
positivist's answer: by the use of manipulative methodology that controls
for both, and empirical methods that place the point of decision with
nature rather than with the inquirer. The most appropriate methodology is
thus empirical experimentalism, or as close an approximation there to as
can be managed." - methodology
Same answer
Even if effect size is 0.2, that effect
represents something real
17
GUBA: POST-POSITIVISM, SOME RELEVANT
QUOTES AND INSIGHTS
P. 20: "...although a real world driven by real natural causes exist, it
is impossible for humans truly to perceive it with their imperfect
sensory and intellective mechanisms” - ontology
There is a real effect out there – but we can never ‘know’ if our theories are true
We can only compare validity of theory with validity of another theory, never Not necessarily
direct access to the world itself same answer
18
POPPER ON TRUTH (EPISTEMOLOGY)
https://youtu.be/li0ciaqJ0m0?t=155
19
GUBA: CONSTRUCTIVISM, SOME
RELEVANT QUOTES AND INSIGHTS
P. 25: ”…facts are only within some theoretical framework…basis for discovering ”how
things really are”…is lost”.
P. 26: ”Realities are multiple”. You can’t step outside your theoretical, constructed
framework - ontology
P. 26: "...the results of an inquiry are always shaped by the interaction of inquirer and
inquired into.”
p. 25: “…no unequivocal explanation is ever possible. There can be many constructions,
and there is no foundational way to choose among them.”
P. 26: ”…it makes the findings of an inquiry not a report about what is ”out there” but the
residue of a process that literally creates them [the findings]” - epistemology
P. 26: ” If realities exist only in respondents' minds, subjective interaction seems to be the
only way to access them..” Concepts are constructed (e.g. motivation, inflation), let's be
explicit about it and investigate how and why we have constructed it. Notice: not a direct
quote. - methodology
Epistemology We can observe the world as Objectivity the ideal, theories Theories / constructs are constructed,
it is. Have access to theories / / constructs are (inter)subjective knowledge, findings
constructs, which represent approximations, can’t see the are creation of interaction between
and correspond to the world = world as it is ≈ Popper’s inquiry / inquired into
objective knowledge position
Methodology Uncover truth, fly on the wall, RCTs are good, but valuable We can’t go outside our theories,
(or investigator) see ‘through’ a microscope, to add ‘naturalistic settings’, don’t aim to see unfiltered world, re-
RCTs the best many sources reduces construct the world, intersubjective
distortions, aim to remove approach, interesting world is in the
filter, see ‘with’ a microscope minds of those we research
ON CONSTRUCTIVISM I
https://twitter.com/alexgaggio/status/1023131258065436672?s=11
Each row orders the color chips for one language: Colors
farther left are easier to communicate, those farther to the
right are harder to communicate.
https://images.theconversation.com/files/186385/original/file-20170918-8258-
oz7qnm.jpg?ixlib=rb-1.1.0&q=45&auto=format&w=1000&fit=clip
ON CONSTRUCTIVISM II: TYPICAL
MISCONCEPTIONS
• Just because something is constructed – doesn’t mean it is made up
• Constructivism is about the process of constructing theories – not that the world is made up
• You can’t build any kind of theory – the world limits
• You can’t build a building without materials – similar with theories (there is a world/building material)
• Buildings can be easily destroyed – our institutionalized theories cannot (easily)
https://twitter.com/research_tim/status/993756255423430656?s=11
WHERE DOES KUHN FIT INTO GUBA’S FRAMEWORK?
VEVOX.APP 143-136-950 (CLICK ON IMAGE)
Epistemology We can observe the world as Objectivity the ideal, theories Theories / constructs are constructed,
it is. Have access to theories / / constructs are (inter)subjective knowledge, findings
constructs, which represent approximations, can’t see the are creation of interaction between
and correspond to the world = world as it is ≈ Popper’s inquiry / inquired into
objective knowledge position
Methodology Uncover truth, fly on the wall, RCTs are good, but valuable We can’t go outside our theories,
(or investigator) see ‘through’ a microscope, to add ‘naturalistic settings’, don’t aim to see unfiltered world, re-
RCTs the best many sources reduces construct the world, intersubjective
distortions, aim to remove approach, interesting world is in the
filter, see ‘with’ a microscope minds of those we research
KUHN AND GUBA
“However, the classification of Kuhn into the paradigm of constructivism is not perfect. A central claim of constructivism is
that due to the undeterminaiton of theory, i.e. that “no theory can ever be fully tested”, “no unequivocal explanation is
ever possible. There can be many constructions, and there is no foundational way to choose among them.” (Guba 1990,
pp. 25) This stands somewhat in contrast to Kuhn’s perspectives on scientific development. Indeed, while he argues that
the incommensurability of paradigms implies that both may be right because they cannot be compared; “The
proponents of different theories are like the members of different language-culture communities. Recognizing the
parallelism suggests that in some sense both groups may be right.” (Kuhn 1962, pp. 205) However, Kuhn does then go on
to argue – somewhat – that scientific progress means later theories are better for solving scientific “puzzles” than earlier
theories; “…scientific development is, like biological, a unidirectional and irreversible process. Later scientific theories are
better than earlier ones for solving puzzles in the often quite different environments to which they are applied.” (Kuhn
1962, pp. 206) This implies that Kuhn is not entirely aligned with the subjectivist epistemology of constructivism despite
seemingly being aligned with the relativist ontology of the paradigm. Thus, it seems that Kuhn does not adhere perfectly
to any of the three paradigms defined by Guba (1990), but that his perspectives on ontology places him closest to the
constructivist view on the social science world.”
26
NELSON 2016 AND PARADIGMS
(REFLECT 2 MIN)
”That is, the number, and the assumptions behind its construction, comes to be treated as what the phenomenon is,
rather than as a useful if limited measure of that phenomenon.” (Nelson 2016, p. 1695).
“The numbers used in the social and behavioral sciences almost always are, by themselves, somewhat limited and
imprecise characterizers of the phenomena they are designed to measure, and need to be understood as a part of a
broader and more detailed if qualitative characterization (Nelson 2016, p. 1695).
”And in these fields the kind of precise law-like relationships that physics has identified simply may not exist, even in
the sense of tightly stable stochastic relationships.” (Nelson 2016, p. 1693)
“If these phenomena are to be associated with numbers, some proxies need to be invented or some quantitative
indicators constructed that relate to phenomena of interest, but are not natural measures of that.” (Nelson 2016, p.
1696)
Guba 1990, p.26: ”First, it renders the distinction between ontology and epistemology obsolete; what can be known
and the individual who comes to know it are fused into a coherent whole””
A GOOD ANSWER TO 2020 Q5, PART 1
To categorise Nelson’s take on business economics as a science as per Guba (1990), I will need to first establish what Guba’s positions
are, as to be able to place Nelson in one (or several) of them. In his 1990 paper titled “The Alternative Paradigm Dialog”, Guba
discusses the three paradigms of positivism, post-positivism and constructionism. He discusses how the different paradigms will have
different answers to the ontological, epistemological and methodological questions (Guba 1990, page 18).
Positivism has the ontology that concepts and theories are real, i.e. there is one real “truth”. The epistemology is that the theories used
then represents the world as it is, and is therefore objective knowledge.
Post-positivism has almost the same ontology as positivism, in that there is one real “truth”. However, they are more critical about one’s
access to it, as they argue it is not reachable. Their epistemology is that objectivity then is the ideal, but we will never be able to see the
world as it is, and therefore our best bet is theories that approximate the real “truth” as closely as possible.
Constructivism has a completely different ontology, as this paradigm argues that there is several “truths”. As Guba puts it: “ “Reality”
exists only in the context of a mental framework (construct) for thinking about it” (Guba 1990, page 25). This also means, in relation to
constructivism’s epistemology, that theories are constructed, and that the constructed reality only then can be “seen” through this
window of constructed theory (Guba 1990, page 25).
At first, I establish why I do NOT consider Nelson to be a positivist. As described above, a positivist will argue that the theories they use
are a window straight into how the world really is, and is therefore objective knowledge. However, in his paper, Nelson several times
mentions that many variables in social science does not make sense to evaluate at face value: “A good part of the reason why one or
a few numbers alone generally do not cover adequately the subject matter being studied, and the numbers themselves often are
somewhat fuzzy, is that the phenomena studied have several aspects and each of these has blurry edges.” (Nelson 2016, page 1696).
Nelson argues that we cannot just use one number to define a variable such as unemployment, as unemployment as a subject is
inherently fuzzy. You can get different arguably correct answers to what one considers unemployment. Whether it be people out of
work, people out of work but who are willing to work etc. But you cannot get different answers to what the weight of a specific atom is
(obviously given the respondent knows the answer). And therefore, it does not make sense to view numbers or theories in the social
science field as a window into reality, as there needs to be accounted for their inherent fuzziness.
A GOOD ANSWER TO 2020 Q5, PART II
A quote that could place Nelson on the more constructivist side is the following: “Many of the particular subject
matters treated by the social and behavioural sciences not only are quite heterogeneous, but the general conception
of the nature of the phenomenon often has uncertain boundaries.” (Nelson, page 1695). In the quote, Nelson argues
that the nature of phenomena has uncertain boundaries, which I think hints at him meaning the phenomena are a
construct of the created data and the given researcher, and should thus be treated as such. However, one could also
argue that he says the uncertain boundaries means that the data are only an approximation to reality, which would
then seem like him being a post- positivist
In general, I would argue the way he addresses quantifiable measures makes him seem like a post-positivist, as it
seems as though some data/numbers do have a face value on their own, such as in the following quote: “The
numbers used in the social and behavioural sciences almost always are, by themselves, somewhat limited and
imprecise characterizers of the phenomena they are designed to measure, and need to be understood as part of a
broader and more detailed if qualitative characterization.” (Nelson 2016, page 1695). However, the way he
addresses that data needs to be understood as part of a broader qualitative characterization makes him seem more
constructivist, as it seems he relies more on the creation between the data and the researcher to create meaning.
To conclude, I would argue that Nelson’s 2016 paper is definitely not positivistic, as he does not believe theories are a
window of the one “true” reality. However, when evaluating whether he is post-positivistic or constructivist, I would
argue that the discussion gets blurrier, and that he is somewhere in between.
MORE ON METHODOLOGY
FOR EXAMPLE, HOW TO DO INTERVIEWS I
31
HOW TO DO INTERVIEWS – AS A
CONSTRUCTIVIST? II
A constructivist would
• probably not hide identity – but participate openly
• not focus on ‘contamination’ and ‘purity’
• seek out multiple, different understandings – and
understand why participants think/behave
• contrast responses from one with responses from
others (dialectic, p. Guba 26-27)
• while in PP aim of standardized approach is to be
able to smoothly compare respondents’ responses
– in C one accepts and uses their different realities
• rely on more open-ended, flexible approaches &
questions
33
CHATBOT ANSWER: “CAN YOU GIVE EXAMPLES OF HOW INTERVIEW QUESTIONS
WOULD DIFFER IN A POST-POSITIVISTIC VS CONSTRUCTIVIST APPROACH?
Certainly! Here are examples of interview questions that might differ between a post-positivist and constructivist approach:
Post-positivist approach: In a post-positivist approach, researchers attempt to identify patterns and relationships while
acknowledging the limitations of human knowledge. They strive for objectivity and often try to minimize their own biases. Interview
questions in a post-positivist approach might include:
Can you describe the steps involved in your decision-making process?
What factors do you consider when evaluating a new opportunity or project?
How do you measure the success of a particular initiative or project?
Constructivist approach: In a constructivist approach, researchers focus on understanding the subjective experiences and
meaning-making processes of the participants. They embrace the co-construction of knowledge between the researcher and the
participant and acknowledge their own role in shaping the research findings. Interview questions in a constructivist approach
might include:
Can you tell me about a time when you faced a difficult decision, and how you made sense of the situation?
How do your personal values and beliefs influence your evaluation of new opportunities or projects?
In what ways has your understanding of success evolved throughout your experiences with various initiatives or projects?
The main difference between the questions lies in the emphasis on the participant's subjective experiences and meaning-making
processes in the constructivist approach. While both approaches use interviews to explore participants' perspectives, the
constructivist approach places more emphasis on the individual's interpretation of events and personal context. In contrast, the
post-positivist approach focuses more on identifying patterns and relationships across participants' experiences, while still
acknowledging the limitations of human knowledge and the potential influence of context.”
34
POST-POSITIVISM VS CONSTRUCTIVISM: A FEW NOTES
Typical (PP) aim: Find out what is really the case
Most (?) bachelor projects: Some variant of post-positivism
• What is aim: To discover / carve out information (like a treasure-hunt/archeologist) – or to
produce/construct knowledge?
• If risk of interfering: Aiming to be as distant / invisible as possible – or can inteviewer also be
engaged and passionate (e.g. in interview situation)?
• Interviewguide: Standardized and how strictly followed?
• Considered necessary to quantitatively test (qualitative) insights, afterwards?
• How do you assess objectivity? ‘Truth’ vs trustworthiness
• ….
35
ANOTHER EXAMPLE: STUDYING THE VARYING
KINDS OF IMPACT OF BREASTFEEDING I
https://w0.pngwave.com/png/791/656/medical-laboratory-hospital-physician-science-science-png-clip-art.png
37
BREASTMILK IS GOOD FOR
THE BABY!
Breast milk ------- > Good for baby
https://lh3.googleusercontent.com/proxy/AawPg7iPTMOJez2tm
0RTBo6qJdfX0V7d4fuLpbFLUTRRdi_KB1qr_T7ahed0pxlzBTBFEIu
_jj9WHGP7PDKPgnPKUA
38
BUT…IT IS NOT SO SIMPLE
WHAT IS THE RESEARCH QUESTION?
Breastfeeding can be very painfull, and basically
impossible potentially loss of sleep & self-
esteem, + feeling of shame by mother
https://media.istockphoto.com/id/1346197512/vector/rate-us-bar-from-1-to-10-best-for-website-design-app-ui-isolated-on-white-
background-eps-10.jpg?s=612x612&w=0&k=20&c=RrUgDdl7SLsP2snleuivNDyslYbWbAZ91Xl3t4zGfjc=
39
• Norms of self-esteem and when to feel shame
vary across cultures ≠ physiological impact
https://media.istockphoto.com/id/1346197512/vector/rate-us-bar-from-1-to-10-best-for-website-design-app-ui-isolated-on-white-
background-eps-10.jpg?s=612x612&w=0&k=20&c=RrUgDdl7SLsP2snleuivNDyslYbWbAZ91Xl3t4zGfjc=
40
WHAT IS THE RESEARCH QUESTIONS (RQ)
PHILOSOPHY OF SCIENCE ASSUMPTIONS
When RQ shifts from physiological components to psychological
considerations
What paradigm applies – thus also depends on the RQ and the phenomena
the RQ seeks to investigate
Also: If one has certain (paradigm) assumptions, one addresses certain RQs
FINAL PERSPECTIVES ON
THEORY
Self-falsification? Did saying a theory out
loud change behavior? (Price and Wolfers
2010; Pope et al. 2013)
42
PARADIGMS AND METAPHORS OF THEORY I
44
PHILOSOPHY OF SCIENCE
LESSON 10
HUMAN ‘RATIONALITY’ AND BIASES
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
PHILOSOPHY OF SCIENCE ASSUMPTIONS ABOUT
HUMAN (RESEARCHERS)
Philosophy of Science: Usually focuses on assumptions about
research designs, what a theory is and what world we investigate
RD /
Humans?
Methods Social world
Theory
2
https://cdn-images-1.medium.com/max/2000/1*kQdOtDv-7KolqkWPyQD0kw.png
AGENDA FOR LESSON 10
• Where does ‘irrationality’ of economic thinking come from?
• Introducing system 1 & system 2
• What You See Is All There IS (cf. chapter 7)
• Systematic cognitive biases (e.g. confirmation bias) – not just random errors
• Efficient yet imperfect cognitive systems
• How to solve the flaws of system 1 & 2 – in general and in science
KAHNEMAN: THINKING, FAST AND SLOW (2011). NOBEL PRIZE
WINNER WHOSE WORK LED TO ”BEHAVIORAL ECONOMICS”
4
THE RISE OF ‘BEHAVIORAL ECONOMICS’ (≈ KAHNEMAN
INSPIRED LITERATURE) IN ECONOMIC LITERATURE
https://www.google.com/url?sa=i&url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3D
GdujVdjAKUg&psig=AOvVaw1RCiiuI24tFsbW-
1j3t2Jh&ust=1581924792471000&source=images&cd=vfe&ved=0CAIQjRxqFwoTCOCO1KDH
1ecCFQAAAAAdAAAAABAD
https://www.bloomberg.com/view/articles/2018-08-02/how-economics-went-from-philosophy-to-science
WHERE DOES THIS LACK OF RATIONALITY ORIGIN? I
https://image.slidesharecdn.com/behaviourchangeivovla
evclahrc-150515122817-lva1-app6892/95/what-is-
unique-about-behavioural-economics-prof-ivo-vlaev-30-
638.jpg?cb=1431693095
https://www.google.com/url?sa=i&url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DGdujVdjAKUg&psig=AOvVaw1RCiiuI24tFsbW-
1j3t2Jh&ust=1581924792471000&source=images&cd=vfe&ved=0CAIQjRxqFwoTCOCO1KDH1ecCFQAAAAAdAAAAABAD
6
WHERE DOES THIS LACK OF RATIONALITY COME FROM? II
NO EMOTIONS INVOLVED, BUT STILL NOT QUITE RATIONAL
https://image.slidesharecdn.com/goldpresentation-130107111506-phpapp01/95/how-behavioral-
economics-informs-website-design-5-638.jpg?cb=1357719312
7
WHAT DOES LINDA DO?
(REFLECT 1 MIN, VEVOX.APP 143-136-950)
“Linda is thirty-one years old, single, outspoken, and very bright,” they wrote. “She
majored in philosophy. As a student, she was deeply concerned with issues of
discrimination and social justice, and also participated in antinuclear
demonstrations.” Then they went around asking people the same question:
https://janav.files.wordpress.com/2013/11/lindafixed.jpg
‘IRRATIONALITY’ OF HUMAN THINKING
WHERE DOES IT COME FROM? (P. 8)
https://www.google.com/url?sa=i&url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3F
v%3DGdujVdjAKUg&psig=AOvVaw1RCiiuI24tFsbW-
1j3t2Jh&ust=1581924792471000&source=images&cd=vfe&ved=0CAIQjRxqFwoTCOCO
1KDH1ecCFQAAAAAdAAAAABAD
KAHNEMAN (2011): SYSTEM 1 AND 2
10
SYSTEM 1 VS. SYSTEM 2 (S. 19-22): TWO DIFFERENT
SYSTEMS / FORMS OF HUMAN THINKING I
2+2 = 4 17*24 = ?
19
SYSTEM 1 VS
SYSTEM 2
• ”[some] mental activities become
fast and automatic through
prolonged practice” (p. 22)
13
HOWEVER, THESE SYSTEMS LEAD TO SYSTEMATIC
MISTAKES / TENDENCY TO MAKE ERRORS
System 1 shortcuts + system 2 laziness errors
https://thumbs.dreamstime.com/z/biases-human-mind-pictured-as-word-inside-head-to-
symbolize-relation-psyche-d-illustration-172338048.jpg
14
WHAT DO WE SEE HERE?
15
THE ‘STROOP’ TEST: WE CAN FAIL AT TASKS THAT SEEM QUITE SIMPLE
Kahneman 2011, ch 7, p. 85
Kahneman 2011, ch 7, p. 87
18
19
WE DON’T JUST SEE THE WORLD AS IT IS
ATTENTION SHAPES PERCEPTION TOP DOWN FILTERING OF INFORMATION
WYSIATI
https://twitter.com/Foone/status/1014273847518851073?s=20
20
TOP DOWN ‘FILTERING’ I: EXPECTATIONS MAKE US MORE
READY FOR SOME KIND OF INFORMATION
https://twitter.com/page_eco/status/14
61306098757947396?s=20&t=qZ2hAz8r
Zxex3lDLwXmBMw
21
TOP DOWN ‘FILTERING’ II
https://twitter.com/page_eco/status/1155077311399489536?s=20
SLIDE BORROWED FROM JACOB ORQUIN
KAHNEMAN ON THINKING FAST AND SLOW
https://www.youtube.com/watch?v=iVrLsN5zM_A
24
SYSTEM 1 VS. SYSTEM 2: OVERVIEW
http://ameritest.wordpress.com/2013/01/29/completing-the-picture-part-ii-system-1-vs-system-2-thinking0/
Irrationality does not only come from lack of cognitive effort or interrupting emotions –
but are an integrated feature of our brain’s cognitive setup
Kahneman: Important to understand the systematic nature of these mistakes, i.e. biases
SYSTEM 1: NOT JUST RANDOM ERRORS, BUT
SYSTEMATIC COGNITIVE BIASES
https://www.castoredc.com/blog/cognitive-bias/
26
NON-SYSTEMATIC LIST OF COGNITIVE BIASES
(EMERGING FROM SYSTEM 1, LACK OF SYSTEM 2 ACTIVATION)
http://lifehacker.com/this-graphic-explains-20-cognitive-biases-that-affect-y-1730901381?sidebar_promotions_icons=testingon&utm_expid=66866090-
67.e9PWeE2DSnKObFD7vNEoqg.2
ANCHORING BIAS & CONFIRMATION BIAS
• Scenario: Two headlines on a scientific study appear on your
feed—one confirms, the other contradicts your beliefs.
• Inclination: We naturally gravitate towards information that
supports our existing views.
• Impact: This bias affects not only our perception but also the
information we choose to share and endorse.
Shapes what information we process & thus conclusions
CONFIRMATION BIAS (NOT JUST EMOTIONS, SOCIAL CONFORMITY,
CLASSICAL IRRATIONALITY THAT DRIVES BIAS)
https://www.all-about-psychology.com/selective-perception.html
https://twitter.com/dsquintana/status/1488815799427022849?s=12
30
WHY DOES OUR SOPHISTICATED BRAIN ‘SUFFER’
FROM CONFIRMATION BIAS?
System 1 is lazy:
• We prefer to process information that matches
expectations - easier for new bit of information to find ”its
place” in the brain
• Processing information that challenges beliefs, requires
more energy and is ‘risky’
31
CONFIRMATION BIAS: NEGATIVE VS. POSITIVE
FEEDBACK PROCESSED DIFFERENTLY
Emotions and / or
Cognition?
32
THIS IS NOT NEW:
HUMANS HAVE KNOWN
ABOUT BIASES FOR
CENTURIES
34
BIASES - AND PRIOR PHILOSOPHY OF SCIENCE
CURRICULUM II
Kahneman 2011, ch 7, p. 81
35
HOWEVER: BIASES AREN’T JUST BAD
36
SYSTEM 1 AND 2: EFFICIENT YET IMPERFECT
COGNITIVE SYSTEMS
• Why these systems? Why is (lazy) system 1 default?
• Evolutionary perspective
• When humans ran away from wild animals quick system 1
decision necessary
• Confirmation bias strenghtens groups
• System 2 requires energy, energy is expensive
• Thus, we can’t constantly question ourselves (p. 28) efficient
heuristics needed.
• The system is efficient – just has biases/errors
• …
38
SO, WE HAVE SYSTEMATIC BIASES. HOW TO REDUCE
BIASES AND ‘IRRATIONALITY’? I (REFLECT 1 MIN)
39
HOW TO REDUCE BIASES? (PRE-KAHNEMAN)
• Pre-Kahneman, homo economicus assumption: Irrationality stems from lack of
incentives / information or emotions
40
INCENTIVES CAN’T EASILY FIX IT THOUGH
41
POST-KAHNEMAN: DIFFERENT PROPOSALS ON
HOW TO FIX BIASES
How to reduce risk of “falling” for system 1’s biases via focus on cognition / thinking
2) Nudges / choice architecture: Change the way options and information is presented
42
1) DEBIASING: ENHANCING CRITICAL
THINKING (IN SCIENCE)
a) Proactively seek differing viewpoints to
challenge your own (disconfirm)
b) Rely on evidence, not mere intuition, to support
claims (be aware of (your) bias)
c) Carefully analyse information to ensure it
substantiates your claims (correlation vs.
causation e.g.)
d) Aim for objectivity by slowing down thinking &
controlling emotions (system 2 activation)
e) Clearly define terms to reduce
misunderstandings (our world is ambiguous, cf.
Nelson)
…
43
2) NUDGING: CHANGE THE DEFAULT
SETTING OF A CHOICE SITUATION
• Other books:
• Thaler: Misbehaving (Nobel Prize 2017 winner)
• Galef: The Scout Mindset (2021)
LESSON 11
SCIENTIFIC PROCESSES AND OBJECTIVITY
Confirmation bias:
Both about how we collect /
process / obtain information
2
LESSON 11: AGENDA
• Aschwanden 2015
Lesson 7-9
• Science is difficult I – politics example The principles of why it is difficult
• Is science broken due to cheating & biases? to reach true objectivity
• Science is difficult II – soccer example
• Is there a right scientific analysis of a study? Lesson 11
Practice-oriented look at the
• Science as a collective process & what kind of truth?
objectivity challenge and options
to solve it
• Harrer 2021
• What is a meta-analysis
• Strenghts and pitfalls
3
ASCHWANDEN 2015: IS SCIENCE BROKEN?
Science troubled by
• Replication crisis
• Fraudelent papers getting published
• Researchers have biases
• Inconsistent results
• ….
4
MANY ORIGINAL STUDIES DO NOT REPLICATE?
(WHEN REDONE BY OTHERS)
http://theneuroeconomist.com/2016/03/is-the-
replication-crisis-in-psychology-real/
https://science.sciencemag.org/conte
nt/349/6251/aac4716
WHY THIS LACK OF REPLICATION?
Shouldn’t science have institutions, peer review processes and criteria that ensures that only true /
good research gets published?
6
PEER REVIEW CAN’T FIX EVERYTHING: CHEATING,
(BAD?) INCENTIVES AND HUMAN BIASES
Partially due to fraud / cheating (see lesson 12)
• Some papers get retracted (0.02%) most due to fraud
https://twitter.com/lakens/status/133522764
5135564800?s=20
7
P-(VALUE) HACKING
“The variables in the data sets you used to test your hypothesis had 1,800 possible combinations. Of these, 1,078 yielded
a publishable p-value,1 but that doesn’t mean they showed that which party was in office had a strong effect on the
economy. Most of them didn’t.” (Aswchwanden 2015, https://fivethirtyeight.com/features/science-isnt-broken/#part1)
8
CHARLES BABBAGE, IN 1830:
REFLECTIONS ON THE DECLINE
OF SCIENCE IN ENGLAND
An old problem!
https://twitter.com/lakens/status/1305153073258561542?s=20
9
ASCHWANDEN 2015: SCIENCE IS NOT BROKEN
Yet also better reasons for why science studies do
not always replicate:
https://fivethirtyeight.com/features/science-isnt-broken/#part1
HOWEVER: WHY SAME DATA, DIFFERENT CONCLUSIONS?
IS THE CONCLUSION TRUE? ARE DARK-SKINNED PLAYERS
MORE LIKELY TO GET A RED CARD? (VEVOX.APP 143-136-950)
• Yes, true
• No, not all studies show this result. Maybe they were (politically,
cognitively, emotionally) biased?
• Well, what does ‘truth’ mean?
NO BIAS? WELL, NO PRIOR SYSTEMATIC DIFFERENCES
BETWEEN THE 29 TEAMS
For example
Research team’s prior (personal) beliefs about an expected relation unrelated to result
Fig. 3. Point estimates (clustered by analytic approach) and 95% confidence intervals for the effect of soccer players’ skin tone on the
number of red cards awarded by referees. Reported results, along with the analytic approach taken, are shown for each of the 29
analytic teams. The teams are clustered according to the distribution used in their analyses; within each cluster, the teams are listed in
order of the magnitude of the reported effect size, from smallest at the top to largest at the bottom. The asterisks indicate upper bounds
that have been truncated to increase the interpretability of the plot (see Fig. 2). OLS = ordinary least squares; WLS = weighted least
squares; Misc = miscellaneous.
DOI: (10.1177/2515245917747646)
IS THERE EVEN SUCH A THING AS THE RIGHT ANALYSIS?
WELL, SOME CHOICES ARE BETTER THAN OTHERS
https://twitter.com/lakens/status/1166049199470850049?s=20
DIFFERENT INTERPRETATION OF RESEARCH
QUESTION DIFFERENT CHOICES
”In 2018…published an article in which 29 teams analyzed the same research question with the same
data: … In this article, we investigate why results diverged so much. We argue that the main reason was
an unclear research question: Teams differed in their interpretation of the research question and
therefore used diverse research designs and model specifications. We show by reanalyzing the data
that with a clear research question, a precise definition of the parameter of interest and theory-guided
causal reasoning, results vary only within a narrow range. The broad conclusion of our reanalysis is that
social science research needs to be more precise in its “estimands” to become credible.”
https://doi.org/10.1177/23780231211024421
17
ANOTHER EXAMPLE: MANY ANALYSTS, MANY (QUITE?)
DIFFERENT RESULTS
https://twitter.com/tomstafford/status/1377131265405505546
18
SCIENTISTS ALSO AGREE ON MANY QUESTIONS
IMPRESSIVE WHEN THEY DO, GIVEN POTENTIAL FOR VARIATION…
https://en.wikipedia.org/wiki/Scientific_cons
ensus_on_climate_change
19
SAME (RED CARD) DATA, MOSTLY SIMILAR CONCLUSIONS
• “On the other hand, it also suggests there’s
a there there. It’s hard to look at that data
and say there’s no bias against dark-
skinned players.”
(Aschwanden 2015)
SCIENTIFIC DEVELOPMENT:
COLLECTIVE PROCESS
"Since people are fallible, and often obstinate and overly fond of
their own ideas, the objectivity of the process which tests
conjectures lies not in the emotional detachment and
impartiality of individual scientists, but rather in the scientific
community being organized in certain ways, with certain
institutions, norms and traditions, so that individuals’ prejudices
more or less wash out (Popper, 1945, Chapters 23–24)."
http://www.stat.columbia.edu/~gelman/research/published/philosophy.pdf ≈ Kahneman, Kuhn
23
KUHN VS ASCHWANDEN: WHAT DOES THE WORD
‘TRUTH’ ACTUALLY MEAN?
But the scientists have not realized that Kuhn’s claim is based on a philosopher’s
definition of “truth,” which is not achieved (if ever) until there is exact
knowledge of the ultimate constituents of matter on the quantum-gravity scale.
Scientists use a less demanding definition for the word “truth,” in which
measured parameters can be subject to nonzero error bars.”
Kenneth G. Wilson, in Physics Today 54, 3, 53 (2001)
Kuhn
Aschwanden
24
A RELATED METAPHOR, ON TRUTH
"The empirical basis of science has nothing 'absolute'
about it. Science does not rest upon solid bedrock. The
bold structure of its theories rise, as it were above a
swamp. It is like a building erected on piles. The piles
are driven down from above into the swamp, but not
down to any natural or 'given' base; and if we stop
driving the piles deeper, it is not because we have
reached firm ground. We simply stop when we are
satisifed that the piles are firm enough to carry the
structure at least for the time being.”
(Popper 1959, p. 111 quoted from Duberley & Johnnson 2000, )
25
SCIENCE AS UNCERTAINTY REDUCTION I
WE OFTEN SHOULDN’T WAIT FOR 100% CERTAINTY
"Science is not a magic wand that turns everything it touches to truth. Instead, “science
operates as a procedure of uncertainty reduction,” said Nosek, of the Center for Open
Science. “The goal is to get less wrong over time.” (Aschwanden 2015)
https://twitter.com/erikleejohnson/status/1357413415606755334?s=20
Reflekter 2-3 min: Hvad er Aschwandens (2015) primære budskab?
Yes, researchers have biases, some cheat, some are sloppy. But
Question 3 (20%)
i) Please identify a quote in Aschwanden (2015) that you think has
the best fit with the overall message in Nelson (2016). Provide an
explanation for how the quote fits into Nelson's overall message. ii)
Please also identify a quote in Aschwanden (2015) that seems to be
in disagreement with Nelson (2016).
29
ASCHWANDEN AND NELSON: 2021 GRADING GUIDE
Several quotes can be applied in i). The following lists some of the most obvious ones. Again, others can be applied. Arguments should be
presented, both on Nelson’s position (not elaborated here), but also on how the quote links to Nelson.
• ”That’s because answering even a simple scientific question — which party is correlated with economic success — requires lots of
choices that can shape the results. This doesn’t mean that science is unreliable. It just means that it’s more challenging than we
sometimes give it credit for.” - this is probably the most ideal quote
• ”Even the most skilled researchers must make subjective choices that have a huge impact on the result they find”
• ”If we are going to rely on science as a means for reaching the truth – and it’s still the best tool we have – it’s important that we
understand and respect just how difficult it is to get a rigorous result.”
• “As a society, our stories about how science works are also prone to error. The standard way of thinking about the scientific method is:
ask a question, do a study, get an answer. But this notion is vastly oversimplified”
• ”The important lesson here is that a single analysis is not sufficient to find a definitive answer.”
• ”The uncertainty inherent in science doesn’t mean that we can’t use it to make important policies or decisions. It just means that we
should remain cautious and adopt a mindset that’s open to changing course if new data arises.”
Some rely on the following quote: “The state of our science is strong but it’s plagued by a universal problem: Science is hard - really fucking
hard.” Nelson would not disagree as such, but since this quote is not specific, using this quote requires one to specify what exactly
Aschwanden means, in order to enable a direct comparison to Nelson’s position.
31
FROM ASCHWANDEN TO
HARRER ET AL. 2021:
META-ANALYSES
https://www.stevestewartwilliams.com/p/exercise-can-be-more-effective-than
HOW TO MAKE SENSE OF A
DIVERSE RESEARCH FIELD?
https://www.youtube.com/watch?v=i675gZNe3MY
COMPARING APPLES AND ORANGES? (P. 7) I
COMPARING APPLES AND ORANGES? II
https://datacolada.org/103
But what kind of ”choice architecture interventions” (cf. Kahneman) are we comparing?
Intervention in one study: Menu presentation Another study: Serving sizes Very different
interventions / variables
(cf. Nelson 2016)
https://datacolada.org/105
COMPARING APPLES AND ORANGES? III
Scenario C….
One can check how symmetric published results are: more Result: Published studies are biased
results just barely significant, than just barely non-
significant?
A post-positivistic approach
Focused on quantitative challenges
Modified from:
https://www.researchgate.net/publication/336191285/figure/fig3/AS:8437196705054
73@1578169853483/ASPAN-hierarchy-of-evidence-ASPAN-American-Society-of-
PeriAnesthesia-Nurses-O-Laura.jpg
SUMMARY LESSON 11
• Science is complicated and produced by biased researchers that sometimes cheat
• Not all science results replicate non-trustworthy science? Can we make anything up?
• Aschwanden: How one dataset can be analyzed in many ways if most agree,
probably inching towards truth
• From individual subjectivity to collective intersubjectivity / ‘objectivity
• While multiple choices might be right, many will be wrong
• Trust many studies, not one – yet, how to interpret many studies?
LESSON 12
SCIENTIFIC INTEGRITY
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
META-ANALYSIS VS. INDIVIDUAL STUDIES:
REDUCING RISK OF SELECTION/CONFORMITY BIAS
https://twitter.com/NBTiller/status/1765376104125038940/photo/1
2
OVERVIEW: REST OF THE COURSE
‘Lesson 16’: Wednesday the 20th of March I am available on-site + Zoom meetings.
First hour of lesson in Fuglesangssalen, second hour of lesson on Zoom:
https://aarhusuniversity.zoom.us/my/bergenholtz (see info on Brightspace)
3
CONTINUOUS Q&A: VEVOX.APP 143-136-950
4
LESSON 12: AGENDA
iv) Flowlock+
5
Following introducion of scientific misconduct also
available on video here:
https://youtu.be/uIRFJlp0V2o
6
WHY IS IT – PARTICULARLY - IMPORTANT THAT SCIENCE
HAS HIGH INTEGRITY?
Science
Good science Bad science
7
INFAMOUS EXAMPLE: SCIENTIFIC MISCONDUCT’S
SERIOUS CONSEQUENCES
Vaccine and autism – a correlation? No!!
• Wakefield 1998 Lancet study showing correlation – later retracted
• Personal economic interests + fabricated data + distorted message + important impact on
science = scientific misconduct
http://tallguywrites.livejournal.com/148012.html
8
https://www.wsj.com/articles/SB10001424127887323300004578555453881252798
GOOD SCIENCE VS BAD SCIENCE VS SCIENTIFIC
MISCONDUCT/DISHONESTY I
Bad science
Misconduct=videnskabelig uredelighed
• Kuhn: Paradigms and norms develop research freedom essential. What seems like
misconduct to some, might be ok to others?
Scientific freedom refers in particular to the freedom to choose research topic, the
freedom to ask questions, the freedom to choose materials and methods to find the
answers, and freedom to publicly present hypotheses, results and reasoning
https://forskerportalen.dk/en/freedom-of-research/
9
BAD SCIENCE VS. SCIENCE MISCONDUCT/DISHONESTY II
”Section 3: The Committees shall not be entitled to consider cases involving the validity or
truth of scientific theories or cases involing the research quality of a scientific product.”
Intention
We often change/omit
data – has to be mislead
to be misconduct
Misleading
Plagiarism
12
EXAMPLE OF SCIENTIFIC MISCONDUCT I: PENKOWA CASE
• Scientific misconduct
• Fabricated data
• Cheated with funding sources
13
DOES THIS EXAMPLE CONSTITUTE SCIENTIFIC MISCONDUCT?
(DISCUSS 1-2 MIN)
https://x.com/literalbanana/status/1744406652680622205?s=20
14
https://slate.com/technology/2018/02/how-brian-wansink-forgot-the-difference-
between-science-and-marketing.html
WANSINK’S FOOD RESEARCH – PROBABLY MISCONDUCT/DISHONEST
https://www.vox.com/science-and-health/2018/9/19/17879102/brian-wansink-
cornell-food-brand-lab-retractions-jama
https://slate.com/technology
/2018/02/how-brian-
wansink-forgot-the-
difference-between-science-
and-marketing.html
http://news.cornell.edu/media-relations/advisories/cornell-university-statement-regarding-questions-about-professor-brian
15
A RESULT BASED ON SCIENTIFIC MISCONDUCT CAN STILL BE TRUE
Scientific misconduct is about the process, not the result
https://x.com/mbialek82/status/1720472868641735148?s=20
16
WHAT ONE SHOULD KNOW (TRAFFIC RULES VS. ALL
STATISTICAL METHODS)
17
WHAT ARE THE POTENTIAL CONSEQUENCES OF
MISCONDUCT?
USA
19
NOTICE ANYTHING
PARTICULAR ON THIS LIST?!
20
https://retractionwatch.com/the-retraction-watch-leaderboard/
GINO: AN ILLUSTRATIVE EXAMPLE
THE DIFFICULTY OF ASSESSING MISCONDUCT
21
DATA COLADA 2023 I
Investigation of study 1 (of a paper ON HONESTY !) with 4 studies, where study 3 has already been
shown to have been fabricated
22
DATA COLADA 2023 II
THE EXCEL FILE: MANUALLY ADDED DATA
• Some data entries are out of sequence / duplicates
• All those entries strongly (!) support the hypothesis
• Excluding those entries = no significant result
23
DATA COLADA III
COULD THIS BE EXTREME RANDOMNESS?
24
GINO’S DEFENSE
https://x.com/justsomeoneDK/status/1687064352522657793?s=20
25
STATUS OF THE CASE
• Gino is currently on academic leave from Harvard University
• All papers where she has collected data are being investigated
• Gino has sued Data Colada for 25 million dollars (money alreay been crowdfunded)
• Someone fabricated data & thus cheated. Who. Gino? A research assistant?
incredibly difficult to prove gross negligence / disohnest intentions
26
HOW TO REDUCE SCIENTIFIC MISCONDUCT?
SOME SUGGESTIONS
Data transparency
https://www.aeaweb.org/journals/data/data-code-policy
“Plagiarism is using someone else's material as if it were your own (books, articles, films, web pages,
tables, figures, etc.), without referring explicitly to the original source. It is plagiarism when you
knowingly cheat – if, for example, you borrow someone else's work and submit it as your own - but it is
also plagiarism if your references to sources are imprecise and incomplete.”
https://library.au.dk/en/students/plagiarism
28
PLAGIARISM IN MUSIC I
https://www.youtube.com/watch?v=R5sO9dhPK8g 29
PLAGIARISM IN MUSIC II
NOT THE SAME, BUT STILL TOO SIMILAR
https://www.youtube.com/watch?v=wUgNJ7Qdst8 30
PLAGIARISM IS DISTORTING THE MESSAGE
“Plagiarism involves using someone else’s words or ideas without giving proper credit…to
the author of the original…’the false assumption of authorship’” (Carter 2000).
“Using texts, ideas etc. produced or conceived by others does not in itself constitute
plagiarism. Plagiarism only arises when the examiner(s) are led to believe that you are the
author or originator.” https://studerende.au.dk/fileadmin/studerende.au.dk/studerende/Avoid_cheating_on_exams.pdf
31
2 FORMS OF PLAGIARISM
32
EXAMPLE OF REFERENCING PROPERLY
You could reference this exam by e.g. writing (Philosophy of Science Brightspace: 2020 exam question 4 answer, p. 9)
33
EXAMPLE FROM CARTER 2000 I (VEVOX.APP 143-136-950, 2 MIN)
The major events in Indian-white relations in Major events in Native American-settler relations
these years were military and legal: more than during these years were military and legal, and
two hundred pitched battles, not to speak of were shaped in important ways by both the US
guerilla warfare in outlying Western regions, Army and the Federal legislature. There were more
and a reservation policy promulgated in 1887, than two hundred pitched battles as well as guerilla
to remain in effect until the 1930's. In each case, warfare in remote Western areas, and a reservation
military action and legal solution, economic policy which remained in effect from 1887 until the
and cultural issues figured as unspoken but vital 1930's. In the case of both legal solution and military
imperatives. The drama of these years was action, cultural and economic issues were unspoken
played largely on the Great Plains, but against but vital imperatives. The drama of the period,
a scenery put in place by the forced removal or which took place mostly on the Great Plains, had as
expulsion of tribes East of the Mississippi in the its context a scenery put in place by the ante-
decades before the Civil War. bellum expulsion or forced removal of tribes living
(original) East of the Mississippi River.
34
EXAMPLE FROM CARTER 2000 II (VEVOX.APP 143-136-950, 1-2 MIN)
The major events in Indian-white relations in these Alan Trachtenberg notes in The Incorporation
years were military and legal: more than two of America that the major events in Indian-
white relations in these years were military and
hundred pitched battles, not to speak of guerilla
legal, with over two hundred pitched battles as
warfare in outlying Western regions, and a well as guerilla warfare in outlying Western
reservation policy promulgated in 1887, to remain regions. In each case, military action and legal
in effect until the 1930's. In each case, military solution, he adds, economic and cultural issues
action and legal solution, economic and cultural were latent but crucial imperatives. Though
played out largely on the Great Plains, the
issues figured as unspoken but vital imperatives.
drama of these years took place 'against a
The drama of these years was played largely on scenery put in place by the forced removal or
the Great Plains, but against a scenery put in place expulsion of tribes East of the Mississippi in the
by the forced removal or expulsion of tribes East of decades before the Civil War.' 1
Footnote linking to original source
the Mississippi in the decades before the Civil War.
(original)
35
EXAMPLE FROM CARTER 2000 III (VEVOX.APP 143-136-950, 1-2 MIN)
The major events in Indian-white relations in these The relationship between Native Americans
years were military and legal: more than two and settlers in the trans-Mississippi West
hundred pitched battles, not to speak of guerilla during the late nineteenth century was
marked by military conflict and legal
warfare in outlying Western regions, and a
coercion. As Alan Trachtenberg notes, there
reservation policy promulgated in 1887, to remain were “more than two hundred pitched
in effect until the 1930's. In each case, military battles, not to speak of guerilla warfare in
action and legal solution, economic and cultural outlying Western regions, and a reservation
issues figured as unspoken but vital imperatives. policy promulgated in 1887, to remain in
effect until the 1930’s.” (Trachtenberg 1982, p.
The drama of these years was played largely on
29). But if the relationship was played out on
the Great Plains, but against a scenery put in place the battlefield and in response to legislative
by the forced removal or expulsion of tribes East of action, it was also shaped in significant ways
the Mississippi in the decades before the Civil War. by economic and cultural concerns. 2
Footnote linking to original source
(original)
36
(FORMER) HARVARD UNIVERSITY PRESIDENT
(NOW RESIGNED, DUE TO THIS SCANDAL)
38
PLAGIARIZE: ”THEY’LL NEVER NOTICE?” - OURIGINAL
• Automatic system to find original sources
• https://medarbejdere.au.dk/en/administration/au-studies-
administration/sagsbehandlingsportal/eksamen/plagiarism-detection
• Serious sanctions
• Don’t want to create unnecessary anxiety!
• Reference your points and ideas, if needed
• Produce independent wording
• Please do get inspired
• No need to provide reference for every term
Remember: Cite group notes! Might be used word-by-word by someone else (applies to notes
found online or shared among few people)
E.g. write: (Study group notes 2024)
SHORT BSS VIDEO (BY ME) ON AVOIDING PLAGIARISM
English: https://youtu.be/BrJ0UnAgZn0
Danish: https://youtu.be/Hx5vBIFezg4
40
2020 SPRING EXAM, QUESTION 6 (DISCUSS 3-4 MIN)
Spørgsmål 6 (10%)
Aschwanden (2015) beskriver et studie, hvor 29 teams af forskere analyserede det samme datasæt for at vurdere, om
spillere med en mørkere hudfarve fik flere røde kort. De 29 teams genererede forskellige resultater, om end de generelt
indikerede den samme tendens; spillere med en mørkere hudfarve havde en tendens til at få flere røde kort. Læs venligst
følgende hypotetiske beskrivelser og vurder – baseret på den angivne information – om beskrivelserne implicerer
videnskabelig uredelighed. Gør rede for din vurdering for hver af de 2 beskrivelser.
A) One of the teams had a strong, prior assumption that a darker skin colour would not lead to more red cards. They considered
several different models, and were told by multiple colleagues during the process that the model they selected was probably
inferior, although the team disagreed. Just before they submitted their article, they realised they had left out a major chunk of the
data, which meant they had only analysed some of the data. They submitted the article after all, not disclosing their mistake, since
including the missing data would not have changed their results. They ended up arguing for a model that was statistically
insignificant.
B) One of the teams was funded by an international soccer organisation. The research team would only get the funding if they
could show that there was no relation between skin colour and getting a red card. The team did not disclose their funding source,
and ended up picking a model that fitted their expectations the best. In order to achieve the results they had to leave out parts of
the data that did not fit their interpretation. They ended up arguing for a statistically significant model that very clearly showed there
was no relation.
Good exam answer will be uploaded to Brightspace
41
FROM 2020 ‘CORRECTION GUIDE’
All answers should explicitly rely on a relevant conceptual framework, from the curriculum. The definition
Denmark / the curriculum has worked with, is that scientific dishonest behavior a) is intentional or grossly
negligent and b) misleads the scientific message. Also fine if an answer has referred to Resnik and Stewart
(2000), in particular their ‘decision tree’ figure, that includes a definition of misconduct. It should be clear that
misconduct does not equal errors nor disagreement.
In order to pass, the answer should identify that case b) demonstrates intentional, biased behavior and
leaving out the data, misleads the scientific message. We know from Aschwanden 2015, that out of the 29
research teams, none found a significant negative effect, so the researchers in scenario b) found a quite
particular finding.
Case a) could be argued to imply intentional or grossly negligent behavior, but the behavior does not
mislead the scientific message. According to the Danish rules, one should probably not categorize the
behavior as scientific misconduct/dishonest (videnskabeligt uredeligt), even though it is certainly a
questionable scientific practice.
42
SUMMING UP: SCIENTIFIC MISCONDUCT
• Dishonest & gross negligence: How to demonstrate researcher knew / should have known?
• Gino: Dishonesty can be very difficult to prove
• Plagiarism: “Using texts, ideas etc. produced or conceived by others does not in itself
constitute plagiarism. Plagiarism only arises when the examiner(s) are led to believe that
you are the author or originator.” (Carter 2000).
43
GDPR: GENERAL DATA PROTECTION REGULATION
https://commission.europa.eu/law/law-topic/data-protection/data-protection-eu_en
https://studerende.au.dk/en/it-support/information-
security/data-protection-gdpr
44
SOME EXAM INFO I
• 23th of March: 9-14 (don’t hand in a second late!)
• Be familiar with the Flowlock+ system (only PDF
documents, no internet) – more info in 2 slides
45
(SOME) GENERAL CRITERIA FOR A GOOD EXAM ANSWER
(I’LL RETURN TO THIS LATER)
• Read the question carefully. And then re-read it.
• Make sure you answer the question (surprisingly many forget/don’t)
• Rely on the curriculum, avoid common sense approach (if you could have written it without relying on the curriculum,
probably not relevant answer)
• Identify what the question is about (if ”strength and weaknesses of an experiment” probably validity concepts)
• Key challenge for poor exams: Too superficial, so be specific and provide reasons (e.g.: ”External validity is
limited because they relied on students”.)
• Be specific and detailed: a) link directly to the question, b) show which part of the curriculum is relevant
for answering a given question (don’t just repeat general terminology – show how is it relevant for THIS situation / question?)
• Signal you have read the curriculum, by quoting / referencing curriculum when possible (maybe even quote a
snippet not shown on a slide)
46
REMINDER: FLOWLOCK+ SETUP
NOTES: PDF + OWN COMPUTER
Why Flowlock+? Only way to ensure you can access notes / documents
47
ACCESSING WISEFLOW
NEW UPDATE -> LOG IN AGAIN
• AU implemented new Wiseflow system March 1: https://europe.wiseflow.net/login
• Same link: Therefore, log out of current/old Wiseflow (need to use different browswer/delete history?)
48
TRY OUT FLOWLOCK+ SYSTEM I
ACCESSING THE WISEFLOW SYSTEM (SEE ALSO BRIGHTSPACE FOR THE
FOLLOWING INFORMATION)
49
TRY OUT FLOWLOCK+ SYSTEM II
OVERVIEW OF FUNCTIONS
50
TRY OUT FLOWLOCK+ SYSTEM III
ACCESSING PDF DOCUMENTS
51
COPYING FROM PDF DOCUMENTS
52
TRY OUT FLOWLOCK+ SYSTEM IV
WHITE-LISTED WEB PAGES (DICTIONARIES)
53
PHILOSOPHY OF SCIENCE
LESSON 13
COMPLEXITY IN SOCIAL SCIENCE
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
FULL EXAM ANSWERS UPLOADED
2
SUMMING UP SCIENTIFIC INTEGRITY
Scientific misconduct?
• Assess intention / gross negligence
3
LESSON 13: AGENDA
4
PUZZLE: CURRENT ECONOMIC PARADIGMS HAVE
LIMITED EXPLANATORY / PREDICTIVE POWER – WHY?
Many economic phenomena and outcomes can’t be predicted, or fully explained – why?
Because
• The unexplained part is random/noise?
• Simply have not discovered true relation
yet?
• Or need entirely new approach?
5
WATTS 2007: P 489
SOCIAL SCIENCE IS REALLY
DIFFICULT/HARD
6
ASCHWANDEN 2015 ANSWER: SOCIAL SCIENCE IS
COMPLICATED / HARD
Complicated due to many variables / links between
“The variables in the data sets you used to test your hypothesis had 1,800 possible combinations. Of these, 1,078 yielded
a publishable p-value,1 but that doesn’t mean they showed that which party was in office had a strong effect on the
economy. Most of them didn’t.” (Aswchwanden 2015, https://fivethirtyeight.com/features/science-isnt-broken/#part1)
WATTS 2007 ANSWER
(P. 489)
9
TERMINOLOGY: COMPLICATED VS COMPLEX I
10
THREE CHARACTERISTICS OF A COMPLEX SYSTEM
SULLIVAN 2011
• Heterogeneous agents: Decisions evolve and adapt over time
• Interaction: Agents are interdependent
• Emergence: The whole is greater than the sum of its parts
Aim of lecture: Explaining
not completely predictable, not complete chaos these characteristics and
their implications
11
COMPARISON: COMPLICATED VS COMPLEX SYSTEMS I
• Independence vs interdependence
Complex system: We can’t just look at the parts, we need to understand interactions and changes of
behavior ≈ understand system dynamics
13
WEATHER AS A COMPLEX SYSTEM
BETWEEN CHAOS AND UNPREDICTABILITY
https://www.weather-atlas.com/en/denmark/copenhagen-
weather-december
https://www.youtube.com/watch?v=M7Hf6VfsJ0U
14
COMPLEX SYSTEMS: ECONOMIC EXAMPLES
Market Economies:
• Heterogeneous agents: Consumers and producers with diverse and adaptive preferences and
production techniques.
• Interaction: Market transactions, price mechanisms and social influence
• Emergence: Economic dynamics that are at times unpredictable
Innovation Ecosystems:
• Heterogeneous agents: Startups, investors, universities, and R&D departments.
• Interaction: Collaboration, competition, and knowledge transfer.
• Emergence: Technological advancements and industry disruptions beyond single entities
15
IF STUDYING INCENTIVES AND PERFORMANCE AS A
COMPLICATED SYSTEM
• Starting point: Other factors probably also impact incentives & performance
• Yet, if the system is equal to the sum of its parts, we can
• Investigate sub components of an economic system independently
• Investigate incentives and performance
• E.g. investigate leadership style and peformance
• …
• And then combine all results into overall insight into how organizations work.
17
EXAMPLE I: SIMULATION OF ANTS THAT FOLLOW
SIMPLE RULES
We can’t look at each kind of ant
individually, and then simply combine
their behavior & then explain the system
http://netlogoweb.org/launch#http://netlogoweb.org/assets/modelslib/Sample%20Models/Biology/Ants.nlogo
Sullivan 2011, p. 2
19
HOW ABOUT AN EVEN SIMPLER SYSTEM?
EXAMPLE II: GAME OF LIFE I
• Very structured, complex behavior emerges from
very (!!) simple rules, in a system of black and white
dots that are interconnected in a feedback loop
https://www.youtube.com/watch?v=C2vgICfQawE
EXAMPLE II: GAME OF LIFE II
https://www.youtube.com/watch?v=C2vgICfQawE
21
BUT WHAT ABOUT COMPLEXITY IN HUMAN SYSTEMS?
To sum up
• Complexity is more than just ”very complicated”:
• Interdependence (interactions) feedback loops adaptive/changing behavior
• Complex, sophisticated systems not coordinated by centralized leader
• Emerging behavior and system: The whole is more than the sum of its parts
• But, how about human, economic systems?
https://www.youtube.com/watch?v=ysa5OBhXz-Q
WOLVES AND YELLOWSTONE PARK II (DISCUSS 2 MIN)
25
TRADITIONAL STATISTICAL ANALYSIS
(E.G. REGRESSION) VS. COMPLEXITY
Why would a traditional statistical (e.g. linear
regression) analysis not work well?
26
A LEGO BRICK METAPHOR
https://brickarchitect.com/2018/overcoming-builders-block/
Linear regression approach: Predicting Lego building based on bricks, colour etc.
- a lego brick is a lego brick, stays the same limits what you can build
What if bricks could adapt: change shape, behavior based on interacting with other bricks?
Dynamics would be complex, much more difficult to predict (whole more than sum of its parts)
27
PEOPLE AND ECONOMIC SYSTEMS?
28
EXAMPLE: WATTS, MUSIC AND LINEAR
REGRESSIONS I
• If we want to predict and explain how well music numbers will do in a hitlist.
• We just let people listen to music numbers, let them rate / buy it result?
• Well…what about interdependence and feedback listeners get from others?
https://en.wikipedia.org/wiki/Billboard_Year-End_Hot_100_singles_of_2022
29
EXAMPLE: WATTS, MUSIC AND LINEAR REGRESSIONS II
http://www.nytimes.com/2012/08/05/business/of-luck-and-success-economic-view.html
WATTS 2007 ON COMPLEXITY IN SOCIAL SCIENCE
KEY QUOTE!
“It is hard to understand, for example, why even a single organization behaves the way it
does without considering
a) the individuals who work in it
b) the other organizations with which it competes, cooperates and compares itself to
c) the institutional and regulatory structure within which it operates
d) the interactions between all these components.” (p. 489)
Watts 2007: ”…interactions have consequences for the choices of others" (p. 489)
31
EXAMPLE: INCENTIVES AND PERFORMANCE’
• Holst-Hansen & Bergenholtz (cf. Bergenholtz 2024b) studied how isolated individuals
responded to a single variable (incentives, 1 vs. 5 DKK) change
• What about interactions with other individuals?
• What about different management styles, organizational cultures?
• What about effects from competitors, who might implement similar system – will
the labour supply change?
• What about national cultures, financial crises etc.?
In complex system:
• Adaptive behavior, variables not fixed
• Interdependent variables
• Sum is more than the parts
34
BARRETT-FELDMAN 2021 I
MECHANISTIC VS COMPLEXITY MINDSET
BARRETT-FELDMAN 2021 II
MECHANISTIC VS COMPLEXITY MINDSET
• Maybe effect of incentives depends on: The particular task, if the task length is 10 or 20 min,
recent economic news, weather, room temperature etc. + the combined interaction
• Every single factor maybe basically unimportant – but all interconnected and combined?
Diener E, Northcott R, Zyphur M, West SG (2022) Beyond Experiments. Perspectives on Psychological Science 17(4):1101-1119.
37
EVALUATION OF THE COURSE
https://go.blueja.io/hcsMgsb5QkqSxrDin9ZdIA
38
LINKING BACK TO PRIOR CURRICULUM I
COMPLEXITY AND KUHN’S PARADIGM SHIFTS
39
LINKING BACK TO PRIOR CURRICULUM II
COMPLEXITY AND GUBA: WHICH PARADIGM?
Very difficult question – wouldn’t be exam question
40
LINKING BACK TO PRIOR CURRICULUM III
COMPLEXITY AND CASE STUDIES
Case studies could via in-depth study of unique context capture
• How elements are connected / interdependent
• Feedback loops
• Adaptive, changing behavior
41
Some of the following points are covered (but with
different examples) in this video
https://www.youtube.com/watch?v=_Z2DVh2ws2I
42
INTERDEPENDENT OR INDEPENDENT SYSTEMS?
HOW TO ANALYSE ECONOMIC PROBLEMS
43
EXAMPLE FROM LESSON 7: BUSINESS ECONOMICS
- SIMPLE CAUSATION FROM A AND B?
https://twitter.com/graykimbrough/status/1108403322690654208?s=11
WATER DOES RUN UPHILL? FEEDBACK LOOPS
CHANGE BEHAVIOR AND ECONOMIC SYSTEM
“Why did so many economists get this wrong? They looked at the micro and ignored the macro.
In the basic model of supply and demand, an increase in the cost of something reduces
demand. This may be true in an isolated laboratory setting. But when minimum wages went up
in the real world, it affected not just the parties to that transaction, but the regional economy.
By considering only the relationship between employer and employee, the dismal-science set
was focusing too narrowly. The critics failed to consider the impact of lower-wage employees
earning more money; these folks typically spends almost everything they earn, which means
that when they’re paid more it goes right back into the local economy.”
https://www.bloomberg.com/view/articles/2016-10-21/doomsayers-keep-getting-it-wrong-on-higher-minimum-wages
FROM IDENTIFYING WHAT A COMPLEX
SYSTEM IS TO HOW TO ANALYZE I
In a regression analysis:
• Collect data on main variables (including potential confounders) and
run regressions (cross-sectional or longitudinal)
• Reductionist approach: If you understand the parts, you have
understood the whole
Complexity perspective:
• Understand how variables interact and influence each other
• Study system’s behavior over time, capturing patterns and changes
• Aim to model system behavior, over time
• Holistic, system approach: Need to understand the parts, and their
interactions, to understand the whole
46
FROM IDENTIFYING WHAT A COMPLEX SYSTEM IS
TO HOW TO ANALYZE II - EXAMPLE
Regression approach
Example: Explaining consumer spending through isolated factors like income levels,
interest rates, and consumer confidence, assuming direct and predictable relationships.
• Establish direct causal relationships?
• An equation can be written
Complexity approach
Example: Investigating how agents (e.g. individuals and institutions) interact and
influence each others’ adaptive behavior
• Requires tracking all individuals’ consumer spending behavior + how they are
connected + the changes in behavior of interacting
• overall system level interdependencies and dynamics
• No equation can be written
47
2021 SPRING EXAM QUESTION (DISCUSS 2-3 MIN)
Question 4 (17%)
Complexity researchers like Duncan Watts (the author of Watts 2007) do carry out experiments and
find great value in this approach. However, what challenges might the complexity perspective
identify when doing experiments - in particular lab-experiments? When referring to the complexity
perspective you can rely on the curriculum or the Covid-19 digital exercise we engaged in, in the
live part of lesson 12. Please explain your answer in detail.
48
SUMMING UP COMPLEXITY
49
A FEW COMMENTS ABOUT THE EXAM
Suggestion for preparation
• Read curriculum + reflect on exercises/questions in class + study previous exams
• You’ll be too busy to read the curriculum in the exam situation
• Practice Research Design challenges + look at former exam questions and consider how you would
approach them. This year’s question will be similar in style
• This course doesn't call for a template approach, as some other courses.
• Not a course where one (only) learns a model or table or technique: critical thinking and reflection on
what science and truth is about is needed
• Please identify the weaknesses of each of the three studies and explain […]
• Four statements are listed below. Has XXX written all these? Please be specific in your explanation of why you think
some (if any) of the sentences might not match the content provided in XXX.
• As discussed in the lectures, XXXs interpretation of YYY arguably shares common ground with ZZZs philosophy of
science. Please explain in detail how ZZZs philosophy of science is similar to and differs from XXXs presentation of
YYY.
FLOWLOCK+ (= W0+) EXAM
Aids: Physical books, physical notes and articles. PDF files on computer and
access to dictionary.
• Test-flow has been open this week. Hope you all have ensured you can use Flowlock+ and
can access PDF notes / documents.
• Notice, you might have to log out & in at the exam, the 23rd of March!
51
PHILOSOPHY OF SCIENCE
LESSON 14
STUDY CAFE 2
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
Agenda
2
Assessing quality of RCT research design: What to
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
Not your personal assessment: Curriculum based assessment of the quality of the RCT
3
AARHUS UNIVERSITY
SCHOOL OF BUSINESS AND SOCIAL SCIENCES
5
PLEASE DISCUSS THE FOLLOWING 3
EXERCISES (30 MIN IN TOTAL)
• Do ask us questions
(you don’t have time to write a full answer, but I suggest you think about a) what
criteria you will rely upon when assessing external validity, b) how you would
structure your answer, and c) maybe think about how the external validity
compares to some of the other examples we have heard about in class)
6
A) 2023 SPRING EXAM QUESTION (18-20 MIN)
Question 1 (25%)
A recent paper provided evidence for a relationship between recessions and job satisfaction. The authors argue that during an economic
recession, when the job market is less robust and the economy is suffering, people will perceive their jobs to be more satisfying, as
alternative job prospects appear less attractive. The authors of the paper set out to test this hypothesis through three separate studies.
Please identify the weaknesses of each of the three studies and explain how a given study's weakness can be remedied by the other
studies.
Study 1 involved a survey sent out to adults in different regions of South America. The questions in the survey covered a broad range of
variables, including the respondents' perspective on the current economic situation, their job security, job satisfaction and a range of
demographic variables. At the time of data collection, some participants lived in regions experiencing high growth, while others lived in
regions that experienced a more neutral economic development. The total sample size consisted of 12.300 participants, across all
regions and ages. The statistical analysis found that people perceive their jobs as more satisfying if the economic conditions are poorer,
compared to a high growth environment, thus supporting the overall hypothesis.
Study 2 aimed to replicate the findings of Study 1, relying on surveys of adults living in three regions in the United States. The sample
size was 8.500 participants. The country (and thus all regions) was in the middle of a very serious economic crisis when the data were
collected. The sample size of Study 2 was 8.150 participants, across all regions and ages. Some participants were more exposed to the
crisis than others were, e.g. by working in industries that suffered the most economically. The survey consisted of the same question as
those in Study 1. This study also found that people that were the most highly exposed to the economic crisis were the ones reporting the
highest job satisfaction. This provides further support for the hypothesis that economic conditions influence people's job satisfaction.
Study 3 aimed for a different approach to answer the research question. The study, which had a sample size of 512 participants, sought
to investigate the mechanism behind the observed relationship between economic conditions and job satisfaction. Participants were
from the United States. The study created a treatment and control condition (and randomised participants into each of them), by
making participants read either a New York Times article that praised the current economy or an article from the New York Times that
highlighted the negative characteristics of the current economy (both were actual articles, printed in the New York Times). Being in the
"bad economy" group (having read the negative article) led to a more negative assessment of the current economy, compared to the
average assessment of the economy by the people in the other group (who had read the article praising the economy). After being
exposed to the treatment (or control), the participants answered questions about their job and their job satisfaction. As predicted and in
line with the former studies, participants in the 'bad economy' condition reported greater job satisfaction.
While previous research has shown that job satisfaction can be influenced by factors within organisations, these studies suggest that
external factors such as economic conditions can also play a role in how people evaluate their jobs. The findings of the three studies
support the hypothesis that people perceive their jobs to be more satisfying during economic downturns, which may have implications
for employers and policymakers seeking to improve job satisfaction levels during times of economic instability.
7
B) 2023 SPRING EXAM QUESTION (12-14 MIN)
Question 2 (15%)
Imagine a large Danish company that offers telemarketing services. The company has 60 different
units across Scandinavia, each with at least 60 employees. It has for years been fairly profitable but
also realised an increasing level of competition. The top management team is worried about its future
earnings and decides to implement a new system that integrates a chatbot (a computer program that
uses advanced technology to imitate human conversations) in the sales process. The idea is that
telemarketing employees would be able to get inspiration from the chatbot on how to optimise their
sales. 30 of the 60 unit locations have been selected to try out this new feature. The selection was
partially related to logistical challenges and the availability of IT expertise. The team that
implemented the chatbot visited each of the 30 different units during a span of 15 weeks. Subsequent
analysis of sales performance shows that the implementation of the chatbot, on average, led to a
subsequent and immediate increase in performance, compared to the units that have not (yet)
implemented the chatbots. The analysis controlled for prior performance, size of the unit and the
experience of the manager in charge of each of the units.
You have now received some limited information on the above outlined research design. Based on
the provided information one can infer that it is not a randomised controlled trial. What information
would you have liked to know about this case in order to identify if the setup constitutes a quasi-
experiment or not? Please be specific when explaining what information you need.
8
9
PLEASE DISCUSS THE FOLLOWING 3
EXERCISES (30 MIN IN TOTAL)
• Do ask us questions
(you don’t have time to write a full answer, but I sugest you think about
a) what texts and perspectives you would rely upon, b) what the main
points of these texts are, c) how you would structure an answer, and d)
remember the questions asks for counterargumentS – plural.
10
C) 2023 SPRING EXAM QUESTION (12-14 MIN)
Question 3 (20%)
In Guba (1990), the positions post-positivism and constructivism are outlined. Please
identify both similarities and differences across these two positions and explain these
similarities and differences in detail. Your explanation should use at least two different
quotes from Guba.
11
D) 2021 SPRING EXAM QUESTION (10-12 MIN)
Question 4 (15%)
Please rely on the curriculum and course material to explain what features of complexity
that are not covered in the following brief presentation of the concept. Provide at least
one quote from the curriculum to support your argument.
In the context of economics or business administration, complexity refers to the degree of
difficulty involved in understanding and managing the dynamics of a particular system or
organisation. A complex system or organisation is characterised by a large number of
interconnected elements or variables that relate to each other in often complicated ways.
Managing complexity in business administration involves analysing and understanding the
various elements and their interconnectedness, identifying potential risks and
opportunities, and developing strategies to optimise performance and mitigate negative
impacts. This requires a multidisciplinary approach. Overall, understanding and managing
complexity is a crucial skill for businesses and organisations that operate in dynamic and
uncertain environments, and it requires a deep understanding of the underlying
mechanisms and dynamics of the system.
12
E) ASSESSING RESEARCH DESIGNS: WHAT IS
THE MAIN CHALLENGE HERE? (5 MIN)
As stated in the text below, the aim of the study was to figure out if running increases the risk of hip and
knee problems. Please look at the information below and based on our curriculum identify the primary
obstacle to trusting the study's findings.
Dr. Hartwell and his colleagues conducted the prospective cohort study, “Does Running Increase the Risk for
Hip and Knee Arthritis? A Survey of 3,804 Chicago Marathon Runners” …
An electronic survey was completed by 3,804 participants registered for the 2019 and 2021 Chicago
Marathons. The mean age of respondents was 43.9. Participants had run an average of five or fewer
marathons, and they had been running for an average of 14.7 years.
The survey featured 30 questions targeted at assessing:
• Demographic information (age, sex, height, weight, country of origin and occupation)
• Running history (number of years running; average weekly mileage over the past year; average running
pace during training runs; number of marathons completed, including ultra-marathons;
and participation with cross training)
• Hip/knee health (occurrence of hip or knee pain over the past year that prevented running; a history of
hip or knee injuries that required a break in running; a family history of hip and/or knee arthritis;
surgical procedures performed on the hip or knee; and receiving a hip or knee arthritis diagnosis)
https://aaos-annualmeeting-presskit.org/2023/research-news/new-study-demonstrates-that-running-does-not-increase-risk-of-arthritis/ (slightly
modified)
13
14
A) 2023 SPRING EXAM QUESTION (17-19 MIN)
Overall comments
• Cross-sectional studies (1 & 2) imply the risk of confounding variables, or omitted
variable bias (or similar terminology). It is not adequate to write the generic
statement that cross-sectional studies have low internal validity. One should
specify the given weakness (e.g. that even though control variables are included
in study 1 & 2, some variables might be confounding (or maybe there are
variables missing, constituting an omitted variable bias).
• The RCT (study 3) can eliminate the risk of such confounding variables/omitted
variable bias (in study 1-2), but implies lower external validity.
• Study 1 + 2 complement each other and imply higher external validity compared
to 3.
Note also
• Different economic contexts in study 1 & 2
• All relevant control variables collected? Might some be confounding?
• Common methods bias in study 1 & 2?
Make sure to explain how one could remedy [=afhjælpe] the weaknesses.
15
B) 2023 SPRING EXAM QUESTION (12-14 MIN)
Not a RCT: No randomization between treatment and control
As good as random?
Key questions: How was treatment allocated? Might the units in the different conditions be as
similar as possible, that is is there only one difference (the treatment) between them.
Highlight that IT expertise and logistics were partial reasons for allocation of units to treatment
and condition
Maybe Hawthorne effect: Treatment group might be happy about access to new technology
– vice versa, control condition might be unhappy about being left out. Related to managerial
visit.
16
C) 2023 SPRING EXAM QUESTION (12-14 MIN)
Similarities
• This is probably the most relevant quote and will be provided by many: “Even post-positivists
have conceded that objectivity is not possible, the results of an inquiry are always shaped by the
interaction of inquirer and inquired into” (both PP & C)
• Some will refer to a clip of “Yes Prime Minister” that I showed in class. It illustrates a problem that
both PP and C agree we have (neutral observations of the world not possible). They have
different solutions to the problem though.
• Complete Truths with capital T is unachievable, due to human limitations, biases. This similarity
leads into differences (either modified objectivity, or subjectivity).
• Both agree qualitative methods and multiple perspectives could be useful
Differences
• PP assumes that there are regularities to uncover, and seeks approximate truths to uncover
those ≠ C
• PP argues we should aim to reduce biases, by relying on triangulation, distanced
• C: ”…it makes the findings of an inquiry not a report about what is ”out there” but the residue of a
process that literally creates them[findings]” ≠ PP
• PP: "...although a real world driven by real natural causes exist, it is impossible for humans truly
to perceive it with their imperfect sensory and intellective mechanisms” ≠ C
• C argues that realities are multiple ≠ PP
17
D) 2021 SPRING EXAM QUESTION (10-12 MIN)
• Please note that the questions asked you to identify which featureS are missing
from the given description.
• The exam set provides some information that is relevant, such as:
interconnectedness, analyzing various elements, deep understanding of
dynamics of the system, very difficult systems etc. This does not fully capture the
essence of complexity though. Notice in particular the text says: “A complex
system or organisation is characterised by a large number of interconnected
elements or variables that relate to each other in often complicated ways.”
Relevant features of complex systems not clearly explained in Q4 text.
• Interdependence: Elements/components in a system interact and adapt to
each other. This constitutes feedback loops and change in the behavior that can
be difficult to predict.
• The whole is more than the sum of its parts (emergence)
• Watts quote, p. 489: “…which it operates the interactions between all these
components.” (p. 489)
18
E) ASSESSING RESEARCH DESIGNS: WHAT IS THE
MAIN CHALLENGE HERE? (5 MIN)
As stated in the text below, the aim of the study was to figure out if running increases the risk of hip and knee problems.
Please look at the information below and based on our curriculum identify the primary obstacle to trusting the study's
findings.
Dr. Hartwell and his colleagues conducted the prospective cohort study, “Does Running Increase the Risk for Hip and Knee
Arthritis? A Survey of 3,804 Chicago Marathon Runners” …
An electronic survey was completed by 3,804 participants registered for the 2019 and 2021 Chicago Marathons. The mean
age of respondents was 43.9. Participants had run an average of five or fewer marathons, and they had been running for
an average of 14.7 years.
The survey featured 30 questions targeted at assessing:
• Demographic information (age, sex, height, weight, country of origin and occupation)
• Running history (number of years running; average weekly mileage over the past year; average running pace during
training runs; number of marathons completed, including ultra-marathons; and participation with cross training)
• Hip/knee health (occurrence of hip or knee pain over the past year that prevented running; a history of hip or knee
injuries that required a break in running; a family history of hip and/or knee arthritis; surgical procedures performed on
the hip or knee; and receiving a hip or knee arthritis diagnosis)
https://aaos-annualmeeting-presskit.org/2023/research-news/new-study-demonstrates-that-running-does-not-increase-risk-
of-arthritis/ (slightly modified)
19
PHILOSOPHY OF SCIENCE
LESSON 15
Q&A
Carsten Bergenholtz (cabe@mgmt.au.dk)
Department of Management, Aarhus BSS
Aarhus University
AGENDA: VEVOX.APP QUESTIONS
• Flowlock+ / references
• Evaluation: Follow up
• Misc questions
• Overview of the course
• Illustrating how/when Phil of Science is relevant
• Chatbot survey(s)
2
FLOWLOCK+ EXAM
• 5 hours: 5 questions
• No internet access
• Access to PDF documents (notes, texts) on own PC (don’t just
save online) + physical/printed notes allowed
• Access to dictionary
• About references (quotes, chatbot, group notes, general
references)
• In particular when you refer to / quote other sources than
curriculum provide reference
• How to reference -> see prior exam answers
https://studerende.au.dk/studier/fagportaler/businessadm/eksamen/retning
slinjer-for-skriftlige-opgaver-paa-hacandmerc 3
EVALUATION: FOLLOW UP I
RESPONSE RATE: 27%
5
OVERVIEW OF THE COURSE / CONNECTIONS
6
ILLUSTRATING HOW / WHEN THE COURSE
CONTENT IS RELEVANT
7
APPLYING PHIL OF SCIENCE INSIGHTS I:
ASSESSING CONSULTANCY REPORTS
8
CHALLENGES WITH THE REPORT I:
REVERSE CAUSALITY?
https://maycontainlies.com/discernment-matters-even-more/
9
CHALLENGES WITH THE REPORT II:
HETEROGENEOUS VARIABLES
10
CHALLENGES WITH THE REPORT III:
HETEROGENEOUS VARIABLES
11
CHALLENGES WITH THE REPORT IV:
CAUSALITY? & CONFOUNDING VARIABLES?
12
APPLYING PHIL OF SCIENCE INSIGHTS II:
CORRELATION & CAUSALITY IN MEDIA STORY ABOUT SCIENCE
13
14
CHATBOT SURVEY(S)
15