Professional Documents
Culture Documents
Psychological Assessment 5th Edition PDF
Psychological Assessment 5th Edition PDF
EDITION
PSYCHOLOGICAL
ASSESSMENT
FIFTH EDITION
PSYCHOLOGICAL
ASSESSMENT
6 Developing a psychological
measure 6.1 Introduction
6.2 Steps in developing a measure
6.2.1 The planning phase
6.2.1.1 Establish a test-development team
6.2.1.2 Specify the aim, target population, and nature of the measure
6.2.1.3 Define the content of the measure
6.2.1.4 Other test specifications to include in the
plan 6.2.2 Item development, creation or sourcing
6.2.2.1 Develop the items
6.2.2.2 Review the items
6.2.3 Assemble and pre-test the experimental version of the
measure 6.2.3.1 Arrange the items
6.2.3.2 Finalise the length
6.2.3.3 Answer mechanisms and protocols
6.2.3.4 Develop administration instructions
6.2.3.5 Pre-test the experimental version of the
measure 6.2.4 The item-analysis phase
6.2.4.1 Classical Test-Theory item analysis: Determine item difficulty (p)
6.2.4.2 Classical Test-Theory item analysis: Determine discriminating power
6.2.4.3 Item-Response Theory (IRT)
6.2.4.4 Identify items for final pool
6.2.5 Revise and standardise the final version of the
measure 6.2.5.1 Revise the items and test
6.2.5.2 Select items for the final version
6.2.5.3 Refine administration instructions and scoring procedures
6.2.5.4 Administer the final version
6.2.6 Technical evaluation and establishing norms
6.2.6.1 Establish validity and reliability
6.2.6.2 Establish norms, set performance standards or
cut-scores 6.2.7 Publish and refine continuously
6.2.7.1 Compile the test manual
6.2.7.2 Submit the measure for classification
6.2.7.3 Publish and market the measure
6.2.7.4 Revise and refine continuously
6.3 Evaluating a measure
6.4 Conclusion
7 Cross-cultural test adaptation, translation, and tests in multiple
languages 7.1 Introduction
7.2 Reasons for adapting measures
7.3 Test adaptation
7.3.1 Types of test adaptation
7.3.2 Challenges related to test adaptation in South Africa
7.4 Equivalence of adapted versions of measures used in multicultural and multilingual
contexts 7.4.1 Equivalence of measures adapted for cross-cultural use
7.4.2 Exploring linguistic equivalence
7.4.2.1 Judgemental approaches
7.4.2.2 Empirical investigations
7.5 Statistical approaches to establish equivalence
7.5.1 Bias and equivalence
7.5.2 Measurement bias and equivalence
7.5.2.1 Bias and equivalence at the item level
7.5.2.2 Bias and equivalence in factor structures between groups
7.5.2.3 Evaluation of measurement invariance (MI) approaches
7.5.3 Prediction bias and equivalence
7.5.3.1 Procedures to establish prediction bias and equivalence
7.5.4 Construct bias and construct equivalence
7.5.5 Nomological bias
7.5.6 Method bias
7.6 Steps for maximising success in test adaptations
7.7 Multicultural test development
7.8 Conclusion
10 Assessment of cognitive
functioning 10.1 Introduction
10.2 Theories of intelligence: A brief history and
overview 10.2.1 Background
10.2.2 Defining intelligence
10.2.3 Theories of intelligence
10.3 Interpreting the intelligence score and diversity issues
10.3.1 The meaning of the intelligence score
10.3.2 Individual differences and cultural diversity
10.4 Measures of general cognitive functioning
10.4.1 Individual intelligence measures
10.4.1.1 Description and aim
10.4.1.2 The application of results
10.4.2 Group tests of intelligence
10.4.2.1 Description and aim
10.4.2.2 The application of results
10.5 Measures of specific abilities
10.5.1 Aptitude measures
10.5.1.1 Description and aim
10.5.1.2 The application of results
10.5.2 Measures of specific cognitive functions
10.5.2.1 Description and aim
10.5.2.2 The application of results
10.6 Scholastic tests
10.7 Current and new trends in cognitive assessment
10.8 Conclusion
11 Measures of well-
being 11.1 Introduction
11.2 Well-being
11.2.1 Defining well-being
11.2.2 Positive psychology and well-being
11.2.3 Domains of well-being
11.2.3.1 Subjective well-being
11.2.3.2 Other dimensions of well-being
11.2.3.3 Mental illness
11.3 Well-being in the workplace
11.3.1 Why well-being matters
11.3.2 The cost of ill health
11.3.3 Well-being in the workplace
11.3.4 Wellness programmes
11.4 Well-being among university students
11.5 Measures of well-being
11.5.1 Assessment of well-being in diverse contexts
11.5.2 Assessment of well-being in the work context
11.6 Conclusion
12 Personality assessment
12.1 Introduction
12.2 A conceptual scheme for personality assessment
12.3 Level 1: The assessment of relatively stable personality traits
12.3.1 The 16 Personality Factor Questionnaire (16PF)
12.3.2 The Big Five model of personality traits
12.3.3 The Basic Traits Inventory
12.3.4 The Hogan Personality Inventory
12.3.5 The Occupational Personality Questionnaire
12.3.6 The Myers-Briggs Type Indicator
12.3.7 The Minnesota Multiphasic Personality Inventory (MMPI)
12.3.8 Cross-cultural use of structured personality assessment measures
12.3.9 The South African Personality Inventory (SAPI)
12.4 Level 2: Assessment of motives and personal concerns
12.4.1 Explicit motives
12.4.2 Implicit motives
12.4.3 Other projective methods of personality assessment
12.5 Personality and counterproductive behaviour at work
12.5.1 Work-related Risk and Integrity Scales (WRISc)
12.5.2 Hogan Development Survey
12.6 Uses and applications of personality tests
12.7 Conclusion
13 Career-counselling
assessment 13.1 Introduction
13.2 The person-environment fit approach
13.2.1 Assessing intelligence
13.2.2 Assessing aptitude
13.2.3 Assessing vocational interests
13.2.4 Assessing personality
13.2.5 Assessing values
13.2.6 Evaluation of the person-environment fit approach to career counselling
assessment 13.3 The developmental approach to career counselling assessment
13.3.1 Career assessment in the developmental approach
13.3.2 Evaluation of the developmental approach to career counselling
assessment 13.4 The systems approach to career counselling assessment
13.4.1 Career assessment in the systems approach
13.4.2 Evaluation of the systems approach to career counselling assessment
13.5 Career construction theory and life-design counselling
13.5.1 Assessment in career construction theory and life-design counselling
13.5.2 Evaluation of career construction theory and life-design counselling
13.6 Career counselling in a changing environment
13.7 Conclusion
The editors would like to thank the following authors for their
contributions to previous editions of Introduction to Psychological
Assessment: Prof. Diane Elkonin, Caroline Davies, Dr Jenny
Jansen, and Prof. Anil Kanjee.
INTRODUCTION
The start of an exciting journey
Hi there! You are about to embark on an interesting, challenging
journey that will introduce you to the world of psychological
assessment. Interesting? Challenging? ‘That’s not what I have heard
about assessment,’ you might say. Many people view courses and
books on psychological assessment as dull, boring, and not relevant
for students and professionals in people-oriented careers. Even if you
have some doubts that you are going to enjoy learning about
assessment, we beg you to give us a chance to share our enthusiasm
about assessment with you and to show you why our enthusiasm is
shared by countless others in South Africa and, indeed, all over the
world. As you read this book, we hope that you will discover that
assessment is an essential, integral part of psychology and other
people-oriented professions. You will be exposed to the fact that
assessment can be used in many different contexts, such as
educational, counselling, clinical, psychodiagnostic,
psycholegal/forensic, industrial (occupational), and research contexts.
Furthermore, you will discover that assessment can be used for many
different purposes (e.g. diagnosing learning problems, assessing
whether a mass murderer is fit to stand trial, determining whether
therapeutic intervention has been successful, and making job
selection, placement, and training decisions). Does assessment still
sound dull, boring, and irrelevant to you? No ways! Well then, let the
journey begin.
Concepts
To help you to identify the really important concepts in a chapter, they
will always be presented in boldface or italics.
We hope that you find your journey through this book as interesting
as we found the conceptualisation and preparation stages. We would
also like to thank all those who contributed to making this book a
reality, either by way of writing sections or on the editorial,
typesetting, and production side. The contributors, in particular, need
to be thanked for their time and effort in preparing their insightful,
scholarly contributions.
Finally, we would like you to let us know your impressions of this
book, and any suggestions you have for improving it. Send your
comments to Cheryl at the following email address:
Cheryl.Foxcroft@mandela.ac.za.
1 Assessment overview
Part 2 Historical perspectives
3 Basic measurement concepts
1 4 Reliability
5 Validity
6 Developing a measure
7 Cross-cultural perspectives
An overview of assessment:
Definition and scope
CHERYL FOXCROFT AND GERT ROODT
Chapter CHAPTER OUTCOMES
1 By the end of this chapter you will be able to:
❱ distinguish between tests, assessment measures, testing and
psychological and competency-based assessment
❱ name the characteristics of assessment measures
❱ understand the value of adopting an Africa-centred approach
in South Africa
❱ recognise the limitations of assessment results
❱ explain assessment as a multidimensional process.
1.1 Introduction
We constantly have to make decisions in our everyday lives: what to
study, what career to pursue, whom to choose as our life partner,
which applicant to hire, how to put together an effective team to
perform a specific task, and so on. This book will focus on the role of
psychological assessment in providing information to guide
individuals, groups, and organisations to understand aspects of their
behaviour and make informed and appropriate decisions. In the
process you will discover that assessment can serve many purposes.
For example, assessment can help to identify strengths and
weaknesses, map development or progress, inform decisions
regarding suitability for a job or a field of study, identify training and
education needs, or it can assist in making a diagnosis. Assessment
can also assist in identifying intervention and therapy needs,
measuring the effectiveness of an intervention programme, and in
gathering research data to increase psychology’s knowledge base
about human behaviour or to inform policy-making.
South Africa is situated in Africa, and professionals and
researchers are increasingly recognising the importance and value of
adopting a more African-centred approach. As South African society
is multicultural and multilingual in nature, this book also aims to raise
awareness about the role of culture and language in assessment, and
to suggest ways of addressing them from an African-centred
perspective. As we begin our journey in this chapter you will be
introduced to key terms, the characteristics of assessment measures,
the multidimensional nature of the assessment process, and the
limitations of assessment results.
Another way more fun way to check your progress is to see how you
respond to the questions posed after reading through the following
case study:
Questions
(a) What would have guided the RC in deciding how to address the
referral question of providing Lindiwe with career guidance?
(Hint: what do you need to know about Lindiwe’s context?)
(b) In what ways did the RC demonstrate responsible use of
assessment measures in this case study?
(c) Is there evidence of irresponsible assessment practices in this case
study that could have resulted in negative consequences for Lindiwe?
1.6 Conclusion
This chapter introduced you to key terms in the field of psychological
assessment. You learned about the tools (assessment measures or
tests) used in psychological testing and assessment, their
characteristics, and the psychometric properties (measurement
standards) required of them. You further learned that test results are
only one source of information in an assessment process and that the
conclusions reached from them are always approximate in nature. The
multidimensional nature of the assessment process, and the
competing values and motives of those involved in the assessment
process, were also highlighted. The chapter concluded with the
proposition that for culturally appropriate assessment to be conducted
in South Africa and the African continent, a fusion of African-centred
and Western-orientated approaches is needed.
ADVANCED READING
Mkhize, N. (2004). Psychology: An African perspective. In K. Ratele, N. Duncan, D.
Hook, P. Kiguwa, N. Mkhize & A. Collins (Eds.), Self, community & psychology
(pp. 4.1–4.29). Lansdowne, South Africa: University of Cape Town Press.
Makhubela, M. (2016). ‘From psychology in Africa to African psychology’: Going
nowhere slowly. Psychology in Society (PINS), 52, pp. 1–18. Retrieved on
12 April 2018 from http://dx.doi.org/10.17159/2309-8708/2016/n52a1.
Moletsane, M. (2016). Understanding the role of indigenous knowledge in
psychological assessment and intervention in a multicultural South African
context. In R. Ferreira (Ed.), Psychological assessment: Thinking innovatively in
contexts of diversity (pp. 20−36). Cape Town: Juta and Company.
Mpofu, E., Ntinda, K. & Oakland, T. (2012). Understanding human abilities in sub-
Saharan African settings. Online Readings in Psychology and Culture, 4(3).
Retrieved on 12 April 2018 from https://doi.org/10.9707/2307-0919.1036.
Nsamenang, A.B. (2007). Origins and development of scientific
psychology in Afrique Noire. In M.J. Stevens and D. Wedding (Eds.,
under the supervision of John G. Adair). Psychology: IUPsyS Global
Resource (Edition 2007). London: Psychology Press.
Nwoye, A. (2015). What is African psychology the psychology of?
Theory & Psychology, 25(1), pp. 96–116.
Ratele, K. (2017). Four (African) psychologies. Theory &
Psychology, 27(3), pp. 313– 327.
Valchev, V.H., Van de Vijver, F.J.R., Nel, J.A., Rothman, S. & Meiring,
D. (2013). The use of traits and contextual information in free
personality descriptions across ethnocultural groups in South Africa.
Journal of Personality and Social Psychology, 104, pp. 1077−1091.
(Can be accessed online at http://dx.doi.org/10.1037/a0032276).
Van der Merwe, K., Ntinda, K & Mpofu, E. (2016). African perspectives of personality
psychology. In L.J. Nicholas (Ed.), Personality Psychology (pp. 32–50). Cape
Town, South Africa: Oxford University Press Southern Africa (Pty) Ltd.
Zuilkowski, S.S., McCoy, D.C., Serpell, R., Matafwali, B. & Fink, G. (2016).
Dimensionality and the development of cognitive assessments for children in
sub-Saharan Africa. Journal of Cross-Cultural Psychology, 47(3), pp. 341–354.
(Can be accessed online at http://dx.doi.org/10.1177/0022022115624155).
Psychological assessment: A
brief retrospective overview
CHERYL FOXCROFT, GERT ROODT, AND FATIMA ABRAHAMS
Chapter CHAPTER OUTCOMES
2 By the end of this chapter you will be able to :
❱ understand how assessment has evolved since ancient times
❱ appreciate the factors that have shaped psychological
assessment in South Africa
❱ develop an argument about why assessment is still valued in
modern society.
2.1 Introduction
At the start of our journey into the field of psychological assessment, it
is important to gain a perspective on its origins. This is the focus of
this chapter. Without some idea of the historical roots of the discipline
of psychological assessment, the great progress made by modern
assessment measures cannot be fully appreciated. In this chapter you
will also be introduced to some of the key concepts that we will be
elaborating on in the Foundation Zone of this book.
As you journey through the past with us, you should be on the
lookout for the following:
how difficult it was in ancient times to find an objectively verifiable way
of measuring human attributes (ask yourself the reasons for this)
how the most obvious things in the world of ancient philosophers and
scientists (such as the human hand, head, and body, as well as
animals) were used in an attempt to describe personal attributes
the stepping stones that some of the ancient ‘measures’ provided for
the development of modern psychological assessment
the factors both within and outside of the discipline of psychology that
have shaped the development of modern psychological assessment
the factors that shaped the development and use of psychological
assessment in South Africa.
2.2.1 Astrology
Most people are aware of the horoscopes that appear in daily
newspapers and popular magazines. The positions of planets are
used to formulate personal horoscopes that describe the personality
characteristics of individuals and to predict what might happen in their
lives. The origin of horoscopes can be traced back to ancient times,
possibly as early as the fifth century BCE (McReynolds, 1986). Davey
(1989) concludes that scientists, on the whole, have been scathing in
their rejection of astrology as a key to understanding and describing
personality characteristics. Do you agree? Briefly state your reasons.
2.2.2 Physiognomy
McReynolds (1986) credits Pythagoras for being perhaps the earliest
practitioner of physiognomy, in the sixth century BCE. Later on,
Aristotle also came out in support of physiognomy, which attempted
to judge a person’s character from the external features of the body
and especially the face, in relation to the similarity that these features
had to animals. Physiognomy was based on the assumption that
people who shared physical similarities with animals also shared
some psychic properties with these animals. For example, a person
who looked like a fox was sly, or somebody who looked like an owl
was wise (Davey, 1989). What is your view on this?
Many application forms for employment positions or for furthering your studies
require that a photograph be submitted. It is highly unlikely that selection and
admission personnel use these photographs to judge personal attributes of the
applicants, as physiognomists would have done. So why do you think that
photographs are requested? Try to interview someone in the human resources
section of a company, or an admissions officer at an educational institution, to
see what purpose, if any, photographs serve on application forms.
2.2.3 Humorology
In the fifth century BCE, Hippocrates, the father of medicine,
developed the concept that there were four body humours or fluids
(blood, yellow bile, black bile, and phlegm) (McReynolds, 1986).
Galen, a physician in ancient Rome, took these ideas further by
hypothesising four types of temperament (sanguine, choleric,
melancholic, and phlegmatic), corresponding to the four humours
(Aiken & Groth-Marnat, 2005). The problem with the humoral
approach of classifying personality types into one of four categories
was that it remained a hypothesis that was never objectively verified.
Today the humoral theory mainly has historical significance.
However, based on the views of Hippocrates and Galen, Eysenck
and Eysenck (1958) embedded the four temperaments within the
introversion/extroversion and the emotionally stable/emotionally
unstable (neurotic) personality dimensions that they proposed. Of
interest is the fact that Eysenck and Eysenck’s (1958) two personality
dimensions still form the basis for modern personality measures such
as the Myers-Briggs Type Indicator and the 16 Personality Factor
Questionnaire. You can read more about these measures in Chapter
12.
2.2.4 Phrenology
Franz Gall was the founder of phrenology, the ‘science’ of ‘reading
people’s heads’ (McReynolds, 1986). Phrenologists believed that the
brain consisted of a number of organs that corresponded with various
personality characteristics (e.g. self-esteem, cautiousness, firmness)
and cognitive faculties (e.g. language, memory, calculation). By
feeling the topography of a person’s skull, phrenologists argued that it
was possible to locate ‘bumps’ over specific brain areas believed to be
associated with certain personality attributes (Aiken & Groth-Marnat,
2005). The fundamental assumptions underlying phrenology were
later demonstrated to be invalid in research studies. Consequently, no
one really places any value on phrenology today.
2.2.6 Graphology
Graphology can be defined as the systematic study of handwriting.
Handwriting provides graphologists with cues that are called
‘crystallised gestures’ that can be analysed in detail. As handwriting is
a type of stylistic behaviour, there is some logic to the argument that it
could be seen to be an expression of personality characteristics.
Graphologists hypothesise that people who keep their handwriting
small are likely to be introverted, modest, humble, and shun publicity.
Large handwriting, on the other hand, shows a desire to ‘think big’,
which, if supported by intelligence and drive, provides the ingredients
for success. Upright writing is said to indicate self-reliance, poise,
calm and self-composure, reserve, and a neutral attitude (Davey,
1989). Davey (1989) concluded that efforts of graphologists to
establish validity of such claims have yielded no or very few positive
results.
Although there are almost no studies in which it has been found
that handwriting is a valid predictor of job performance, graphology is
widely used in personnel selection to this day (Simner & Goffin, 2003).
It is especially used in France, but also in other countries such as
Belgium, Germany, Italy, Israel, Great Britain, and the United States
(US). This has prompted Murphy and Davidshofer (2005) to ask why
graphology remains popular. What reasons do you think they
unearthed in attempting to answer this question?
Murphy and Davidshofer (2005) concluded that there were three
main reasons that fuelled the popularity of handwriting analysis in
personnel selection:
It has high face validity, meaning that to the ordinary person in the
street it seems reasonable that handwriting could provide indicators of
personality characteristics, just as mannerisms and facial expressions
do.
Graphologists tend to make holistic descriptions of candidates such as
‘honest’, ‘sincere’, and ‘shows insight’, which, because of their
vagueness, are difficult to prove or disprove.
• Some of the predictions of graphologists are valid. However,
Murphy and Davidshofer (2005) cite research studies that reveal
that the validity of the inferences drawn by the graphologists was
related more to what they gleaned from the content of an
applicant’s biographical essay than the analysis of the handwriting!
Despite having found reasons why graphology continues to be used in
personnel selection, Murphy and Davidshofer (2005) concluded that,
all things considered, there is not sufficient evidence to support the
use of graphology in employment testing and selection. Simner and
Goffin (2003) concur with this and argue that the criterion-related
validity of graphology is lower and more variable than that of more
widely known and less expensive measures. For example, whereas
the criterion-related validity of graphology varies between .09 and .16
(Simner & Goffin, 2003), the criterion-related validity of general mental
testing and structured interviews in job selection has been found to be
.51, & when used in combination, the validity coefficient increases to
.63 (Schmidt & Hunter, 1998). Simner and Goffin (2003) thus caution
that the continued use of graphology for personnel selection could
prove to be costly and harmful to organisations.
2.2.7 Summary
All the avenues explored by the early philosophers, writers, and
scientists did not provide verifiable ways of measuring human
attributes. The common thread running through all these attempts (but
probably not in the case of graphology), is the lack of proper scientific
method and, ultimately, rigorous scientific measurement.
New item types have emerged as The need to ensure that the focus
has the use of simulations and remains on the appropriateness of
gaming in assessment (see these new item types for the
Chapter 14 for a fuller discussion). intended assessment purpose and
the intended test-takers, rather
than on simply wanting to
advance the application of new
technologies in assessment.
Among the important trends in this era that would continue into
subsequent eras were:
• the development of measures in response to a need that
existed within a certain political dispensation
• the notion that people who are unfamiliar with the concepts in a
measure should be familiarised with them before they are assessed
• the use of overseas measures and their norms without investigating
whether they should be adapted/revised for use in South Africa.
The Employment Equity Act (EEA) has since been amended. The
following was added to the Employment Equity Amendment Act, Act
47 of 2013: ‘and (d) has been certified by the Health Professions
Council of South Africa or any other body which may be authorised by
law to certify those tests or assessments’ (Republic of South Africa,
2013, section 4).
The EEA and the Employment Equity Amendment Act have major
implications for assessment practitioners in South Africa because
many of the measures currently in use, whether imported from the US
and Europe or developed locally, have not been investigated for bias
and have not been cross-culturally validated here (as was discussed
in Section 2.3.5). The impact of this act and its amendment on the
conceptualisation and professional practice of assessment in South
Africa in general is far-reaching as assessment practitioners and test
publishers are increasingly being called upon to demonstrate, or prove
in court, that a particular assessment measure does not discriminate
against certain groups of people. It is thus not surprising that there has
been a notable increase in the number of test bias studies since the
promulgation of the EEA in 1998 (e.g. Abrahams & Mauer, 1999a,
1999b; Lopes, Roodt & Mauer, 2001; Meiring, Van de Vijver,
Rothmann & Barrick, 2005; Meiring, Van de Vijver, Rothmann &
Sackett, 2006; Schaap, 2003, 2011; Schaap & Basson, 2003; Taylor,
2000; Van Zyl & Visser, 1998; Van Zyl & Taylor, 2012; Visser &
Viviers, 2010).
A further consequence of the EEA and its amendment is that there
is an emerging thought that it would be useful if test publishers and
distributors could certify a measure as being ‘Employment Equity Act
Compliant’ as this will aid assessment practitioners when selecting
measures (Lopes, Roodt & Mauer, 2001). While this sounds like a
very practical suggestion, such certification could be misleading. For
example, just because a measure is certified as being compliant does
not protect the results from being used in an unfair way when making
selection decisions. Furthermore, given the variety of cultural and
language groups in this country, bias investigations would have to be
conducted for all the subgroups on whom the measure is to be used
before it can be given the stamp of approval. Alternatively, it would
have to be clearly indicated for which subgroups the measure has
been found to be unbiased and that it is only for these groups that the
measure complies with the act.
The addition of clause (d) in the amendment of the EEA was
controversial. While the control and use of psychological tests falls
under the Professional Board for Psychology, the Association for Test
Publishers (ATP) instituted legal proceedings and argued that the
Board only has procedures in place to classify psychological tests, not
to issue certificates, as is required in section 8(d) in the amended act.
The court found in favour of the ATP and section 8(d), which required
that tests used in industry should be certified, was rescinded (ATP v.
the President of SA, the Minister of Labour and the Health Professions
Council of South Africa, 2017). Nonetheless, the onus is still on the
assessment practitioner to comply with 8(a), (b) and (c) − namely, that
the psychological tests used are valid and reliable, applied fairly, and
are not biased.
The advent of the Employment Equity Act has forced assessment
practitioners to take stock of the available measures in terms of their
quality, cross-cultural applicability, the appropriateness of their norms,
and the availability of different language versions. To this end, the
Human Sciences Research Council (HSRC) conducted a survey of
test use patterns and needs of practitioners in South Africa (Foxcroft,
Paterson, le Roux & Herbst, 2004). Among other things, it was found
that most of the tests being used frequently are in need of adapting for
our multicultural context or require updating, and appropriate norms
and various language versions should be provided. The report on the
survey, which can be viewed at www.hsrc.ac.za, provides an agenda
for how to tackle the improvement of the quality of measures and
assessment practices in South Africa as well as providing a list of the
most frequently used measures, which should be earmarked for
adaptation or revision first.
Having an agenda is important, but having organisations and
experts to develop and adapt tests is equally important. As was
discussed in Section 2.4.3, the HSRC, into which the NIPR and IPER
were incorporated, developed a large number of tests during the
1960s through to the 1980s. For at least three decades the HSRC
almost exclusively developed and distributed tests in South Africa.
However, at the start of the 1990s the HSRC was restructured,
became unsure about the role that it should play in psychological and
educational test development, and many staff with test-development
expertise left the organisation. Consequently, since the South African
adaptation of the Wechsler Adult Intelligence Scales-III (WAIS-III) in
the mid-1990s, the HSRC has not developed or adapted any other
tests and its former tests that are still in circulation are distributed by
Mindmuzik Media now. According to Foxcroft and Davies (2008), with
the HSRC relinquishing its role as the major test developer and
distributor in South Africa, other role players gradually emerged to fill
the void in South African test development. Some international test
development and distribution agencies such as SHL (www.shl.co.za)
and Psytech (www.psytech.co.za) have agencies in South Africa.
Furthermore, local test agencies such as Jopie van Rooyen &
Partners SA (www.jvrafrica.co.za) and Mindmuzik Media
(www.mindmuzik.com) have established themselves in the
marketplace. Each of these agencies has a research and
development section, and much emphasis is being placed on adapting
international measures for the South African context. However, there
is still some way to go before all the tests listed on the Web sites of
the agencies that operate in South Africa have been adapted for use
here and have South African norms.
The other encouraging trend is that there is a greater involvement
of universities (e.g. Unisa, Nelson Mandela, North-West, Rhodes,
Witwatersrand, Pretoria, Johannesburg, Stellenbosch, and KwaZulu-
Natal universities) in researching and adapting tests, developing local
norms and undertaking local psychometric studies, and even
developing indigenous tests (e.g. Taylor & De Bruin, 2003).
Furthermore, some organisations such as the South African Breweries
and the South African Police Services (SAPS) that undertake large-
scale testing have undertaken numerous studies to provide
psychometric information on the measures that they use, investigate
bias, and adapt measures on the basis of their findings (e.g. Meiring,
2007).
South African assessment practitioners and test developers have
thus not remained passive in the wake of legislation impacting on test
use. Although the future use of psychological assessment, particularly
in industry and education, still hangs in the balance at this stage, there
are encouraging signs that progress is being made to research and
adapt more measures appropriate for our context and to use them in
fair ways to the benefit of individuals and organisations. The way in
which psychologists and test developers continue to respond to this
challenge will largely shape the future destiny of psychological testing
here.
2.4.4.3 Professional practice guidelines
According to the Health Professions Act (No. 56 of 1974), the
Professional Board for Psychology of the Health Professions Council
of South Africa (HPCSA) is mandated to protect the public and to
guide the profession of psychology. In recent years, the Professional
Board has become increasingly concerned about the misuse of
assessment measures in this country, while recognising the important
role of psychological assessment in the professional practice of
psychology as well as for research purposes. Whereas the
Professional Board for Psychology had previously given the Test
Commission of the Republic of South Africa (TCRSA) the authority to
classify psychological tests and oversee the training and examination
of certain categories of assessment practitioners, these powers were
revoked by the Board in 1996. The reason for this was that the
TCRSA did not have any statutory power and, being a section 21
company that operated largely in Gauteng, its membership base was
not representative of psychologists throughout the country. Instead,
the Professional Board for Psychology formed the Psychometrics
Committee, which, as a formal committee of the Board, has provided
the Board with a more direct way of controlling and regulating
psychological test use in South Africa. The role of the Psychometrics
Committee of the Professional Board and some of the initiatives that it
has launched will be discussed more fully in Chapter 8.
The further introduction of regulations and the policing of
assessment practitioners will not stamp out test abuse. Rather,
individual assessment practitioners need to make the fair and ethical
use of tests a norm for themselves. Consequently, the Psychometrics
Committee has actively participated in the development of
internationally acceptable standards for test use in conjunction with
the International Test Commission’s (ITC) test-use project (see
Section 2.3 and Chapter 8) and has developed certain competency-
based training guidelines (see www.hpcsa.co.za). Furthermore, the
various categories of psychology professionals who use tests have to
write a national examination under the auspices of the Professional
Board before they are allowed to practise professionally. Such a
national examination helps to ensure that professionals enter the field
with at least the same minimum discernible competencies. In Chapter
8, the different categories of assessment practitioners in South Africa
will be discussed, as will their scope of practice and the nature of their
training.
To complicate matters, the introduction of computer-based and
Internet-delivered testing has led to a reconceptualisation of the
psychology professional’s role, especially in terms of test
administration. There are those who argue that trained test
administrators (non-psychologists) can oversee structured, computer-
based testing and that the test classification system in South African
should make allowance for this. Indeed, in the landmark court case
that hinged on the definition of what test use means, the honourable
Judge Bertelsmann concluded that the Professional Board needs to
publish a list of tests that are restricted for use by psychologists as it
‘is unlikely that the primarily mechanical function of the recording of
test results should be reserved for psychologists’ (Association of Test
Publishers of South Africa; and Another, Saville and Holdsworth South
Africa v. The Chairperson of the Professional Board of Psychology,
2010, p. 20). Whether it is correct to conclude that test administration
is purely mechanistic will be critically examined in Chapter 9.
As you look back over this section, you should become aware that psychological
assessment in South Africa has been and is currently being shaped by:
• legislation and the political dispensation of the day
• the need for appropriate measures to be developed that can be used in a
fair and unbiased way for people from all cultural groups in South Africa
• the role that a new range of test development and distribution
agencies and universities are playing to research, adapt, and
develop measures that are appropriate for our multicultural context
• the need for assessment practitioners to take personal
responsibility for ethical test use
• training and professional practice guidelines provided by statutory (e.g. the
Professional Board for Psychology) and other bodies (e.g. PsySSA, SIOPSA).
Spend some time finding the common threads or themes that run through the
historical perspective of assessment from ancient to modern times, internationally
and in the South African context. You might also want to read up on the
development of psychological testing in Africa (see for example Foxcroft, 2011b).
You have been asked to appear on a TV talk show in which you need to convince
the South African public that psychological assessment has an important role to
play in the democratic, post-apartheid South Africa. Write down what you plan to
say on the talk show and then make a recording of your talk using your cell
phone. Get a classmate to watch the recording and give you feedback.
CHECK YOUR PROGRESS 2.1
2.1 Define the following terms:
• Phrenology
• Graphology
• Psychometrics
• Standardised
• Standards
2.2 Describe how the following has impacted on psychological test
use and development in South Africa:
• Apartheid
• The Employment Equity Act and its amendment.
Questions:
(a) Is it appropriate for Sam to use a translator? Motivate your answer.
(b) Will the use of a translator be sufficient to ensure that the
measures are not biased against test-takers who do not have
English as a home language? Motivate your answer.
(c) How else could Sam have approach this dilemma?
(Hint: consult the Ethical Rules of Conduct for Professionals Registered under the
Health Professions Act, 1974 (Department of Health, 2006), which you can source
from www.hpcsa.co.za)
2.6 Conclusion
This chapter introduced you to the history and development of the field
of psychological assessment from ancient to modern times,
internationally and in the South African context. Attention was given to
the factors that shaped the development of psychological assessment
in South Africa from the early years, in the apartheid era, and in the
democratic era. The chapter concluded with a discussion on whether
psychological assessment fulfils a useful function in modern society.
So far we are about one quarter of the way in our journey through
the Foundation Zone on our map. Have you picked up some of the
fundamental assessment concepts already? You should be
developing an understanding that:
psychological assessment is a complex process
• psychological measures/tests represent a scientific approach to
enquiring into human behaviour; consequently, they need to be
applied in a standardised way and have to conform to rigorous
scientific criteria (i.e. it must be empirically proved that they are
reliable and valid and, especially in multicultural contexts, that they
are not biased).
Let’s move on to the next chapter, where you can explore some of the
foundational assessment concepts further.
Basic measurement and
statistical concepts
GERT ROODT AND FRANCOIS DE KOCK
Chapter CHAPTER OUTCOMES
3 The next three chapters are the next stops in the Foundation Zone and they
introduce you to the basic concepts needed for understanding the properties
of psychological measures and how test results are interpreted. The topics
covered in these three chapters include basic measurement and scaling
concepts, reliability, and validity. In the first chapter the following themes,
namely levels of measurement, measurement errors, measurement scales,
basic statistical concepts, and test norms, are introduced.
Once you have studied this chapter you will be able to:
❱ describe the three distinguishing properties of measurement levels
❱ describe and give an example of the four different measurement levels
❱ describe and give an example for each of the different measurement scales
❱ explain for each of the different scaling methods what type
of data (measurement level) it will generate
❱ define and describe the three basic categories of statistical
measures, namely location, variability, and association
❱ name and describe the different types of test norms
❱ apply measurement concepts and evaluate your understanding.
3.1 Introduction
It is important to take some time to consider measurement issues in
psychological assessment. Recall the words by Thorndike (1918)
introduced in Chapter 2: ‘Whatever exists at all exists in some amount.
To know it thoroughly involves knowing its quantity as well as its
quality’ (p. 16). In psychology, when we measure the intangible
characteristics of people, such as their intelligence, personality, or
depression, we place a great deal of emphasis on measurement. In
psychological measurement (psychometrics), we are mostly dealing
with the measurement of abstract, intangible constructs. Although
these constructs are not visible or tangible, it does not mean that they
do not exist. We can still measure their existence in the different ways
that they manifest themselves. For instance, we can measure
intelligence by assessing people’s ability to solve complex problems,
by assessing their mental-reasoning ability, or sometimes even by the
way people drive.
Guilford (1936) referred to the great German philosopher
Immanuel Kant who once asserted that the sine qua non of a science
is measurement and the mathematical treatment of its data. Kant’s
view was that psychology could never rise to the dignity of a natural
science because, at the time, it was not possible to apply quantitative
methods to its data. Guilford concluded that if Kant today were to
browse through one of the contemporary journals of psychology, he
would at least be forced to conclude that psychologists as a group are
expending a large amount of energy in maintaining the pretence that
their work is science. Following this, Guilford then posed the question:
What would be the motive of modern-day scientists for spending this
arduous amount of energy to express their findings in terms of
statistical probability and significance coefficients? He then concluded
it is all in our ‘struggle for objectivity. Objectivity is after all the
touchstone of science’ (1936, p. 1) (own emphasis in italics). This
objectivity can be achieved through effective measurement. Runyon
and Haber define measurement as follows:
EXAMPLE A
Totally 1 2 3 4 5 Totally
disagree agree
EXAMPLE B
1 2 3 4 5
formula 3.1
The mean is thus calculated by the sum (∑) of all scores divided by
the number (N) of the scores.
The median presents the score where, when the data are
arranged in numerical order, 50 per cent of the distribution is below
that point and the other 50 per cent is above that point. As such, the
median is also called the 50th percentile. The median (Me) is not that
easy to derive by formula (see Price, 2017, p. 34) but can be readily
obtained from standard statistical software (e.g. MS Excel, SPSS,
etc.). The median is a very useful statistic because it is not influenced
as much by one or a few extreme low or high scores. In this way, it is
known as a less biased estimator of the population mean as
compared to the sample mean.
The median position (or location) is the location of the median in
an ordered series, which simply means that the median is the xth
number in an ordered series. The formula for calculating the median
position (Med) is as follows (Howell, 1995):
formula 3.2
Table 3.2 Aptitude and job performance scores of machine operators
B 8 10 80 18 8
C 7 7 49 14 10
D 8 9 72 17 8
E 7 6 42 13 6
F 6 8 48 14 9
G 5 4 20 9 6
H 8 7 56 15 9
I 6 5 30 11 8
J 2 3 6 5 6
Calculate the means for the test scores provided in Table 3.2 using the formulae
provided above (please note that upper- and lower-case descriptors are used):
(a) What is the mean score for column xo
(b) What is the mode for scores in column xe
(c) What is the median position for scores in column X?
You can find the answers at the end of the chapter.
Can you determine the following ranges of the different sets of scores in Table 3.2?
(a) for the scores in column X?
(b) for the scores in column Y?
You can find the answers at the end of the chapter.
The variance gives an indication of how scores vary about the mean
or how they deviate from the mean . You will find that half of
these scores have positive values while the other half have negative
values, with the result that the sum of these deviations always equals
zero. By squaring the deviations one can overcome the problem of the
zero sum (squared minuses become a plus) and the outcome will
therefore always have a positive value. The sample variance (denoted
as s2) can now be calculated as the sum of the squared deviations
about the mean, divided by n − 1. According to Howell (1995) the
formula for the variance is as follows:
formula 3.3
formula 3.4
formula 3.5
Runyon and Haber (1980) propose that this formula can be simplified
further by replacing the variances and covariances with their
computational formulae and by resolving the equation. This formula is
also known as the Pearson product-moment correlation:
formula 3.6
formula 3.7
The value of the slope (b) and the intercept (a) can be calculated once
we know the value of r (Howell, 1995).
formula 3.8
Since we know the value of b, we can now proceed by calculating the
value of a by using the following formula (Howell, 1995):
formula 3.9
In the case of multiple predictors (X1 and X2), the equation slightly
changes to become a multiple regression equation:
formula 3.10
In the case of the multiple regression equation the b1 and b2 denotes
the respective regression weights of the predictor variables X1 and X2
with a as the value of the intercept.
3.6 Norms
We alluded earlier in the book to the fact that some measures may be
norm-referenced while others are criterion-referenced. We pointed out
that with norm-referenced measures each test-taker’s performance is
interpreted with reference to a relevant standardisation sample or
norm group. Criterion-referenced measures on the other hand
compare the test-taker’s performance to the attainment of a defined
skill or content. The last of the basic concepts in testing that you need
to understand before studying the rest of the book is norms. We will
discuss definitions and types of norms in this section. Other aspects
concerning the use of norms and interpretation of test scores are
covered in Chapters 5 and 16.
formula 3.11
The disadvantages are that these scores are less evident than
ordinary z-scores and are statistically more complex. Three frequently
used standard scores are McCall’s T-score, stanine scales, and sten
scales.
formula 3.13
Sten scale: The rationale for the sten scale (a contraction of ‘standard
10’) is the same as for the stanine scale except that it consists of 10
scale units. The mean is 5.5 and the standard deviation 2. The normal
distribution curve percentages that fall into each of the 10 categories
are given in Table 3.4. The sten scale has the same advantages and
disadvantages as the stanine scale. Standard scores, z, are
transformed to sten scores as follows:
formula 3.14
Table 3.3 Normal curve percentages for the stanine scale
STANINE 1 2 3 4 5 6 7 8 9
PERCENTAGE 4 7 12 17 20 17 12 7 4
Table 3.4 Normal curve percentages for the sten scale
STEN 1 2 3 4 5 6 7 8 9 10
PERCENTAGE 2 5 9 15 19 19 15 9 5 2
16–20 10 90 10
11–15 8 80 20
6–10 15 30 70
Below 5 10 10 90
3.7 Conclusion
This chapter introduced you to the basic concepts in psychological
measurement. You learned about the four types (levels) of
measurement, and different methods of measurement scaling. You
were also introduced to the basic statistical concepts of displaying
data; i.e. measures of central tendency, variability, and association.
The chapter ended with a discussion of the different types of norm
scores, and procedures for applying norms in assessment settings.
What you have learned in this chapter will enable you to establish
whether your data meet the criteria for applying particular statistical
analyses (e.g. in normally distributed data versus distribution-free
data). This is an important consideration that is often overlooked in
psychological and/or social research. Besides this, measurement data
should also meet reliability and validity requirements. These two
concepts will be discussed in more detail in the next two chapters.
ADVANCED READING
Cronbach, L.J. 1990. Essentials of psychological testing. 5th ed.
New York: Harper & Row.
Gregory, R.J. 2007. Psychological testing: History, principles and applications. 5th ed.
Boston: Pearson Education.
Groth-Marnat, G. 2003. Handbook of psychological assessment. 4th ed. Hoboken, NJ:
John Wiley & Sons.
Kerlinger, F.N. & Lee, H.B. 2000. Foundations of behavioral research. 4th ed. Fort
Worth, TX: Harcourt College Publishers.
Miller, L.A., Lovler, R.L. & McIntire S.A. 2011. Foundations of psychological testing: A
practical approach. 3rd ed. Thousand Oaks, CA: Sage.
Murphy, K.R. & Davidshofer, C.O. 2005. Psychological testing: Principles and
applications. 6th ed. Upper Saddle River, NJ: Prentice Hall.
Reliability: Basic concepts
and measures
GERT ROODT AND FRANCOIS DE KOCK
Chapter CHAPTER OUTCOMES
4 Once you have studied this chapter you will be able to:
❱ define ‘reliability’
❱ name and describe the different types of reliability
❱ interpret a reliability coefficient
❱ explain which factors may affect reliability coefficients.
4.1 Introduction
As discussed in the previous chapter, all measures used in
psychological assessment contain some form of measurement error.
Even if the characteristic we are trying to measure with Test A, for
example, remains constant, scores from this test are likely to fluctuate.
Let’s look at some example data to illustrate. As a psychologist at a
packaging organisation you wish to use an aptitude measure
consisting of 20 items to assist you in the selection of machine
operators. You decide to assess 10 current operators to determine the
reliability of the measures. On both the aptitude test and the
performance evaluation, higher scores indicate more favourable
results. In Table 4.1 (on p. 60) we have a fictitious data set of 10
machine operators. The shaded column on the extreme left presents
the individual machine operators (A–J), the second and third column
their aptitude scores for odd- and even-numbered items, as well as
their combined (total) scores for times one (T1) and two (T2) in the
fifth and sixth columns respectively. We will use this example to do
some very basic calculations to illustrate the different concepts that we
want to introduce.
One of the best ways to get to grips with the topics discussed in
this chapter is to try out some of the calculations. In order to gain a
better understanding of the concepts that we introduce in this and the
following chapter you can use the data provided in Table 4.1 to do the
calculations yourself.
Table 4.1 Aptitude scores of machine operators
OPERATOR APTITUDE TEST SCORES
A 5 6 30 11 12 132
B 8 10 80 18 16 288
C 7 7 49 14 15 210
D 8 9 72 17 17 289
E 7 6 42 13 14 182
F 6 8 48 14 13 182
G 5 4 20 9 9 81
H 8 7 56 15 14 210
I 6 5 30 11 13 143
J 2 3 6 5 6 30
formula 4.2
Reliability (rxx) can be defined as the ratio of true score variance to
observed score variance:
formula 4.3
In Table 4.1 you will find two sets of measures for the aptitude total
scores (at time 1 and at time 2).
(a) Calculate the test-retest reliability (coefficient of stability) for the
aptitude measure.
(b) Is this coefficient ‘acceptable’ in your view?
formula 4.4
in which:
rhh = the correlation coefficient of the half-tests.
In Table 4.1 you will find the scores for the odd-numbered (xo) and
the even-numbered items (xe). These two sets of scores represent
two equivalent halves of the aptitude test.
(a) Calculate the split-half reliability (coefficient of internal consistency) of the
aptitude measure.
(b) Use the Spearman-Brown formula above for correcting your
internal consistency coefficient.
(c) Is this coefficient acceptable?
formula 4.5
where:
rtt = reliability coefficient
n = number of items in test
St2 = variance of total test score
pi = proportion of test-takers who answered item i correctly qi
= proportion of test-takers who answered item i incorrectly.
formula 4.6
where:
α = reliability coefficient
n = number of items in test
st2 = variance of total test score
∑si2 = sum of individual item variances.
formula 4.7
where:
α = inter-scorer reliability coefficient
n = number of items or rating dimensions in a test
St2 = variance on all the raters’ summative ratings (total scores)
∑Si2 = the sum of the raters’ variances across different rating
dimensions (sub-scores).
4.2.2.6 Intra-scorer reliability
Whereas the inter-scorer reliability refers to the consistency of ratings
between raters, the intra-scorer reliability coefficient refers to the
consistency of ratings for a single rater. Repeated ratings or scores by
the same rater would give an indication of the degree of error variance
between such ratings for that particular rater. The same Coefficient
Alpha formula as above is then slightly adapted in the following
manner:
formula 4.8
where:
α = intra-rater reliability coefficient
n = number of items or rating dimensions in a test
St2 = variance of a rater’s scores on different individual’s
summative ratings (total scores)
∑Si2 = the sum of the raters’ variances on each rating dimension
(sub-scores) for different individuals being assessed.
4.2.2.7 Contemporary approaches to estimating reliability
In addition to the methods outlined above, a vast array of
contemporary approaches to reliability estimation has been
developed. Even Cronbach, one of the main early protagonists of
Coefficient Alpha, had serious reservations about the limits of
mainstream methods we heavily rely on today to estimate reliability: ‘I
no longer regard the alpha formula as the most appropriate way to
examine most data’ (Cronbach, 2004, p. 403). In response to these
limitations, various alternative indices have been developed, reflecting
either the classic tradition (based on Classical Test Theory), the
random effects model tradition (based on generalisability theory), the
confirmatory factor analytic tradition or, finally, the Item Response
Theory (IRT) tradition (Putka, 2017; Price, 2017). Estimating reliability
is surely important and this field is developing rapidly.
formula 4.9
Where:
st = SD of the test score
rtt = reliability of the test scores.
Like any standard deviation, SME can be interpreted in terms of the
normal distribution frequencies (see Section 3.5). If, for example, SME
= 4, the chances are 68 per cent that a person’s true score on the
test will be 4 points on either side of his or her measured/observed
score (e.g. between 26 and 34 if the measured score is 30).
Questions
(a) Can Dr Mthembu just assume that the managerial aptitude test
will also be reliable and valid in the South African context based
on the US validation studies?
(b) Based on these test results, should Dr Mthembu just proceed with
the selection of the best candidates for the programme or should he
first try to establish the causes for the group differences in reliability?
(c) What could these causes for group differences in reliability potentially be?
(d) How may Dr Mthembu potentially increase the levels of reliability?
4.3 Conclusion
This chapter introduced and defined the concept ‘reliability’, which is
an important building block in working with and understanding
psychological measurement data. This chapter introduced different
types of reliability, as well as factors (individual and administrative)
that affect the reliability of measures, and finally how reliability should
be interpreted. What you have learned in this chapter will enable you
to determine whether the data you intend to use is reliable or meets
reliability requirements. But instruments and measures can never be
valid without being reliable. Validity will be discussed in more depth in
the next chapter.
(b) Assuming that internal and external conditions remain equal and
that no transfer effects occur, there is high test-retest reliability.
4.2 (a) The split-half reliability (coefficient of internal consistency) is:
(rounded to two decimals: .82)
(b) Using the Spearman-Brown formula, the corrected internal
consistency coefficient is:
(c) Yes this coefficient is highly acceptable.
4.3 (a) The KR 20 is: (rounded to two decimals: .94)
(b) Yes, it can be used because there is little error variance (high reliability).
4.4 (a) The Coefficient Alpha is: (1.091)(.7) = .764 (rounded to two decimals: .76)
(b) Yes, the questionnaire is moderately-highly reliable (considering
the amount of error variance in the questionnaire).
4.5 (a) The SME is:
(b) A person’s true score will vary +/- 3.12 about the obtained score.
ADVANCED READING
Anastasi, A. & Urbina, S. 1997. Psychological testing. 7th ed. Upper Saddle River, NJ:
Prentice-Hall.
Cronbach, L.J. 1951. Coefficient alpha and the internal structure of
tests. Psychometrika, 16(3): pp. 297–334. doi:10.1007/bf02310555
Cronbach, L.J. 1990. Essentials of psychological testing. 5th ed.
New York: Harper & Row.
Gregory, R.J. 2010. Psychological testing: History, principles and applications. 5th ed.
Boston: Pearson Education.
Groth-Marnat, G. 2003. Handbook of psychological assessment. 4th ed. Hoboken, NJ:
John Wiley & Sons.
Kerlinger, F.N. & Lee, H.B. 2000. Foundations of behavioral research.
4th edition. Fort Worth, TX: Harcourt College Publishers.
Miller, L.A., Lovler, R.L., & McIntire S.A. 2011. Foundations of psychological testing: A
practical approach. 3rd ed. Thousand Oaks, CA: Sage.
Murphy, K.R. & Davidshofer, C.O. 2005. Psychological testing: Principles and
applications. 6th ed. Upper Saddle River, NJ: Prentice Hall.
Nunnally, J.C. & Bernstein, I.H. 1994. Psychometric theory. New York: McGraw-Hill.
Price, L.R. 2017. Psychometric methods: Theory into practice. New
York: Guilford Press.
Putka, D. 2017. Reliability. In J. Farr, & N. Tippins (Eds.), Handbook of
employee selection. 2nd ed. (pp. 3-33). Mahwah, NJ: Lawrence Erlbaum.
Validity: Basic concepts and measures
GERT ROODT AND FRANCOIS DE KOCK
Chapter CHAPTER OUTCOMES
5 Once you have studied this chapter you will be able to:
❱ define ‘validity’
❱ name and describe the three broad types of validity in the
validity triangle and the sub-categories under each of them
❱ provide examples of indices of validity as well as their interpretation
❱ explain which factors may affect validity coefficients.
5.1 Introduction
In the previous chapter we mentioned that the Employment Equity Act
(No. 55 of 1998) specifies that psychological measures and other
assessments should be reliable and valid. Reliability and validity are
two important criteria that all measuring instruments have to meet.
Now let us return to the practical example that we introduced in the
previous two chapters.
As a psychologist at a packaging organisation you wish to use an
aptitude measure consisting of 20 items to assist you in the selection
of machine operators. You decide to assess 10 current operators to
determine the validity of the measures. On both the aptitude test and
the performance evaluation, higher scores indicate more favourable
results.
In Table 5.1 on the next page we have a fictitious data set of 10
machine operators. The shaded column on the extreme left presents
the individual machine operators (A–J), the second and fourth column
their mechanical aptitude scores for Test X and B respectively, and
their job-performance scores (Y) in the sixth column. We will use this
example to do some very basic calculations to illustrate the different
concepts that we want to introduce.
In order to get to grips with the topics that we introduce in this
chapter, you can try out some of the calculations yourself by using the
data provided in Table 5.1 on the next page.
Table 5.1 Aptitude and job-performance scores of machine operators
OPERATOR MECHANICAL APTITUDE TEST SCORES JOB
(TEST X AND TEST B) PERFORMANCE
SCORES (Y)
A 11 121 15 225 7 49
B 18 324 27 729 8 64
D 17 289 10 100 8 64
E 13 169 17 289 6 36
F 14 196 13 169 9 81
G 9 81 8 64 6 36
H 15 225 11 121 9 81
I 11 121 7 49 8 64
J 5 25 7 49 6 36
∑X∑Y= ∑X ∑B ∑B ∑Y
9779 = =
16510 10010
5.2 Validity
At this point you should already know what reliability is. One other
requirement of any measure is that it must be valid. In this section we
will focus on the concept of validity.
The validity of a measure concerns what the test measures and how
well it does so. As such, measurement validity reflects the accuracy of
measurement.
Predictive validity
If we use both the aptitude tests’ scores in the prediction of the job-
performance scores:
(a) Would Aptitude Test X explain more variance than Aptitude Test B?
(b) What do the results suggest?
formula 5.1
Where:
sy = standard deviation (SD) for scale Y
r2xy =coefficient of determination for scales X and Y
The standard error of estimation is interpreted in the same way as a
standard deviation. If, for example, rxy = .60 and SEest = .80, where sy
= 1, it can be accepted with 95 per cent certainty that the true criterion
score will not deviate (plus or minus) more than (1.96 × .8) = 1.57
from the predicted criterion score.
With reference to our example (Exercise 5.2), aptitude test scores (x) were
used to predict performance scores within a 95 per cent confidence interval:
(a) Compute the standard error of estimation of the job
performance criterion where rXY = 0.61, sY = 1.42.
(b) How much would these scores vary about an obtained score of 30?
formula 5.2
formula 5.3
where:
X1, X2, X3 =predictors 1, 2, 3
b1, b2, b3 = weights of the respective predictors
b0 = intercept (i.e. the a in simple regression).
You have conducted the correlation and regression between the Aptitude Scores
(X) and the Job Performance Scores (Y) in Exercise 5.3. If b = .22 and a
= 4.91: What would the predicted value of Ŷ be for value X = 16?
Questions
(a) Should Dr Mthembu proceed to use the predictive model as
sufficient evidence for the validation of his selection model?
What is the basis for your argument?
(b) On the basis of what you know about the learning potential and the
managerial aptitude tests, would the regression slopes and the
intercepts of the regression lines be the same for the different
culture/race groups in predicting programme performance?
(c) Is Dr Mthembu’s selection model equitable and fair for the
different culture groups?
5.3 Conclusion
This chapter introduced and defined the concept ‘validity’. Three broad
types of validity were introduced and within these three categories
different variations of validity were explained. The rest of the chapter
focused on the application of validity, factors that affect validity
coefficients, and the different validity indices used in practice. What
you have learned thus far in Chapter 5 will enable you to establish or
assess whether measures are valid or generate data that is valid.
It would also be helpful if you could get hold of the Guidelines for
the Validation and Use of Assessment Procedures compiled by the
Society for Industrial and Organisational Psychology of South Africa
(SIOPSA, 2006b). These guidelines were developed to assist
psychologists and assessment practitioners in industry to adhere to
the requirements of the Employment Equity Act (No. 55 of 1998) to
ensure that assessment practices are reliable, valid, unbiased, and
fair.
(b) The true score will (with 95% certainty) not deviate more
than 2.2 from the predicted score of 30, that is vary between
27.8 to 32.2 (1.96 × 1.124 = 2.20)
5.7 (a) Predicted value of Y where X = 16 is: 8.43
ADVANCED READING
Anastasi, A. & Urbina, S. 1997. Psychological testing. 7th ed. Upper Saddle River, NJ:
Prentice-Hall.
Cronbach, L.J. 1990. Essentials of psychological testing. 5th ed.
New York: Harper & Row.
Gregory, R.J. 2010. Psychological testing: History, principles and applications. 5th
edition. Boston: Pearson Education.
Groth-Marnat, G. 2003. Handbook of psychological assessment. 4th ed. Hoboken, NJ:
John Wiley & Sons.
Miller, L.A., Lovler, R.L., & McIntire S.A. (2011). Foundations of psychological testing: A
practical approach. 3rd ed. Thousand Oaks, CA: Sage.
Murphy, K.R. & Davidshofer, C.O. 2005. Psychological testing: Principles and
applications. 6th ed. Upper Saddle River, NJ: Prentice Hall.
Nunnally, J.C. & Bernstein, I.H. 1994. Psychometric theory. New York: McGraw-Hill.
Developing a psychological measure
CHERYL FOXCROFT
Chapter CHAPTER OUTCOMES
6 By the end of this chapter you will be able to:
❱ list the steps involved in developing a measure
❱ explain the essential psychometric properties and item characteristics
that need to be investigated when developing a measure
❱ discuss how technology has revolutionised test development
❱ explain what a test manual should contain
❱ explain how to evaluate a measure.
6.1 Introduction
Developing a psychological measure is a time-consuming and people-
intensive process. Using technology has improved the efficiency of the
test-development process but has also added additional aspects such
as designing multimedia items that involve drawing on the expertise of
graphic designers and technologists. It takes three to five years to
develop a new measure from the point at which the idea is conceived
to the moment when the measure is published and becomes available
to the consumer. In-depth, advanced psychometric knowledge and
test-development expertise is required by those who initiate and guide
the development process. It is beyond the scope of this book to
provide detailed information regarding the development of a measure.
Instead, this chapter will give you a brief glimpse of the steps involved
in developing a measure. The main purpose in outlining the steps in
developing tests is to enable you to evaluate the technical quality of a
measure and the information found in a test manual.
As you can see from the above, should all these people be included,
the team could be large, which could reduce the efficiency of the test-
development process. Figure 6.1 is an example of a more efficient
approach, in which a core team takes overall responsibility for
directing and managing all aspects of the test-development process.
The core team will normally be decided on first, and then they will
identify the other teams and specialists needed to undertake aspects
of the development process. These specialist teams are activated at
relevant points in the process. For example, the construct domain
team(s) could be consulted to map out the domains and subdomains
that the measure will tap. This team could also be involved in item
development, sometimes in collaboration with other teams (e.g. the
design and technology specialists). Once the experimental version of
the measure has been administered, and item and factor analysis
results are available, the construct domain team can identify
problematic items and revise or replace these items. The core team, in
consultation with the construct domain team, will also sign off on all
the items in the final version of the measure.
Figure 6.1 An efficient approach to manage the relationship between the core and expert
test-development teams
6.2.1.2 Specify the aim, target population, and nature of the measure
The core test-development team develops a test plan that contains
the specifications of the test and will guide its development. As a first
step in developing the plan, the following aspects need to be clearly
stated:
the purpose or aim of the measure
what attribute, characteristic, or construct it will measure
whether the measure is to be used for screening purposes or for in-
depth diagnostic assessment, or for competency-based selection and
training purposes
what types of decisions could be made based on the test scores
for which population (group of people) the measure is intended (e.g.
young children, adults, people with a Grade 12 education). Included in
this should be consideration of whether the measure is to be used in a
multicultural and multilingual context and, if so, which cultural and
language groups it could be used with
whether the measure can be individually administered and/or
administered in a group context
whether the measure is paper-based, computer-based or
performance-based (see also 6.2.1.4)
whether it is a normative measure (where an individual’s performance
is compared to that of an external reference or norm group), an
ipsative measure (where intra-individual as opposed to interindividual
comparisons are made when evaluating performance), or a criterion-
referenced measure (where an individual’s performance is interpreted
with reference to performance standards associated with a clearly
specified content or behavioural domain such as managerial decision-
making or academic literacy).
The box below provides an example of the wording of the aim of a
fictitious psychological measure as it would appear in the test plan
and manual.
The last step in defining the content of the measure is to specify the
number of items per construct and subconstruct relative to other
biographical variables (e.g. age groups) and/or psychometric
considerations (e.g. range of difficulty levels, cognitive load), for
example.
Develop key aspects of the test plan for the SADST that was outlined in
6.2.1.2 and 6.2.1.3. Expand the table/concept map that you developed to
map the construct domains and subdomains and add the number of
items for each subdomain, remembering that the SADST covers the age
range from birth to 5 years. Also think about the format of the items and
the nature of the responses that will be required.
ADVANCED READING
Zuilkowski et al. (2016, p. 341) argue that while Westernised children ‘are exposed
to extensive two-dimensional materials during early childhood, such as picture
books and photographs, most rural African children are not’. Most Western-oriented
tasks of non-verbal reasoning involve books of line drawings and visual patterns that
need to be completed, for example. Such two-dimensional stimuli could be foreign to
(rural) African children and result in the assessment results being biased. To create
more appropriate stimuli to tap non-verbal rea-soning in Zambian children,
Zuilkowski et al. (2016) developed the Object-based Pattern Reasoning Assessment
(OPRA). Instead of drawings, OPRA items use familiar objects such as beans,
stones, and bottle caps. The assessment practitioner uses a sheet of paper or
cardboard with a grid of six squares on, and puts five red stones in five squares of
the grid, leaving one square with no stone in. The child is then given a choice of four
objects, one of which will be a red stone. ‘The child can answer by pointing to their
chosen item, naming their selection verbally, or by placing the item in the correct cell
in the grid’ (p. 346). When comparing the performance of children on two-
dimensional stimuli, performance on the OPRA was found to be more valid. The
work done to develop the culturally sensitive stimuli for OPRA reminds us of the
importance of investing time and creativity into
developing tests containing indigenous stimuli and tasks.
Source: Zuilkowski, S.S., McCoy, D.C., Serpell, R., Matafwali, B., and Fink,
G. 2016. Dimensionality and the development of cognitive assessments for
children in sub-Saharan Africa. Journal of Cross-Cultural Psychology,
47(3), pp. 341-354. DOI: 10.1177/0022022115624155
ADVANCED READING
Drozdick, L.W., Getz, K., Raiford, S. E., & Zhang, O. 2016. WPPSI-
IV: Equivalence of Q-interactive and paper formats. Q-
interactive Technical Report 14. Bloomington, MN: Pearson.
Retrieved 3 May 2018 from
http://www.helloq.com/content/dam/ped/ani/us/helloq/media/WPP
Qi-Tech-Report-14-FNL.pdf
Gelderblom, J.H. 2008. Designing technology for young children:
Guidelines grounded in a literature investigation on child
development and children’s technology. Unpublished doctoral
dissertation, University of South Africa.
Pade, H. 2014. The evolution of psychological testing: Embarking on
the age of digital assessment. In R. Valle & J. Klimo (Eds.), The
changing faces of therapy: Innovative perspectives on clinical
issues and assessment. Retrieved 3 May 2018 from
https://www.helloq.com/content/dam/ped/ani/us/helloq/media/The
Parshall, C. G., Harmes, J.C., Davey, T. & Pashley, P. 2010.
Innovative items for computerised testing. In W.J. van der
Linden & C.A. W. (Eds.), Computerised adaptive testing: Theory
and practice (2nd ed). Norwell, MA: Kluwer Academic
Publishers.
Scalise, K. & Gifford, B. 2006. Computer-based assessment in e-
learning: A framework for constructing ‘intermediate constraint’
questions and tasks for technology platforms. Journal of
Technology, Learning, and Assessment, 4(6). Retrieved 21
January 2018 from https://files.eric.ed.gov/fulltext/EJ843857.pdf
The higher the percentage of correct responses, the easier the item;
the smaller the percentage of correct responses, the more difficult the
item. However, p is a behavioural measure related to the frequency of
choosing the correct response, and therefore tells us nothing about
the intrinsic characteristics of the item. Furthermore, the difficulty
value is closely related to the specific sample of the population to
which it was administered. A different sample might yield a different
difficulty value. Nonetheless, one of the most useful aspects of the p-
value is that it provides a uniform measure of the difficulty of a test
item across different domains or dimensions of a measure. It is for this
reason that difficulty values are often used when selecting final items
for a measure.
6.2.4.2 Classical Test-Theory item analysis: Determine
discriminating power One of the purposes of item analysis is to
discover which items best measure the construct or content domain
that the measure aims to assess. Good items consistently measure
the same aspect that the total test is measuring. Consequently, one
would expect individuals who do well in the measure as a whole to
answer a good item correctly, while those who do poorly on the
measure as a whole would answer a good item incorrectly. The
discriminating power of an item can be determined by means of the
discrimination index and item-total correlations.
To compute the discrimination index (D), the method of extreme
groups is used. Performance on an item is compared between the
upper 25 per cent and the lower or bottom 25 per cent of the sample.
If the item is a good discriminator, more people in the upper group will
answer the item correctly.
To compute the discrimination index, the upper and lower 25 per
cent of test-takers need to be identified. Next, the percentage of test-
takers in the upper and lower groups who passed the item is
computed, and the information is substituted into the following
formula:
formula 6.1
where:
U = number of people in the upper group who passed the item
nu= number of people in the upper group
L = number of people in the lower group who passed the item
nl = number of people in the lower group
A positive D-value indicates an item that discriminates between the
extreme groups, while a negative D-value indicates an item with
poor discriminatory power.
An item-total correlation can be performed between the score on
an item and performance on the total measure. A positive item-total
correlation indicates that the item discriminates between those who do
well and poorly on the measure. An item-total correlation close to zero
indicates that the item does not discriminate between high and low
total scores. A negative item-total correlation is indicative of an item
with poor discriminatory power. In general, correlations of .20 are
considered to be the minimum acceptable discrimination value to use
when it comes to item selection.
The advantage that using item-total correlations brings with it is
that correlation coefficients are more familiar to people than the
discrimination index. Furthermore, the magnitude of item-total
correlations can be evaluated for practical significance (e.g. .4 =
moderate correlation) and can be expressed in terms of the
percentage of variance of the total score that an item accounts for
(e.g. if r = .40 the item accounts for 16 per cent of the total variance).
There is a direct relationship between item difficulty and item
discrimination. The difficulty level of an item restricts the discriminatory
power of an item. Items with p-values around .5 have the best
potential to be good discriminators.
6.2.4.3 Item Response Theory (IRT)
By using Item Response Theory, the difficulty level and discriminatory
power of an item can be even more accurately determined. In IRT the
item parameters are sample invariant. That is, the parameters are not
dependent on the ability level of the test-takers responding to the item
and are thus ‘a property of the item and not of the group that
responded to it’ (De Beer, 2005, p. 723).
In IRT an item-response curve (see Figure 7.1 in Chapter 7) is
constructed by plotting the proportion of test-takers who gave the
correct response against estimates of their true standing on a latent
trait (e.g. ability) (Aiken & Groth-Marnat, 2005). The estimates are
derived through logistic equations, with the specific equations used
varying according to the estimation procedures used. In the one-
parameter or Rasch model, an item-response curve is constructed by
estimating only the item difficulty parameter (b). In the two-parameter
model, an item-response curve is constructed by estimating both
difficulty (b) and discrimination (a) parameters. This is the most
popular model, especially since it makes it possible to compare the
difficulty level and discriminating power of two items, for example
(Aiken & Groth-Marnat, 2005). The three-parameter model allows for
variations among the items in terms of difficulty level (b),
discrimination (a), and for guessing (c) on multiple-choice items (De
Beer, 2005).
Especially when you are developing a measure in a multicultural,
multilingual country such as South Africa, it is useful to investigate
item bias in the early stages of developing a measure. IRT also
makes it possible to explore differential item functioning (DIF) in a
much more sophisticated way to identify items that may be biased or
unfair for test-takers from certain groups. One of the common
procedures for detecting DIF is to calculate the area between the item
characteristic curves for two groups (e.g. males and females, or
isiXhosa and isiZulu speakers). The larger the area, the more likely it
is that the item shows DIF. However, as De Beer (2005) points out, it
remains a subjective decision regarding how large the area should be
for DIF to be evident. Usually a combination of visual inspection and
empirically estimated cut-off points is used to flag items that exhibit
DIF. Kanjee (2007) also points out that DIF procedures are only
applicable when two groups are compared. He thus suggests that
logistic regression can be used to simultaneously explore bias when
multiple groups need to be compared. If you are interested in reading
more about how IRT is used for item-analysis purposes during test
development, De Beer (2005) provides an example of how the items
of the LPCAT were analysed and refined using IRT. DIF will also be
discussed in more detail in Chapter 7.
6.2.4.4 Identify items for final pool
Classical test theory, IRT and DIF analyses can be used to determine
item parameters based on which items should be included in the final
version of the measure, or which should be discarded. Available
literature should be consulted to set item-parameter values that must
be met for inclusion or exclusion. For example, if the three-parameter
IRT model is used, a-values should fall between 0.8 and 2.0 while c-
values should fall between 0.0 and 0.3 for an item to be included
(Hambleton & Swaminathan, 1985).
A word of caution, however, is necessary. Item analysis is a
wonderful aid that the test developer can use to identify good and bad
items and to decide on the final item pool. Performance on an item is,
however, impacted on by many variables (e.g. the position of an item
or the administration instructions). Consequently, item-analysis
statistics should not necessarily replace the skill and expertise of the
test developer. At times, the item-characteristic statistics may be poor,
but the test developer could motivate the inclusion of the item in the
final pool on logical or theoretical grounds. In a similar vein, the item
statistics may be very impressive, but on closer inspection of the item,
the test developer may feel that there is no logical or theoretical
reason for retaining the item.
Questions
(a) Why do you think that the test agency approached Jeremy? [Hint: consult
the Health Professions Act and the Ethical Rules for psychologists
regarding the nature and development of psychological tests.]
(b) How should Jeremy respond?
6.3 Evaluating a measure
Test developers and publishers have the responsibility of ensuring
that measures are developed following a rigorous methodology, and
of providing information to verify the measure’s validity and reliability.
Assessment practitioners have the responsibility of evaluating the
information provided about a measure and weighing up whether the
measure is valid and reliable for the intended purpose. Assessment
practitioners can use the information provided in Section 6.2 on the
development of a measure to identify the key technical and
psychometric aspects that they should keep in mind when they
evaluate a measure. When evaluating a measure, assessment
practitioners should, among other things, consider the following:
How long ago a measure was developed, as the appropriateness of
the item content and the norms changes over time
The quality and appeal of the test materials. For computerised tests,
an evaluation of the quality of the test materials includes looking at
how easy it is to make a response; the clarity of graphical and digital
content; the quality of design of the user interface and what the
usability studies indicate; and how robust the design of the software is
(e.g. are responses stored on an ongoing basis so that if the power
fails or Internet connection is lost the test-taker restarts the test at the
point that they were at when the problem occurred?)
The quality of the manual contents (e.g. does it indicate the purpose of
the measure, to whom it should be applied, what the limitations are,
how to administer, score, and interpret the measure)
The clarity of the test instructions (especially where there are different
language versions or it is a computerised test that will be administered
in an automated way)
The cultural appropriateness of the item content and the constructs
being tapped (e.g. has there been an empirical investigation into item
bias, differential item functioning, and construct equivalence for
different groups?)
If the measure is to be used in a multi-cultural and/or multilingual
context, whether bias investigations provide evidence that the
performance of various groups on the measure is equivalent
The adequacy of the psychometric properties that have been
established
• The nature of the norm groups and the recency of the norms.
6.4 Conclusion
In this chapter you have learned about the steps involved in
developing a psychological test, and the psychometric properties and
item characteristics that need to be investigated in the process. The
chapter concluded with a discussion on how to evaluate a measure.
How’s the journey going? There’s only one more stop left in the
Foundation Zone. In the last stretch of uphill climbing for a while, you
will be introduced to issues related to the development and adaptation
of measures in cross-cultural contexts.
Cross-cultural test
adaptation, translation, and
tests in multiple languages
FRANCOIS DE KOCK AND CHERYL FOXCROFT
Chapter CHAPTER OUTCOMES
7 By the end of this chapter you will be able to:
❱ list reasons why assessment measures are adapted
❱ describe what test adaptation is and the different types of adaptation
❱ explain the concepts of bias and equivalence, which are fundamental to
understanding the different designs and methods used to adapt measures
❱ describe statistical methods for detecting bias and equivalence
❱ discuss challenges encountered when adapting measures in
South Africa and suggest possible solutions
❱ describe the simultaneous multicultural test-development process.
7.1 Introduction
As you saw in Section 2.4 in Chapter 2, the history of psychological
test development in South Africa was significantly shaped by our
racially segregated society that resulted from apartheid times as well
as earlier colonial influences. This resulted in measures that have
been developed generally not being available in the range of
languages spoken by South African test-takers, or not being
sufficiently culturally sensitive. In addition, not only are measures
largely Western-oriented rather than African-centred in the nature of
the constructs they assess, many of the measures that were
developed in Europe or the United States have been uncritically
adopted for use here. The challenge of providing linguistically and
culturally relevant measures in a multicultural and multilingual society
is not limited to South Africa. With increased globalisation, it is an
international challenge. As was discussed in Section 2.3.5 in Chapter
2, there are two ways that this challenge can be addressed. On the
one hand, tests developed in one culture can be adapted for use for
test-takers from other cultural and language groups. On the other
hand, many non-Western countries are developing their own
indigenous measures in their cultural contexts. This chapter will
largely focus on test adaptation, with a short section on multicultural
test development. The reason for this focus is that the adaptation of
assessment measures is essential in South Africa if test results are to
be valid and reliable for test-takers from diverse backgrounds.
Adapting measures is certainly a daunting and challenging task
that is not only costly and time-consuming, but also requires high
levels of expertise. In this chapter, we will discuss the reason for
adapting measures, and the aspects to consider when doing so, as
well as how to investigate test equivalence and bias. The challenges
in adapting measures for the multicultural and multilingual South
African context will further be highlighted and some solutions will be
proposed. In addition, the possibility of developing indigenous
multiple-language versions of a measure, using a simultaneous
multicultural approach, will be explored.
Statistical approaches to
7.5
establish equivalence
As noted in Section 7.4.1, it is important that measures used in
multiple populations are equivalent. In other words, psychological
measures should yield scores that mean the same thing across
different groups. We should consistently examine whether an
assessment measure functions similarly across groups, for reasons
that may be leg-ally, practically, socially or theoretically important
(Aguinis, Culpepper & Pierce, 2010; Zickar, Cortina & Carter, 2010). In
this section, we will start out by giving you a broad overview of the
concepts of bias, equivalence, and invariance, before we address the
main statistical techniques used to determine equivalence.
7.5.1 Bias and equivalence
To meaningfully interpret differences on psychological measures
between test-takers from different groups, we have to consider the
threat of bias and non-equivalence (Van de Vijver & Matsumoto,
2011). Bias refers to any systematic* error or noise in test scores that
is caused by irrelevant factors (i.e. any factor other than the construct
that we are trying to measure with a particular test). In other words,
bias is any construct-irrelevant source of variance that results in
systematically higher or lower scores for identifiable groups of test-
takers (American Educational Research Association, American
Psychological Association & National Council for Measurement in
Education, 1999). Bias is closely related to the concept of
equivalence. Johnson (1998) reviewed a wide variety of
interpretations of equivalence and concluded that they can be grouped
into those that focus on equivalence of meaning, and others that focus
on measures and procedures used to make comparisons between
groups of test-takers. Van de Vijver and Leung (2011) suggest an
integrative taxonomy of equivalence and bias. They interpret bias and
equivalence as opposite sides of the same coin, albeit with
distinguishing foci: ‘Bias threatens the equivalence of measurement
outcomes across cultures, and it is only when instruments are free
from bias that measurement outcomes are equivalent, and have the
same meaning within and across cultures’ (p.19). In their view, cross-
group equivalence requires a lack of bias, and the presence of cross-
group bias inevitably results in some form of equivalence. Bias is a
broad term that encompasses various construct-irrelevant sources of
variance, whereas equivalence generally refers to a lack of bias that is
due to group-related effects, specifically. As such, equivalence may be
better understood as a lack of group-related bias, i.e. where bias in
measurement lies primarily in some systematic difference in
measurement between comparison groups. However, in South African
practice, the term ‘equivalent’ normally is used to mean that a test is
free from group bias (in how it measures a particular construct across
populations).
Finally, the terms equivalence and invariance are often used
interchangeably depending on the functional discipline from which the
researcher approaches the statistical analysis. In cross-cultural
psychology, the term equivalence is more often used, although its
implied meaning often includes more substantive cross-cultural issues
(such as the meaning of constructs that may or may not exist across
groups) and not only measurement aspects. In literature that focuses
on equivalence from a purely psychometric point of view, the term
‘invariance’ is more popular. To be clear, when we use the term
‘invariance’ in this chapter, it means ‘equivalence’. In other words, a
measure that lacks invariance between two groups is deemed to be
equivalent.
Various typologies of bias exist (e.g. Murphy & Davidshofer, 2005;
Van de Vijver & Leung, 2011), but for now, let us distinguish broadly
between measurement bias and predictive bias (American
Psychological Association, 2003):
Figure 7.1 Item characteristic curves for two items in same group of test-takers
Figure 7.2 Item characteristic curves for biased item (X) for two groups
Look at the ICCs in Figure 7.2. Assume that Group 1 represents the
ICC for males and Group 2 the ICC for females for an item. Explain
to a friend why this item is biased against females.
Configural invariance only requires that the same items load on the
same factors for both groups (Zickar, Cortina & Carter, 2010).
However, we cannot yet say whether differences may exist between
the groups’ respective factor loadings, item factor intercepts, unique
variances, correlations between factors, and so forth. Often, the model
that is tested here is used as a baseline model, against which more
restrictive successive models are tested, and additional parameters
are constrained in a step-by-step fashion.
Metric invariance: A ‘stronger’ form of equivalence exists when
metric invariance is established.
When item intercepts are set to be the same across groups, and the
model still fits, it is called scalar invariance (Vandenberg & Lance,
2000), intercept invariance (Cheung & Rensvold, 1998, 1999, 2000),
or full-score equivalence (Fontaine, 2005; Van de Vijver & Leung,
1997). If one wants to compare the mean scores of two groups, scalar
invariance needs to be supported.
Scalar equivalence simply means that the measurement scale is
identical across versions of the assessment (Sireci, 2011). For
example, scalar equivalence across Afrikaans- and English-language
versions of a test would mean that a score of 80 on one version of the
test ‘means the same thing’ as the same score on the other language
version of the test.
Apart from CFA techniques that assess scalar equivalence, item
bias analyses, such as differential item functioning (DIF) analysis, may
also be used to establish scalar equivalence. In such cases, DIF is
typically used in conjunction with exploratory factor analysis – EFA
identifies the equivalent factors, and DIF is used to examine the
appropriateness of items in more detail (Van de Vijver & Leung, 2011).
Figure 7.3 No predictive bias, because lines of prediction coincide (slopes and intercepts are equal)
Figure 7.4 Predictive bias in intercepts only
ADVANCED READING
7.8 Conclusion
In the context of a multicultural society like South Africa, the
adaptation of measures and the detection and elimination of bias from
measures plays a vital role in the transformation process. To this end,
it is important that rigorous methods and approaches (such as those
outlined in this chapter) are used if the information obtained from
assessment measures is to be valid and reliable. As we pointed out in
Chapters 1 and 2, assessment practitioners must accept responsibility
for ensuring that the development and use of measures, and the
interpretation and reporting of test information, are non-discriminatory,
unbiased, and fair towards all South Africans.
And so … you have finished the long climb through the Foundation
Zone. You should give yourself a little check-up to see if you have
mastered the basic concepts in Chapters 1 to 7 that will be used in the
remainder of this book. The set of questions below will help you check
that you can move on with confidence to the Assessment Practice
Zone. If, however, you stumble over a few of these questions, do
some backtracking down the hill, just to familiarise yourself with some
of the more difficult concepts.
The last question (7.12) sets the scene for the next zone of our
journey, and our next stop in particular.
* It is important to note that bias implies systematic variation in test scores due to one or a
number of influences, and not random variation (Cohen & Swerdlik, 2010).
Assessment Practice Zone
Now that we are out of the Foundation Zone, things are going to become a
little less theoretical and a little more practical or applied. In the
Assessment Practice Zone you will get a view of good assessment
practices in general, as well as the administration of measures in particular.
8 In this chapter you will learn more about who is legally allowed to use
psychological assessment measures in South Africa, why the use of
assessment measures needs to be controlled, and the ethical
practice standards to which assessment practitioners should aspire.
Please note: In this chapter the term ‘use’ relates to whoever may
administer, score, interpret, and report on assessment measures and
the outcome of assessment, and not to other users of assessment
information such as human resource managers, teachers, and so on.
By the end of this chapter you will be able to:
› give reasons why the use of psychological measures should be controlled
› explain the statutory control of psychological assessment in South Africa
› list the different categories of assessment practitioners in South
Africa, indicate the types of measures they are allowed to use,
and discuss their training and registration requirements
› understand why and how psychological measures are classified
› describe good, ethical assessment practices, and the
roles and responsibilities of the various stakeholders.
8.1 Introduction
In this chapter we introduce you to the nature of the statutory control
of psychological measures in South Africa and the reasons why
measures are classified. We will outline the different types of
psychology professionals that may use measures and the purposes
for and/or contexts in which they may use measures. Finally we will
discuss fair and ethical assessment practices.
As a psychologist, you find yourself testifying in court at a murder trial. You assessed
the person accused of the murder on a battery of measures, which included a
measure of general intelligence as well as a personality measure. While you are
testifying regarding your assessment findings, the judge asks you to show some of
the intelligence test items to the court so that they can get a more concrete idea of
what was assessed. How would you respond to this request? What factors would
you weigh up in your mind to help you to formulate your response?
You will realise from Table 8.2 that there is some overlap between
the goals and values of the role players in assessment contexts in
industry, but perhaps for different underlying reasons (motives). For
example, all role players involved in assessment value accurate, valid,
and reliable assessment. However, assessment practitioners value
valid assessment measures as the use of such measures helps them
to pursue good practice standards; test developers and publishers, on
the other hand, value valid assessment measures because
assessment practitioners are more likely to buy such measures.
Table 8.2 The utility value of psychological assessment for stakeholders/role players in industry
Questions
(a) Who is responsible for this situation, and why?
(b) In retrospect, what could have been done to avoid such a situation?
A policy on testing will need to cover most, if not all, the following issues:
• proper test use
• security of materials and scores
• who can administer tests, score, and interpret tests
• qualification requirements for those who will use the tests
• test-user training
• test-taker preparation
• access to materials and security
• access to test results and test-score confidentiality issues
• feedback of results to test-takers
• responsibility to test-takers before, during, and after test session
• responsibilities and accountability of each individual user.
You are working as a psychologist in a large organisation. You tested all the
trainee managers in your organisation when you were developing and testing
items for a new leadership test a year ago. As a result of this testing, the
content of the test was considerably refined as it was found that many of the
original items had questionable validity and reliability. The human resources
director now approaches you with the following request: ‘Please forward me
the leadership scores for each of the trainee managers whom you tested a
year ago, as we want to appoint the ones with the most leadership potential as
managers now.’ Write a memo to the human resources director in response to
this request. Motivate your response clearly.
8.5 Conclusion
You have now completed your first stop in the Assessment Practice
Zone. In this chapter you have learned why psychological assessment
is controlled by law and which categories of assessment practitioners
may use such measures. You have learned further why and how
psychological assessment measures are classified. You have been
introduced to ethical assessment practices for psychology
professionals. In the process you have probably realised that there are
many different aspects to consider when you are trying to develop an
understanding of what constitutes good and fair assessment practices.
Administering psychological
assessment measures
LOURA GRIESSEL
Chapter CHAPTER OUTCOMES
9.1 Introduction
The basic rationale of assessment involves generalising the sample of
behaviour observed in the testing situation to behaviour manifested in
other non-test situations. A test score should help to predict, for
example, how the test-taker will achieve or how he or she will perform
on the job. However, assessment measurement scores are influenced
by many factors. Any influences that are specific to the test situation
constitute error variance and may reduce test validity. Moreover, when
scores are obtained, there is a tendency to regard the observed score
as being representative of the true ability or trait that is being
measured. In this regard, the concepts of reliability and measurement
error are mostly concerned with random sources of error that may
influence test scores. In the actual application of assessment
measures, many other sources of error should be considered,
including the effects of test-taker characteristics, the
characteristics of the assessment practitioner, and the testing
situation (see also Chapter 17).
The procedure to be followed in administering an assessment
measure depends not only on the kind of assessment measure
required (individual or group, timed or non-timed, cognitive or
affective), but also on the characteristics of the people to be examined
(chronological age, education, cultural background, physical and
mental status). Whatever the type of assessment measure, factors
related to the individuals, such as the extent to which the test-takers
are prepared for the assessment and their level of motivation, anxiety,
fatigue, health, and test wiseness, can affect performance (see also
Chapter 17).
The skill, personality, and behaviour of the assessment practitioner
can also affect test performance. Administration of most assessment
measures requires that the assessment practitioner be formally
accredited by the Health Professions Council of South Africa. Such
requirements help to ensure that assessment practitioners possess
the requisite knowledge and skills to administer psychological
assessment measures.
Situational variables, including the time and place of assessment,
and environmental conditions such as noise level, temperature, and
illumination, may also contribute to the motivation, concentration, and
performance of test-takers.
In order to systematically negotiate these variables and the role
they play in administering assessment measures, we will discuss the
administering of psychological assessment measures in three
stages. Firstly, we will discuss the variables that the assessment
practitioner has to take into account prior to the assessment session;
secondly, the practitioner’s duties during the assessment session;
and finally, the duties of the assessment practitioner after the
assessment session (see Figure 9.1). Our discussion will deal mainly
with the common rationale of test administration rather than with
specific questions of implementation.
As we progress through our discussion, the reader would be well
advised to keep in mind that the key ethical testing practices that
assessment practitioners should follow and adhere to are highlighted
in the International Guidelines for Test-use developed by the
International Test Commission (ITC, 2001). The Guidelines can be
found on the ITC’s Web site (www.intestcom.org).
Against the internationally accepted backdrop of the core ethical
administration procedures, the discussion will also focus on the
administration of assessment measures in a multicultural South Africa.
❱❱ ETHICAL CONSIDERATIONS
The core ethical responsibility during testing and assessment is that nothing
should be done to harm the client. This implies that the test administrator
should be sensitive to test-takers’ cultural backgrounds and values.
In this regard it is important to note that the impact of certain types of
tests (i.e. self-report questionnaires and self-reflection tests, speed versus
power tests) and general familiarity with the laws governing testing could
differ in a multicultural South Africa. Expand on this viewpoint.
Source: Foxcroft, C.D. 2011a. Ethical issues related to psychological testing in Africa:
What I have learned (so far). Online Readings in Psychology and Culture, 2(2).
Retrieved from http://scholarworks.gvsu.edu/orpc/vol2/iss2/7.
https://doi.org/10.9707/2307-0919.1022
ADVANCED READING
Foxcroft, C.D. 2011a. Ethical issues related to psychological testing in Africa: What I
have learned (so far). Online Readings in Psychology and Culture, 2(2). Retrieved
from http://scholarworks.gvsu.edu/orpc/vol2/iss2/7.
https://doi.org/10.9707/2307-0919.1022
Paterson, H., & Uys, K. 2005. Critical issues in psychological test use in the South
African workplace. SA Journal of Industrial Psychology, 31(3), pp. 12–22.
❱❱ ETHICAL CONSIDERATIONS
What is, in your opinion, the main ethical challenge that assessment
practitioners face in Africa?
Source: Foxcroft, C.D. 2011a. Ethical issues related to psychological testing in Africa:
What I have learned (so far). Online Readings in Psychology and Culture, 2(2).
Retrieved from http://scholarworks.gvsu.edu/orpc/vol2/iss2/7.
https://doi.org/10.9707/2307-0919.1022
Lerato Molopo is 19 years old and completed Grade 12. She hails
from the rural areas of KwaZulu-Natal. Lerato wants to study at a
university and is applying for a bursary from the organisation where
you are employed as a psychometrist. She is referred for evaluation.
Critically reflect on the factors that would have to be taken into account
to make the assessment fair and ethical from a language perspective.
9.2.8 Planning how to address test sophistication
The effects of test sophistication (test-wiseness), or sheer test-
taking practice, are also relevant in this connection. In studies with
alternate forms of the same test, there is a tendency for the second
score to be higher. The implication is that if a test-taker possesses
test-sophistication, and especially if the assessment measure
contains susceptible items, the combination of these two factors can
result in an improved score; in contrast, a test-taker low in test-
sophistication will tend to be penalised every time he or she takes a
test that includes test-wise components. Individuals lacking exposure
to specific test materials or test formats may be at a disadvantage.
Brescia and Fortune (1989) have reported that children from a
Native American culture may not exhibit behaviours conducive to
successful test-taking due to a lack of experience. This includes not
reading tests items accurately or not having opportunities to practise
tasks required by the test. Test scores are therefore meaningless if
the test-taker has not had exposure to an optimal educational
environment (Willig, 1988) associated with specific test-taking. A test-
taker who has had prior experience in taking a standardised test thus
enjoys a certain advantage in test performance over one who is taking
his or her first test.
It has been proven that short orientation and practice sessions
can be quite effective in equalising test sophistication. Such
familiarisation training reduces the effects of prior differences in test-
taking experiences. Issues of test sophistication become imperative in
a multicultural and diverse South Africa where individuals may have
had differing experiences with standardised testing procedures.
Orientation sessions should be distinguished from coaching sessions
(Anastasi & Urbani, 1997).
❱❱ ETHICAL CONSIDERATIONS
Source: Foxcroft, C.D. 2011a. Ethical issues related to psychological testing in Africa:
What I have learned (so far). Online Readings in Psychology and Culture, 2(2).
Retrieved from http://scholarworks.gvsu.edu/orpc/vol2/iss2/7.
https://doi.org/10.9707/2307-0919.1022
❱❱ ETHICAL CONSIDERATIONS
Source: Levin, M.M., & Buckett, A. 2011. Discourses regarding ethical challenges in
assessment: Insights through a novel approach. SA Journal of
Industrial Psychology, 37(1), pp. 251–263.
Suppose you must ask a four-year-old child to copy a circle. Explain how
you would obtain the cooperation of, and establish rapport with, the child.
ADVANCED READING
ADVANCED READING
Foster, D.F. 2010. Worldwide testing and test security issues: Ethical
challenges and solutions. Ethics & Behavior, 20(3–4), pp. 207–228.
Page back to Chapter 2 (Section 2.4.4.3), where the question of who may use
psychological tests was raised in relation to a section in a judgment that stated that it
is ‘unlikely that the primarily mechanical function of the recording of test results
should be reserved for psychologists’ (Association of Test Publishers of South Africa
and Another, Saville and Holdsworth South Africa v. The Chairperson of the
Professional Board of Psychology, 2010, p. 20). Now that you have worked through
this chapter and have an idea of what administering a psychological test entails, what
is your view on whether psychological test administration is ‘primarily a mechanical
function’, which by implication can be performed by a non-psychology professional?
Do you agree or disagree with the judge, and why?
ADVANCED READING
If you want to know more about test administration visit the following Web site:
www.intestcom.org
9.5 Conclusion
In this chapter we introduced you to the importance of administering
psychological measures in a standardised way. Furthermore, we
outlined the responsibilities of the practitioner before, during, and after
assessment, and we highlighted ethical practices and dilemmas.
We have travelled a little way through the Assessment Practice
Zone now. You should have a broad understanding of what fair
assessment practices are, as well as some of the specific practices
that need to be adhered to when administering measures. Before we
journey further into assessment practice territory, we first need to gain
some insight into the different types of measures that are available.
So, we will be crossing into the Types of Measures Zone for a while.
Once we have visited the different types of measures, we will return to
the Assessment Practice Zone and consider issues related to
interpreting measures, reporting, factors that affect assessment
outcomes, as well as look ahead to the future of assessment.
Types of Measures Zone
In this zone you will be introduced to a variety of measures that can
be used to assess psychological constructs. You will be making
many stops along the way. Look out for the following signposts:
10.1 Introduction
What do we mean by cognitive functioning, how do we go about
measuring it, and how should we interpret the scores? These are
some of the key issues in the somewhat controversial field of cognitive
assessment discussed in this chapter. Individual differences in general
cognitive ability has been the subject of much scientific interest and
‘individual variation in general cognitive ability (GCA) matters greatly
for many life outcomes, including educational and occupational
success’ (Yeo, Ryman, Pommy, Thoma & Jung, 2016, p. 93).
The measurement of cognitive functioning has been an integral
part of the development of psychology as a science. Although
measurement has a much longer history, the main body of cognitive
assessment and measurement procedures as we know it today
developed since the beginning of the last century. Some argue that
the lack of theoretical consensus on intelligence and, consequently,
also on the measurement of cognitive ability, and in particular the
cross-cultural measurement of cognitive ability, endangers the status
of this field as a science (Helms-Lorenz, Van de Vijver & Poortinga,
2003). Also, the fact that different measures provide different scores
has often been used as evidence in attacks on cognitive measures in
particular. However, the fact that, even in the physical sciences, there
are still differences of opinion after centuries of research (Eysenck,
1988) should guide and inspire the theory and practice of cognitive
assessment. In other words, much can be achieved that is useful,
even if the ideal situation has not yet been reached. This offers the
advantage that it stimulates further research and new developments.
One of the reasons why the measurement of cognitive ability has
remained on the forefront of psychological assessment is that it is a
clear and significant predictor of success in the work and educational
contexts. Research results from meta analyses – which show
stronger evidence of trends in research results because they are
based on the collation of large numbers of similar studies – have
indicated that cognitive assessment results are generally valid
predictors of job performance and training results with predictive
validity correlations in the region of .5 to .6, thus indicating both
statistical and practical significance (Bertua, Anderson & Salgado,
2005). In a 100-year meta-analytic review of research on selection
methods, Schmidt, Oh, and Shaffer (2016) confirmed that general
cognitive ability remains the best predictor of job performance, and
job-related learning and training results.
This chapter provides a brief overview of the theoretical
underpinnings of intelligence. This serves as an important backdrop to
understand the assessment of cognitive functioning and the measures
used to do so, which is the main focus of this chapter.
10.2Theories of intelligence: A
brief history and overview
10.2.1 Background
A clear cyclical or pendulum movement can be traced throughout the
history of cognitive ability assessment. Initially, the focus was mainly
on psychophysical measures with the emphasis on objective methods
of measuring simple sensory processes, with the work of Galton and
Wundt’s psychological laboratory in 1879 as a good example of this
era in testing (Gregory, 2010). Hoffman and Deffenbacher (1992)
showed how applied cognitive psychology uses methods and theories
of experimental psychology to gain understanding of and solve real-
world problems regarding cognitive functioning. This field of applied
psychology evolves from the earliest experimental research in
psychology from the likes of Binet, Cattell, Munsterberg, and Wundt,
amongst others.
The French researchers Alfred Binet and Theodore Simon used
more complex tasks of reasoning and thinking to measure intelligence
when they were commissioned in 1904 to develop techniques for
identifying children in need of special education (Locurto, 1991) (as
was pointed out in Chapter 2). According to Gould (1981), Binet
wanted the test scores to be used practically and not as the basis for a
theory of intellect. He wanted test scores to be used to identify
children who needed special help and not as a device for ranking
children according to ability. Lastly, he hoped to emphasise that
intellectual ability can be improved and modified through special
training. However, this was not the way in which the Binet-Simon
measure was used in practice. Recently there has been a renewed
focus on this more applied approach to cognitive assessment in what
is referred to as dynamic assessment (De Beer, 2013).
Another modern method that reflects the Binet-Simon approach to
assessment is computerised adaptive testing (CAT). As we will
point out in Chapter 14, in adaptive assessment the difficulty level of
items is calculated beforehand and the items are then interactively
selected during assessment to match the estimated ability level of the
test-taker. A South African test that utilises this technology is the
Learning Potential Computerised Adaptive Test (LPCAT) (De Beer,
2005, 2010). The modern version of this approach is based on Item
Response Theory (IRT) and the use of interactive computer
programmes. However, in principle, Binet and Simon applied these
same principles and used the assessment practitioner’s judgement to
decide which entry-level item the child should get first. They made use
of a ceiling level by terminating the measure if a certain number of
items at a specific level were answered incorrectly (Reckase, 1988).
During World War I and II, important contributions were made by
psychologists with regard to assessment for selection and placement
decisions (Hoffman & Deffenbacher, 1992). The influence of
computers and information processing approaches were seen in the
1960s. However, some researchers warned that social, cultural,
motivational, and emotional factors should not be neglected when
attempting to fully understand cognition.
Some exciting new developments in the use of computerised
adaptive testing include what is referred to as cognitive diagnostic
assessment (CDA) (Gierl & Zhou, 2008; Xu, Wang & Shang, 2016).
This involves promising applications in the educational domain, where
diagnostic assessment can be linked to remedial instruction in areas
where lower performance levels are indicated (Gierl & Zhou, 2008).
Multidimensional computerised adaptive testing (MCAT) has been
shown to offer shorter tests, improve test precision, and increase
selection utility compared to traditional testing (Makransky & Glas,
2013). Other new developments are discussed in Section 10.7.
Another recurring theme is the controversy between Spearman’s
and Thurstone’s theories, namely the question of whether intelligence
is influenced by a single general factor or by multiple factors. Both
theories are still actively used in current research today (Carroll, 1997;
Helms-Lorenz et al., 2003; Johnson, Bouchard, Krueger, McGue &
Gottesman, 2004; Major, Johnson & Deary, 2012). The measurement
of emotional intelligence (EQ), which represents the area in which
cognitive and personality assessment are combined, has also been
receiving a lot of attention in the last two decades (Boyle, Humphrey,
Pollack, Hawver & Story, 2011; Matthews, Zeidner & Roberts, 2002).
On the social front, there is still active and ongoing debate on
cognitive ability assessment, the interpretation of test scores, and
particularly the differences between groups. Every now and again, a
new publication appears and, once again, controversy flares up. One
example is Herrnstein and Murray’s book The Bell Curve, published in
1994. Some earlier books that have had a similar effect are Gould’s
The Mismeasure of Man (1981), Fancher’s The Intelligence Men
(1985), and Eysenck and Kamin’s The Intelligence Controversy
(1981).
Here in South Africa, our country is adapting to the integration of
different cultures and guidelines on equity and affirmative action have
been introduced in our legislation. In practice, this legislation also
affects the use of psychological assessment. We may well see a
repeat of the American pattern, that is, a movement away from and
subsequent return to assessment, as well as possible legal cases
about the fairness of measures for various groups. One of the ongoing
debates is the appropriateness of norms for various subgroups –
particularly in multicultural contexts such as South Africa
(Shuttleworth-Edwards, 2017).
Emotional intelligence
The measurement of emotional intelligence refers to the behavioural
and interpersonal adjustment of the individual to a particular
environment or situation.
Goleman (1995) argued that our traditional way of looking at
intelligence is much too narrow and does not allow for the role that our
emotions play in thought, decision-making, and eventually also in our
success. He challenged the narrow view that intelligence is merely
‘what intelligence tests measure’ or IQ, and argued that aspects such
as self-control, zeal, and persistence, as well as the ability to motivate
oneself, are important factors determining success in life. A distinction
can be made between the socio-emotional approach to the study of
emotional intelligence (a broad model that includes abilities as well as
a series of personality traits) and the view that links emotions with
reasoning (a view that focuses on abilities only). The first approach
includes the initial work by Salovey and Mayer (1990), and the work
by Goleman (1995) and Bar-On (1997), while the latter refers to the
adaptations by Mayer and Salovey (1997). Studies involving various
measures of emotional intelligence have indicated that the construct is
related to cognitive variables such as thinking styles as well as
personality variables such as leadership effectiveness and stress
management (e.g. Herbst & Maree, 2008; Murphy & Janeke, 2009;
Ramesar, Koortzen & Oosthuizen, 2009). Examples of measures of
emotional intelligence include objective measures, such as the
Multifactor Emotional Intelligence Scales (MEIS) and the Mayer,
Salovey and Caruso Emotional Intelligence Test (MSCEIT), and self-
report measures such as the Bar-On Emotional Quotient Inventory
(EQ-I).
Emotional intelligence (EI) has been linked to ‘emotional labour’
(O’Boyle, Humphrey, Pollack, Hawver & Story, 2011), which is
required by jobs with a heavy interpersonal component such as
service work and leadership positions. EI measures have been
reported as predicting job performance equally as well as or better
than a combination of cognitive ability and personality and add unique
additional predictive power over and above that provided by cognitive
and personality (O’Boyle et al., 2011).
Figure 10.1 Example of cognitive ability items (left) inspired by African art (right)
SOURCE: Bekwa (2016). Reproduced by permission. Photograph supplied by Kaetana,
Shutterstock.
Measures of general
10.4
cognitive functioning
10.4.1 Individual intelligence measures
Every year, large numbers of students who have just finished school apply for
admission to highereducation institutions such as universities and universities of
technology. Government fund these institutions on the basis of the number of
students who pass, so they prefer to admit students who will be successful in
their studies. Admission testing programmes are often used to select students
who are most likely to benefit from what is offered educationally and who are
most likely to meet the criteria of success as defined by each institution.
We call the sound that a MOUSE makes a SQUEAK and the sound that a
FROG makes a CROAK. MOUSE stands to SQUEAK as FROG stands to
CROAK. CROAK should therefore go in the place of the question mark.
The letter above CROAK is C, therefore the answer to Question 1 is C.
SOURCE: Copyright © by the Human Sciences Research Council. Reproduced by permission.
Consider the comment by Claassen, De Beer, Hugo et al. (1991, p. 25) above – that
a group intelligence test measures ‘a student’s present level of reasoning ability in
respect of scholastic material’ – to which a further comment is added that a group
intelligence test score can only be interpreted as an intelligence score for test-takers
who do not come from environmentally disadvantaged contexts. Would
these comments also apply to individual intelligence test scores?
Reread Section 10.4.1.2 to help you.
Questions:
(a) Was the psychologist acting in an ethically responsible way in
administering the SSAIS-R to Thabo and in reaching the
conclusions that he did? Motivate your answer.
(b) Advise the psychologist how he or she can avoid transgressing
ethically in a similar situation in the future.
Numerical Reasoning ✔ ✔
Verbal Reasoning ✔ ✔
Inductive Reasoning ✔
Deductive Reasoning ✔ ✔
Mechanical ✔
Comprehension
Spatial Ability ✔ ✔
Checking ✔
Calculation
Reading
Comprehension
APTITUDE TESTS
10.8 Conclusion
To understand why the theories and measurement of cognitive
functioning have sparked such intense debate and emotional
reactions, we have to consider the psychosocial issues involved. In
most Western societies, either directly or indirectly, various
‘advantages’ are associated with a specific level of cognitive
functioning – ‘intelligence’ as it is defined in psychometric terms.
These include aspects such as a good education, a corresponding
higher level of employment, higher standing in the society, a higher
level of income, and so the list goes on. At a country level, IQ has
been shown to correlate positively with the Gross Domestic Product
(GDP) of countries, attained education of countries as well as other
social and economic factors (Roivainen, 2012). Furthermore, socio-
economic status (SES) is positively associated with both cognitive
performance and academic achievement of young children and
adolescents (Von Stumm, 2012). However, after studying this chapter,
you should realise that measures of cognitive functioning differ in
terms of the underlying assumptions. The interpretation of results is
limited to what each test measures and should always be used in the
context for which a specific measure was intended. Technical
features, such as the reliability and validity of the measure, should be
considered, as well as the population or norm group for whom the
measure was standardised. In addition, research on differences within
and between groups should make us aware of the need to invest in
people through social or educational programmes.
11.1 Introduction
In this chapter you will be introduced to the concept of ‘well-being’ in
general as well as the theoretical underpinnings of well-being. You will
also be introduced to measures that can be used to assess well-being,
and its subdimensions.
11.2 Well-being
The overriding impression one gets from reading the available
literature on well-being is of the vastness of the concept. Different
authors provide different definitions, and also emphasise different
aspects of well-being. Roodt (1991) attributes this to the numerous
disciplines that investigate well-being. These disciplines include
psychology, sociology, medicine, and economics. He also points out
that the use of numerous terms (including well-being, psychological
well-being, life satisfaction, happiness, and quality of life) further
complicate understanding of well-being. There does seem to be
consensus, however, that well-being is a multidimensional construct
with numerous domains. Moreover, these different domains exercise
mutual influence on each other. In terms of schools of thought within
psychology, well-being is most at home within positive psychology.
In the following sections attention will be given to providing a fairly
inclusive definition of well-being, explaining its link to positive
psychology, and discussing some of the domains of well-being.
Think of a person that you know that, in your opinion, exemplifies well-
being in all its aspects. Based on your knowledge of and experiences with
this person, develop a model of well-being that encompasses both
hedonistic and eudaimonic understandings and perceptions thereof. Your
model should also indicate the interrelatedness of concepts of well-being.
Well-being among
11.4
university students
A particular population in which well-being is very important is
university students. Studying at a tertiary institution is usually a life-
changing, positive experience. Nevertheless, it also brings challenges
of its own. Some of the challenges faced by tertiary students are the
pressures to succeed and achieve in academics, the dynamics of
adult romantic and social relationships, and financial pressures. In
addition, other life challenges are also still present, including family
dysfunction, mental illnesses such as depression, and chronic
physical illnesses. Among South African students there are also
unique factors that impact well-being, amongst others our socio-
political context (cf. Taljaard-Plaut & Strauss, 1998). The latter relates
to elements like violence, poverty, and inequality.
Psychological wellbeing occurs in, and is influenced by, multiple
life domains. These include intrapersonal, interpersonal, social, and
spiritual domains. To investigate well-being of tertiary students, it is
therefore important to view it systemically. All domains of a student’s
life can influence his or her well-being and should therefore be taken
into consideration when addressing student wellness.
Research shows support for the application of a holistic wellness
model when considering student well-being (e.g. van Lingen, Douman,
Wannenburg, 2011). In addition, recent studies applied a positive
psychology approach in terms of student well-being. In this regard,
Van Zyl and Rothmann (2012) investigated flourishing among
university students, while Morris-Paxton, Van Lingen and Elkonin
(2017a, 2017b) explored a salutogenic lifestyle programme among
disadvantaged university students. Salutogenesis refers to the belief
that it is possible to reach a state of well-being (and not illness)
(Morris-Paxton et al., 2017a). Important features of a salutogenic
approach are (1) an emphasis on assessing health rather than
diagnosing illlness; (2) considering stress and stressors as normal; (3)
emphasising that which will enhance hardiness and adaptability; and
(4) paying particular attention to factors that will enhance well-being in
its broadest sense (Morris-Paxton et al., 2017a). A salutogenic
lifestyle approach is therefore an approach that will work from a
positive perspective to enhance health and well-being in all spheres of
living.
Van Lingen et al. (2011) identified the following domains that
contribute to a holistic view of student well-being: physical, career,
intellectual, environmental, social, emotional, and spiritual wellness.
The table below is adapted from Van Lingen et al. (2011, p. 401) and
offers descriptors for each of the domains.
Table 11.1 Domains of a holistic view of student well-being
11.6 Conclusion
The concept of ‘well-being’ is central to the discipline and profession
of psychology. Well-being and its subdimensions were considered
from a theoretical perspective as well as in terms of the range of
available measures used to assess them in this chapter.
Questions
1. To what extent should the information you have about Hannah’s psychiatric
condition affect your final report and the ranking of the candidates? Indicate all
the aspects that would influence you in this regard, including ethical aspects.
2. Should you disclose your knowledge about Hannah’s psychiatric
condition (a) in the written report to the executive and/or (b)
verbally to the executive concerned? Frame your answer in
terms of good, ethical assessment practice guidelines.
ADDITIONAL READING
Sheldon, K.M., Kashdan, T.B., & Steger, M.F. 2011. Designing positive psychology:
Taking stock and moving forward. Oxford: Oxford University Press.
Cooper, C.L. 2011. Organisational health and wellbeing. Los Angeles: Sage.
Bakker, A.B. & Leiter, M.P. 2010. Work engagement: A handbook of essential theory
and research. Hove: Psychology Press.
Seligman, M.E.P. 2003. Authentic happiness: Using the new positive psychology to
realise your potential for lasting fulfilment. London: Nicholas Brealey.
Csikszentmihalyi, M. & Csikszentmihalyi, I.S. 2006. A life worth living: Contributions to
positive psychology. New York: Oxford University Press.
Feldman, F. 2010. What is this thing called happiness? New York:
Oxford University Press.
Diener, Ed. 2009. Assessing well-being: The collected works of Ed Diener. New York:
Springer.
Leiter, M.P. & Maslach, C. 2005. Banishing burnout: Six strategies for improving your
relationship with work. San Francisco, California: Jossey-Bass.
Did you find this an interesting stop in the Types of Measures Zone?
The next stop should prove to be just as stimulating, as you will be
given an opportunity to explore how personality functioning can be
assessed. Assessing what makes people tick (personality make-up) is
fundamental to the discipline and profession of psychology, so it is
something that most psychology students are keen to learn more
about.
Personality assessment
NICOLA TAYLOR AND GIDEON P. DE BRUIN
Chapter CHAPTER OUTCOMES
12 By the end of this chapter you should be able to:
❱ have a conceptual understanding of personality and personality assessment
❱ distinguish between structured and projective personality-assessment
methods
❱ critically describe the main assumptions underlying the use of the structured
assessment methods and discuss the challenges associated with the cross-
cultural use of such methods
❱ critically describe the main assumptions underlying the use of projective
tests such as the Thematic Apperception Test (TAT) and discuss the
problems associated with the cross-cultural use of the TAT.
12.1 Introduction
Psychology professionals in industrial, psychodiagnostic, forensic,
educational, and other contexts are often requested to assess an
individual’s personality (see Chapter 15). Usually the purpose of such
an assessment is to understand the uniqueness of the individual.
Psychologists also use personality assessment to identify the
individual’s characteristic strengths and weaknesses, and his or her
typical way of interacting with the world and the self. There are many
different methods and procedures that psychologists may use for
personality assessment. The choice of method usually depends on (a)
the reason for the assessment, (b) the psychologist’s theoretical
orientation, and (c) the psychologist’s preference for particular
methods and procedures. In this chapter you will be introduced to how
to conceptualise personality and its subdimensions, then different
methods used to assess personality, and specific assessment
measures of personality functioning that psychologists typically use in
practice.
In our example we used only eight variables and even a casual glance
at the words suggested they could be grouped according to two
dimensions. However, when the aim is to identify a comprehensive set
of basic personality dimensions, a much larger number of variables
that are representative of the total personality domain are needed. In
such a study it will not be possible to identify basic personality
dimensions by visual inspection of the variables alone. Factor analysis
helps researchers to identify statistically the basic dimensions that are
necessary to summarise the original variables. Factor analysis also
helps us understand the nature of the dimensions that underlie the
original variables. This is achieved by inspecting the nature or
meaning of the variables that are grouped together in the dimension or
factor. Although the aims of factor analysis can be relatively easily
understood on a conceptual level, the mathematics underlying the
procedure is quite complex. Fortunately, computers can perform all
the calculations, although researchers should have a thorough
knowledge of the mathematical model in order to use factor analysis
properly. Interested readers are referred to Gorsuch (1983) and Kline
(1994).
Primary Factors
Second-order Factors
Note: The factor labels and descriptors were taken from the
16PF5 (Conn & Rieke, 1998).
Personality and
12.5
counterproductive behaviour at work
When selecting people for positions in the workplace, most of the
focus is on the positive characteristics that a person possesses that
would enable him or her to be successful in that position. However,
there is also a need to ensure that the person being hired will not pose
a risk to the organisation or the people that they work with due to
negative behaviours. These negative behaviours could be described
as counterproductive work behaviours (CWB) and could either be
deliberate or unconscious responses to stress in the workplace. Types
of CWB could range from seemingly innocuous behaviours like being
manipulative or blaming others when things go wrong to interpersonal
abuse or absenteeism, and finally to serious crimes such as theft,
fraud, and destruction of company property. Two assessments look at
these negative behaviours from a personality perspective, namely the
Work-related Risk and Integrity Scales (WRISc), and the Hogan
Development Survey (HDS).
12.5.1 Work-related Risk and Integrity Scales (WRISc)
Van Zyl and De Bruin (2016) developed the Work-related Risk and
Integrity scales (WRISc) in response to a need for a locally
constructed assessment of potential counterproductive work
behaviour. Tests used for the assessment of CWB are generally
referred to as integrity tests (Berry, Sackett & Wiemann, 2007) and
could either be overt (asking directly whether a person has or Is likely
to engage in CWB) or covert (based on personality where the purpose
of the assessment is not obvious). The WRISc is a covert measure,
and identifies 12 narrow personality constructs that are empirically
linked to engaging in CWB.
The 12 scales of the WRISc are: Aggression, Callous affect,
Cynicism, Egotism, External locus of control, Impulsivity, Low effortful
control, Manipulation, Negative affect, Pessimism, Risk-taking and
Rule defiance. Several of these scales can also be combined to give
an indication of complexes or toxic personality profiles that can have a
profoundly negative impact on the organisation.
The WRISc is a South African-developed test, so has local norms
and local research. The test has good internal consistency reliability,
with Cronbach alpha values above .74 for all scales. The structure of
the assessment appears sound, and there was no practical evidence
of bias in any of the scales. Correlations with measures of
counterproductive behaviours show good concurrent validity (Van Zyl
& De Bruin, 2016).
12.7 Conclusion
In order to understand the uniqueness of the individual, it is necessary
to examine personality on all three levels of McAdams’ scheme. The
first level of the scheme provides insight into the individual’s typical
ways of interacting with the world and characteristic patterns of
behaviour. Structured personality assessment tools such as the 16PF,
BTI, OPQ, MMPI and SAPI may help psychologists to understand and
describe these typical patterns of interaction and behaviour. The
information gained by using such tools may also be useful in
predicting how an individual is likely to behave in the future. However,
these assessment measures do not provide insight into the reasons or
motives for a person’s behaviour. In order to achieve this,
psychologists need to move to the second level of McAdams’ (1995)
scheme where the focus falls on personal concerns, projects, and
motives. Projective techniques, such as the TAT, may be useful in
gaining insight into implicit and unconscious aspects of personality.
This may provide insight into the reasons for a person’s behaviour. By
combining information gained on both these levels, and by integrating
it with an individual’s life story, the psychologist can construct an
integrated and rich picture of the person who is being assessed.
Questions:
1. What are the possible ethical dilemmas with regards to the use
of personality assessments in this case study?
2. What are the possible ethical dilemmas with regards to the
assessment process?
3. Are there any professional or legal issues that would relate to
personality assessment (e.g. scope of practice)?
4. What assessments and process would you recommend if you
were in Thembi’s situation?
Cheung, F.M. & Fetvadjiev, V.H. 2016. Indigenous approaches to testing and
assessment. In F.T. Leong, D. Bartram, F.M. Cheung, K.F. Geisinger, & D Iliescu.
(Eds.), The ITC international handbook of testing and assessment (pp. 333−346).
New York, NY: Oxford University Press.
Cheung, F.M., van de Vijver, F.J.R. & Leung, F.T.L. 2011. Toward a new approach
to the study of personality in culture. American Psychologist, 66, pp. 593−603.
Fetvadjiev, V.H., Meiring, D., Van de Vijver, F.J., Nel, J.A. & Hill, C.
2015. The South African personality inventory (SAPI): A culture-
informed instrument for the country’s main ethnocultural
groups’, Psychological Assessment, 27(3), pp. 827−837
Iliescu, D. & Ispas, D. 2016. Personality assessment. In F.T. Leong, D. Bartram, F.M.
Cheung, K.F. Geisinger, & D Iliescu., (Eds.), The ITC international handbook of
testing and assessment (pp. 134−146). New York, NY: Oxford University Press.
Valchev, V.H., Nel, J.A., Van de Vijver, F.J.R., Meiring, D., De Bruin, G.P., &
Rothman, S. 2013. Similarities and differences in implicit personality
concepts across enthnocultural groups in South Africa. Journal of Cross-
Cultural Psychology, 44, pp. 365−388. DOI: 10.1177/0022022112443856.
Valchev, V.H., Van de Vijver, F.J.R., Nel, J.A., Rothman, S. & Meiring, D. 2013.
The use of traits and contextual information in free personality descriptions
across ethnocultural groups In South Africa. Journal of Personality and
Social Psychology, 104, pp. 1077− 1091. DOI: 10.1037/a0032276.
There have been quite a few stops to make in the Types of Measure
Zone. You only have two left. At the next stop in this zone, you will
take a look at some of the measures used in career assessment.
Career-counselling assessment
GIDEON P. DE BRUIN AND KARINA DE BRUIN
Chapter CHAPTER OUTCOMES
13 Once you have studied this chapter you will be able to:
› identify the central theoretical concepts and assumptions of four
approaches to career-counselling assessment: the person-environment fit,
developmental, systems, and career construction/life-design
› identify the major differences between these four approaches
› identify points of criticism against the approaches
› name the most important measures and techniques associated with each of
the four approaches.
13.1 Introduction
Career counselling can be defined broadly as the process in which a
professional career counsellor guides clients in identifying their current
and potential skill strengths and career options that provide them with
the best opportunities to use these strengths (Ryan Krane & Tirre,
2005). Career counselling is a comprehensive process and includes
any counselling activity related to the career decisions a client has to
make (Zunker, 2016). More and more, individuals do not make a
career choice for life but, rather, make numerous career decisions.
The current work environment, which is marked by volatility,
uncertainty, complexity, and ambiguity, forces individuals to be flexible
to adapt to these challenges often resulting in career decisions that in
turn result in regular job and career changes.
Zunker (2016) describes several career-counselling models, each
consisting of a number of stages. Although these models use a wide
range of techniques, the stages appear to be very similar. Career
counselling usually includes an intake interview, assessment,
diagnosis, and problem identification. These steps are followed by a
counselling process that focuses on a collaborative relationship with
the client, intervention strategies, and the evaluation of outcomes and
future plans.
Although standardised assessment does not dominate career
counselling in all these models, it plays an important role in gathering
information about the client. In this chapter we aim to provide an
overview of career-counselling assessment from four different
theoretical points of view. The first two, namely the person-
environment fit and developmental approaches, represent traditional
assessment approaches. The last two, the systems and career
construction/life-design approaches, are strongly influenced by
constructivism and represent a qualitative assessment approach.
There are at least three general reasons why assessment in career
counselling is important: ‘(1) to stimulate, broaden, and provide focus
to career exploration; (2) to stimulate exploration of self in relation to
career; and (3) to provide what-if information with respect to various
career choice options’ (Prediger, 1974, p. 338). Savickas, Nota,
Rossier, Dauwalder, Duarte, Guichard, Soresi, Van Esbroeck, Van
Vianen and Annelies (2009) and Maree (2010a) add that career
counselling/assessment is useful in assisting clients to design their
lives in the new postmodern world of work. Assessment in career
counselling is likely to be important irrespective of which theoretical
approach the career counsellor follows. However, as you will see in
the sections that follow, each of the theoretical approaches
emphasises different aspects of the assessment process. Individual
career counsellors often have a preference for certain assessment
tools and techniques. Some find the use of standardised
psychological tests useful, whereas others may feel more comfortable
with qualitative methods. For others, a combination of standardised
psychological tests and qualitative assessment techniques may be
more appealing. To best provide for the needs of the client, a
comprehensive and collaborative approach to career assessment –
including both standardised psychological tests and qualitative
assessment procedures – is often the most desirable approach
(Maree & Morgan, 2012).
James Mhlangu is 18 years old and in Grade 12. He comes from the rural areas
of KwaZulu-Natal. He was awarded a scholarship to study engineering at any
South African higher-education institution. Although James is very excited about
this opportunity, his parents are not convinced that this is the right career path for
James. The three of them approach you for help with this important career
decision. What career counselling and assessment approach would you follow,
and which measures would you use to assist James to make this difficult
decision? Why would you use the specific assessment tools that you choose?
Career counselling in a
13.6
changing environment
We described two traditional approaches as well as two emerging
approaches to career counselling assessment in this chapter. Each
one of these approaches has a different emphasis. The person-
environment fit approach emphasises the fit between a person and his
or her environment. The developmental approach adds to this by
emphasising the readiness of an individual to make a career choice.
The systems approach provides for the influence of various
subsystems on the individual’s career development. Career
construction and life-design counselling explore how people fit their
career into their life stories and how they adapt to changes in their
careers.
All four approaches can be valuable in the career-counselling
context, depending on the particular needs of a client. An emphasis on
one of these at the expense of another may be seen as limiting in the
career-assessment process. The main purpose of career assessment
is to enhance self-awareness. Traditional tools such as personality
and interest inventories may serve this purpose very well. In keeping
with a constructivist and narrative approach, the information gained
from the inventories is not taken as absolute and fixed, but as
information that must be given meaning within the broader system of
influences and life themes of a client. Career counsellors are
encouraged to further explore these and other approaches; in that
way, they will be able to tailor their assessment and interventions
according to what is best for the client, and not according to that with
which they are most comfortable and familiar.
Career counsellors are also encouraged to recognise the
pervasive social and technological changes that are influencing the
way in which work is structured and perceived. Individuals no longer
enter organisations and devote their entire career to that organisation.
Often, they do not only change from one organisation to another, but
to whole new fields of work during their career. More and more so,
workers have to be multiskilled and flexible. Career-counselling
assessment can be valuable in this regard by helping individuals
recognise their work-related strengths, skills, and abilities that can be
applied in many different contexts. Values and interests also remain
important, because workers continuously seek to find working
environments that satisfy their psychological needs.
13.7 Conclusion
In this chapter you have been introduced to different approaches to
career counselling, the changing role of assessment in career
counselling, and a range of career-counselling assessment measures.
Questions
(a) At what point should Dr Jansen decide whether he will take on
this client? If he does continue with the career-counselling
assessment, what ethical risks is he facing?
(b) If Dr Jansen decides to refer the client to his colleague, Ms Mosia, who is
an experienced career counsellor, what challenges might she face in
terms of her available assessment measures? What alternative process
would you suggest to provide appropriate career counselling to the client?
ADVANCED READING
Gysbers, N.C., Heppner, M,J. & Johnston, J.A. 2014. Career counseling: Holism,
diversity and strengths 4th ed. Alexandria, VA: American
Counseling Association.
Heppner, M.J. & Jung, A. 2012. When the music changes, so does the
dance: With shifting U.S. demographics, how do career centers
need to change? Asian Journal of Counselling, 19(1&2), pp. 27–51.
Osborn, D.S. & Zunker, V.G. 2016. Using assessment results for
career development 9th ed. Boston, MA: Cengage Learning.
McMahon, M. & Watson, M. Eds. 2015. Career assessment: Qualitative approaches.
Rotterdam: Sense.
Prince, J.P. & Heiser, L.J. 2000. Essentials of career interest assessment. New York:
John Wiley.
Seligman, L. 1994. Developmental career counseling and
assessment. Thousand Oaks, CA: Sage.
Stead, G.B. & Watson, M.B. Eds. 2017. Career psychology in the
South African context 3rd ed. Pretoria: Van Schaik.
Computer-based and Internet-
delivered assessment
NANETTE TREDOUX
Chapter CHAPTER OUTCOMES
14 In this chapter we will outline the current state of development in computer-
based assessment, considering the advantages and potential disadvantages of
computer-based and Internet testing as well as the ethical and practical issues
that the practitioner should bear in mind when using such assessment
technology, with the consideration of guidelines for good practice and
decision-making.
By the end of this chapter you will be able to:
› understand the features of different types of technology-supported
assessment and critically evaluate new developments in computer-based
and Internet-delivered assessment systems
› critically assess the effect of different computerised or Internet-delivered test
administration systems on the standardisation and fairness of the
assessment process, and use computer-generated assessment reports in a
professionally responsible and discerning manner
› understand your legal, ethical, and professional responsibilities when using
computer-based and Internet-delivered assessment systems, and be aware
of best-practice guidelines regarding the use of computer-based and
Internet-delivered assessments.
14.1 Introduction
The use of computer technology has had a profound effect on
assessment practices. Initially, these changes were seen as
revolutionary but they have now become so commonplace that
traditional pencil-and-paper assessments are now the exception,
rather than the default, for testing in occupational and many
educational, counselling, and clinical settings.
While computer technology contributes to efficiency in test
construction and assessment practice and feedback (Gregory, 2010),
and even makes innovative new means of assessments possible, it
also presents technical and ethical challenges to the practitioner.
Technological advances are proceeding so rapidly that they threaten
to outpace attempts to regulate technology-supported assessment
practices. Practitioners need to make an effort to keep up to date with
technical developments and also exercise critical judgment regarding
the degree to which they choose to make use of new technologies.
Even so, they will often need to make decisions about new technology
and measures that they do not fully understand.
14.4Computer-based testing in
the twenty-first century
14.4.1 A wider variety of tests
Since the dawn of the twenty-first century, there has been a huge
increase in the number of test providers offering Internet-delivered
assessment technology in South Africa. Test users have many new
tests, both foreign-developed and local, at their disposal. In some
cases, the track records of the tests and the test providers are not
known to the person contemplating the use of the measures. For
instance, a practitioner may find a new test using a Google search,
which may seem ideal for the intended use, but the test is not
represented by any distributor in South Africa and the Web site is
hosted in another country. Such a test would not have been tested on
or adapted for South Africans, and may even be available to persons
who are not qualified to use psychological tests in South Africa.
14.4.8 Summary
The field of computerised assessment is developing at an ever-
increasing pace. Whereas early computerised assessments offered
improved standardisation compared to individually administered tests
(some even claimed to guarantee total standardisation), the diversity
of devices that are now used, the use of item-banked and adaptive
tests, and the increasing prevalence of unsupervised testing in
uncontrolled settings, mean that test-takers are not all experiencing
the assessment process in the same way.
Computerised assessments are appearing in new forms, and are
also targeted at populations that were not formerly included, such as
children and people with mental and emotional problems. The
emergence of artificial intelligence, simulations and gamified
assessments, and sophisticated computer-generated reports further
complicate the responsibilities and decisions of the assessment
practitioner. Because of the huge scale of many assessment projects,
many people are affected, which means that the assessment
practitioner carries a very heavy responsibility.
Questions
1. What are Sindi’s ethical and professional responsibilities
regarding the implementation of this system?
2. What can Sindi do to ensure that the system is used legally and
professionally in the company?
3. Specifically consider the fairness implications of introducing
such a system in South Africa.
The Employment Equity Act requires that tests used in industry should
be reliable, valid, not biased against a particular group or individual,
and that they must be able to be used fairly (Department of Labour,
2014). The requirement that such tests should also be certified was
rescinded (ATP v. the President of SA, the Minister of Labour and the
HPCSA, 2017). This means that if assessment practitioners should
decide to use a new or specialised computer-based test that has not
been classified by the Psychometrics Committee, they need to verify
that the psychometric properties of the test comply with the
Employment Equity Act. The code of conduct for the profession
requires the use of classified tests.
In terms of the Scope of Practice for the Profession of Psychology
(Department of Health, 2008), testing and assessment of a long list of
psychological constructs are reserved for the profession of
psychology, and it is a criminal offence for unregistered people to
perform this activity. It is also no longer allowed for psychology
professionals to delegate psychological acts to unregistered people.
The question of whether the administration of a structured test, or, by
extension, a computerised test, constitutes a psychological act is still
not entirely clear. In his 2010 judgment, Judge Bertelsmann refrained
from declaring that it is not illegal for an unregistered person to
administer psychological tests. However, at the time there was no
gazetted list of classified tests. The judge expressed the opinion that
the mere administration of tests did not constitute the use of just tests
– only when tests are interpreted for the purpose of assessing
psychological constructs does it become a psychological act − but
nevertheless he did not issue the declaratory order to confirm this
(Bertelsmann, 2010). At the time of writing this chapter, new
regulations regarding the use and control of psychological tests are in
development, and if these are passed the situation may change.
Psychometrists have a vested interest in protecting the administration
of tests as something reserved for registered professionals.
With regards to computer-based assessment, practitioners should
ask themselves whether administration of the tests, or the testing
project, requires professional skill and judgment. If it is totally
straightforward, it may be acceptable to delegate this task. But if there
are complex considerations that affect the fairness of the assessment
process, it should not be delegated. Even if a routine action is
delegated, professional responsibility can never be delegated to an
unregistered person. In this regard it should be noted that internet-
delivered assessment projects may be becoming more complex where
assessment is done in different geographical locations, on people with
different degrees of technological familiarity, and on different devices.
The code of conduct for the profession of psychology (Department
of Health, 2006) specifies that assessment must always take place
within a defined professional relationship. Informed consent must
always be obtained when assessment is done using automated or
Internet-based testing, even if it is a routine assessment such as a job
application where consent could be implied. Informed consent also
involves the clarification of why the assessment is being done, the
possible consequences and risks of doing the assessment, and the
consequences of refusing to do it or withdrawing from the process
once it has begun. The practitioner is also obliged to discuss the
limitations of the data obtained from such assessments with clients
(respondents themselves and also possible third parties such as
employers). Practitioners should take personal responsibility for the
reports and feedback done on such assessments. This may mean that
computer-generated reports need to be further edited to add
background information, or to triangulate information to ensure that
important decisions are not based on single assessments.
Establishing the professional relationship when Internet-based
testing is used could be a difficult process, especially when one
practitioner is responsible for the assessment of a large number of
people in different geographical locations. It may be possible to
establish the relationship via a telephone call or an electronic text
conversation such as Skype, or via an online meeting platform where
the assessment professional can address a group of people at a time.
Even with the use of such technologies, this can be a challenge if
there are hundreds of respondents and only one practitioner. There is
at present no easy solution, and this aspect is often overlooked with
assessment systems that, with one click, send out hundreds of emails
containing links to tests to test-takers who the assessment practitioner
has never even met.
All ethical considerations regarding psychological services apply to
computer-based assessment. The practitioner should treat clients as
individuals with dignity and personal worth. All clients should be
treated equally, even if some find the completion of the tests a
frustrating technical challenge.
Although computer-based assessments represent a significant
labour-saving technology, it is difficult to do them properly and
professionally. Assessment practitioners should guard against
becoming complicit in a dehumanising and excessively automated
assessment process where the rights of the test-takers may be
compromised. This may require them taking a stand in their places of
employment, and towards the suppliers of assessment technologies,
against the de-personalisation and de-professionalisation of
psychological assessments.
14.6 Conclusion
In this chapter a brief overview of the rapidly developing field of
computerised assessment was provided. The history of computer-
based assessment in South Africa was presented. It was noted that
computerised assessments are appearing in new forms (e.g.
computerised simulation assessments), a variety of devices are being
used, unsupervised testing in uncontrolled settings is commonplace,
and computer-generated reports have become more sophisticated.
These further add to the responsibilities of the assessment practitioner
and the decisions required. The chapter ended with a discussion
about the ethical and legal use of computer-based and Internet-
delivered testing in South Africa.
Consider the following statement from the Ethical Rules of Conduct (Department of
Health, 2006, paragraph 44(7)):
‘When any electronic, Internet or other indirect means of
assessment is used, the psychologist concerned shall declare this
and appropriately limit the nature and extent of his or her findings’.
What is your reaction to this statement? In what way does computer-
based or Internet-delivered assessment limit the nature of the assessment
findings, and is this any different to the limitations experienced when
performing traditional paper-based or performance-based testing?
This was the last stop in the Types of Measure Zone. Here are a few
Web sites where you can obtain additional information on, and
evaluations/reviews of, assessment measures: https://buros.org
(Buros Center for Testing); http://ericae.net (ERIC Clearinghouse on
Assessment and Evaluation); and http://cresst.org (CRESST, the
National Center for Research on Evaluation, Standards, and Student
Testing).
As you look back over this zone, there are a number of critical
concepts that you should have grasped in Chapters 10 to 14. By
answering the following questions you will be able to assess whether
you have mastered the critical concepts in this zone.
The final stretch of the journey lies ahead. You will be travelling in the
Assessment Practice Zone again, where you will learn to apply what
you have learned to the everyday practice of assessment.
Assessment Practice Zone (2)
We are now in Part 2 of the Assessment Practice Zone and you are
entering the final section of your journey! There is still quite a lot to
be seen and digested at the remaining stops.
15 You will explore how the different types of measures to which you
Part have been introduced are applied in a variety of contexts.
16 You will explore issues related to interpreting measures and
4 reporting on assessment results.
17 You will investigate the factors that influence the outcome of an assessment.
18 And finally, you will be able to take stock of progress being made regarding
making assessment more meaningful and African-centred, and take
a peek at some of the next steps. Enjoy the rest of the journey!
The use of assessment measures
in various applied contexts
CHERYL FOXCROFT, LOUIS STROUD, FRANCOIS DE KOCK, AND GERT ROODT
Chapter CHAPTER OUTCOMES
15 In this chapter, you will be required to draw on your knowledge of fair and
ethical assessment practices (covered in Chapters 1, 7, 8, 9, and 14) and the
cognitive, affective, personality, interest, and career measures that were cove red
in Chapters 10 to 13 of the Types of Measure Zone. In this chapter, we will discuss
how these and other measures are applied in certain contexts.
At the end of this chapter you will be able to:
› explain how psychological assessment is applied in industry
› explain what developmental assessment is, what functions are assessed, and
the different types of measures used
› describe some potential assessment innovations for children emerging from
therapeutic contexts
› explain how assessment measures are used in educational contexts
› understand how psychological assessment is used for psychodiagnostic
purposes in a variety of contexts, including psycholegal areas.
15.1 Introduction
In this chapter you will be introduced to how psychological
assessment is applied in industry, for the purposes of developmental
assessment, in educational contexts, and for psychodiagnostic
purposes. In the process some of the measures discussed in previous
chapters will be referred to and you will also be introduced to other
assessment measures.
You are the in-house psychologist for Company A. You are requested to
assemble a test battery for selecting candidates from diverse groups for
a management-development course. You have to make use of aptitude,
ability, interests, and personality measures. Based on what you have
learned so far, which criteria will you apply for selecting measures from
the different lists above? Motivate your choice of criteria.
ADVANCED READING
Aguinis, H. (2013). Performance management. 3rd edition. Englewood Cliffs, NJ:
Prentice-Hall
Schlebusch, S. & Roodt, G. (2008). Assessment centres: Unlocking potential for
growth. Randburg: KnowRes Publishers.
Scott, J.C. & Reynolds, D.H. (2010). Handbook of workplace assessment. Wiley:
Hoboken.
ADVANCED READING
These readings provide you with African-centred assessment
tasks and more information about the use of creative expressive
arts in therapy, and for assessment and observation purposes.
Fritz, E. 2016. Flowing between assessment and therapy through
creative expressive arts. In R. Ferreira (Ed.), Psychological
assessment: Thinking innovatively in contexts of diversity (pp.
103−209). Cape Town: Juta and Company.
Knoetze, J. 2013. Sandworlds, storymaking, and letter writing: The therapeutic
sandstory method. South African Journal of Psychology, 43, pp. 459−469.
Mpofu, E. 2002. Indigenization of the psychology of human intelligence
in sub-Saharan Africa. In W.J. Lonner, D.L. Dinnel, S.A. Hayes &
D.N. Sattler (Eds.), Online Readings in Psychology and Culture (Unit
5, Chapter 2). Center for Cross-Cultural Research, Western
Washington University, Bellingham, Washington USA.
Richards, S.D., Pillay, J. & Fritz, E. 2012. The use of sand-tray techniques
by school counsellors to assist children with emotional and behavioural
problems. The Arts in Psychotherapy, 39, pp. 367−373.
Roos, V. 2008. The Mmogo-method: Discovering symbolic
community interactions. Journal of Psychology in Africa, 18(4), pp.
659−668.
Roos, V. 2012. The Mmogo-method: An exploration of experiences through
visual projections. Qualitative Research in Psychology, 9(3), pp. 249−261.
Faith, who is 5 years old (i.e. 60 months), was assessed on the Griffiths III after
the psychologist obtained informed consent from Jenny, Faith’s caregiver. Faith’s
parents both work full time, so Jenny took her to the assessment.
The assessment was started in English, but when it became
apparent early on that Faith found tasks challenging, the language
of assessment was switched to Afrikaans to deliver instructions.
The tester, who was a psychologist, was sufficiently fluent in both
English and Afrikaans to ensure that translations were appropriate.
Faith appeared quite anxious throughout the assessment and
continually sought reassurance from the tester administering the
items whenever she was unsure. Her weak grip strength was noted
and appeared to affect her hand and eye coordination the most.
Faith was assessed on all five subscales of the Griffiths III, namely: Foundations of
Learning (Subscale A); Language and Communication (Subscale B); Eye and Hand
Coordination (Subscale C); Personal, Social, and Emotional development (Subscale D);
and Gross Motor Skills (Subscale E), and her general development was also calculated.
On Subscales A, B, C, and D, Faith performed at a developmental
age of 3 years, which is two years below her chronological age of 5 years. Her
performance on these four subscales can be described as extremely low when
compared to other children of the same age. Faith’s best performance was on
Subscale E. She performed at a developmental age of 4 years on this
subscale, which is one year below her chronological age. Her performance
can be described as borderline in comparison to children of the same age.
Overall, Faith obtained a general developmental age of 3 years,
which is one year below her chronological age. Her general
developmental functioning can therefore be described as extremely
low in comparison to other children of the same age.
When interpreting the two profiles on the left (which reflect the
Scaled Scores and Developmental Quotients that Faith achieved),
the dotted line indicates average functioning, while a score below
the dotted line indicates below-average functioning.
Questions
(a) Did the psychologist employ good, ethical assessment
practices? Motivate your answer.
(b) Were there any ethical dilemmas? How did the psychologist
negotiate the dilemma(s)?
(c) Why do you think the psychologist used a diagnostic, as
opposed to a screening measure?
(d) Summarise Faith’s level of development and indicate what her
relative strengths and weaknesses are.
You have been asked to address teachers at a primary school about the use
of psychological assessment in schools. Prepare the talk that you will give.
15.4.3 Uses of tests in higher-education contexts
Entrance to universities and universities of technology is usually
dependent on meeting certain entrance requirements. Traditionally,
matriculation results have been used when determining admission to
higher education programmes. However, there is a growing body of
research that points to the unreliability of matriculation results in
predicting academic success in higher education, especially for
learners from disadvantaged educational backgrounds (e.g. Foxcroft,
2004a; Skuy, Zolezzi, Mentis, Fridjhon & Cockroft 1996). Furthermore,
there is a need to broaden access to higher education, especially for
historically disadvantaged South Africans. In addition, national and
international research findings suggest that the combination of
entrance assessment results with high-school performance leads to
significantly better predictions of academic performance (Anastasi &
Urbina, 1997; Foxcroft, 2004a, 2005; Griesel, 2001; McClaran, 2003).
Consequently, as happens elsewhere in the world, a number of
universities and universities of technology in South Africa have
instituted assessment programmes to supplement matriculation
information. Such an assessment needs to tap generic or basic
academic competencies such as verbal, numerical, or reasoning
abilities, which are seen as the foundation for success in higher
education programmes. In addition to these generic competencies,
subject-specific content (e.g. chemistry or accountancy) may also be
assessed. However, it is not sufficient to simply tap academic
competencies, as research has indicated that there are personal
attributes and behaviours that are equally important for academic
success. Fields of interest, motivation, study skills, having career
goals, academic self-belief, and leadership ability have inter alia been
found to be related to academic success.
Apart from assisting in admissions decisions, another advantage of
assessing learners on entry to higher-education programmes is that
the resultant information can be used to guide programme planning
and the development of student support initiatives. Furthermore,
assessment results can be used to counsel students regarding the
most appropriate programme placement for them as well as specific
aspects of their functioning that require more development.
Assessment at entry to higher education thus has the potential to go
beyond simply trying to select learners with the potential to succeed. It
can, in fact, serve a more developmental purpose, which ultimately
benefits the learner.
Psychological assessment measures are also used to assess
individual and groups of students during their university studies to
assess barriers to learning (e.g. the Learning Enhancement
Checklist), learning styles and strategies, well-being and adjustment
(e.g. Sommer & Dumont, 2011; Van Lingen, Douman & Wannenburg,
2011), career development, and so on. This assessment information is
used to plan interventions to enhance academic performance and
personal development and well-being.
A further application of psychological assessment in higher
education is to aid in learning design and materials development. For
example, based on the established relationship between spatial
perception and academic success in science and engineering
programmes, Potter, et al. (2009) developed and implemented ‘high-
imagery course material for engineering students’ (p. 109). Significant
gains in three-dimensional spatial perception were found, which
impacted significantly on academic success.
Tebogo, who is 48, has recently been retrenched. Since then, he has
struggled to sleep at night, has lost his appetite, sits at home staring into
space and sighing, and he complains of feeling very down in the dumps.
His wife becomes increasingly concerned about his lack of drive and energy and
his depressed mood. Tebogo is referred to you (as a psychologist) by his general
practitioner to find out what his problem is, as there is no physical cause for his
symptoms. What type of questions would you ask Tebogo in the initial interview?
What aspects of functioning would you want to tap when you plan an assessment
battery? If you were a registered counsellor or a psychometrist (in independent
practice), would you be able to perform this assessment? Motivate your answer.
15.6 Conclusion
Psychological assessment can be used in a range of contexts, some
of which were highlighted in this chapter. You probably noticed that in
many cases the same measures are used in different contexts (e.g.
intelligence tests are used in industry, education, and
psychodiagnostic assessment), although the purpose for which the
measure is used in different contexts might differ. Conversely, you
were also introduced to measures in this chapter that are only used in
a specific context (e.g. developmental assessment measures and
work-related aptitude measures). At the end of the chapter you were
reminded that the conclusions reached from an assessment are open
to alternative interpretations but that the task of the psychology
professional is to reach as informed a conclusion as possible.
By now you should have quite a few snapshots of the way in which
assessment measures are used in different contexts. At our next stop
in the Assessment Practice Zone, you will have an opportunity to
explore issues related to the interpretation of measures, as well as
how to report on the results of an assessment.
Interpreting and reporting
assessment results
CHERYL FOXCROFT AND KATE W. GRIEVE
Chapter CHAPTER OUTCOMES
16 This chapter deals with the different ways of interpreting assessment results
and reporting or conveying these to the people concerned.
By the end of this chapter you will be able to:
› explain different methods of interpreting assessment results
› explain the ethical considerations that need to be adhered to when
conveying assessment results
› explain the basic methods that can be used to convey assessment results
› understand the purpose and nature of an assessment report.
16.1 Introduction
Assessment measures are used for two main purposes: research and
applied practice. In the research situation, tests are used in a variety
of ways, depending on the purpose of the research. When you are
faced with tests that measure a similar domain, your choice of test will
be determined by the research problem. For example, if you were
interested in assessing the home environments of malnourished
children, you might decide to use the Home Screening Questionnaire
(Coons, Frankenburg, Gay, Fandal, Lefly & Ker, 1982), because it
takes all aspects of a young child’s environment into account, rather
than another measure, such as the Family Environment Scale (Moos
& Moos, 1986), that focuses on relationships between family
members.
In practical applications, assessment measures are used to help
professionals to gain more information about a person and to facilitate
therapeutic intervention. For example, assessment measures may be
used for child and adult guidance, to evaluate scholastic achievement,
to investigate psychopathology or determine individual strengths and
weaknesses, and for selection, placement, and training of people in
industrial or commercial organisations (see Chapter 15). Assessment
measures are generally used to answer specific questions about a
person. For example, if a child is not performing well at school, you
may wonder if this is due to limited intellectual ability, emotional
problems, or environmental difficulties. Assessment measures can
help us understand people’s functioning and indicate where
intervention is required.
Whatever the purpose of the assessment, the results of an
assessment need to be interpreted so that a person’s functioning can
be meaningfully described and conclusions can be reached. This
chapter outlines methods for interpreting assessment results.
Furthermore, the chapter will introduce you to good practices for
conveying assessment results and ethical dilemmas faced when doing
so. The chapter concludes with guidelines for writing effective
assessment reports.
16.2 Interpretation
Once you have administered a measure and obtained a score, you
have to decide what that result means for the specific person who was
assessed. In order to reach the best possible decision, detailed
information is required about the person concerned. Why is this so? A
test score on its own is meaningless (you will learn more about this in
Chapter 17). A test score only becomes meaningful when it is viewed
in light of information about the particular measure used, as well as all
aspects of the particular person, and the specific purpose of the
assessment. The assessment practitioner therefore has to interpret a
test score in order for it to be meaningful.
While you are administering a test to Thabo (who is 10 years old), he confides in
you that he is a victim of sexual abuse and that the perpetrator is a wellrespected
member of the family. What should you do? The situation in South Africa dictates
that if you suspect abuse of any kind, you are required by law to report it. In
matters of sexual abuse, it is generally best to encourage the person
concerned to report the problem themselves. In the case of a child, this is
more difficult. If you decide to report the matter yourself, make sure to have all
the facts first, to take steps to protect the child (this requires simultaneous
multidisciplinary intervention), and to ensure that professional services are
available to the family to deal with reactions to the matter. If adequate steps
are not taken, serious complications may arise (e.g. innocent people can be
traumatised, relationships in families can be irreparably damaged, and the
psychological practitioner can be sued for breach of confidentiality).
When someone is referred for assessment, it is easy for the assessment practitioner
to adopt a position of authority and lose sight of the individual concerned. Part of
professional accountability is to consider the rights of test-takers (Rogers, 1997).
It is often helpful to ask the test-taker about his or her own knowledge
or feelings about the aspect of behaviour that was assessed. This may
serve as a starting point for further elaboration. Discussing the test-
taker’s experience of the measures can provide information that can
confirm the validity of the results or throw light on ambiguous findings.
The test-taker should also be given the opportunity to express feelings
about the assessment procedure and to ask for further information.
Assessment results should be conveyed in a way that will best
serve the original purpose for which the test was administered. If, for
example, you collect test data with the aim of revising a school
curriculum, the test results should be interpreted and conveyed with
that purpose in mind.
As far as possible, assessment results should be conveyed in
general terms, in descriptive form rather than as numerical scores.
Test scores, such as those obtained on intelligence measures, are
often misunderstood and may lead to labelling, which can stigmatise a
person. This is particularly important when the test-taker differs from
the members of the norm group, in which case the score obtained
may not be an adequate reflection of the test-taker’s capabilities or
characteristics. In addition, the assessment practitioner should use
language that is understandable and avoid technical terminology or
jargon.
Probably the most important point is that communicating
assessment results should always occur in a context that includes
consideration of all relevant information about the individual who takes
the measure. We will discuss the importance of this point in Chapter
17. Contextual information provides a framework that makes any test
result richer and more meaningful.
16.5 Conclusion
In this chapter, you have learned that there are different ways of
interpreting and conveying assessment results, according to the
purpose of assessment, the individual being assessed, and the
context in which that individual functions. Interpretation of assessment
results relies on the validity of the particular measures used and
methods of interpretation vary according to the type of measures
applied. In the face of all this variation, it is the human element that is
the most important. The trained professional makes assessment
results meaningful by interpreting them in an ethically acceptable and
personally accountable way, based on training and expert knowledge.
In this way, numbers (test scores) can be transformed into valuable
information that serves the best interests of the individual.
ADVANCED READING
Page back to Ethical Dilemma Case Study 15.1 in Chapter 15. A case study was
presented of a Faith, a 5-year-old, who was assessed at the suggestion of a
psychologist and with the consent of her caregiver, Jenny. The assessment
findings on the five subscales of the Griffiths III were provided. Write a report for
Faith’s parents on the assessment results using the principles outlined above. Do
not use psychological jargon; focus on Faith’s strengths, and make practical
suggestions regarding aspects of Faith’s development that need to be stimulated.
Questions
(a) Was it appropriate for the RC to email the report to Lucas?
Motivate your answer.
(b) If a psychology professional forwards a copy of a client’s report electronically,
what does he or she need to do to ensure that confidentiality is not breached?
17.1 Introduction
Psychological assessment is based on the notion that people have
inherent abilities and characteristics that can be measured. These
broad traits or abilities are partly inherited but are not static, because
the way they are manifest can vary according to different contexts and
conditions. It is because we assume that traits permeate behaviour in
many contexts that we undertake psychological measurement and
make assumptions about behaviour. However, it is important to
remember that a test score is just a number. It is not equivalent to the
ability that is being measured. When we measure two kilograms of
potatoes we know exactly what we are measuring; with psychological
measures, however, we cannot measure an attribute or ability directly
and we therefore have to infer it from a person’s test performance. In
so doing, we give the person a score (or number) on the basis of the
response given. A score is therefore an inferred or indirect measure of
the strength of that attribute or ability. You learned about
measurement error in Chapter 3, and will therefore know that there
are factors in addition to a person’s inherent ability that can influence
scores and the meaning of those scores. In this chapter, we will look
at some important factors that should be considered when interpreting
performance on psychological measures.
Imagine that you have selected a suitable measure, administered it
carefully, and scored the protocol. You have probably also worked out
a scaled score with reference to a normative group (see Chapter 3).
But what does that score actually mean? You may be puzzled by this
question, so let us look at some examples.
Our first case study concerns John, who is 34 years old. He is the
managing director of an international company and a consultant for
several other business concerns. He has a pilot’s license and plays
chess on a national level. He was head boy of his school, obtained
distinctions in all his subjects, and was voted the boy most likely to
succeed by his class. John was referred for assessment because he
has recently had difficulties coping with his business commitments.
Our second case study involves Thandi, who is 20 years old and a
prospective university student. Her father is a manual labourer, she
comes from a disadvantaged environment, and the quality of her high
school education was poor. Thandi requested help in deciding
whether she would be capable of doing a science degree.
John and Thandi are assessed with a standardised intelligence
scale and both obtain an IQ of 105. Does this score mean the same
for both individuals? We will return to discuss these examples later,
but they serve as an introduction to this chapter in which we discuss
some of the important factors that have to be taken into account when
interpreting a test score.
Being a developing country, South Africa has a higher rate of traumatic brain injury
than the developed Western countries (Naidoo, 2013). There are different types of
brain injury and the consequences of the injuries vary as a result of individual
differences in physiological make-up, personal temperament and ability factors, as
well as social circumstances. When there is a focal injury (a blow to one particular
part of the brain), the impairment is usually limited to those functions served by the
specific area of the brain that is injured. When injury is widespread, there may be
diffuse axonal injury that results in microscopic damage throughout the brain.
This type of injury is frequently sustained in car accidents due to abrupt
acceleration-deceleration and rotational forces, and is often more debilitating
than focal injuries, because a larger area of the brain is involved.
People who have sustained diffuse brain injury, particularly if it is a severe
injury, usually have residual cognitive difficulties in the areas of attention and
concentration, verbal functioning and memory, as well as motor slowing (Lezak,
1995). There are often behavioural difficulties, such as increased irritability and
aggression, decreased motivation, and emotional disorders, such as depression.
There is no one pattern of symptoms and the effects of brain injury vary with age,
the location and severity of the injury, and the personal and social circumstances of
the person concerned. However, the effects of brain injury are generally seen in
poor performance on intelligence tests and neuropsychological measures designed
to assess different functions and abilities. An intelligence measure alone is often not
sufficient to identify brain impairment, because a previously intelligent person may
be able to rely on previously well-learned and established patterns of behaviour and
abilities and still be able to perform adequately while, at the same time, being unable
to cope with new situations and the demands of daily living.
Did you know that mild traumatic brain injury can occur when sports
players suffer numerous concussions? This is why there are concussion-
management programmes in place and sports players are sent for concussion
tests, which often include the ImPACT computerised neurocognitive test
(Shuttleworth-Edwards, Whitefield-Alexander, Radloff, Taylor & Lovell, 2009).
17.2.3.2 Language
Language is generally regarded as the most important single
moderator of performance on assessment measures (Nell, 2000). This
is because performance on assessment measures could be the
product of language difficulties and not ability factors if a measure is
administered in a language other than the test-taker’s home language.
At some time in our lives, we have probably all had to read or study
something written in a language different to our own and, even though
we may have a working knowledge of that language, we experience
the feeling that it takes longer to process the material in another
language. We can think and discuss so much better in our own
language. By the same token, we would probably therefore perform
better in a test that is written and administered in our home language
than in a test in a second language. In this way, language becomes a
potential source of bias (see Chapters 6, 7 and 9). When a test is
written in a different language, it may present a range of concepts that
are not accessible in our home language.
You may think that the problem can be solved by translating the
measure. Although this is the solution in some instances, this is not
always the case because there are other difficulties (see Chapter 7).
Some languages do not have the concepts and expressions required
by measures and an equivalent form of the measure cannot be
translated. Measures of verbal functioning pose a particular problem.
For example, you cannot merely translate a vocabulary test into
another language because the words may not be of the same level of
difficulty in the two languages.
There is also the problem of what is the most suitable language to
use (see Chapter 9). Let us consider an example. Itumeleng speaks
Northern SeSotho at home, together with a little bit of English and
some isiZulu he has learned from his friends. From Grade 4 his
lessons at school have been presented in English and he is currently
studying through the medium of English at university. He has acquired
most of his knowledge and experience in English and he has a
working knowledge of concepts in English. Would it be better to
assess Itumeleng’s level of ability in English or in his home language?
There is no simple answer to this question. Many students prefer to be
assessed in English but do not have the advantages of people whose
home language is English; they are therefore not comparable. It is
unfortunate that this situation often results in a double disadvantage in
that their competence in both languages is compromised. For
example, students may have some knowledge of English but it may
not be very good and, at the same time, having to study in English can
detract from further development of their home language. This would
have an influence on assessment performance and may mask the
actual ability that the measure is designed to assess. In addition, De
Sousa, Greenop, and Fry (2010) point out that exposure to
monolingual and bilingual-literacy learning contexts can impact on
cognitive processing which, in turn, could result in differing
performance on some cognitive tasks.
An additional problem arises from the fact that a combination of
languages, referred to as a ‘township patois’, is commonly used in the
residential areas surrounding cities and a pure version of one
language is seldom spoken. A child who grows up speaking the patois
would be at a disadvantage if assessed with a formally translated
measure. The solution may lie in bilingual or multilingual assessment
and the possible use of drawings where a test-taker can capture
visually what they are unable to express verbally (Moletsane & Eloff,
2006). The moderating impact of language on assessment
performance remains a challenge for the field of psychometrics in
South Africa (see Chapters 6, 7, and 9).
Furthermore, where an assessment measure is only available in
English but the measure will be applied to test-takers for whom
English is their second or third language, an option is to make use of
language/text simplification, which includes the use of simpler
English words and breaking down the syntax of complex sentences so
that the language used in the measure does not introduce bias or
construct irrelevant variance (Aluisio, Specia, Gasperin & Scarton,
2010). This approach is commonly used in international assessment
programmes such as the Trends in International Mathematics and
Science Study (TIMSS) and the Programme for International Student
Assessment (PISA). In South Africa, the English version of the
Learning Potential Computerised Adaptive Test (LPCAT) employs
language simplification. The test-taker can ‘hover over’ the English
item with their cursor and simplified wording appears.
It is not only the language background of the test-taker that adds to
the complexity of psychological assessment interpretation in South
Africa. Very often, the psychology professional is not proficient in the
home language of the test-taker, which makes communication difficult.
It is thus not surprising that in a survey of intern psychologists, almost
76 per cent indicated language difficulties with clients as being
problematic (Pillay & Johnston, 2011). Language differences not only
provide communication challenges but also impact on test
performance. For example, Kamara (1971) reports that when children
from Sierra Leona were tested by an assessment practitioner who
spoke their home language their performance was better than when
this was not the case.
17.2.3.3 Culture
Our culture has a pervasive influence on the way we learn, think about
things, and behave. In fact, culture cannot be considered as a
separate factor, exerting an influence apart from other factors in the
person’s environment. It is an integral part of that environment.
Although there may be cultural differences in test scores, there is no
decisive evidence that culture influences competence rather than
performance. In other words, although socio-cultural factors may
affect manifestations of underlying cognitive and emotional processes,
it is unlikely that cultural differences exist in the basic component
cognitive processes (Miller, 1987). For example, empirical studies
have shown that the pattern and sequence of acquisition according to
the Piagetian stages of cognitive development is universal, although
the rate of acquisition may vary among different cultures
(Mwamwenda, 1995). This is why it is important to take cultural
differences into account when interpreting test performance.
The content of any measure will reflect the culture of the people
who designed the measure and the country in which it is to be used
(see Chapters 1, 6, and 7). Clearly, people who do not share the
culture of the test developers will be at a disadvantage when taking
that particular measure. For example, a child’s performance on a
memory task is likely to vary according to the way in which the task is
constructed – the more familiar the nature of the task, the better the
performance (Mistry & Rogoff, 1994). Similarly, Malda, Van de Vijver,
and Temane (2010) developed cognitive tests in which item content
was drawn from both the isiXhosa and Setswana cultures equally. The
two versions of the tests were administered to samples of Afrikaans-
and Setswana-speaking children. They found that the children
performed better on the tests developed for their cultural/language
group, especially when the tests tapped more complex functions.
Despite several attempts, there is no such thing as a culture-free
measure but assessment practitioners can be sensitive to culture
fairness in assessment. For example, items should be examined for
cultural bias (see Chapters 6 and 7). When a measure is administered
to a person who does not share the culture of the people who
developed it and the country in which it was standardised, the
assessment practitioner should always consider that the test score
obtained may not be a good reflection of the person’s true level of
functioning due to the influence of cultural factors. An additional
consideration is that measures and their resultant scores cannot be
assumed to have equivalent meaning for different cultures and
countries (see Chapter 7). For example, a Spanish version of the
Wechsler Adult Intelligence Scales has been developed in Spain, but
there is doubt about its suitability for Spanish-speaking people living in
the US. When considering adopting a measure developed elsewhere,
validation studies have to be undertaken first (see Chapter 7).
According to the belief that behaviour is shaped by prescriptions of
the culture, its values and guiding ideals, it is accepted that cultural
experience influences the meaning of events for an individual and,
therefore, responses will differ among cultures. This is particularly
important in a projective measure, such as the Thematic Apperception
Test, that rests on the assumption that the stories an individual tells
based on the pictorial stimulus material will reveal something of the
person’s personality disposition. For example, a card showing a
picture of a man working on a farm may elicit different responses
according to the culture of the person being tested. One person may
say that the man is poor because he is working the land himself and
has no labourers to help him, while another may say the man is rich
because he has his own land that he can farm. A different response
should not be seen as a deficient (poorer or abnormal) one but should
be viewed as a function of the person’s culture.
In addition to the number of different cultures in South Africa, we
also have the problem of variations in acculturation. Acculturation
refers to the process by which people become assimilated into a
culture. This process occurs at different speeds and is not the same
for all facets of behaviour. We see, for example, that many people
who move from rural areas into cities generally let go of values
associated with a more traditional way of life and adopt a more
Western way of living. But not everything changes at the same pace.
There might be quite a quick change in the way the person dresses or
in habits such as watching television, but the person might still adhere
to more traditional beliefs such as consulting a faith healer rather than
a psychologist when there are problems. There are wide interpersonal
differences in the degree and pace of acculturation. We cannot, for
example, presume that someone who has lived in a city for five years
has adopted Western values. Differences in acculturation have
important implications for performance on assessment measures due
to the impact of the familiarity with the culture on which the measure is
based (see for example Kennepohl, Shore, Nabors & Hanks, 2004).
The closer the person’s personal values and practices are to the
culture, the more appropriate the measure. In turn, if the measure is
appropriate, the person’s test performance is more likely to be an
accurate indication of his or her actual level of functioning. The effects
of acculturation can be seen in the closing of the gap in intelligence
test scores of black and white South African students who have
benefitted from good quality education (Shuttleworth-Edwards et al.,
2004).
In some cases, it becomes very difficult to know which measure is
most appropriate to use. Let us consider the following example.
Nomsa is 12 years old and her home language is isiXhosa. She
attends a school in a middle-class suburb, speaks English at school,
and has friends from different cultural groups. Should she be
assessed with the Individual Scale for isiXhosa-speaking Pupils
because of her home language or with the SSAIS-R (see Chapter 10)
because of her situation? There is no easy answer to this question.
Each person’s situation has to be considered individually, according to
his or her familiarity with the predominant language and culture.
The most important point to remember is that without measures
with culturally relevant content and appropriate norms, fair testing
practices may be compromised. The use of potentially culturally
biased measures may have enormous implications for making
important decisions in people’s lives. For example, people may be
denied educational and occupational opportunities due to a test score
that is not a valid indicator of their ability. However, to add points to a
score to compensate for cultural differences (so-called ‘affirmative
assessment’) does not provide the answer. For instance, if a
psychologist used a particular test in an attempt to determine a
specific deficit, adding points may mask the presence of a deficit and
this would not be responsible practice. The potential contribution of
cultural influences to assessment performance should always be kept
in mind.
17.2.3.4 Environmental factors
Performance on measures of ability is influenced by variation of
environmental factors. We need to consider distal factors, or those
relating to the wider environment, such as socio-economic status and
the degree of exposure to an enriching social environment, as well as
proximal factors, or those relating to the individual’s immediate
environment, such as socialisation experiences in the home. Why is
this important? Environmental factors determine the types of learning
experiences and opportunities to which we are exposed and this, in
turn, affects our level of ability and the extent to which we are able to
use that ability. This, of course, plays a role in test performance.
The home environment
There are certain child-rearing practices that have been shown to
promote the development of competence and cognitive abilities that
are tapped by traditional measures of development. The kinds of child-
rearing practices shown to promote children’s development are
parental responsivity (including fathers’ involvement) and the provision
of home stimulation in the form of the mother’s active efforts to
facilitate cognitive and language skills, allowing the child to explore the
possibilities of the environment, encouraging interaction with the
environment, and supporting the development of new skills (Grieve &
Richter, 1990). However, there are also indications that universal
assumptions about development do not apply equally to the process
of development in all cultures. Certain culturally sanctioned child-
rearing practices may appear to be contradictory to mainstream
Western practices but should not necessarily be seen as deviant or
detrimental (Garcia Coll & Magnuson, 2000). Therefore, in assessing
children, it is important to understand children and their families in
relation to both the immediate and larger socio-cultural environment
and the ways in which cultural factors operate in the child’s
environment (September, Rich & Roman, 2016). Individual differences
within cultures should also be taken into account. For example,
despite living in disadvantaged communities in South Africa, there are
mothers who are able to structure their children’s environments in
order to promote their children’s mental development, irrespective of
cultural practices and impoverished environments (Richter & Grieve,
1991).
There are certain characteristics of household structure that can
also influence children’s abilities. Overcrowding, for example, has
generally been found to have an adverse affect on cognitive
development as well as on the educational performance of young
children. A study of South African children from previously
disadvantaged communities has suggested that overcrowding
negatively influences children’s cognitive milieus by, for example,
limiting and interrupting exploratory behaviour, decreasing the number
of intimate child-caretaker exchanges, and increasing noise levels
(Richter, 1989). Overcrowding does not only have a negative effect on
children’s development but has been shown to have a detrimental
effect on adults as well. However, overcrowding does not necessarily
have a negative impact; it all depends on whether the person
interprets the situation as stressful. People who have grown up in
crowded homes and are used to crowded places generally find them
less unpleasant than people who are used to a greater degree of
privacy. Overcrowded homes may even offer an advantage to young
children, as a result of the possibility of multiple caretakers to provide
affection and see to their needs, particularly in cases where their
mothers work away from home.
Socio-economic status
Socio-economic status (SES) refers to the broader indices of a
person or family’s social standing. The major indicators of SES are
education, occupation, and income. You may wonder why this would
affect performance on psychological measures. A person’s SES is
important because it determines the type of facilities that are available
(such as schools, libraries, clinics, and other social services), the
opportunities that present themselves, and the attitudes generally
exercised by others in the same socio-economic group. The socio-
economic aspects of the environment thus have a significant influence
on the experiences that moderate abilities, attitudes, and behaviour. It
has been found, for example, that mothers with relatively high levels of
education are more likely to provide stimulating home environments
than mothers who are less educated. The educated mothers are more
likely to read magazines and newspapers (and therefore are more
informed about issues), play actively with their children, help them
learn rhymes and songs, and take them on outings (Richter & Grieve,
1991). These types of activities help to promote children’s
development and well-being. Similarly, talking about current issues
and having books at home help to promote a culture of learning.
Some of the socio-economic factors that are associated with
educational disadvantage are poverty (see Box 17.4), poor health,
malnutrition, inadequate school facilities, poor social adjustment,
absence of cultural stimulation, lack of books and the other aspects of
modern civilisation that contribute to the development of Western
ways of thinking, which is what is measured in traditional
psychological measures. People from disadvantaged environments
often do not perform well on psychological measures because the
tasks are unfamiliar and may be regarded as unimportant.
Achievement in many cognitive measures is a function of the ability to
develop strategies to solve problems in the test situation, as well as of
being motivated to do well.
Despite the fact that poverty rates are decreasing In South Africa, there are still a
large number of children living In poverty (Barnes, Hall, Sambub, Wright & Zembe-
Mkabile, 2017). For a child, being poor means being at risk for developing both
physical and psychological problems. What are these risks? In comparison with
children from higher socio-economic backgrounds, children from impoverished
backgrounds experience, among other things, a greater number of perinatal
complications, higher incidence of disabilities, greater likelihood of being in foster
care or suffering child abuse, greater risk for conduct disorders, and arrest for
infringements of the law (Barnes, Hall, Sambub, Wright & Zembe-Mkabile, 2017).
One of the major reasons for developmental risk is believed to be associated
with the fact that poverty increases the risk of inattentive or erratic parental care,
abuse, or neglect because a significant proportion of parents living in poverty
lack the personal and social resources to meet their chilsdren’s needs (Halpern,
2000). Many young adults living in impoverished inner-city communities have
personal experiences of parental rejection or inadequate nurturance, disrupted
family environments, parental substance abuse, family violence, and difficulties
at school. These adverse conditions have a negative effect on their own abilities
as parents as well as their ability to form adult relationships, the capacity to
complete schooling, to obtain and keep jobs, and to manage their households. In
addition, overstressed parents can adversely influence their children in an
indirect way by communicating their despondency and despair. This does not
mean that all parents in impoverished circumstances are unable to fulfil parenting
roles successfully. However, it suggests that consideration should be given to the
fact that the stress of poverty can play a major role in directly undermining the
efforts of even the most resourceful parents and can therefore detract from the
quality of childcare.
Urbanisation
It is generally found that urban children show superiority over rural
children in terms of cognitive performance (Mwamwenda, 1995). The
reason for this may be that the enriched and invigorating urban
environment stimulates those aspects of cognition usually assessed
by formal psychological measures. It is also more likely that provision
is made for formal schooling in urban rather than rural environments
and access to education at an early age is considered to be the single
most powerful factor facilitating cognitive development. Urbanisation is
also associated with higher parental levels of education and we
pointed out earlier that better-educated mothers are more likely to
provide the kind of home environments that are believed to be
beneficial for children’s cognitive development.
The effects of urbanisation on the changes in people’s ability as
measured by psychological measures can be seen in South African
studies that enable the comparison of test performances of children
over a period of 30 years. When the New South African Group Test
(NSAGT) was standardised in 1954, it was found that English-
speaking pupils were better than Afrikaans-speaking pupils at all age
and education levels by an average of 10 IQ points. When the
successor of the NSAGT, the General Scholastic Aptitude Test
(GSAT), was administered in 1984, a difference of five IQ points in
favour of English-speaking children aged 12, 13, and 14 years was
reported. The reduction of the difference between the groups is
ascribed to the fact that, during the period between 1954 and 1984,
Afrikaans-speaking whites went through a period of rapid urbanisation
during which their level of education and income increased
significantly. The gradual convergence of the scores corresponded
with the more general process of cultural and economic convergence
between the populations, facilitated by the process of urbanisation.
Urbanisation must, however, be considered as a global variable
that cannot offer a complete explanation for the psychological
mechanisms which account for differences in ability.
17.2.3.5 Test-wiseness
All the factors that we have discussed here contribute to a state of
test-wiseness or test-taking skills (Nell, 2000) (see Chapters 8 and
9). Such skills include assuming that you need to work as fast and as
accurately as you can, and that items get progressively more difficult
(Van Ommen, 2005). Many illiterate adults, those with little schooling,
or those who live in isolated communities are not test-wise. This
means that they do not have the skills for, and understanding of, the
requirements for successful test performance. Often they do not
understand the motivation for doing the assessment measure, and are
unfamiliar with the ethic of working quickly and accurately.
Clearly, socio-cultural experiences can affect the quality of
responses to a psychological measure. Assessment techniques
developed for Western children do not always have the same
applicability to non-Western children, depending on the ability that one
is assessing. Culture can determine the availability or priority of
access to the repertoire of responses that the person has, irrespective
of the responses that the examiner requires. Not only language
differences, but also demands posed by test requirements create
problems, because they relate to social skills, the development and
expression of which are encouraged in Western society. Experiences
within particular socio-cultural contexts can significantly influence a
person’s understanding of the meaning of a task, and may place
greater emphasis on competencies and cognitive processes that differ
from those expected in the assessment situation. It must be noted,
however, that, while culture plays a contributory role, it cannot be
isolated from the social and personal contexts in which people live.
As the use of technology in testing is becoming the norm (see
Chapter 14), an additional aspect has been introduced into
considerations of test-wiseness.
Questions:
(a) Was it appropriate for Ashwin to administer the measures in
English? Motivate your answer.
(b) Could the measures have been biased? Were the test results
used fairly? Motivate your answer.
(c) How could Ashwin have approached this dilemma differently to
avoid ethical transgressions?
17.4 Conclusion
What have we learned from all this? In this chapter, you have been
introduced to several ideas about test scores and assessment results.
Perhaps the most important is that a test score on its own is limited in
what it can tell us about a person’s behaviour. The test score
represents a person’s performance. The person’s ability or trait is
inferred from the score. In order for scores to be meaningful, we need
to look at the scores in context and in terms of the psychometric
properties of the measure. Many problems and tricky issues have
been raised, particularly with regard to applying psychological
measures in a heterogeneous society such as South Africa. Given
these difficulties, you may be wondering whether psychological
assessment is desirable, or even whether it is worth using
psychological measures at all. The answer, a qualified ‘Yes’, lies in the
ability of the person administering and interpreting the measures to do
so competently and responsibly. Psychological assessment is a
necessary condition for equity and efficient management of personal
development. By not using measures, you could be denying many
individuals access to resources or opportunities. By not using
measures responsibly, you could be denying individuals their rights.
Write a short article for a local newspaper in which you motivate the continued
use of psychological measures against the backdrop of the numerous factors
that affect the accuracy and reliability of the assessment results.
18.1 Introduction
For now, we have arrived at the final stop of the journey through
psychological assessment territory. In this chapter we will reflect on
the extent to which progress is being made on the road towards
psychological assessment becoming more culturally meaningful.
Thereafter, a roadmap to guide the next steps of this journey will be
outlined. Lastly, concluding remarks will be made regarding the
resilience and determination needed to continue this daunting journey
of integrating African-centred and Western-oriented approaches to
ensure that psychological assessment is applied in culturally
meaningful and ethical ways in South Africa.
Figure 18.1 Core aspects needed to expand culturally meaningful assessment in South Africa
18.3.1.1 Routes to developing indigenous measures and
adapting Western-oriented measures
A common understanding has generally been reached that the best
approach to create a pool of culturally appropriate measures is to both
develop indigenous, African-centred measures and adapt appropriate
Western-oriented measures for the South African context. How do we
set about undertaking such development and adaptation?
As was pointed out in Chapters 6 and 7, whether developing an
indigenous measure or adapting an imported, international one, a
multidisciplinary team needs to be assembled, which should include
cultural experts from the communities concerned as well as language
experts.
Chapter 6 provides a detailed account of how to develop a
psychological assessment measure, while Section 7.7 in Chapter 7
provides more depth regarding the development of tests for a
multicultural context. But you now have a broader understanding of
etic and emic approaches in cross-cultural assessment. One of two
routes can be followed when using a combined emic-etic approach to
conceptualise and delineate the construct domain of an indigenous
test in the development phase. The first route is termed ‘etic-driven’
(Iliescu & Ispas, 2016, p. 137). Here a universal conceptualisation of
the construct domain is applied to which emic, culture-specific traits or
sub-domains are added. The second route is termed ‘emic-driven’
(Iliescu & Ispas, 2016, p.137). Following this route requires that the
construct domain of interest be explored ‘from the ground up’ so that a
culture-specific (emic) understanding of a construct develops. The
resultant structure that emerges of the construct can then be
compared to structures of the construct that are thought to be
universal (etic). This provides insight into aspects of the construct that
are similar across cultures and those that are unique.
When figuring out which constructs to tap in an indigenous
measure, the findings of research studies could also prove useful. For
example, Cockroft, Bloch, and Moolla (2016) compared the
performance of South African school beginners from high and low
socio-economic groups on working memory and vocabulary
measures. They found that ‘socio-economic status accounts for
considerable variance in vocabulary measures, while it explains only
small amounts of variance in working memory measures’ (p. 199).
Thus, when developing an indigenous cognitive test battery, including
working memory as one of the sub-domains can result in a fairer
assessment if children from differing socio-economic groups are to be
assessed on the battery.
In Table 7.1 in Chapter 7, eight kinds of test adaptation were
indicated (Van de Vijver, 2016). When adapting a measure for use in
the multicultural and multilingual South African context, the first step is
to decide which types of adaptation will be needed. It is considered
good practice to include a combination of types of adaptation (Van de
Vijver, 2016). The subsequent steps to adapt a measure were outlined
in Chapter 7 (especially in 7.3.1 and 7.6). Readers are also referred to
Van de Vijver (2016) and the International Test Commission’s
Guidelines for Translating and Adapting Tests − 2nd edition
(International Test Commission, 2016) for further guidance on how to
adapt a measure.
Whether an indigenous measure is to be developed or an imported
one adapted, progress needs to be made regarding assessment
measures being available in multiple languages. There is growing
evidence that more measures are being adapted and translated,
which results in different language versions of a measure being
available. But to have 11 different language versions of a measure
available is quite challenging, costly, and time-consuming. Given that
many South African children are educated in a language that is not
their mother tongue, the possibility of bi- and even tri-lingual
assessment should be explored (see 9.2.7 in Chapter 9).
An innovation that has been seen in the research arena has been
the development of parallel versions of indigenous cognitive measures
for different cultural or language groups. The same construct is
measured, but some of the content differs to allow for greater
familiarity for each specific group, and scores are standardised across
the groups to allow for direct comparisons (Malda, Van de Vijver &
Temane, 2010). Given the multicultural and multilingual nature of our
country, it is hoped that parallel versions of indigenous measures will
be developed for use in applied clinical, counselling, educational, and
organisational contexts as such assessment will be fairer and more
culturally meaningful.
18.3.1.2 Inspiration for African-centred item development
An aspect that can be a challenge in indigenous test development is
where to seek inspiration from for developing African-centred item
content. Some ideas that surfaced in this regard during our
assessment journey in this book are:
Turn to African artwork to develop geometric shapes used in problem-
solving cognitive assessment tasks (Bekwa, 2016; Bekwa & De Beer,
2012) (see Figure 10.1 in Chapter 10).
Study the context so that appropriate emic tasks can be developed.
For example, Reuning and Wortley’s (1973) study of San in the
Kalahari Desert resulted in the development of a three-dimensional
maze presented on a wooden board that had to be solved. This was a
more familiar task than the unfamiliar one of solving a paper-based
maze using a pencil (see 2.4.3 in Chapter 2).
Instead of using two-dimensional, visually represented patterns on a
page, use familiar three-dimensional objects such as beans, stones,
and caps from bottles that can be packed out into patterns that need
to be completed (e.g. Zuilkowski, et al., 2016) (see 6.2.2.1 in Chapter
6).
Where drawing something using pencil and paper is unfamiliar,
drawing something using a stick on wet clay might be a more familiar
task, as could be producing an object visually by making it out of clay
or string rather than drawing it (e.g. Mpofu, 2002) (see 15.3.4.1 in
Chapter 15).
Activities undertaken in therapy can take on the form of an
assessment technique. For example, creative arts such as painting,
using clay, music, dance and movement, and using sand-trays and
sand play, or using the Mmogo-method® (Roos, 2008, 2012) provide
opportunities to gain insight into a child’s behaviour and level of
functioning (see 15.3.4.5 in Chapter 15). Other than the insights
gained, it is possible that these creative activities can spark ideas for
African-centred item development. But we have to find ways to
harvest these ideas. Perhaps test-development teams could consult
therapists who use these creative activities during therapy to be
inspired and develop ideas that can be translated into items?
In career-construction and life-design counselling, clients develop
stories and narratives about themselves based on prompting
questions and tasks. From this, the counsellor and client then ‘co-
construct a life-career portrait’ (e.g., Hartung, 2015, p. 116) (see
13.5.1 in Chapter 13). Given the strong oral tradition in African
cultures, test tasks that require storytelling and even listening to
stories could prove to be meaningful and familiar activities.
18.3.1.3 Delineating norm groups
The best way or ways to delineate and compile norm groups must be
found (Grieve & Van Eeden, 1997; Huysamen, 1986; Nell, 1997;
Schepers, 1999). Grieve and Van Eeden (1997) suggest that the
factors that moderate test performance in a heterogeneous society
provide a useful starting point when trying to identify homogenous
norm groups. Factors such as language proficiency, cultural
background, extent of urbanisation, socio-economic level, extent and
quality of formal education, as well as ‘test-wiseness’ have been found
to be important moderator variables on performance in psychological
measures in South Africa (Nell, 1997; Shuttleworth-Edwards, 2017).
Huysamen (1986) furthermore argues that, other than developing
separate norms for different target groups, which is fairly costly, the
other route would be to select only those items that do not unfairly
discriminate between the different groups and to then use only one
(common) norm group. In this instance, the more unique aspects of a
particular group may be discarded in favour of what they have in
common with other groups. Another potential solution is to use the
typical performance of particular subgroups of the larger population as
a criterion or standard against which a test-taker’s test performance
can be compared (Raiford, Weiss, Rolfhus & Coalson, 2005, 2008).
The outcome of the comparison could assist in reaching a diagnosis
or developing an intervention. The Advanced Reading block in Section
10.4.1.2 in Chapter 10 problematises the matter of delineating norm
groups in a multicultural context.
18.3.1.4 Qualitative measures
There is a strong call for the development and use of qualitative
measures as an alternative approach to quantitative, psychometric-
based psychological assessment. While some of these calls come
from those who focus on psychotherapy, some innovative work has
been done in this regard in the field of career counselling (see Section
13.5 in Chapter 13). Getting clients to narrate career stories or to
reflect on career influences in structured ways while still allowing for
flexibility has resulted in a range of qualitative career-assessment
measures (McMahon et al., 2003). Given the structured approached, it
is possible to use the same assessment tasks and activities with
different clients, and the outcomes or outputs (e.g. stories) lend
themselves to being used as research data. Going forward, capacity-
building is needed in structuring or systematising qualitative
assessment tasks to produce qualitative assessment measures. In
this regard, McMahon et al. (2003) provide suggestions to guide the
development of qualitative career-assessment measures (see Section
13.4.1 in Chapter 13). A further matter that will require attention is the
standards or requirements that qualitative assessment measures must
meet. Some possible minimum requirements could be:
The identification and delineation of the constuct(s) tapped by the
measure needs to be evaluated and found to be credible and
trustworthy by a panel of experts. This provides qualitative ‘evidence’
of construct validity.
Sensitivity reviews of items or tasks, as well as of the process of
presenting and responding to the assessment tasks for designated
subgroups of interest (DSI), needs to be undertaken using expert
panels and focus groups with samples from the target population. This
should ensure that the assessment tasks or activities are not biased
against any of the DSIs. The usual procedures used in qualitative
research to establish the trustworthiness and consistency of the
findings from the sensitivity review should be used. Consult Aston
(2006) for an example of conducting a sensitivity review of the WAIS-
III.
• Guidelines for making sense of the output or artefact produced and
assisting clients to ‘connect the dots’ and, in so doing, co-
developing a holistic sense of their development, career pathways,
and so on, could be helpful.
What other requirements can you think of?
18.3.1.5 Creating repositories of research studies into
assessment measures While progress is being made, it is still on a
relatively small scale as researchers often focus on locally/regionally
available samples and adapt tests for cultures and language groups in
a region. To derive maximum benefit from these initiatives, cross-
cultural research databases on commonly used South African
measures should be established and Web sites should be developed
to assist researchers to gain access to these databases easily. In this
way, information on specific psychological measures from research
studies across the country can be pooled.
It should be noted that if such repositories are to be created, the
informed consent obtained by psychological assessment practitioners
from test-takers must include the use of assessment results for
research purposes. Test-takers will need to be assured that their
assessment data will remain anonymous in this case. This can be
done by not including the test-taker’s name, identification number, and
address in the repository.
Questions
(a) Under what circumstances is it ethically acceptable to use an
imported test in clinical assessment practice?
(b) Can psychological test data be used for research purposes
without the consent of the client? Motivate your answer.
(c) If Cyril submits his database to a national project aimed at pooling research
studies into specific tests across the country, what should he do to ensure that
he does not breach the confidentiality assurance that he gave to his clients?
18.4 Conclusion
When he was freed from his lengthy period of imprisonment, the late
Nelson Mandela knew that the long journey to a democratic South
Africa was not yet over. He thus had the following to say: ‘The long
walk continues’ (Hatang & Venter, 2011, p. 117).
In this chapter we have reflected on the progress that we have
made in transforming psychological assessment in South Africa to
become more culturally meaningful. While steady progress is being
made, there is undoubtedly still a long road ahead. We have outlined
some of the core aspects that need to be tackled in the next phase of
the journey, some of which are quite daunting. However, having taken
up the challenge of transforming psychological assessment in South
Africa, we cannot stop now. The long walk has to continue until
psychological assessment is perceived as being meaningful, relevant,
and of value for all cultural groups in South Africa.
But, for the moment, we have reached the end of our journey into
the world of psychological assessment in this book. We hope that you
have enjoyed the journey and that you share our enthusiasm for the
valuable role that psychological assessment can play in assisting
individuals, organisations, educational and training institutions, and
the broader South African society. However, given the sensitive nature
of psychological assessment and the potential for misuse, we hope
that this book has provided you with some of the more important
signposts regarding fair and ethical assessment practices in the
multicultural and multilingual South African context. When you begin,
or continue, your real-life journey as an assessment practitioner, we
hope that the lessons that you have learned will transcend ‘book
knowledge’ and will make a difference to the way in which you
practise psychological assessment and contribute to its transformation
in South Africa.
References
Abrahams, F. (1996). The cross-cultural comparability of the 16
Personality Factor Inventory (16PF). Unpublished doctoral
dissertation, University of South Africa.
Abrahams, F. (2002, December). The (un)fair usage of the 16PF
(SA92) in South Africa: A response to C.H. Prinsloo and I.
Ebersohn. South African Journal of Psychology, 32(3), 58−61.
Abrahams, F. & Mauer, K.F. (1999a). The comparability of the
constructs of the 16PF in the South African Context. SA Journal of
Industrial Psychology, 25(1), pp. 53–59.
Abrahams, F. & Mauer, K.F. (1999b). Qualitative and statistical
impacts of home language on responses to the items of the 16
Personality Factor Questionnaire (16PF) in South Africa. SA
Journal of Psychology, 29, pp. 76–86.
Abubakar, A., Holding, P., Van de Vijver, F., Bomu, G., & Van Baar, A.
(2010). Developmental monitoring using caregiver reports in a
resource-limited setting: The case of Kilifi, Kenya. Foundation Acta
Pædiatrica/Acta Pædiatrica, 99(2), pp. 291–297.
doi:10.1111/j.1651-2227.2009.01561.x
Adler, D.A., McLaughlin, T.J. & Rogers, W.H. (2006). Job
performance deficits due to depression. The American Journal of
Psychiatry, 163(9), pp. 1569–1576.
Aguinis, H. (2013). Performance Management (3rd edition).
Englewood Cliffs, NJ: Prentice-Hall.
Aguinis, H., Culpepper, S. A., & Pierce, C.A. (2010). Revival of test
bias research in preemployment testing. Journal of Applied
Psychology, 95(4), pp. 648–680. doi: 10.1037/a0018714.
Aguinis, H., & Stone-Romero, E.F. (1997). Methodological artifacts in
moderated multiple regression and their effects on statistical
power. Journal of Applied Psychology, 82, pp. 192–206.
Ahmed, R. & Pillay, A.L. (2004). Reviewing clinical psychology training
in the post-apartheid period: Have we made any progress? SA
Journal of Psychology, 34(4), pp. 630–656.
Aiken, L.R. & Groth-Marnat, G. (2005). Psychological testing and
assessment (12th edition). Massachusetts: Allyn and Bacon, Inc.
Ajayi, O.R., Matthews, G.B., Taylor, M., Kvalsvig, J.D., Davidson, L.,
Kauchali, S. & Mellins, C. (2017). Structural equation modeling of
the effects of family, preschool, and stunting on the cognitive
development of school children. Front. Nutr, May 15, 4:17. doi:
10.3389/fnut.2017.00017
Allen, M.J. & Yen, W.M. (1979). Introduction to measurement theory.
Brooks/Cole Publishing Company: Monterey, California.
Aloia, M., Arnedt, J., Davis, J., Riggs, L., & Byrd, D. (2004).
Neuropsychological sequelae of obstructive sleep apnea-
hypopnea syndrome: A critical review. Journal of the International
Neuropsychological Society, 10(5), pp. 772–785.
Alonso-Arbiol, I. & Van de Vijver, F.J.R (2010). A historical analysis of
European Journal of Psychological Assessment. European Journal
of Psychological Assessment, 26(4), pp. 238–247.
Alpern, G.D., Boll, T.J., & Shearer, M. (1980). The Developmental
Profile II. Aspen, CO: Psychological Development Publications.
Aluisio, S., Specia, L., Gasperin, C. & Scarton, C. (2010). Readability
assessment for text simplification. Proceedings of the NAACL HLT
2010 Fifth Workshop on Innovative Use of NLP for Building
Educational Applications, pages 1–9, Los Angeles, California, June
2010. Association for Computational Linguistics. Retrieved on
15 July 2012 from
http://rer.sagepub.com/content/74/1/1.full.pdf+html
Aluja, A., & Blanch, A. (2004) Replicability of first-order 16PF-5
factors: An analysis of three parcelling methods. Personality and
Individual Differences, 37, pp. 667–677.
American Educational Research Association, American Psychological
Association, & National Council for Measurement in Education.
(1999). Standards for educational and psychological testing.
Washington, D.C., NY: Author.
American Psychological Association. (2003). Principles for the
validation and use of personnel selection procedures (4th ed.).
Washington, D.C., NY: Author.
Anastasi, A. (1986). Evolving concepts of test validation. Annual
Review of Psychology, 37, pp. 1–15.
Anastasi, A. & Urbina, S. (1997). Psychological testing (7th edition).
Upper Saddle River, NJ: Prentice-Hall.
Angoff, W.H. (1993). Perspectives on differential item functioning
methodology. In P.W. Holland & H. Wainer (Eds.), Differential item
functioning (pp. 3–23). Hillside, NJ: Lawrence Erlbaum
Antonovsky, A. (1979). Health, stress and coping. San Francisco:
Jossey-Bass.
Antonovsky, A. (1987). Unravelling the mystery of health: How people
manage stress and stay well. San Francisco: Jossey-Bass.
Antonovsky, A. (1996). The salutogenic model as a theory to guide
health promotion. Health Promotion International, 11(1), pp. 11–
18.
Argyle & Crossland, (1987). Dimensions of positive emotions. British
Journal of Social Psychology, 26, pp. 127–137
Arthur, W., Day, E.A., Mcnelly, T.L. & Edens, P.S. (2003). A meta‐
analysis of the criterion‐related validity of assessment center
dimensions. Personnel Psychology, 56(1), pp. 125−154.
doi:10.1111/j.1744-6570.2003.tb00146.x
Arthur, W., Doverspike, D., Muňoz, J., Taylor, J.E. & Carr, A.E.
(2014). The use of mobile devices in high-stakes remotely
delivered assessments and testing. International Journal of
Selection and Assessment, 22(2), pp. 113−123.
Austin, J.T., & Villanova, P. (1992). The criterion problem: 1917–1992.
Journal of Applied Psychology, 77, pp. 836–874.
Asiwe, D.N., Jorgensen, L.I., & Hill, C. (2014). The development and
investigation of the psychometric properties of a burnout scale
within a South African agricultural research institution. SA Journal
of Industrial Psychology, 40(1), Art. #1194, af pages. doi:
http://dx.doi.org/10.4102/sajip.v40i1.1194
Association of Test Publishers of South Africa and Another, Saville
and Holdsworth South Africa v The Chairperson of the
Professional Board of Psychology, unreported case number
12942/09 North Gauteng High Court, 2010 (p. 1–22). Retrieved on
21 November 2012 from
http://atp.org.za/assets/files/JUDGMENT.PDF.
Aston, S. (2006). A qualitative bias review of the adaptation of the
WAIS-III for English-speaking South Africans. Unpublished
master’s dissertation, Nelson Mandela Metropolitan University.
Retrieved on 13 September 2018 from
http://hdl.handle.net/20.500.11892/159016
ATP (Association of Test Publishers of South Africa) v the President
of SA, the Minister of Labour and the HPCSA, case number 89564
(North Gauteng High Court, May 2, 2017).
Badenhorst, F.H. (1986). Die bruikbaarheid van die Senior Suid-
Afrikaanse Indivduale Skaal vir die evaluering van Blanke
Afrikaanssprekende, hardhorende kinders (The suitability of the
Senior South African Indivdual Scale for evaluating white
Afrikaans-speaking children who are hard of hearing). Unpublished
master’s dissertation, University of Stellenbosch.
Bagnato, S.J. & Neisworth, J.T. (1991). Assessment for early
intervention: Best practices for professionals. New York: The
Guilford Press.
Baicker, K., Cutler, D., & Song, Z. (2010). Workplace wellness
programs can generate savings. Health Affairs, 29(2), pp. 304–
311.
Bakker, A. B., & Demerouti, E. (2007). The job demands-resources
model: State of the art. Journal of Managerial Psychology, 22(3),
pp. 309–328.
Bakker, A., & Demerouti, E. (2008). Towards a model of work
engagement. Career Development Journal, 13, pp. 209–223.
Bakker, A.B. & Leiter, M.P. (2010). Work engagement: A handbook of
essential theory and research. Hove: Psychology Press.
Bakker, A.B., & Oerlemans, W.G.M. (2010). Subjective well-being in
organizations. In Cameron, K., & Spreitzer, G. (Eds.). Handbook of
positive organizational scholarship (pp. 1–31). Oxford University
Press.
Baran, S. (1972). Projective Personality Test/Projektiewe Persoonlik-
heidstoets. Johannesburg: National Institute for Personnel
Research.
Barnes, H., Hall, K., Sambub, W., Wright, G. & Zembe-Mkabile, W.
(2017). Review of research evidence on child poverty in South
Africa. Children’s Institute, University of Cape Town. Retrieved on
13 September 2018 from
http://www.ci.uct.ac.za/sites/default/files/image_tool/images/367/pu
Baron, E. C., Davies, T., & Lund, C. (2017). Validation of the 10-item
Centre for Epidemiological Studies Depression Scale (CES-D-10) in
Zulu, Xhosa and Afrikaans populations in South Africa. BMC
Psychiatry, 17(1), 6.
Bar-on, R. (1997). BarOn Emotional Quotient Inventory Technical
Manual. Toronto: Multi-Health Systems Inc.
Barry, C.L., Horst, S.J., Finney, S.J., Brown, A.R., & Kopp, J.P.
(2010). Do examinees have similar test-taking effort? A high-
stakes question for low-stakes testing. International Journal of
Testing, 10(4), pp. 342–363.
Bartram, D. (2003). Policy on the design of test-related products for
executive users (Version 1.0). London: SHL.
Bartram, D. (2004). Assessment in organisations. Applied Psychology:
An International Review, 53(2), pp. 237−259.
Bartram, D. (2008a). Global versus local personality norms: Logical
and methodological issues. Paper presented at the 6th Conference
of the International Test Commission, Liverpool, UK, 14–16 July
2008. (Can be accessed online at www.intestcom.org).
Bartram, D. (2008b). An ISO standard for assessment in work and
organizational settings. In J. Bogg (Ed.), Testing International, 20,
December 2008, pp. 9–10. (Can be accessed online at
www.intestcom.org).
Bartram, D. (2012). Concluding thoughts on the internationalization of
test reviews. International Journal of Psychological Testing, 12, pp.
195–201.
Bartram, D. & Coyne, I. (1998). The ITC/EFPPA survey of testing and
test use in countries world-wide: Narrative report. Unpublished
manuscript, University of Hull, United Kingdom.
Bass, C. & Fleury, D. (2010). The (not so) hidden costs of member
absences. Slide presentation at the meeting of Global Benefits
Outsourcing Conference 2011, Mercer.
Bayley, N. (1969). Manual for the Bayley Scales of Infant
Development. New York: The Psychological Corporation.
Beck, A.T., Steer, R.A. & Brown, G.K. (1996). BDI II Manual. Boston:
The Psychological Corporation.
Becker, J. (2013). Preliminary psychometric investigation of the
Minnesota Multiphasic Personality Inventory-2TM in the South
African context. Johanneburg, South Africa: JvR Psychometrics.
Beckham, J., Crawford, A., & Feldman, M. (1998). Trail Making Test
performance in Vietnam combat veterans with and without post
traumatic stress disorder. Journal of Traumatic Stress, 11(4), pp.
811–820.
Beery, K.M., Buktenica, N.A., & Beery, N.A. (2005). Beery™-
Buktenica Developmental Test of Visual-Motor Integration
(Beery™-VMI) (5th edition). Florida: Psychological Assessment
Resources.
Bekhet, A.K., Zauszniewski, J.A. & Nakhla, W.E. (2008). Happiness:
theoretical and empirical considerations. Nursing Forum, 43(1), pp.
12–24.
Bekwa, N.N. (2016). The development and evaluation of Africanised
items for multicultural cognitive assessment. Unpublished doctoral
thesis, University of South Africa.
Bekwa, N.N., & De Beer, M. (2012). Multicultural cognitive
assessment: Exploring the use of African artefacts as inspiration
for the development of new items. Poster presented at the 21st
International Association for Cross-cultural Psychology, 17–21
July, 2012, Stellenbosch, South Africa.
Berry, J.W. 2000. Cross-cultural psychology: A symbiosis of cultural
and comparative approaches. Asian Journal of Social Psychology,
3, pp. 197–205. doi: 10.1111%2F1467-839X.00064
Berry, J. W., Poortinga, Y. H., Segall, M. H., & Dasen, P.R. 2002.
Cross-cultural psychology: Research and applications 2nd ed.
Cambridge, MA: Cambridge University Press.
Berry, C.M., Sackett, P.R. & Wiemann, S. (2007). A review of recent
developments in integrity test research. Personnel Psychology,
60(2), pp. 271−301.
Bertelsmann, E. (2010, May 19). Association of Test Publishers of
South Africa and Saville and Holdsworth, case number 12942/09,
Pretoria, Gauteng, Republic of South Africa: North Gauteng High
Court.
Bertua, C., Anderson, N., & Salgado, J.F. (2005). The predictive
validity of cognitive ability tests: A UK meta analysis. Journal of
Occupational and Organizational Psychology, 78, pp. 387–409.
Biddle Consulting Group, Inc. (2005). Manual for Test Analysis and
Validation Program (TVAP) v4.1. Folsom, CA: Author.
Biesheuvel, S. (1943). African intelligence. Johannesburg: South
African Institute of Race Relations.
Biesheuvel, S. (1949). Psychological tests and their application to
Non-European peoples. In The year book of education, pp. 87–
126. London: Evans.
Biesheuvel, S. (1952). Personnel selection tests for Africans. South
African Journal of Science, 49, pp. 3–12.
Birkenstock, C. (2017). Gender differences in work-family conflict as a
predictor of well-being in the transport sector. Unpublished
master’s thesis. University of Johannesburg, Johannesburg.
Blanchflower, D.G. & Oswald, A.J. (2002). Well-being over time in
Britain and the USA. Journal of Public Economics, 88, pp. 1359–
1386.
Bloom, S. (2008, August). Employee wellness programs, Professional
Safety, pp. 41–42.
Bond, T. G., & Fox, C.M. (2007). Applying the Rasch Model:
Fundamental measurement in the human sciences. Mahwah:
Lawrence Erlbaum Associates.
Bosman, J., Rothmann, S., & Buitendach, J.H. (2005). Job insecurity,
burnout and work engagement: The impact of positive and
negative affectivity. SA Journal of Industrial Psychology, 31(4), pp.
48–56.
Bostwick, J. (1994). Neuropsychiatry of depression. In J. Ellison, C.
Weinstein, & T. Hodel- Malinofsky (Eds.), The psychotherapist’s
guide to neuropsychiatry, pp. 409–432. Washington, DC: American
Psychiatric Press.
Botha, C. & Pienaar, J. (2006). South African correctional occupational
stress: The role of psychological strengths. Journal of Criminal
Justice, 34, pp. 73–84.
Bracken, B.A. & Barona, A. (1991). State of the art procedures for
translating, validating and using psychoeducational tests in cross-
cultural assessment. School Psychology International, 12, pp.
119–132.
Brescia, W. & Fortune, J.C. (1989). Standardized testing of American
Indian students. College Student Journal, 23, pp. 98–104.
Brislin, R.W. (1986). The wording and translation of research
instruments. In W.J. Lonner & J.W. Berry (Eds.), Field methods in
cross-cultural psychology, pp. 137–164. Newbury Park, CA: Sage
Publishers.
Brody, N. (2003a). Construct validation of the Sternberg Triarchic
Abilities Test: comment and reanalysis. Intelligence, 31, pp. 319–
329.
Brody, N. (2003b). What Sternberg should have concluded.
Intelligence, 31, pp. 339–342.
Brooks-Gunn, J. (1990). Identifying the vulnerable young children. In
D.E. Rogers and E. Ginsberg (Eds.), Improving the life chances of
children at risk, pp. 104–124. Boulder, CO: Westview Press.
Brown, A., & Bartram, D. (2018). The Occupational Personality
Questionnaire revolution: Applying Item Response Theory to
questionnaire design and scoring. Corporate Executive Board.
Retrieved on 13 September 2018 from
https://www.cebglobal.com/content/dam/cebglobal/us/EN/regions/uk
personality-quesionnaire-white-paper-uk.pdf
Budoff, M. (1987). Measures for assessing learning potential. In C.S.
Lidz (Ed.). Dynamic assessment: An interactional approach to
evaluating learning potential, pp. 173–195. New York: Guilford
Press.
Bunderson, C.B., Inouye, D.K., & Olsen, J.B. (1989). The four
generations of computerized educational measurement. In R.L.
Linn (Ed.). Educational measurement (3rd edition). New York:
Macmillan.
Burke, A. (2008). Could anxiety, hopelessness and health locus of
control contribute to the outcome of a kidney transplant? SA
Journal of Psychology, 38(3), pp. 527–540.
Burke, E. (2006). Better practice for unsupervised online assessment.
SHL White Paper. London: SHL Group. (www.shl.com)
Busbin, J.W. & Campbell, D.P. (1990). Employee wellness programs:
A strategy for increasing participation. Journal of Health Care
Marketing, 10(4), pp. 22–30.
Busichio, K., Tiersky, L., DeLuca, J., & Natelson, B. (2004).
Neuropsychological deficits in patients with chronic fatigue
syndrome. Journal of the International Neuropsychological Society,
10(2), pp. 278–285.
Busseri, M.A., Sadavan, S.W. & DeCourville, N. (2007). A hybrid
model for research on subjective well-being: Examining common –
and component-specific sources of variance in life satisfaction,
positive affect and negative affect. Social Indicators Research,
83(3), pp. 413–445.
Butcher, J.N., Hass, G.A. & Paulsen, J.A. (2016). Clinical
assessments in international settings. In F.T. Leong, D. Bartram,
F.M. Cheung, K.F. Geisinger, & D Iliescu. (Eds.), The ITC
international handbook of testing and assessment (pp. 217−230).
New York, NY: Oxford University Press.
Byrne, B.M., Oakland, T., Leong, F.T., Van de Vijver, F.J.R.,
Hambleton, R.K., Cheung, F.M., & Bartram, D. (2009). A critical
analysis of cross-cultural research and testing practices:
Implications for improved education and training in psychology.
Training and Education in Professional Psychology, 3(2), pp. 94–
105. doi: 10.1037/a0014516
Caffrey, E., Fuchs, D., & Fuchs, L.S. (2008). The predictive validity of
dynamic assessment. The Journal of Special Education, 41(4), pp.
254–270.
Campbell, D.P. & Fiske, D.W. (1959). Convergent and discriminant
validation by the multitrait-multimethod matrix. Psychological
Bulletin, 56, pp. 81–105.
Campion, M. C., Campion, M. A., Campion, E. D., & Reider, M.H.
(2016). Initial investigation into computer scoring of candidate
essays for personnel selection. Journal of Applied Psychology,
101(7), pp. 958−975.
Campione, J.C. & Brown, A.L. (1987). Linking dynamic assessment
with school achievement. In C.S. Lidz (Ed.). Dynamic assessment:
An interactional approach to evaluating learning potential, pp. 82–
115. New York: Guilford Press.
Carroll, J.B. (1997). Psychometrics, intelligence and public perception.
Intelligence, 24(1), pp. 25–52.
Carvalho, A. 2005. Meaning of work and life role salience in a South
African context: A cross-cultural comparison. Unpublished Masters
dissertation, University of Johannesburg, Johannesburg.
Cascio, W.F. (2006). Safety, health and employee assistance
programs. (7th Ed.). In Managing Human Resources: Productivity,
quality of work life, profits (pp.584–623). Boston: McGraw-Hill.
Cascio, W.F. & Aguinis, H. (2005). Applied Psychology in Human
Resource Management. Upper Saddle River, New Jersey:
Pearson.
Casillas, A. & Robbins, S.B. (2005). Test adaptation and cross-cultural
assessment from a business perspective: Issues and
recommendations. International Journal of Testing, 5(1), pp. 5–21.
Cattell, R.B. & Cattell, M.D. (1969). The High School Personality
Questionnaire. Champaign, IL: Institute for Personality and Ability
Testing.
Cattell, R.B., Cattell, A.K., & Cattell, H.E. (1993). The Sixteen
Personality Factor Questionnaire, Fifth Edition. Champaign, IL:
Institute for Personality and Ability Testing, Inc.
Cattell, R.B., Eber, H.W., & Tatsuoka, M.M. (1970). Handbook for the
Sixteen Personality Factor Questionnaire. Champaign, IL: Institute
for Personality and Ability Testing.
Cattell, R.B., Saunders, D.R. & Stice, G.F. (1950). The 16 personality
factor questionnaire. Champaign, Il: Institute for Personality and
Ability Testing.
Cattell, R.B. & Scheier, I.H. (1961). The meaning and measurement of
neuroticism and anxiety. New York: Ronald Press.
Cattell, R.B., Scheier, I.H., & Madge, E.M. (1968). Manual for the
IPAT Anxiety Scale. Pretoria: HSRC.
CEB. (2006). OPQ32 technical manual. Surrey, UK.
CEB. (2013−2016). Assessment catalogue. Pretoria: CEB Inc.
Chambers, L.W. (1993). The McMasters Health Index Questionnaire:
An Update. In S.R. Walker & R.M. Rosen (Eds), Quality of life
assessment: Key issues of the 1990s. Lancaster, England: Kluwer
Academic Publishers.
Charlesworth, W.R. (1992). Darwin and Developmental Psychology:
Past and Present. Developmental Psychology, 28(1), pp. 5–6.
Cheung, F. M., & Cheung, S.F. (2003). Measuring personality and
values across cultures: Imported versus indigenous measures. In
W.J. Lonner, D.L. Dinnel, S.A. Hayes, & D.N. Sattler (Eds.), Online
Readings in Psychology and Culture (Unit 6, Chapter 5),
(http://www.wwu.edu/~culture), Center for Cross- Cultural
Research, Western Washington University, Bellingham,
Washington USA.
Cheung, F.M. & Fetvadjiev, V.H. (2016). Indigenous approaches to
testing and assessment. In F.T. Leong, D. Bartram, F.M. Cheung,
K.F. Geisinger, & D. Iliescu. (Eds.), The ITC international
handbook of testing and assessment (pp.333−346). New York, NY:
Oxford University Press.
Cheung, F. M., Leung, K., Fan, R., Song, W. Z., Zhang, J. X., &
Zhang, J.P. (1996). Development of the Chinese Personality
Assessment Inventory (CPAI). Journal of Cross-Cultural
Psychology, 27, pp. 181–199.
Cheung, G. W., & Rensvold, R.B. (1998). Testing measurement
models for factorial invariance: A systematic approach.
Educational and Psychological Measurement, 58, pp. 1017–1034.
Cheung, G. W., & Rensvold, R.B. (1999). Testing factorial invariance
across groups: A reconceptualization and proposed new method.
Journal of Management, 25(1), pp. 1–27.
Cheung, G. W., & Rensvold, R.B. (2000). Assessing extreme and
acquiescence response sets in cross-cultural research using
structural equations modeling. Journal of Cross-Cultural
Psychology, 31(2), pp. 187–212.doi:
10.1177/0022022100031002003.
Cheung, G. W., & Rensvold, R.B. (2002). Evaluating goodness-of-fit
indexes for testing measurement invariance. Structural Equation
Modeling: A Multidisciplinary Journal, 9(2), pp. 233–255. doi:
10.1207/s15328007sem0902_5.
Cheung, F. M., Leung, K., Fan, R., Song, W. Z., Zhang, J. X., &
Zhang, J.P. (1996). Development of the Chinese Personality
Assessment Inventory (CPAI). Journal of Cross-Cultural
Psychology, 27, pp. 181–199.
Cheung, F.M., Van de Vijver, F.J.R., & Leung, F.T.L. (2011). Toward a
new approach to the study of personality in culture. American
Psychologist, 66, pp. 593−603.
Claassen, N.C.W. (1990). The comparability of General Scholastic
Aptitude Test scores across different population groups. SA
Journal of Psychology, 20, pp. 80–92.
Claassen, N.C.W. (1995). Cross-cultural assessment in the human
sciences. Paper presented at a work session on the meaningful
use of psychological and educational tests, HSRC, 24 October
1995.
Claassen, N.C.W. (1996). Paper and Pencil Games (PPG): Manual.
Pretoria: HSRC.
Claassen, N.C.W. (1997). Cultural differences, politics and test bias in
South Africa. European Review of Applied Psychology, 47(4), pp.
297–307.
Claassen, N.C.W., De Beer, M., Ferndale, U., Kotzé, M., Van Niekerk,
H.A., Viljoen, M., & Vosloo, H.N. (1991). General Scholastic
Aptitude Test (GSAT) Senior Series. Pretoria: HSRC.
Claassen, N.C.W., De Beer, M., Hugo, H.L.E., & Meyer, H.M. (1991).
Manual for the General Scholastic Aptitude Test (GSAT). Pretoria:
HSRC.
Claassen, N.C.W. & Hugo, H.L.E. (1993). The relevance of the
general Scholastic Aptitude Test (GSAT) for pupils who do not
have English as their mother tongue (Report ED-21). Pretoria:
HSRC.
Claassen, N.C.W., Krynauw, A., Paterson, H., & wa ga Mathe, M.
(2001). A standardisation of the WAIS-III for English-speaking
South Africans. Pretoria: HSRC.
Claassen, N.C.W. & Schepers, J.M. (1990). Groepverskille in
akademiese intelligensie verklaar op grond van verskille in sosio-
ekonomiese status (Group differences in academic intelligence
that can be explained by differences in socio-economic status). SA
Journal of Psychology, 20, pp. 294–302.
Claassen, L., Schepers, J.M. & Roodt, G. (2004). Werkwaardes van
akademici [Work values of academics]. SA Journal of Industrial
Psychology, 30(4), pp. 82–92.
Claassen, N.C.W., Van Heerden, J.S., Vosloo, H.N., & Wheeler, J.J.
(2000). Manual for the Differential Aptitude Tests Form R (DAT-R).
Pretoria: HSRC.
Clarke, M.M., Maduas, G.F., Horne, C.L., & Ramos, M.A. (2000).
Retrospective on educational testing and assessment in the 20th
century. Journal of Curriculum Studies, 32(2), pp. 159–181.
Cleary, T.A. (1968). Test bias: prediction of grades of Negro and white
students in integrated colleges. Journal of Educational
Measurement, 5, pp. 115–124.
Cloete, P.A. (2015). The impact of an employee wellbeing programme
on return on investment in terms of absenteeism and employee
psychosocial functioning. Unpublished master’s dissertation,
University of Pretoria, South Africa.
Cockcroft, K., Alloway, T., Copello, E., & Milligan, R. (2015). A cross-
cultural comparison between South African and British students on
the Wechsler Adult Intelligence Scales Third Edition (WAIS-III).
Front. Psychol. doi: 10.3389/fpsyg.2015.00297
Cockcroft, K., Amod, Z., & Soellaart, B. (2008). Maternal level of
education and performance on the 1996 Griffiths Mental
Development Scales. African Journal of Psychiatry, 11, pp. 44−50.
Cockcroft, K. & Blackburn, M. (2008). The relationship between Senior
South African Individual Scale – Revised (SSAIS-R) subtests and
reading ability. SA Journal of Psychology, 38(2), pp. 377–389.
Cockroft, K., Bloch, L. & Moolla, A. (2016). Assessing verbal
functioning in South African school beginners from diverse
socioeconomic backgrounds: A comparison between verbal
working memory measures and vocabulary measures. Education
as Change, 20(1), pp. 199−215. doi:
http://dx.doi.org/10.17159/1947-9417/2016/559
Cockcroft, K. & Israel, N. (2011). The Raven’s Advanced Progressive
Matrices: A comparison of relationships with verbal ability tests. SA
Journal of Psychology, 41(3), pp. 363–372.
Coetzee, C.J.H., Fourie, L., & Roodt, G. (2002). The development and
validation of the communication-for-change questionnaire. SA
Journal of Industrial Psychology, 28(3), pp. 16–25.
Coetzee, N. & Vosloo, H.N. (2000). Manual for the Differential
Aptitude Tests Form K (DAT-K). Pretoria: HSRC.
Coetzer, W. J., & Rothmann, S. (2004, August). A model of work
wellness of employees in an insurance company. Poster presented
at 28th International Congress of Psychology, Beijing, China.
Cohen, R.J., & Swerdlik, M.E. (2010). Psychological testing and
assessment: An introduction to tests and measurement (7th ed.).
Boston: McGraw-Hill.
Colby, K. M., Watt, J.B. & Gilbert, J.P. (1966). A Computer method of
psychotherapy. The Journal of Nervous and Mental Disease,
142(2), pp. 148−152.
Conn, S.R., & Rieke, M.L. (1998). The 16PF Fifth Edition Technical
Manual (2nd ed.). Champaign, IL: Institute for Personality and
Ability Testing, Inc.
Coons, C.E., Frankenburg, W.K., Gay, E. C., Fandal, A.W., Lefly, D.L.
& Ker, C. (1982). Preliminary results of a combined
developmental/environmental screening project. In N.J.
Anastasiow, W.K. Frankenburg & A. Fandal (Eds.), Identifying the
developmentally delayed child, pp. 101–110. Baltimore: University
Park Press.
Cooper, C.L. (2011). Organizational health and wellbeing. Los
Angeles: Sage.
Cormier, D.C., Bulut, O., McGrew, K.S. & Singh, D. (2017). Exploring
the relations between Catell-Horn-Carroll (CHC) cognitive abilities
and mathematics achievement. Applied Cognitive Psychology, 31,
pp. 530−538.
Corrigall, J., Ward, C., Stinson, K., Struthers, P., Frantz, J., Lund, C.,
Flisher, A.J., & Joska, J. (2007). Decreasing the burden of mental
illness. Western Cape burden of disease reduction project, 4, pp.
1–19.
Cotton, G, (2004). Health in the workplace. Occupational Health,
56(12), pp. 2–6.
Coyne, I. & Bartram, D. (2002). Assessing the effectiveness of
integrity tests: A review. International Journal of Testing, 2(1), pp.
15–34.
Crites, J.O. (1981). Career counselling: Models, methods and
materials. New York: McGraw Hill.
Cronbach, L.J. (1951). Coefficient Alpha and the internal structure of
tests. Psychometrika, 16 (3): pp. 297–334.
doi:10.1007/bf02310555.
Cronbach, L. (1970). Psychological testing (3rd ed.). New York:
Harper & Row.
Cronbach, L.J. (1990). Essentials of Psychological Testing (5th ed.).
New York: Harper & Row.
Cronbach, L.J. (2004). My current thoughts on Coefficient Alpha and
successor procedures. Educational and Psychological
Measurement, 64, pp. 391–418.
Cronbach, L. J., & Meehl, P.E. (1955). Construct validity in
psychological tests. Psychological Bulletin, 52(4), pp. 281–302.
Csapó, B., Molnár, G. & Nagy, J. (2014). Computer-based
assessment of school readiness and early reasoning. Journal of
Educational Psychology, 106(3), pp. 639−650.
Csikszentmihalyi, M. & Csikszentmihalyi, I.S. (2006). A life worth
living: Contributions to positive psychology. New York: Oxford
University Press.
Cui, Y. & Guo, Q. (2016, July). Statistical classification for cognitive
diagnostic assessment: An artificial neural network approach.
Educational Psychology, 36(6), pp. 1065−1082.
Dannhauser, Z. & Roodt, G. (2001). Value disciplines: Measuring
customer preferences. Journal of Industrial Psychology, 27(1), pp.
8–16.
Dai, B., Zhang, M. & Li, G. (2016). Exploration of item selection in
dual-purpose cognitive diagnostic computerized adaptive testing:
Based on the RRUM. Applied Psychological Measurement, 40(8),
pp. 625−640.
Das, J.P., Naglieri, J.A., & Kirby, J.R. (1994). Assessment of cognitive
process: The PASS theory of intelligence. Boston: Allyn and
Bacon.
Davey, D.M. (1989). How to be a good judge of character. London:
Kogan Page.
Davidson, C.A., Johannesen, J.K. & Fiszdon, J.M. (2016). Role of
learning potential in cognitive remediation: Construct and
predictive validity. Schizophrenia Research, 171, pp. 117−124.
Dawis, R.V. (2002). Person-environment-correspondence theory. In
D. Brown & Associates (Eds.), Career Choice and Development
(4th ed.; pp. 427–464). San Francisco, CA: Jossey-Bass.
Dawis, R.V., & Lofquist, L.H. (1984). A psychological theory of work
adjustment. Minneapolis, MN: University of Minnesota Press.
De Beer, M. (2000a). Learning potential computerized adaptive test
(LPCAT): User’s manual. Pretoria: Unisa Press
De Beer, M. (2000b). The construction and evaluation of a dynamic,
computerised adaptive test for the measurement of learning
potential. Unpublished D. Litt et Phil dissertation, University of
South Africa.
De Beer, M. (2004). Use of differential item functioning (DIF) analysis
for bias analysis in test construction. SA Journal of Industrial
Psychology, 30(4), pp. 52–58.
De Beer, M. (2005). Development of the Learning Potential
Computerized Adaptive Test (LPCAT). SA Journal of Psychology,
35(4), pp. 717–747.
De Beer, M. (2006). Dynamic testing: Practical solutions to some
concerns. SA Journal of Industrial Psychology, 32(4), pp. 8–14.
De Beer, M. (2010). A modern psychometric approach to dynamic
assessment. Journal of Psychology in Africa, 20(2), pp. 241–246.
De Beer, M. (2013). The Learning Potential Computerised Adaptive
Test in South Africa. In S. Laher & K. Cockcroft (Eds.),
Psychological assessment in South Africa: Research and
applications (pp. 137-157). Johannesburg: Wits University Press.
De Beer, M., Botha, E.J., Bekwa, N.N. & Wolfaardt, J.B. (1999).
Industrial psychological testing and assessment (Study guide for
IOP301-T). Pretoria: University of South Africa.
De Bod, A. (2011). Contemporary Ethical Challenges in Industrial
Psychology Testing in South Africa. Unpublished M Phil
dissertation, University of Johannesburg, South Africa.
De Bruin, G.P. (1996). An item factor analysis of the Myers-Briggs
Type Indicator (Form G Self-scorable). Paper presented at the
International Type User’s Conference. Sandton, Johannesburg.
De Bruin, G. (2000). An inter-battery factor analysis of the Comrey
personality scales and the 16 Personality Factor questionnaire. SA
Journal of Industrial Psychology, 26(3), pp. 4–7.
doi:10.4102/sajip.v26i3.712
De Bruin, G.P. (2002). The relationship between personality traits and
vocational interests. Journal of Industrial Psychology, 28(1), pp.
49–52.
De Bruin, G.P., & Bernard-Phera, M.J. (2002). Confirmatory factor
analysis of the Career Development Questionnaire and the Career
Decision-making Self-Efficacy Scale for South African high school
students. SA Journal of Industrial Psychology, 28(2), pp. 1–6.
De Bruin, G.P. & De Bruin, K. (2017). Career assessment. In G.B.
Stead & M.B. Watson (Eds.), Career psychology in the South
African context (pp. 185−193). Pretoria, South Africa: Van Schaik.
De Bruin, K., De Bruin, G.P., Dercksen, S., & Hartslief-Cilliers, M.
(2005). Can measures of personality and intelligence predict
performance in an adult basic education training programme? SA
Journal of Psychology, 35, pp. 46–57.
De Bruin, G.P. & Nel, Z.J. (1996). ’n Geïntegreerde oorsig van
empiriese navorsing oor loopbaanvoorligting in Suid-Afrika: 1980–
1990 (An integrated overview of empirical research concerning
career counselling in South Africa: 1980–1990). SA Journal of
Psychology, 26, pp. 248–251.
De Bruin, G. P., & Rudnick, H. (2007). Examining the cheats: The role
of conscientiousness and excitement seeking in academic
dishonesty. SA Journal of Psychology, 37(1), pp. 153–164.
De Bruin, G. P., Schepers, J. M., & Taylor, N. (2005). Construct
validity of the 16PF 5th Edition and the Basic Traits Inventory.
Paper presented at the 8th Annual Industrial Psychology
Conference, Pretoria.
De Bruin, G.P. & Taylor, N. (2005). Development of the Sources of
Work Stress Inventory. SA Journal of Psychology, 35(4), pp. 748–
765.
De Bruin, G. P., & Taylor, N. (2008, July). Block versus mixed-order
item presentation in personality inventories: An IRT analysis.
Paper presented at the XXIX International Congress of
Psychology, Berlin, Germany.
Deci, E.L. & Ryan, R.M. (2008). Hedonia, eudemonia, and well-being:
An introduction. Journal of Happiness Studies, 9, pp. 1–11.
DeLamatre, J. E., & Shuerger, J.M. (2000). The MMPI-2 in counseling
practice. In C.E.J. Watkins (Ed.), Testing and assessment in
counseling practice (2nd ed., pp. 15–44). Mahwah, NJ: Erlbaum.
Delgado, M.T., Uribe, P.A., Alonso, A.A. & Diaz, R.R. (2016). TENI: A
comprehensive battery for cognitive assessment based on games
and technology. Child Neuropsychology, 22(3), pp. 276−291.
Demerouti, E., & Bakker, A.B. (2011). The job demands-resources
model: Challenges for future research. Journal of Industrial
Psychology, 37(2), pp. 1–9. doi: 10.4102/sajip.v37i2.974.
Demerouti, E., Bakker, A.B., Nachreiner, F., & Schaufeli, W.B. (2001).
The Job demands-resources model of burnout. Journal of Applied
Psychology, 86(3), pp. 499–512.
Department of Education (2001). Education White Paper 6. Special
needs education: Building an inclusive education and training
system. Pretoria: Department of Education.
Department of Education (2005). Conceptual and operational
guidelines for the implementation of inclusive education: District-
based support teams. Pretoria: Department of Education.
Department of Health (2006). Ethical Rules of Conduct for
Professionals Registered under the Health Professions Act, 1974.
Government Gazette, No. 29079, 4 August 2006.
Department of Health (2008). Regulations Defining the Scope of the
Profession of Psychology. Government Gazette, No. 31433, 16
September 2008. Retrieved 30 November 2012 from
http://www.hpcsa.co.za/downloads/psychology/scope_of_profession
Department of Health (2011). Regulations Defining the Scope of the
Profession of Psychology. Government Gazette, No. 34581, 2
September 2011. Retrieved 30 November 2012 from
http://www.hpcsa.co.za/downloads/psychology/promulgated_scope
Department of Justice (2000) Promotion of Access to Information Act
No. 2 of 2000. Republic of South Africa. Available at
http://www.justice.gov.za/legislation/acts/2000-002.pdf Department
of Justice (2013). Protection of Personal Information Act
No. 4 of 2013. Republic of South Africa. Available at
www.justice.gov.za/legislation/acts/2013-004.pdf
Department of Labour of the Republic of South Africa (2014, January
16). Employment Equity Amendment Act. Government Gazette,
583(37238).
De Ridder, J.C. (1961). The personality of the urban African in South
Africa. New York: Humanities Press.
Desai, F. (2010). The relationship between personality traits and team
culture. Unpublished master’s dissertation. Pretoria, South Africa:
University of South Africa.
De Sousa, D.S., Greenop, K., & Fry, J. (2010). Simultaneous and
sequential cognitive processing in monolingual and bilingual
children in South Africa. SA Journal of Psychology, 40(2), pp. 165–
173.
Dickman, B.J. (1994). An evaluation of the Grover-Counter Test for
use in the assessment of black South African township children
with mental handicaps. Unpublished Doctoral dissertation,
University of Cape Town.
Diener, Ed. (2009). Assessing well-being: The collected works of Ed
Diener. New York: Springer.
Diener, E. & Emmons, R.A. (1984). The independence of positive and
negative effect. Journal of Personality and Social Psychology,
47(5), pp. 1105–1117.
Diener, E., Emmons, R. A., Larsen, R. J., & Griffin, S. (1985). The
Satisfaction With Life Scale. Journal of Personality Assessment,
49, pp. 71–75.
Diener, E., Kesebir, P. & Lucas, R. (2008). Benefits of accounts of
well-being – for societies and for psychological science. Applied
psychology: An international review, 57, pp. 37–53.
Diener, E., Sapyta, J.J. & Suh, E. (1998). Subjective well-being is
essential to well-being. Psychological Inquiry, 9(1), pp. 33–38.
Diener, E., Scollon, C.N. & Lucas, R.E. (2009). The evolving concept
of subjective well-being: The multifaceted nature of happiness. In
E. Diener (ed.), Social indicators research series: Vol. 39.
Assessing well-being: The collected works of Ed Diener (pp.
67−100). New York, NY, US: Springer Science + Business Media.
Dierendonck, D., Diaz, D., Rodriguez-Carvajal, R., Blanco, A. &
Moreno-Jiménez, B. (2008). Ryff’s six-factor model of
psychological well-being. Social Indicators Research, 87(3), pp.
473–479.
Di Fabio, A. & Maree, J.G. (2013). Career counselling: The usefulness
of the Career Interest Profile. Journal of Psychology in Africa,
23(1), pp. 41−49. doi: 10.1080/14330237.2013.10820592
Domino, G. & Domino, M.L. (2006). Psychological Testing: An
introduction. 2nd ed. Cambridge, New York: Cambridge University
Press.
Donadello, I., Spoto, A., Badaloni, S., Granziol, U. & Vidotto, G.
(2017). ATS-PD: An adaptive testing system for psychological
disorders. Educational and Psychological Measurement, 77(5), pp.
792−815.
Dowdeswell, K.& Joubert, T. (2014). Cheating in online unsupervised
assessment & verification trends across Africa. Paper presented at
the 16th annual conference of the Society for Industrial &
Organisational Psychology of South Africa, Pretoria.
Drozdick, L.W., Getz, K., Raiford, S. E., & Zhang, O. (2016). WPPSI-
IV: Equivalence of Q-interactive and paper formats. Q-interactive
Technical Report 14. Bloomington, MN: Pearson. Retrieved 3 May
2018 from
http://www.helloq.com/content/dam/ped/ani/us/helloq/media/WPPSI
Qi-Tech-Report-14-FNL.pdf
Dunn, L.M. & Dunn, L.M. (1981). Peabody Picture Vocabulary Test –
Revised. Circle Pines, MN: American Guidance Service.
Du Toit, H.A. (1986). ’n Voorlopige standaardisering van die SAIS
(Nie-verbaal) vir gehoorgestremdes (A preliminary standardisation
of the SSAIS (Non-verbal) for the hearing disabled). Unpublished
master’s disertation, University of Stellenbosch.
Du Toit, R. & De Bruin, G.P. (2002). The structural validity of Holland’s
R-I-A-S-E-C model of vocational personality types for young black
South African men and women. Journal of Career Assessment, 10,
pp. 62–77.
Du Toit, R. & Gevers, J. (1990). Die Self- ondersoekvraelys ten
opsigte van beroepsbelangstelling (SOV): Handleiding (The self-
examination questionnaire regarding career interest: Handbook).
Pretoria: Raad vir Geesteswetenskaplike Navorsing (HSRC).
Du Toit, W. & Roodt, G. (2003). The discriminant validity of the Culture
Assessment Instrument: A comparison of company cultures. SA
Journal of Human Resource Management, 1(1), pp. 77–84.
Eaton, W. W., Smith, C., Ybarra, M., Muntaner, C., & Tien, A. (2004).
Center for Epidemiologic Studies Depression Scale: Review and
revision (CESD and CESD-R). In M. E Maruish (Ed.), The use of
psychological testing for treatment planning and outcomes
assessment. Volume 3: Instruments for Adults (3rd ed.; pp.
363−377). Mahwah, NJ: Lawrence Erlbaum.
Edwards, S. D., & Edwards, D.J. (2009). Counselling for psychological
well-being in persons living with HIV/Aids. Journal of Psychology in
Africa, 19(4), pp. 561−563.
Ellis, B.B. (1989). Differential item functioning: Implications for test
translation. Journal of Applied Psychology, 74, pp. 912–921.
Ellis, B.B. (1991). Item Response Theory: A tool for assessing the
equivalence of translated tests. Bulletin of the International Test
Commission, 18, pp. 33–51.
Ellis, L.L. (2007). The impact of HIV/Aids on selected business sectors
in South Africa. Studies in Economics and Econometrics, 31(1),
pp. 29−52.
Embretson, S. E., & Reise, S.P. (2000). Item Response Theory for
psychologists. Mahway: Lawrence Erlbaum Associates.
Employment Equity Act, number 55 (1998). Government Gazette, 400
(19370), Cape Town, 19 October.
Encouraging corporate wellness maintenance (29 August 2007).
Retrieved July 15, 2008, from
http://www.bizcommunity.com/Article/196/141/17623.html
England, J. & Zietsman, G. (1995). The HSRC stakeholder survey on
assessment. Internal report prepared for Human Resources:
Assessment and Information Technology, HSRC, Pretoria, South
Africa.
Epic Games Inc. (2017). Unreal engine. Retrieved on 13 September
2018 from: https://unrealengine.com/en-US/blog
Erasmus, P.F. (1975). TAT-Z Catalogue No. 1676. Pretoria: HSRC.
Evers, A. (2012). The internationalization of test reviewing: Trends,
differences, and results. International Journal of Testing, 12, pp.
136–156.
Evers, A., Sijtsma, K., Lucassen, W. & Meijer.R. (2010). The Dutch
review process for evaluating the quality of psychological tests:
History, procedure, and results. International Journal of Testing,
10, pp. 295–317.
Exner, J.E. (2003). The Rorschach: a comprehensive system: Basic
foundations and principles of interpretation. Vol 1 (4th ed.).
Hoboken, NJ: John Wiley & Sons.
Eysenck, H.J. (1988). The concept of ‘intelligence’: Useful or useless.
Intelligence, 12, pp. 1–16.
Eysenck, H.J. & Eysenck, M.W. (1958). Personality and individual
differences. New York: Plenum Publishers.
Eysenck, H.J. & Kamin, L. (1981). The intelligence controversy. New
York: John Wiley and Sons.
Fagbola, T.M., Adigun, A.A. & Oke, A.O. (2013). Computer-based test
(CBT) system for university academic enterprise examination.
International Journal of Scientific and Technology Research, 2(8),
pp. 336−342.
Fancher, R.E. (1985). The intelligence men: Makers of the IQ
controversy. New York: W.W. Norton & Company.
Feldman, F. (2010). What is this thing called happiness? New York:
Oxford University Press.
Ferrett, H. C., Carey, P. D., Thomas, K.G. F., Tapert, S. F., & Fein, G.
(2010). Neuropsychological performance of South African
treatment-naïve adolescents with alcohol dependence. Drug and
Alcohol Dependence, 110, pp. 8–14.
Fetvadjiev, V.H., Meiring, D., Van de Vijver, F.J., Nel, J.A. & Hill, C.
(2015). The South African personality inventory (SAPI): A culture-
informed instrument for the country’s main ethnocultural groups.
Psychological Assessment, 27(3), pp. 827−837.
Feuerstein, R., Rand, Y., Jensen, M.R., Kaniel, S., & Tzuriel, D.
(1987). Prerequisites for assessment of learning potential: The
LPAD model. In C.S. Lidz (Ed.), Dynamic assessment: An
interactive approach to evaluating learning potential, (pp. 35–51).
New York: Guilford Press.
Fick, M.L. (1929). Intelligence test results of poor white, native (Zulu),
coloured and Indian school children and the educational and social
implications. South African Journal of Science, 26, pp. 904–920.
Finkelman, M. D., Lowe, S. R., Kim, W., Gruebner, O., Smits, N., &
Galea, S. (2017). Customized computer-based administration of
the PCL-5 for the efficient assessment of PTSD: A proof-of-
principle study. Psychological Trauma: Theory, Research,
Practice, and Policy, 9(3), pp. 379−389.
Fischer, R., & Fontaine, J.R.J. (2011). Methods for investigating
structural equivalence. In D. Matsumoto & F.J.R. Van de Vijver,
Cross-cultural Research Methods in Psychology (pp. 179–215).
Cambridge, NY: Cambridge University Press.
Fisher, K.N. & Broom, Y.M. (2006). Computerised neuropsychological
testing in South Africa. Paper presented at the 10th national
conference of the South African Clinical Neuropsychological
Association, Cape Town, South Africa.
Folstein, M.F., Folstein, S.E. and McHugh, P.R. (1975). Mini-mental
state. Journal of Psychiatric Research, 12, pp. 189–195.
Fontaine, J. (2005). Equivalence. In Encyclopedia of social
measurement (Vol. 1, pp. 803–813). New York, NY: Elsevier.
Foster, D.F. (2010). World-wide testing and test security issues:
Ethical challenges and solutions. Ethics & Behavior, 20(3–4), pp.
207–228.
Foster, W.H. & Vaughan, R.D. (2005). Absenteeism and business
costs: (does), substance abuse matter? Journal of Substance
Abuse Treatment, 28(1), pp. 27.
Foulke, J. & Sherman, B. (2005). Comprehensive workforce health
management – not a cost, but a strategic advantage. Employment
Relations Today, 32(2), pp. 17–29.
Fourie, L., Rothmann, S., & Van de Vijver, F.J.R. (2008). A model of
work-related well-being for non-professional counsellors in South
Africa. Stress & Health, 24, pp. 35–47.
Foxcroft, C.D. (1997a). Updated policy document on psychological
assessment: The challenges facing us in the new South Africa.
Unpublished manuscript, University of Port Elizabeth.
Foxcroft, C.D. (1997b). Psychological testing in South Africa:
Perspectives regarding ethical and fair practices. European
Journal of Psychological Assessment, 13(3), pp. 229–235.
Foxcroft, C.D. (2004a). Evaluating the school-leaving examination
against measurement principles and methods: From the
Matriculation examination to the FETC. Paper presented at the
Matric Colloquium of the HSRC, Pretoria, November 2004.
Foxcroft, C.D. (2004b). Computer-based testing: Assessment
advances and good practice guidelines for practitioners. Workshop
presented at the 10th national congress of the
Psychological Society of South Africa, Durban, September 2005.
Foxcroft, C.D. (2004c). Planning a Psychological Test in the
Multicultural South African Context. SA Journal of Industrial
Psychology, 30(4), pp. 8–15. Available at:
https://www.google.co.za/url?
sa=t&rct=j&q=&esrc=s&source=web&cd=1&ved=2ahUKEwiP4dqi_u
-SUN2Jj
Foxcroft, C.D. (2005). Benchmark tests: What, why and how. Pretoria:
SAUVCA.
Foxcroft, C.D. (2008). NMMU access testing: Philosophy, principles,
and process. Paper presented at a symposium on admissions
testing at the University of Johannesburg, April 2008.
Foxcroft, C.D. (2011a). Ethical issues related to psychological testing
in Africa: What I have learned (so far). In W.J. Lonner, D.L. Dinnel,
S.A. Hayes, & D.N. Sattler (Eds.), Online Readings in Psychology
and Culture (Unit 5, Chapter 4), Center for Cross-Cultural
Research, Western Washington University, Bellingham,
Washington USA. Retrieved from
http://scholarworks.gvsu.edu/orpc/vol2/iss2/7.
Foxcroft, C.D. (2011b). Some issues in assessment in a developing
world context: An African perspective. Invited stimulus paper for an
internet-based symposium/forum on Issues in Assessment and
Evaluation as part of the ASEASEA International conference held
between 11 and 18 July 2011 and hosted by the Wits Programme
Evaluation Group.
http://wpeg.wits.ac.za/components/com_agora/img/members/3/Fox
paperandintro.pdf
Foxcroft, C.D. (2017). Ethical issues in conducting child development
research in Sub-Saharan Africa. In A. Abubakar & F. Van de Vijver
(Eds.), Handbook of applied developmental science in Sub-
Saharan Africa. New York: Springer-Verlag.
Foxcroft, C.D. & Aston, S. (2006). Critically examining language bias
in the South African adaptation of the WAIS-III. SA Journal of
Industrial Psychology, 32(4), pp. 97–102.
Foxcroft, C.D. & Davies, C. (2008). Historical perspectives on
psychometric testing in South Africa. In van Ommen, C. and
Painter, D. (Eds.), Interiors: A history of psychology in South Africa
(pp. 152–181). Pretoria: University of South Africa Press.
Foxcroft, C.D. & De Bruin, G.P. (2008). Facing the challenge of the
language of assessment and test adaptation. Paper presented at
the XXIX International Congress of Psychology, Berlin, Germany,
July 2008.
Foxcroft, C.D., Paterson, H., Le Roux, N., & Herbst, D. (2004).
Psychological assessment in South Africa: A needs analysis.
Pretoria: HSRC.
Foxcroft, C.D. & Shillington, S.J. (1996). Sharing psychology with
teachers: A South African perspective. Paper presented at the
26th International Congress of Psychology, Montreal, Canada, 16–
21 August 1996.
Foxcroft, C., Watson, M.B., de Jager, A., & Dobson, Y. (1994).
Construct validity of the Life Role Inventory for a sample of
university students. Paper presented at the National Conference of
the Student Counsellors Society of South Africa. Cape Town.
Foxcroft, C.D., Watson, A.S.R., and Seymour, B.B. (2004). Personal
and situational factors impacting on CBT practices in developing
countries. Paper presented at the 28th International Congress of
Psychology, Beijing, China, 8–13 August 2004.
Frankenburg, W.K. & Dodds, J.B. (1990). Denver II: Screening
Manual. Denver, Colorado: Denver Developmental Materials.
Frankenburg, W.K., Dodds, J.B., Archer, P., Bresnick, B., Maschka,
P., Edelman, N., & Shapiro, H. (1990). Denver II: Technical
Manual. Denver, Colorado: Denver Developmental Materials.
Fremer, J. (1997). Test-taker rights and responsibilities. Paper
presented at the Michigan School Testing Conference, Ann Arbor,
Michigan, USA, February 1997.
Fritz, E. (2016). Flowing between assessment and therapy through
creative expressive arts. In R. Ferreira (Ed.). Psychological
assessment: Thinking innovatively in contexts of diversity (pp.
103−209). Cape Town: Juta and Company
Furnham, A., Ndlovu, N. & Mkhize, N. (2009). South African Zulus’
beliefs about their own and their children’s intelligence. South
African Journal of Psychology, 39(2), pp. 157−168. doi:
10.1177/008124630903900202.
Gadd, C. & Phipps, W.D. (2012). A preliminary standardisation of the
Wisconsin Card Sorting Test for Setswana-speaking university
students: Cognitive science and neuropsychology. SA Journal of
Psychology, 42(3), pp. 389–398.
Gallagher, A., Bridgeman, B., & Cahalan, C. (2000). The effect of
computer-based tests on racial/ethnic, gender and language
groups. GRE Board Professional Report No. 96–21P. Princeton,
NJ: Educational Testing Service.
Garcia Coll, C. & Magnuson, K. (2000). Cultural differences as
sources of developmental vulnerabilities and resources. In J.
Shonkoff & S. Meisels (Eds), Handbook of Early Childhood
Intervention (2nd edition), pp. 94–114. Cambridge: Cambridge
University Press.
Gardner, H. (1983). Frames of mind: The theory of multiple
intelligence. New York: Basic Books.
Geerlings, H., Glas, C.A.W., & Van der Linden, W.J. (2011). Modeling
rule-based item generation. Psychometrika, 76(2), pp. 337–359.
Geerlings, H., Van der Linden, W.J., & Glas, C.A.W. (2012). Optimal
test design with rule-based item generation. Applied Psychological
Measurement, 37(2), pp. 140−161.
Geisinger, K. & McCormick, C. (2016). Testing individuals with
disabilities. In F.T.L. Leong, D. Bartram. F.M. Cheung, K.F.
Geisinger & Iliescu, D. (Eds.), The ITC international handbook of
testing and assessment (pp. 259−275). New York, NY: Oxford
University Press.
Gelderblom, J.H. (2008). Designing technology for young children:
Guidelines grounded in a literature investigation on child
development and children’s technology. Unpublished doctoral
dissertation, University of South Africa, South Africa.
Gelman, D., Hager, M., Doherty, S., Gosnell, S., Raine, G. & Shapito,
D. (4 May 1987). Depression. Newsweek, pp. 42–48.
Getz, L.M., Chamorro-Premuzic, T., Roy, M.M., & Devroop, K. (2012).
The relationship between affect, uses of music, and music
preferences in a sample of South African adolescents. Psychology
of Music, 40(2), pp. 164–178.
Gibbs, E. (1990). Assessment of infant mental ability. In E.D. Gibbs
and D.M. Teti (Eds), Interdisciplinary assessment of infants: A
guide for early intervention professionals, pp. 7–89. Baltimore, MD:
Paul H Brookes.
Gierl, M.J., Lai, H., Pugh, D., Touchie, C., Boulais, A., &De
Champlain, A. (2016). Evaluating the psychometric characteristics
of generated multiple-choice test items. Applied Measurement in
Education, 29, pp. 196−210.
Gierl, M.J., & Zhou, J. (2008). Computer Adaptive- Attribute Testing: A
new approach to cognitive diagnostic assessment. Zeitschrift für
Psychologie / Journal of Psychology, 216(1), pp. 29–39.
Gierl, M.J., Zhou, J., Alves, C. (2008). Developing a taxonomy of item
model types to promote assessment engineering. Journal of
Technology, Learning, and Assessment,7(2), pp. 4–50. Retrieved
30 November 2012 from http://www.jtla.org.
Gintiliene, G. & Girdzijauskiene, S. (2008) Test development and use
in Lithuania. In J. Bogg (Ed.), Testing International, 20, December
2008, p. 7–8. (Can be accessed at www.intestcom.org).
Goldberg, D.P. (1978). General Health Questionnaire. London: GL
Assessment Limited.
Goldberg, D.P., & Hillier, V.F. (1979). A scaled version of the General
Health Questionnaire. Psychological Medicine, 9(1), pp. 139–145.
Goldberg, L.R., & Johnson, J.A. (2017, February 22). The 3 320 IPIP
items in alphabetical order. Retrieved on 13 September 2018 from
https://ipip.ori.org/AlphabeticalItemList.htm.
Goleman, D. (1995). Emotional intelligence – why it can matter more
than IQ. London: Bloomsbury.
Gordon, L.V. & Alf, E.F. (1960). Acclimatization and aptitude test
performance. Educational and Psychological Measurement, 20,
pp. 333–337.
Gorsuch, R.L. (1983). Factor analysis. Hillsdale, NJ: Lawrence
Erlbaum.
Gottfredson, L.S. (2003). Dissecting practical intelligence theory: Its
claims and evidence. Intelligence, 31, pp. 343–397.
Gould, S.J. (1981). The mismeasure of man. London: The Penguin
Group.
Govender, S.A. (2008). The relationship between personality and
coping amongst members of the South African Police Service.
Unpublished master’s dissertation. Johannesburg, South Africa:
University of South Africa.
Gradidge, D.H. & De Jager, A.C. (2011). Psychometric properties of
the Wellness Questionnaire for Higher Education. SA Journal of
Psychology, 41(4), pp. 517–527.
Graham, J.R. (2006). MMPI-2: Assessing personality and
psychopathology (4th ed.). New York: Oxford University Press.
Gredler, G.R. (1997). Issues in early childhood screening and
assessment. Psychology in the Schools, 34(2), pp. 98–106.
Greenspan, S.I., & Meisels, S.J. (1996). Toward a new vision for the
developmental assessment of infants and young children. In S.J.
Meisels & E Fenichel (Eds.), New Visions for the developmental
assessment of infants and young children, pp. 11–26. Washington,
DC: Zero to Three: National Center for Infants, Toddlers, and
Families.
Gregory, R.J. (2000). Psychological testing. History, principals, and
applications (3rd ed.). Boston: Allyn and Bacon.
Gregory, R.J. (2007). Psychological testing: History, principles and
applications. (5th ed.). Boston: Pearson Education.
Gregory, R.J. (2010). Psychological testing: History, principles and
applications. (6th edition). Boston: Pearson Education.
Griesel, H. (2001). Institutional views and practices: Access,
admissions, selection and placement. In SAUVCA (2001), The
challenges of access and admissions. Pretoria: South African
Universities Vice-Chancellors Association.
Grieve, K.W. & Richter, L.M. (1990). A factor analytic study of the
Home Screening Questionnaire for infants. SA Journal of
Psychology, 20(4), pp. 277–281.
Grieve, K.W., & Van Eeden, R. (1997). The use of the Wechsler
Intelligence Scales for Adults and the implications for research and
practical application in South Africa. Paper presented at a national
symposium on the future of intelligence testing in South Africa.
HSRC, Pretoria, South Africa, 6 March 1997.
Griffiths, R. (1954). The abilities of babies. London: Child
Development Research Centre.
Griffiths, R. (1970). The abilities of young children. London: Child
Development Research Centre.
Grigorenko, E.L., & Sternberg, R.J. (1998). Dynamic testing.
Psychological Bulletin, 124(1), pp. 75–111.
Groth-Marnat, G. (2003). Handbook of Psychological Assessment (4th
ed.). Hoboken, NJ: John Wiley & Sons.
Groth-Marnat, G. (2009). Handbook of psychological assessment (5th
ed.). Hoboken, NJ, US: John Wiley & Sons.
Guariguata, L., De Beer, I., Hough, R., Bindels, E., Weimers-
Maasdorp, D., Feeley, F.G., & De Wit, T.F.R. (2012). Diabetes,
HIV and other health determinants associated with absenteeism
among formal sector workers in Namibia. BMC Public Health,
12(44), pp. 1–12.
Guilford, J.P. (1936). Psychometric methods. New York: McGraw-Hill
Book Company.
Guichard, J. (2005). Life-long self-construction. International Journal
for Educational and Vocational Guidance, 5(2), pp. 111–124.
Gysbers, N.C.; Heppner, M,J., & Johnston, J.A. (2014). Career
counseling: Holism, diversity and strengths (4th ed.). Alexandria,
VA: American Counseling Association.
Hagtvet, K.A. & Johnsen, T.B. (Eds.). (1992). Advances in test anxiety
research (Vol. 7). Amsterdam: Swets & Zeitlinger.
Hair, J.F., Black, W.C., Babin, B.J., & Anderson, R.E. (2010).
Multivariate data analysis: A global perspective (7th ed.). Upper
Saddle River, NJ: Pearson Prentice-Hall.
Halpern, R. (2000). Early intervention for low-income children and
families. In J. Shonkoff & S. Meisels (Eds.), Handbook of Early
Childhood Intervention (2nd edition), pp. 361–386. Cambridge:
Cambridge University Press.
Hambleton, R.K. (1993). Translating achievement tests for use in
cross-national studies. European Journal of Psychological
Assessment, 9, pp. 54–65.
Hambleton, R.K. (1994). Guidelines for adapting educational and
psychological tests: a progress report. Bulletin of the International
Test Commission in the European Journal for Psychological
Assessment, 10(3), pp. 229–244.
Hambleton, R.K. (2001). The next generation of ITC test translation
and adaptation guidelines. European Journal for Psychological
Assessment, 17, pp. 164–172.
Hambleton, R.K. & Bollwark, J. (1991). Adapting tests for use in
different cultures: Technical issues and methods. Bulletin of the
International Test Commission, 18, pp. 3–32.
Hambleton, R.K., Merenda, P.F., & Spielberger, C.D. (Eds.) (2005).
Adapting Educational and psychological Tests for Cross-Cultural
Assessment. Mahwah, NJ: Lawrence Erlbaum Associates, Inc.
Hambleton, R.K. & Swaminathan, H. (1985). Item Response Theory:
Principles and applications. Boston: Kluwer-Nijhof.
Hambleton, R.K., Swaminathan, H., & Rogers, H.J. (1991).
Fundamentals of Item Response Theory. Newbury, CA: Sage
Publishers.
Maslach, C., Jackson, S.E & Leiter, M.P. (1996). Maslach Burnout
Inventory: Manual (3rd ed.). Palo Alto, CA: Consulting
Psychologists Press.
Maslach, C., Jackson, S.E., Leiter, M.P., Schaufeli, W.B., & Schwab,
R.L. (n.d.). The leading measure of burnout. Retrieved from
http://www.mindgarded.com/products/mbi.htm
Maslach, C. & Jackson, S.E. (1981). The measurement of
experienced burnout. Journal of Occupational Behaviour, 2, pp.
99−113.
Maslach, C., Schaufeli, W.B., & Leiter, M.P. (2001). Job burnout.
Annual Reviews of Psychology, 52, pp. 397–422.
Mason, J.C. (1992, July). Healthy equals happy plus productive.
Management Review, pp. 33–37.
Mason, J.C. (1994, July). The cost of wellness. Management Review,
pp. 29–32.
Matarazzo, J. (1990). Psychological assessment versus psychological
testing: Validation from Binet to the school, clinic and courtroom.
American Psychologist, 45(9), pp. 999–1017.
Matsumoto, D.R. (2000). Culture and psychology: People around the
world (2nd ed.). London: Wadsworth.
Matsumoto, D., & Van de Vijver, F.J.R. (Eds.) (2011). Cross-cultural
research methods in psychology. Cambridge: Cambridge
University Press.
Matthews, G., Zeidner, M., & Roberts, R.D. (2002). Emotional
intelligence: Science and myth. Cambridge, MA: MIT Press.
Mattke, S., Liu, H., Caloyeras, J., Huang, C. Y., Van Busum, K. R.,
Khodyakov, D., & Shier, V. (2013). Workplace wellness programs
study. Rand Health Quarterly, 3(2): 7.
Mayer, J.D. & Salovey, P. (1997). What is emotional intelligence? In
P. Salovey & D. Sluyter (Eds.), Emotional development and
emotional intelligence: Implications for educators, pp. 3–31. New
York: Basic Books.
Mboweni, J.C. (2007). The role of attributional styles, satisfaction with
life, general health and self-esteem on the psychological well-
being of adolescents. Unpublished
doctoral dissertation, North-West University, Mafikeng, South Africa.
McAdams, D.P. (1994). The person: An introduction to personality
psychology. Fort Worth, TX: Harcourt Brace.
McAdams, D.P. (1995). What do we know when we know a person?
Journal of Personality, 63, pp. 365–396.
McCarthy, D. (1944). A study of the reliability of the Goodenough
drawing test of intelligence. Journal of Psychology, 18, pp. 201–
216.
McCaulley, M.H. (2000). The Myers-Briggs Type Indicator in
counseling. In C.E. Watkins, Jr. & V.L. Campbell (Eds), Testing
and assessment in counseling practice (2nd edition), pp. 111–173.
Mahwah, NJ: Lawrence Erlbaum Publishers.
McClaran, A. (2003). From ‘admissions’ to ‘recruitment’: The
professionalisation of higher education admissions. Tertiary
Education and Management, 9, pp. 159–167.
McClelland, D.C., Koestner, R. & Weinberger, J. (1989). How do self-
attributed and implicit motives differ? Psychological review, 96, pp.
690–702. doi: 10.1037/0033-295X.96.4.690
McCrae, R.R., Costa, P.T., Jr., Martin, T.A., Oryol, V.E.,
Rukavishnikov, A.A., Senin, I.G., Hrebícková, M., Urbánek, T.
(2004). Consensual validation of personality traits across cultures.
Journal of Research in Personality, 38, pp. 179–201.
McIntire, S.A. & Miller, L. (2000). Foundations of psychological testing.
Boston: McGraw-Hill Higher Education.
McIntire, S. A., & Miller, L.A. (2006). Foundations of psychological
testing: A practical approach. California: SAGE Publications.
McMahon, M. (2011). The systems theory framework of career
development. Journal of Employment Counseling, 48(4), 170–172.
McMahon, M., Patton, W. & Watson, M.B. (2003). Developing
qualitative career assessment processes. Career Development
Quarterly, 51, pp. 194–202.
McMahon, M. & Watson, M.B. (2008). Systemic influences on career
development: Assisting clients to tell their career stories. The
Career Development Quarterly, 56, pp. 280–288.
McMahon, M. & Watson, M. (Eds.), (2015). Career assessment:
Qualitative approaches. Rotterdam: The Netherlands: Sense.
McMahon, M., Watson, M. & Patton, W. (2005). Qualitative career
assessment: Developing the My System of Career Influences
Reflection Activity. Journal of Career Assessment, 13, pp. 476–
490.
McReynolds, P. (1986). History of assessment in clinical and
educational settings. In Nelson, R.O. and Hayes, S.C. (Eds.),
Conceptual foundations of behavioral assessment. New York: The
Guilford Press.
Mead, A.D., Olson-Buchanan, J.B., & Drasgow, F. (2013).
Technology-based selection. In M.D. Coovert & L.F. Thompson
(Eds.), The psychology of workplace technology (pp. 21−42). New
York: Routledge.
Meade, A.W. (2004). Psychometric problems and issues involved with
creating and using ipsative measures for selection. Journal of
Occupational and Organizational Psychology, 77(4), pp. 531–552.
Meiring, D. (2007), Bias and equivalence of psychological measures
in
South Africa. Ridderkerk, Netherlands: Ridderprint B.V.
Meiring, D., Van de Vijver, F.J.R. & Rothmann, S. (2006). Bias in an
adapted version of the 15FQ+ in South Africa. SA Journal of
Psychology, 36(2), pp. 340–356.
Meiring, D., Van de Viyver, F.J.R., Rothmann, S. & Barrick, M.R.
(2005). Construct, item and method bias of cognitive and
personality measures in South Africa. SA Journal of Industrial
Psychology, 31, pp. 1–8
Meiring, D., Van de Vijver, F.J.R., Rothmann, S. & Sackett, P.R.
(2006). Internal and external bias of cognitive and personality
measures in South Africa. In Meiring, D. (2007), Bias and
equivalence of psychological measures in South Africa, (pp. 67–
102). Ridderkerk, Netherlands: Ridderprint B.V.
Meredith, W. (1993). Measurement invariance, factor analysis and
factorial invariance. Psychometrika, 58(4), pp. 525–543.
Mershon, B. & Gorsuch, R.L. (1988). Number of factors in the
personality sphere. Journal of Personality and Social Psychology,
55, pp. 675–680.
Meyer, J.C. (2003). Manual for the Meyer Interest Questionnaire (MB-
10) and structured vocational guidance. Stellenbosch: University of
Stellenbosch.
Meyer-Rath et al. (2012). Company-level ART provision to employees
is cost saving: A modeled cost-benefit analysis of the impact of
HIV and antiretroviral treatment in a mining workforce in South
Africa, International AIDS Society.
Meyer-Rath, G., Pienaar, J., Brink, B., Van Zyl, A., Muirhead, D.,
Grant, A., Churchyard, G., Watts, C., & Vickerman, P. (2015). The
impact of company-level ART provision to a mining workforce in
South Africa: A cost-benefit analysis. Plos Medicine, 12(9),
e1001869. doi:10.1371/journal.pmed.1001869
Mfusi, S.K. & Mahabeer, M. (2000). Psychosocial adjustment of
pregnant women infected with HIV/AIDS in South Africa. Journal of
Psychology in Africa, South of the Sahara, the Caribbean and
Afro-Latin America, 10(2), pp. 122–145.
Michas, M. (2018). Validation of the Centre for Epidemiologic Studies
Depression Scale: Revised among a non-clinical sample in the
South African workplace. Unpublished master’s dissertation,
University of Johannesburg, South Africa.
Miller, L.A., Lovler, R.L., & McIntire S.A. (2011). Foundations of
Psychological Testing: A practical approach (3rd ed.). Thousand
Oaks, CA: Sage.
Miller, R. (1987) Methodology: A focus for change. In K.F. Mauer &
A.I. Retief (Eds.), Psychology in context: Cross-cultural research
trends in South Africa. Pretoria: HSRC.
Millsap, R.E. (2011). Statistical approaches to measurement
invariance. New York: Routledge/Taylor & Francis.
Mislevy, R.J., Oranje, A., Bauer, M.I., Von Davier, A., Hao, J.,
Corrigan, S., Hoffman, S., DiCerbo, K., & John, M. (2014).
Psychometric considerations in game-based assessments.
Redwood City: Glasslab Institute of Play.
Mistry, J. & Rogoff, B. (1994). Remembering in cultural context. In W.
Lonner & R. Malpass (Eds.), Psychology and culture, pp. 139–144.
MA: Allyn & Bacon.
Mkhize, N. (2004). Psychology: An African perspective. In K. Ratele,
N. Duncan, D. Hook, P. Kiguwa, N. Mkhize & A. Collins (Eds.),
Self, community & psychology (pp. 4.1–4.29). Lansdowne, South
Africa: University of Cape Town Press.
Moch, S.L., Panz, V.R., Joffe, B.I., Havlik, I., & Moch, J.D. (2003).
Longitudinal changes in pituitary-adrenal hormones in South
African women with burnout. Endocrine, 21(3), 267–272. doi:
0969-711X/03/21:267-272.
Moletsane, M.K. (2004). The efficacy of the Rorschach among black
learners in South Africa. Unpublished doctoral thesis, University of
Pretoria. Available at
https://repository.up.ac.za/handle/2263/27930.
Moletsane, M. (2016). Understanding the role of indigenous
knowledge in psychological assessment and intervention in a
multicultural South African context. In R. Ferreira (Ed.),
Psychological assessment: Thinking innovatively in contexts of
diversity (pp. 20−36). Cape Town: Juta and Company.
Moletsane, M., & Eloff, I. (2006). The adjustment of Rorschach
Comprehensive System procedures for South African learners.
South African Journal of Psychology, 36, pp. 880−902.
Moodie, S., Daneri, P., Goldhagen, S., Halie, T., Green, K., &
LaMonte, L. (2014). Early childhood developmental screening: A
compendium of measures for children ages birth to five (OPRE
Report 2014 11). Washington, DC: Office of Planning, Research
and Evaluation, Administration for Children and Families, U.S.
Department of Health and Human Services.
Moodle HQ (2017). About Moodle. Retrieved on 13 September 2018
from: https://docs.moodle.org/34/en/About_Moodle
Moodley, N. (2005, February 9). Absenteeism cost R12bn a year, with
up to R2.2bn due to Aids. Business Report. Retrieved September
10, 2008, from http://www,busrep.co.za/index.php?
fSectionId=611&fArticleId=2403031
Moos, R. & Moos, B. (1986). Family Environment Scale manual (2nd
edition). Palo Alto, California: Consulting Psychology Press.
Moosa, M.Y.H. (2009). Anxiety associeted with colposcopy at Chris
Hani Baragwanath hopsital, Johannesburg. South African Journal
of Psychiatry, 15(2), pp. 48–52.
Morelli, N.A., Mahan, R.P., & Illingworth, A.J. (2014). Establishing the
measurement equivalence of online selection assessments
delivered on mobile versus non-mobile devices. International
Journal of Selection and Assessment, 22(2), pp. 125−138.
Mosam, A., Vawda, N.B., Gordhan, A.H., Nkwanyanat, N., &
Aboobaker, J. (2004). Quality of life issues for South Africans with
acne vulgaris. Clinical and Experimental Dermatology, 30, pp. 6–9.
doi: 10.1111/j.1365-2230.2004.01678.x
Mosdell, J., Balchin, R., & Ameen, O. (2010). Adaptation of aphasia
tests for neurocognitive screening in South Africa. SA Journal of
Psychology, 40(3), pp. 250–261.
Morgan, B. (2010). Career counselling in the 21st century: A reaction
article. Journal of Psychology in Africa, 20(3), pp. 501–504.
Morgan, B., de Bruin, G.P., & de Bruin, K. (2015a). Constructing
Holland’s hexagon in South Africa: Development and initial
validation of the South African Career Interest Inventory. Journal of
Career Assessment, 23(3), pp. 493−511. doi:
10.1177/1069072714547615
Morgan, B., de Bruin, G.P., & de Bruin, K. (2015b). Gender
differences in Holland’s circular/circumplex interest structure as
measured by the South African Career Interest Inventory. South
African Journal of Psychology, 45(3), pp. 349−360. doi:
10.1177/1069072714547615
Morgan, C.D. & Murray, H.A. (1935). A method for investigating
fantasies: The Thematic Apperception Test. Archives of Neurology
and Psychiatry, 34, pp. 289–306.
Moritz, S., Ferahli, S., & Naber, D. (2004). Memory and attention
performance in psychiatric patients. Journal of the International
Neuropsychological Society, 10(4), pp. 623–633.
Morris-Paxton, A. A., Van Lingen, J. M., & Elkonin, D. (2017a). An
evaluation of health information and wellness priorities among
socioeconomically disadvantaged students. Health Education
Journal, 76(3), pp. 271−281.
Morris-Paxton, A. A., Van Lingen, J. M., & Elkonin, D. (2017b).
Wellness and academic outcomes among disadvantaged students
in South Africa: An exploratory study. Health Education Journal,
76(1), pp. 66−76.
Moyo, S., & Theron, C. (2011). A preliminary factor analytic
investigation into the first-order factor structure of the Fifteen
Factor Questionnaire Plus (15FQ+) on a sample of Black South
African managers. SA Journal of Industrial Psychology/SA Tydskrif
vir Bedryfsielkunde, 37(1), Art. #934, 22 pages. doi:
10.4102/sajip.v37i1.934
Mpofu, E. (2002). Indigenization of the psychology of human
intelligence in sub-Saharan Africa. In W.J. Lonner, D.L. Dinnel,
S.A. Hayes, & D.N. Sattler (Eds.), Online readings in psychology
and culture, 5(2). Center for Cross-Cultural Research, Western
Washington University, Bellingham, Washington USA.
Mpofu, E., Ntinda, K. & Oakland, T. (2012). Understanding human
abilities in sub-Saharan African settings. Online readings in
psychology and culture, 4(3). doi: https://doi.org/10.9707/2307-
0919.1036.
Msolo, S. (2017). Campus home to SA’s Serious Games Institute.
North-West University Vaal Triangle News. Retrieved on 13
September 2018 from: www.nwu.ac.za/content/vaal-triangle-
campus-news-campus-home-sa’s-serious-games-institute
Muleya, V.R., Fourie, L., & Schlebusch, S. (2017). Ethical challenges
in assessment centres in South Africa. SA Journal of Industrial
Psychology/SA Tydskrif vir Bedryfsielkunde, 43(0), pp. 1−20,
a1324. doi: https://doi.org/10.4102/sajip.v43i0.1324
Murphy, K.R. & Davidshofer, C.O. (1998). Psychological testing:
principles and applications (4th edition). New Jersey: Prentice-Hall
International, Inc.
Murphy, K. R., & Davidshofer, C.O. (2005). Psychological testing:
principles and applications (6th edition). Upper Saddle River, New
Jersey: Pearson Prentice Hall.
Murphy, A. & Janeke, H.C. (2009). The relationship between thinking
styles and emotional intelligence: An exploratory study. SA Journal
of Psychology, 39(3), pp. 357–375.
Murphy, R., & Maree, D.J.F. (2006). A review of South African
research in the field of dynamic assessment. SA Journal of
Psychology, 36(1), pp. 168–191.
Murphy, E., & Meisgeier, C. (2008). MMTIC(TM) Manual: A guide to the
development and use of the Murphy-Meisgeier Type indicator for
Children(R). Gainesville, FL: Center for Applications of
Psychological Type.
Murstein, B.I. (1972). Normative written TAT responses for a college
sample. Journal of Personality Assessment, 36 (2), pp. 109–147.
Muthén, B.O. (2002). Beyond SEM: General latent variable modeling.
Behaviormetrika, 29(1), pp. 81–117.
Muthén, L.K., & Muthén, B.O. (1998–2017). Mplus User’s Guide (7th
ed.). Los Angeles, CA: Muthén & Muthén.
Mwamwenda, T.S. (1995). Educational Psychology: An African
perspective (2nd ed.). Durban: Butterworths.
Myers, I.B. & McCaulley, M.H. (1985). Manual: A guide to the
development and use of the Myers-Briggs Type Indicator. Palo
Alto, CA: Consulting Psychologists Press.
Naglieri, J.A. (1997). Planning, attention, simultaneous, and
successive theory and the cognitive assessment system: A new
theory-based measure of intelligence. In P. Flanagan, J.L.
Genshaft & P.L. Harrison (Eds.), Contemporary Intellectual
Assessment – Theories, tests and issues, pp. 247–267. New York:
Guilford Press.
Naicker, A. (1994). The psycho-social context of career counselling in
South African schools. SA Journal of Psychology, 24, pp. 27–34.
Naicker, R., & Fouché, C. (2003). The evaluation of an insourced
employee assistance programme. SA Journal of Human Resource
Management, 1(1), pp. 25−31.
Naidoo, D. (2013). Traumatic brain injury: The South African
landscape. The South African Medical Journal, 103(9), pp.
613−614. doi:10.7196/SAMJ.7325.
National Institutes of Health (2017). The NIH toolbox.
HealthMeasures. Retrieved on 13 September 2018 from:
http://www.healthmeasures.net/explore-measurement-
systems/nih-toolbox
Naude, J.L. P., & Rothmann, S. (2004). The validation of the Maslach
Burnout Inventory – Human services survey for emergency
medical technicians in Gauteng. SA Journal of Industrial
Psychology, 30, pp. 21–28.
Nauta, M.M. (2010). The development, evolution, and status of
Holland’s theory of vocational personalities: Reflections and future
directions for counseling psychology. Journal of Counseling
Psychology, 57(1), pp. 11–22.
Nel, J. A., Valchev, V. H., Rothmann, S., van de Vijver, F.J.R.,
Meiring, D. & de Bruin, G.P. (2012). Exploring the personality
structure in the 11 languages of South Africa. Journal of
Personality, 80, pp. 915–948. doi: 10.1111/j.1467-
6494.2011.00751.x
Nell, V. (1994). Interpretation and misinterpretation of the South
African Wechsler-Bellevue Adult Intelligence Scale: a history and a
prospectus. South African Journal for Psychology, 24(2), pp. 100–
109.
Nell, V. (1997). Science and politics meet at last: The insurance
industry and neuropsychological norms. SA Journal of Psychology,
27, pp. 43–49.
Nell, V. (1999). Cross-cultural neuropsychological assessment:
Theory and practice. New Jersey: Lawrence Erlbaum.
Nell, V. (2000). Cross-cultural neuropsychological assessment:
Theory and practice. New Jersey: Erlbaum.
Nell, R.D. (2005). Stress, coping resources and adjustment of married
mothers in the teaching profession. Unpublished master’s treatise,
Nelson Mandela Metropolitan University.
Newborg, J., Stock, J.R., Wnek, L., Guidubaldi, J., & Svinicki, J.
(1984). The Battelle Developmental Inventory. Allen, TX: DLM
Teaching Resources.
Nicholas, L. (2000). An evaluation of psychological reports considered
in the amnesty process of the truth and Reconciliation
Commission. SA Journal of Psychology, 30(1), pp. 50–52.
Nicholls, M., Viviers, A.M., & Visser, D. (2009). Validation of a test
battery for the selection of call centre operators in a
communications company. SA Journal of Psychology, 39(1), pp.
19–31.
Nolte, K., Steyn, B. J., Krüger, P. E., & Fletcher, L. (2016).
Mindfulness, psychological well-being and doping in talented
young high-school athletes. South African Journal for Research in
Sport, Physical Education and Recreation, 38(2), pp. 153−165.
Nolte, P.L., Wessels, J.C., & Prinsloo, C.E. (2009). The effect of
therapeutic recreation activities on students’ appraisals of their
emotions during academically demanding times. Journal for New
Generation Sciences, 7(1), pp. 160–175.
Nsamenang, A.B. (2007). Origins and development of scientific
psychology in Afrique Noire. In M.J. Stevens and D. Wedding
(Eds., under the supervision of John G. Adair). Psychology:
IUPsyS global resource (2007 ed.). London: Psychology Press.
Available at
https://www.unige.ch/fapse/SSE/teachers/dasen/Nsamenang2007.p
Nunnally, J.C. (1978). Psychometric theory. New York: McGraw-Hill.
Nunnally, J. C., & Bernstein, I.H. (1994). Psychometric theory (3rd
ed.). New York: McGraw Hill.
Nuttall, E.V., Romero, I., & Kalesnik, J. (Eds) (1992). Assessing and
screening preschoolers: Psychological and educational
dimensions. United States of America: Allyn and Bacon.
Nwoye, A. (2015). What is African psychology the psychology of?
Theory & Psychology, 25(1), pp. 96–116.
Nzimande, B. (1995). Culture fair testing? To test or not to test? Paper
delivered at the Congress on Psychometrics for Psychologists and
Personnel Practitioners, Pretoria, 5–6 June 1995.
O’Boyle, E.H., Humphrey, R.H., Pollack, J.M., Hawver, T.H., & Story,
P.A. (2011). The relation between emotional intelligence and job
performance: A meta-analysis. Journal of Organizational Behavior,
32, pp. 788–818.
Odendaal, I.E., Brink, M. & Theron, L.C. (2011). Rethinking
Rorschach interpretation: an exploration of resilient black South
African adolescents’ personal constructions. SA Journal of
Psychology, 41(4), pp. 528–539.
OECD (2011). PISA 2009 results: Students on line: Digital
technologies and performance (Volume VI).
http://dx.doi.org/10.1787/9789264112995-en
Olivier, N.M. & Swart, D.J. (1974). Manual for the Aptitude Tests for
School Beginners (ASB). Pretoria: HSRC.
Ones, D. (2008). Recent Assessment Centre research: A recap.
Keynote address presented at the 28th Annual Conference of the
AC study Group, 12–14 March, Protea Hotel, Stellenbosch.
Oosthuizen, R.M. & Koortzen, P. (2007). An empirical investigation of
job and family stressors amongst firefighters in the South African
context. SA Journal of Industrial Psychology, 33(1), pp. 49–58.
Osborn, D.S., & Zunker, V.G. (2016). Using assessment results for
career development (9th ed.). Boston, MA: Cengage Learning.
Outhoff, K. (2010). The pharmacology of anxiolytics. South African
Family Practice, 52(2), pp. 99−105. doi: 10.1080/20786204.2010.
10873947
Owen, K. (1986). Toets en itemsydigheid: Toepassing van die Senior
Aanlegtoetse, Meganiese Insigtoets en Skolastiese Bekwaamheid
Battery op Blanke, swart en kleurling en Indier Technikon
studente. Pretoria: Human Science Research Council.
Owen, K. (1989a). Test and item bias: The suitability of the Junior
Aptitude tests as a common test battery for white, Indian and black
pupils in Standard 7. Pretoria: HSRC.
Owen, K. (1989b). The suitability of Raven’s Progressive Matrices for
various groups in South Africa. Personality and Individual
Differences, 13, pp. 149–159.
Owen, K. (1991). Test bias: the validity of the Junior Aptitude Tests
(JAT) for various population groups in South Africa regarding the
constructs measured. SA Journal of Psychology, 21(2), pp. 112–
118.
Owen, K. (1992a). Test-item bias: Methods, findings and
recommendations. Pretoria: HSRC.
Owen, K. (1992b). The suitability of Raven’s Standard Progressive
Matrices for various groups in South Africa. Personality and
Individual Differences, 13, pp. 149–159.
Owen, K. (1996). Aptitude tests. In K. Owen & J.J. Taljaard (Eds.).
Handbook for the use of psychological and scholastic tests of the
HSRC (pp. 191–270). Pretoria: HSRC.
Owen, K. (1998). The role of psychological tests in education in South
Africa: Issues, controversies and benefits. Pretoria: HSRC.
Owen, K. & Taljaard, J.J. (1995) Handleiding vir die gebruik van
sielkundige en skolastiese toetse van die RGN. Pretoria: HSRC.
Owen, K. & Taljaard, J.J. (Eds) (1996). Handbook for the use of
psychological and scholastic tests of the HSRC (revised edition).
Pretoria: HSRC.
Ozminkowski, R. J., Dunn, R. L., Goetzel, R. Z., Cantor, R. I.,
Murnane, J., & Harrison, M. (1999). A return on investment
evaluation of the Citibank, NA, Health Management Program.
American Journal of Health Promotion, 14(1), pp. 31−43.
Pade, H. (2014). The evolution of psychological testing: Embarking on
the age of digital assessment. In R. Valle & J. Klimo (Eds.), The
changing faces of therapy: Innovative perspectives on clinical
issues and assessment. Available at: http://www.helloq.com
Park, G., Schwartzs, H.A., Eichstaedt, J.C., Kern, M.L., Kosinski, M.,
Stillwell, D.J., Ungar, L.H., & Seligman, M.E. (2015, June).
Automatic personality assessment through social media language.
Journal of Personality and Social Psychology, 108(6), pp.
934−952.
Parshall, C. G., Harmes, J.C., Davey, T., & Pashley, P. (2010).
Innovative items in for computerised testing. In W.J. van der
Linden & C.A.W. (Eds.). Computerised adaptive testing: Theory
and practice (2nd ed.). Norwell, MA: Kluwer Academic Publishers.
Parsons, F. (1909). Choosing a vocation. New York: Houghton Mifflin.
Pasche, S., & Myers, B. (2012). Substance misuse trends in South
Africa. Human Psychopharmacology: Clinical and Experimental,
27(3), pp. 338−341. doi: 10.1002/hup.2228
Paterson, H., & Uys, K. (2005). Critical Issues in Psychological Test
Use in the South African Workplace. SA Journal of Industrial
Psychology, 31(3), pp. 12–22.
Patrick, J. (2000). Mental status examination rapid record form.
Retrieved on 20 January 2009 from
www.nev.dgp.org.au/files/programsupport/mentalhealth/Mental%20
%20form.pdf?
Patton, W. & McMahon, M. (1999). Career development and systems
theory: A new relationship. Pacific Grove, CA: Brooks/Cole.
Patton, W., McMahon, M., & Watson, M.B. (2017). Career
development and systems theory: Enhancing our understanding of
careert. In G.B. Stead & M.B. Watson (Eds.), Career psychology in
the South African context (pp. 85−104). Pretoria, South Africa: Van
Schaik.
Ramsay, L. J., Taylor, N., De Bruin, G. P., & Meiring, D. (2008). The
Big Five personality factors at work: A South African validation
study. In J. Deller (Ed.), Research contributions to personality at
work. Munich & Meiring, Germany: Rainer Hampp Verlag.
Ratele, K. (2004). About black psychology. In D. Hook, N. Mkhize, P.
Kiguwa, A. Collins, E. Burman & I. Parker (Eds.), Critical
Psychology (pp. 389−414). Lansdowne, South Africa: University of
Cape Town Press.
Ratele, K. (2017). Four (African) psychologies. Theory & Psychology,
27(3), pp. 313–327.
Raven, J.C. (1965). Progressive matrices. London: H.K. Lewis.
Raykov, T., Marcoulides, G. A., & Li, C.H. (2012). Measurement
invariance for latent constructs in multiple populations: A critical
view and refocus. Educational and Psychological Measurement,
72(6), pp. 954–974.
Reckase, M.D. (1988). Computerized adaptive testing: a good idea
waiting for the right technology. Paper presented at the meeting of
the American Educational Research Association, New Orleans.
Reeve, C.L. & Lam, H. (2007). The relation between practice effects,
test-taker characteristics and degree of g-saturation. International
Journal of Testing, 7(2), pp. 225–242
Reger, M., Welsh, R., Razani, J., Martin, D., & Boone, K. (2002). A
meta-analysis of the neuropsychological sequelae in HIV infection.
Journal of the International Neuropsychological Society, 8(3), pp.
410–424.
Reis, H.T., Sheldon, K.M., Gable, S,L., Roscoe, J. & Ryan, R.M.
(2000). Daily well-being: the role of autonomy, competence, and
relatedness. Personality and Social Psychology Bulletin, 26(4), pp.
419–435.
Republic of South Africa (1998). Employment Equity Act. Government
Gazette, 400 (19370), Cape Town, 19 October.
Republic of South Africa (2000). Promotion of Equality and Prevention
of Unfair Discrimination Act. Government Gazette, 416 (20876),
Cape Town, 9 February.
Republic of South Africa (2013). Employment Equity Amendment Act
No. 47 of 2013. Government Gazette, 583 (37238), Cape Town, 16
January 2014. Available at:
http://www.labour.gov.za/DOL/downloads/legislation/acts/employme
equity/eea_amend2014.pdf
Retief, A.I. (1987). Thematic apperception testing across cultures:
Tests of selection versus tests of inclusion. SA Journal of
Psychology, 17, pp. 47–55.
Retief, A. (1992). The cross-cultural utility of the SAPQ: Bias or fruitful
difference? SA Journal of Psychology, 22, pp. 202–207.
Reuning, H. & Wortley, W. (1973). Psychological studies of the
Bushmen. Psychologia Africana, Monograph Supplement 7(113),
pp.1−113.
Reynolds, C.R. & Hickman, J.A. (2004). Draw-a-Person Intellectual
Ability Test for children, adolescents, and adults (DAP: IQ). Los
Angeles: Western Psychological Services.
Richards, S.D., Pillay, J. & Fritz, E. (2012). The use of sand-tray
techniques by school counsellors to assist children with emotional
and behavioural problems. The Arts in Psychotherapy, 39, pp.
367−373.
Richardson, J., Martin, N., Danley, K., Cohen, M., Carson, V., Sinclair,
B., Racenstein, J., Reed, R., & Levine, A. (2002).
Neuropsychological functioning in a cohort of HIV infected women:
Importance of antiretroviral therapy. Journal of the International
Neuropsychological Society, 8(6), pp. 781–793.
Richter, L.M. (1989). Household density, family size and the growth
and development of black children: a cross-sectional study from
infancy to middle childhood. SA Journal of Psychology, 19(4), pp.
191–198.
Richter, L.M. & Griesel, R.D. (1988). Bayley Scales of Infant
Development: Norms for Interpreting the performance of black
South African infants. Pretoria: Institute of Behavioural Sciences.
Richter, L.M. & Grieve, K.W. (1991). Cognitive development in young
children from impoverished African families. Infant Mental Health
Journal, 12(2), pp. 88–102.
Richter, L., Mabaso, M., & Hsiao, C. (2016). Predictive power of
psychometric assessments to identify young learners in need of
early intervention: Data from the Birth to Twenty Plus (Bt20+)
cohort, South Africa. South African Journal of Psychology, 46(2),
pp. 175−190.
Ring, L., Hofer, S., McGee, H., Hickey, A. & O’Boyle, C.A. (2007).
Individual quality of life: can it be accounted for by psychological or
subjective well-being? Social Indicator Researcher, 86, pp. 443–
461.
Robertson, G.J. & Eyde, L.D. (1993). Improving test use in the United
States: The development of an interdisciplinary casebook.
European Journal of Psychological Assessment, 9(2), pp. 137–
146.
Robinson, M. (1986). ’n Ondersoek na die waarde van die Junior
Suid-Afrikaanse Indivduele Skale (JSAIS) as diagnostiese
hulpmiddel vir leergestremde kinders met leesprobleme
[Investigating the value of the Junior South African Indivdual
Scales (JSAIS) as diagnostic aid for learning-disabled children with
reading problems]. Unpublished master’s disertation, University of
Pretoria.
Robinson, M. (1989). Bylae tot die Handleiding vir die Junior Suid-
Afrikaanse Individuale Skale (JSAIS): Tegniese besonderhede en
normtabelle vir 6- tot 8-jarige leerlinge [Appendix to the Manual for
the Junior South African Individual Scales (JSAIS): Technical
details and tables of norms for the 6- to 8-year-old pupils]. Pretoria:
HSRC.
Robinson, M. (1994). Manual for the Individual Scale for General
Scholastic Aptitude (ISGSA). Part III: Norm tables. Pretoria: HSRC.
Robinson, M. (1996). Manual for the Individual Scale for General
Scholastic Aptitude (ISGSA). Part I: Background and
standardization. Pretoria: HSRC.
Robinson, M. & Hanekom, J.M.D. (1991). Die JSAIS en ASB as
hulpmiddels by die evaluaring van skoolgereedheid [ The use of
the JSAIS and ASB for the evaluation of school readiness].
Pretoria: HSRC.
Robison, J. (2010). The business case for well-being. Business
Journal, June 2010. Available at
http://news.gallup.com/businessjournal/139373/business-case-
wellbeing.aspx
Roe, R.A. (2008). The impact of testing on people and society. Paper
presented at the 6th Conference of the International Test
Commission, Liverpool, UK, 14–16 July 2008. (Can be accessed
online at www.intestcom.org).
Roger, E.A.S. (1989). The Number and Quantity Concepts subtest of
the Junior South African Individual Scale as predictor of early
mathematical achievement. Unpublished master’s dissertation,
University of Port Elizabeth.
Rogers, T.B. (1997). Teaching ethics and test standards in a
psychological course: A Test Taker’s Bill of Rights. Teaching of
Psychology, 24(1), pp. 41–46.
Roivainen, E. (2012). Economic, educational and IQ gains in eastern
Germany 1990–2006. Intelligence, 40, pp. 571–575.
Roodt, G. (1991). Die graad van werkbetrokkenheid as voorspeller
van persoonlike welsyn: ’n studie van bestuurders [ The degree of
job involvement as predictor of personal well-being: A study of
managers]. Unpublished doctoral dissertation, University of the
Free State.
Roodt, G. (2007). The reliability of structured interviews: Competency
rating interviews as a case in point. Paper presented at the 27th
Annual Conference of the AC study Group, 14–16 March, Protea
Hotel, Stellenbosch.
Roodt, G. (2008). Reliability and validity analyses. In S. Schlebusch &
G. Roodt (Eds.). Assessment Centres: Unlocking potential for
growth. Randburg: KnowRes Publishers.
Roos, V. (2008). The Mmogo-method®: Discovering symbolic
community interactions. Journal of Psychology in Africa, 18(4), pp.
659−668.
Roos, V. (2012). The Mmogo-method®: An exploration of experiences
through visual projections. Qualitative Research in Psychology,
9(3), pp. 249−261.
Rosander, P., Bäckström, M., & Stenberg, G. (2011). Personality traits
and general intelligence as predictors of academic performance: A
structural equation modelling approach. Learning and Individual
Differences, 21(5), pp. 590–596.
Rosen, S., Simon, J.L., Thea, D.M. & Vincent, J.R. (2000). Care and
treatment to extend the working lives of HIV-positive employees:
Calculating the benefits to business. South African Journal of
Science, 96(6), pp. 300–304.
Rosen, S., Simon, J.L., Vincent, J.R., MacLeod, W., Fox, M., & Thea,
D.M. (2003). AIDS is your business. Harvard Business Review,
February, pp. 81–87.
Rosen, S., Vincent, J.R., MacLeod, W., Fox, M., Thea, D.M. & Simon,
J.L. (2004). The cost of HIV/Aids to businesses in southern Africa.
Official Journal of the International AIDS Society, 18(2), pp. 317–
324.
Rosenthal, R. & Rosnow, R.L. (1991). Essentials of behavioral
research: Method and data analysis (2nd edition). New York:
McGraw-Hill.
Ross, S., Krukowski, R., Putnam, S., & Adams, K. (2003). Memory
assessment scales in detecting probable malingering in mild head
injury. Paper presented at the 26th Annual International
Neuropsychological Society Mid-year Conference, Berlin,
Germany.
Roth, M., McCaul, E., & Barnes, K. (1993). Who becomes an ‘at-risk’
student? The predictive value of a kindergarten screening battery.
Exceptional Children, 59(4), pp. 348–358.
Runyon, R.P. & Haber, A. (1980). Fundamentals of behavioral
statistics. Reading, MA: Addison- Wesley Publishing Company.
Ryan Krane, N.E., & Tirre, W.C. (2005). Ability assessment in career
counseling. In S.D. Brown & R.W. Lent (Eds.), Career
development and counseling: Putting theory and research to work
(pp. 330−352). Hoboken, NJ US: John Wiley & Sons Inc.
Ryff, C.D. (1989). Happiness is everything, or is it?: Explorations on
the meaning of psychological wellbeing. Journal of Personality and
Social Psychology, 57, pp. 1069−1081. doi:10.1037//0022-
3514.57.6.1069
Ryff, C.D. & Keys, C.L.M. (1995). The structure of psychological well-
being revisited. Journal of Personality and Social Psychology,
69(4), pp. 719–727.
Ryff, C.D. & Singer, B. (2003). Flourishing under fire: resilience as a
prototype of challenged thriving. In C.L.M. Keyes & J. Haidt.
Flourishing Positive Psychology and the life well-lived.
Washington: American Psychological Association.
SA: 1 in 3 mentally ill. (2007, October). Retrieved September 27,
2008, from
http://www.health24.com/mind/Mental_health_in_SA/1284-
1031,42638.asp
Salovey, P. & Mayer, J.D. (1990). Emotional Intelligence. Imagination,
Cognition and Personality, 9(3), pp. 185–211.
Salgado, J.F. (2017). Personnel selection. Oxford Research
Encyclopedias, Psychology. Oxford University Press: USA.
Retrieved on 13 September 2018 from:
http://dx.doi.org/doi:10.1093/acrefore/9780190236557.013.8
Salgado, J. F., Anderson, N., Moscoso, S., Bertua, C., De Fruyt, F., &
Rolland, J.P. (2003). A meta-analytic study of general mental
ability validity for different occupation in the European community.
Journal of Applied Psychology, 88(6), pp. 1068–1081.
Samelson, F. (1979). Putting psychology on the map: Ideology and
intelligence testing. In A.R. Buss (Ed.), Psychology in social
context. New York: Irvington.
Sampson, J.P. (2009). Modern and postmodern career theories: The
unnecessary divorce. The Career Development Quarterly, 58, pp.
91–96.
Sanderson, K. & Andrews, G. (2006). Common mental disorders in
the workforce: recent findings from descriptive and social
epidemiology. Canadian Journal of Psychiatry, 51(2), pp. 63–76.
Sarason, I.G. (Ed.). (1961). Test anxiety and the intellectual
performance of college students. Journal of Educational
Psychology, 52, pp. 201–206.
Sattler, J.M. (1988). Assessment of Children. (3rd ed.). San Diego,
CA: Author.
Savickas, M.L. (1989). Career-style assessment and counseling. In T.
Sweeney (Ed.), Adlerian counseling: A practical approach for a
new decade (pp. 289−320). Muncie, IN: Accelerated Development
Press.
Savickas, M.L. (2005). The theory and practice of career construction.
In S. Brown, & R. Lent (Eds.), Career development and
counseling: Putting theory and research to work (pp. 42–70).
Hoboken, NJ: John Wiley.
Savickas, M.L. (2010). Foreword: Best practices in career intervention.
In K. Maree (Ed.), Career counselling: Methods that work (pp. xi–
xii). Cape Town, South Africa: Juta.
Savickas, M.L. (2011). Prologue: Reshaping the story of career
counselling. In J.G. Maree (Ed.), Shaping the story: A guide to
facilitating narrative counselling (pp. 1–3). Rotterdam: Sense.
Savickas, M. L., Nota, L., Rossier, J., Dauwalder, J., Duarte, M. E.,
Guichard, J. et al. (2009). Life designing: A paradigm for career
construction in the 21st century. Journal of Vocational Behavior,
75(3), pp. 239–250.
Scalise, K., & Gifford, B. (2006). Computer-based assessment in e-
learning: A framework for constructing ‘intermediate constraint’
questions and tasks for technology platforms. Journal of
Technology, Learning and Assessment, 4(6). Retrieved on 21
January 2018 from http://www.jtla.org.
Schaap, P. (2003). The construct comparability of the PIB/SPEEX
stress index for job applicants from diverse cultural groups in
South Africa. SA Journal of Psychology, 33(2), pp. 95–102.
Schaap, P. (2011). The differential Item functioning and structural
equivalence of a nonverbal cognitive ability test for five language
groups. SA Journal of Industrial Psychology, 37(1), pp. 1–16.
Schaap, P. & Basson, J.S. (2003). The construct equivalence of the
PIB/SPEEX motivation index for job applicants from diverse
cultural backgrounds. SA Journal of Industrial Psychology, 29(2),
pp. 49–59.
Schaufeli, W. B., & Bakker, A. (2003). Utrecht Work Engagement
Scale: Preliminary manual. Occupational Health Psychology Unit,
Utrecht University.
Schaufeli, W. B., & Bakker, A.B. (2004). Job demands, job resources,
and their relationship with burnout and engagement: A multi-
sample study. Journal of Organizational Behavior, 25, pp. 293–
315.
Schaufeli, W.B., Bakker, A.B., & Van Rhenen, W. (2009). How
changes in job demands and resources predict burnout, work
engagement, and sickness absenteeism. Journal of Organizational
Behavior, 30, pp. 893–917. doi: 10.1002/job.595
Schaufeli, W.B., Martínez, I.M., Pinto, A.M., Salanova, M., & Bakker,
A.B. (2002). Burnout and engagement in university students.
Journal of Cross-Cultural Psychology, 33(5), pp. 464–481.
Schaufeli, W.B., Salanova, M., González-Romá, V., & Bakker, A.B.
(2002). The measurement of engagement and burnout: A two
sample confirmatory factory analytic approach. Journal of
Happiness Studies, 3, pp. 71–92.
Schaufeli, W.B., Taris, T.W., & Van Rhenen, W. (2008). Workaholism,
burnout, and work engagement: Three of a kind or three different
kinds of employee well-being? Applied Psychology: An
International Review, 57(2), pp. 173–203.
Scheier, M. F., Carver, C. S., & Bridges, M.W. (1994). Distinguishing
optimism from neuroticism (and trait anxiety, self-mastery, and
self-esteem): A reevaluation of the Life Orientation Test. Journal of
Personality and Social Psychology, 67, pp. 1063–1078.
Schepers, J.M. (1992). Toetskonstruksie: Teorie en Praktyk [Test
Construction: Theory and Practice]. Johannesburg: RAU Drukpers.
Schepers, J.M. (1999). The Bell-curve revisited: a South African
perspective. Journal of Industrial Psychology, 25(2), pp. 52–63.
Schepers, J.M. (2005). Technical manual: Locus of control inventory
(Revised edition 2003). Johannesburg: Jopie van Rooyen &
Partners.
Schepers, J.M., & Hassett, C.F. (2006). The relationship between the
fourth edition (2003) of the Locus of Control Inventory and the
Sixteen Personality Factor Questionnaire (version 5). SA Journal
of Industrial Psychology, 32, pp. 9–18.
Schlebusch, S. & Roodt, G. (Eds.) (2008). Assessment Centres:
Unlocking potential for growth. Johannesburg: KnowRes
Publishing.
Schmidt, C.S. (1997). Engaging the dilemmas of psychological
testing. Paper presented at the Annual Conference of the Society
of Industrial Psychology, CSIR Conference Centre, Pretoria.
Schmidt, F.E. & Hunter, J.E. (1998). The validity and utility of selection
methods in personnel psychology: Practical and theoretical
implications of 85 years of research findings. Psychological
Bulletin, 124, pp. 262–274.
Schmidt, F.L., Oh, I., & Shaffer, J.A. (2016). The validity and utility of
selection methods in personnel psychology: Practical and
theoretical implications of 100 years of research findings. Fox
School of Business research paper. Retrieved on 13 September
2018 from https://ssrn.com/abstract=2853669.
Schmitt, N. W., Arnold, J. D., & Nieminen, L. (2017). Validation
strategies for primary studies. In J.L. Farr & N.T. Tippins (Eds.),
Handbook of employee selection (pp. 51–71) (2nd ed). New York,
NY, US: Routledge/Taylor & Francis Group.
Schuerger, J.M. (2002). The Sixteen Personality Factor Questionnaire
(16PF). In C.E. Watkins, Jr. & V.L. Campbell (Eds.), Testing and
assessment in counseling practice (2nd ed.), pp. 73–110.
Mahwah, NJ: Lawrence Erlbaum Publishers.
Sechrest, L. (1978). Incremental validity. In D.N. Jackson and S.
Messick (Eds.) Problems in human assessment, Huntington, NY:
Robert E. Krieger Publishing Company.
Seligman, L. (1994). Developmental career counselling and
assessment. Thousand Oaks, CA: Sage.
Seligman, M.E.P (2003a). Foreword: The past and future of positive
psychology. In C.L.M. Keyes & J. Haidt. Flourishing: Positive
Psychology and the life well-lived. Washington: American
Psychological Association.
Seligman, M.E.P. (2003b). Authentic happiness: Using the new
positive psychology to realise your potential for lasting fulfilment.
London: Nicholas Brealey.
September, S.J., Rich, E.G. & Roman, N.V. (2016). The role of
parenting styles and socio-economic status in parents’ knowledge
of child development. Early Child Development and Care, 186(7),
pp. 1060−1078, doi: 10.1080/03004430.2015.1076399
Serious Games Institute (2017). About SGI projects. Retrieved from:
http://www.seriousgamesinstitute.co.uk/applied-
research/About-SGI-Projects.aspx
Shabalala, S. & Jasson, A. (2011). PTSD symptoms in intellectually
disabled victims of sexual assault. SA Journal of Psychology,
41(4), pp. 424–436.
Sharf, R.S. (2006). Applying career development theory to
counselling. London: Thomson.
Sheldon, K.M., Kashdan, T.B., & Steger, M.F. (2011). Designing
positive psychology: Taking stock and moving forward. Oxford:
Oxford University Press.
Sherwood, E.T. (1957). On the designing of TAT pictures, with special
reference to a set for an African people assimilating western
culture. Journal of Social Psychology, 45, pp. 161–190.
Shonkoff, J.P., & Meisels, S.J. (2000). Handbook of early childhood
intervention (2nd ed.). Cambridge: Cambridge University Press.
Shuttleworth-Edwards, A.B. (2012). Guidelines for the use of the
WAIS-IV with WAIS-III cross-cultural normative Indications. SA
Journal of Psychology, 42(3), pp. 399–410.
Shuttleworth-Edwards, A.B. (2017). Countrywide norms declared
obsolete: Best practice alert for IQ testing in a multicultural context.
South African Journal of Psychology, 47(1), pp. 3−6. doi:
http://dx.doi.org/10.1177/0081246316684465
Shuttleworth-Edwards, A.B., Gaylard, E.K. & Radloff, S.E. (2013).
WAIS-III test performance in the South African context: Extension
of a prior cross-cultural normative database. In S. Laher & K.
Cockcroft (Eds.), Psychological assessment in South Africa:
Research and applications (pp. 17−32). Johannesburg: Wits
University Press.
Shuttleworth-Edwards, A., Kemp, R., Rust, A., Muirhead, J., Hartman,
N., & Radloff, S. (2004) Vanishing ethnic differences in the wake of
acculturation: A review and preliminary indications on WAIS-III test
performance in South Africa. Journal of Clinical and Experimental
Neuropsychology, 26(7), pp. 903–920.
Shuttleworth-Edwards, A.B., Van der Merwe, A.S., Van Tonder, P. &
Radloff, S.E. (2013). WISC-IV test performance in the South African
context: A collation of cross-cultural norms. In S. Laher & K.
Cockcroft (Eds.), Psychological assessment in South Africa:
Research and applications (pp. 33−47). Johannesburg: Wits Press.
Shuttleworth-Edwards, A. B., Whitefield- Alexander, V. J., Radloff, S.
E., Taylor, A. M., & Lovell, M.R. (2009). Computerized
neuropsychological profiles of South African versus United States
athletes: A basis for commentary on cross-cultural norming issues
in the sports concussion arena. The Physician and Sports
Medicine, 37, pp. 45–52.
Shuttleworth-Jordan, A.B. (1996). On not reinventing the wheel: a
clinical perspective on culturally relevant test usage. SA Journal of
Psychology, 26(2), pp. 96–102.
Sieberhagen, C., Pienaar, J., & Els, C. (2011). Management of
employee wellness in South Africa: Employer, service provider and
union perspectives. SA Journal of Human Resource
Management/SA Tydskrif vir Menslikehulpbronbestuur, 9(1), a305.
doi: 10.4102/sajhrm.v9i1.305
Simner, M.L. & Goffin, R.D. (2003). A position statement by the
International Graphonomics Society on the use of graphology in
personnel selection testing. International Journal of Testing, 3(4),
pp. 353–364.
Sinclair, J. (2012, February). Mercer 2011 national survey of
employer-sponsored health plans. Slide presentation at the
meeting of Mercer, Columbus: Ohio Program.
Sireci, S.G. (2011). Evaluating test and survey items for bias across
languages and cultures. In D. Matsumoto & F.J.R. van de Vijver,
Cross-cultural research methods in psychology (pp. 216–240).
Cambridge, NY: Cambridge University Press.
Sireci, S. & Gandara, F. (2016). Testing in educational and
developmental settings. In F.T. Leong, D. Bartram, F.M. Cheung,
K.F. Geisinger, & D Iliescu. (Eds.), The ITC international handbook
of testing and assessment (pp. 188–202). New York, NY: Oxford
University Press.
Skuy, M., Zolezzi, S., Mentis, M., Fridjhon, P., & Cockcroft, K. (1996).
Selection of advantaged and disadvantaged South African
students for university admission. South African Journal of Higher
Education, 10(1), pp. 110–117.
Skuy, M., Schutte, E., Fridjhon, P.& O’Carroll, S. (2001). Suitability of
published neuropsycho-logical test norms for urban African
secondary school students in South Africa. Personality and
Individual Differences, 30(8), pp. 1413-1425.
Smit, G.J. (1996). Psychometrics: Aspects of measurement. Pretoria:
Kagiso.
Smith, P.B. (2004). Acquiescent response bias as an aspect of
cultural communication style. Journal of Cross-Cultural
Psychology, 35, pp. 50–61.doi: 10.1177% 2F0022022103260380
Smith, S.A. (2002). An evaluation of response scale formats of the
Culture Assessment Instrument. Unpublished D Phil thesis,
Johannesburg: Rand Afrikaans University.
Smith, S.A. & Roodt, G. (2003). An evaluation of response scale
formats of the Culture Assessment Instrument. SA Journal of
Human Resource Management, 1(2), pp. 60–75.
Smith, Z. (2016). Construct validation of the DASS-21 in a non-clinical
sample of working adults. Unpublished master’s thesis, University
of Johannesburg, Johannesburg
Smith, Z., Henn, C. & Hill, C. (2018). Validation of the DASS-21 in a
non-clinical sample of working adults. Manuscript submitted for
publication.
Smoski, W.J., Brunt, M.A., & Tannahill, J.C. (1998). C.H.A.P.S.
Children auditory performance scale. Instruction manual. Tampa,
FL: The Educational Audiology Association.
Snow, E. L., Likens, A. D., & McNamara, D.S. (2016, December).
Taking control: Stealth assessment of deterministic behaviours
within a game-based system. International Journal of Artificial
Intelligence in Education, 26(4), pp. 1011–1032.
SIOPSA: Society for Industrial and Organizational Psychology, Inc.
(2003). Principles for the validation and use of personnel selection
procedures (4th ed.). Bowling Green, OH: Author.
SIOPSA: Society for Industrial and Organisational Psychology in
South Africa (2006a). Code of practice for psychological and other
similar assessment in the workplace. Pretoria: Author (in
association with the People Assessment Initiative (PAI)).
Society for Industrial and Organisational Psychology in South Africa
(2006b). Guidelines for the Validation of Assessment Procedures.
Pretoria: Author (in association with the People Assessment
Initiative (PAI)).
Solano-Flores, G., Turnbull, E. & Nelson-Barber, S. (2002).
Concurrent development of dual language assessments: An
alternative to translating tests for linguistic minorities. International
Journal of testing, 2(2), pp. 107–129.
Sommer, M. & Dumont, K. (2011). Psychosocial factors predicting
academic performance of students at a historically disadvantaged
university. SA Journal of Psychology, 41(3), pp. 386–395.
Spangler, W.D. (1992). Validity of questionnaire and TAT measures of
need for achievement: Two meta-analyses. Psychological Bulletin,
112, pp. 140–154.
Sparrow, S.S., Cicchetti, D.V., & Balla, D.A. (2005). Vineland-II
Adaptive Behavior Scales (2nd edition). Minnesota: AGS
Publishing.
Sparrow, S.S., Cicchetti, D.V. & Saulnier, C.A. (2016). Vineland
Adaptive Behavior Scales, Third Edition (Vineland-3). San Antonio,
Texas: Pearson.
Spector, P.E. (1988). Development of the work locus of control scale.
Journal of Occupational Psychology, 61, pp. 335–340. Spielberger,
C.D., Gorsuch, R.L. & Lushene, R.E. (1970). The State-
Trait Anxiety Inventory: Test Manual for Form X. Palo Alto:
Consulting Psychologists Press.
Spitzer, R. L., Kroenke, K., Williams, J. B., & Löwe, B. (2006). A brief
measure for assessing generalized anxiety disorder: the GAD-7.
Archives of Internal Medicine, 166(10), pp. 1092−1097.
Squires, J., Nickel, R.E., & Eisert, D. (1996). Early detection of
developmental problems: Strategies for monitoring young children
in the practice setting. Developmental and behavioural paediatrics,
17(6), pp. 420–427.
Stark, S., Martin, J. & Chernyshenko, O.S. (2016). Technology and
testing: Developments in education, work and health care. In F.T.
Leong, D. Bartram, F.M. Cheung, K.F. Geisinger, & D. Iliescu
(Eds.), The ITC international handbook of testing and assessment
(pp. 395−407). New York, NY: Oxford University Press.
STATSSA, (2010). HIV prevalence in 2010: Mid-year population
estimates. Retrieved from http://www.statssa.gov.za
StatSoft, Inc. (2012). StatSoft STATISTICA Electronic Statistics
Textbook. Tulsa, OK: StatSoft. WEB:
http://www.statsoft.com/textbook/.
Stead, G.B. & Watson, M.B. (1998). The appropriateness of Super’s
career theory among black South Africans. SA Journal of
Psychology, 28, pp. 40–43.
Stead, G.B., & Watson, M.B. Eds. (2017). Career psychology in the
South African context (3rd ed.). Pretoria: Van Schaik.
Steele, G. & Edwards, D. (2002). The development and validation of
the Xhosa translations of the Beck Depression Inventory, the Beck
Anxiety Inventory, and the Beck Hopelessness Scale. Paper
presented at the 8th annual congress of the Psychological Society
of South Africa (PsySSA), Cape Town, South Africa, 24–27
September 2002.
Steele, G. & Edwards, D. (2008). Development and validation of the
isiXhosa translations of the Beck inventories: 2. Item analysis and
internal consistency. Journal of Psychology in Africa, 18, pp. 217–
226.
Sternberg, R.J. (1984). A contextualist view of the nature of
intelligence. International Journal of Psychology, 19, pp. 307–334.
Sternberg, R.J. (2003). Issues in the theory and measurement of
successful intelligence: A reply to Brody. Intelligence, 31, pp. 331–
337.
Sternberg, R.J. (2012). Toward a triarchic theory of human
intelligence. In J.C. Kaufman and E.L. Grigorenko (Eds), The
essential Sternberg: Essays on intelligence, psychology and
education, pp. 33–70. New York: Springer Publishing Company.
Steyn, C., Howcroft, J.G., & Fouché, J.P. (2011). The psychofortology
of female psychiatric out-patients: living with mood and anxiety
disorders. SA Journal of Psychology, 41(3), 288–299.
Storm, L. & Roodt, G. (2002). Die verband tussen
organisasiesosialisering and organisasieverbondenheid [The
relationship between organisation socialisation and organisational
commitment]. SA Tydskrif vir Bedryfsielkunde, 28(1), pp. 14–21.
Strong, E.K. (1927). Vocational Interest Blank. Stanford, CA: Stanford
University Press.
Stroud, L. (2016). Scale A: Foundations of learning. PowerPoint
presentation on the developments in the Griffiths III on the ARICD
site. Available at https://www.aricd.ac.uk.
Stroud, L., Foxcroft, C., Green, E., Bloomfield, S., Cronje, J., Hurter,
K., Lane, H., Marais, R., Marx, C., McAlinden, P., O’Connell, R.,
Paradice, R., & Venter, D. (2016). Part I: Overview, Development
and Psychometrics Properties. In Griffiths III: Griffiths Scales
Scales of Child Development (3rd Ed.). Hogrefe Ltd. Available at:
https://www.hogrefe.co.uk.
Strümpfer, D.J.W. (1997). Sense of coherence, negative affectivity,
and general health in farm supervisors. Psychological Reports, 80,
pp. 963–966.
Strümpfer, D.J.W. & Wissing, M.P. (1998). Review of the South
African data on the Sense of Coherence Scale as a measure of
fortigenesis and salutogenesis. Paper presented at the 4th Annual
Congress of the Psychological Society of South Africa (PsySSA),
9–11 September 1998, Cape Town, South Africa.
Sue, D, Sue, D & Sue, S. (1997). Understanding abnormal behavior
(5th ed.). Boston: Houghton Mifflin Company.
Sunderaraman, P., Zahodne, L.B., & Manly, J.J. (2016). A
commentary on ‘generally representative is representative of none:
Pitfalls of IQ test standardization in multicultural settings’ by A.B.
Shuttleworth-Edwards. The Clinical Neuropsychologist, 30(7), pp.
999−1005. doi: http://dx.doi.org/10.1080/13854046.2016.1211321
Super, D.E. (1983). Assessment in career guidance: Toward truly
developmental counseling. Personnel and Guidance Journal, 61,
pp. 555–562.
Super, D.E. (1990). A life-span, life-space approach to career
development. In D. Brown & L. Brooks (Eds.), Career choice and
development, pp. 197–261. San Francisco, CA: Jossey-Bass.
Super, D.E. (1994). A life span, life space perspective on
convergence. In M.L. Savickas & R.L. Lent (Eds), Convergence in
career development theories: Implications for science and practice
(pp. 63–74). Palo Alto, CA: CPP Books.
Super, D.E. (1995). Values: Their nature, assessment and practical
use. In D.E. Super, B. Šverko, B., & C.E. Super (Eds) Life roles,
values, and careers (pp. 54–61). San Francisco, CA: Jossey-Bass.
Super, B. Sverko & C.E. Super (Eds.). (1995). Life roles, values, and
careers. San Francisco, CA: Jossey-Bass.
Swart, C., Roodt, G., & Schepers, J.M. (1999). Itemformaat,
differensiële itemskeefheid en die faktorstruktuur van ‘n
selfvoltooiingsvraelys [Item format, differential item skewness and
the factor structure of a self-completion questionnaire]. Tydskrif vir
Bedryfsielkunde, 25(1), pp. 33–43.
Swenson, W.M., Rome, H., Pearson, J., and Brannick, T. (1965). A
totally automated psychological test: Experience in a Medical
Center. Journal of the American Medical Association, 191, pp.
925–927.
Taber, B.J., Hartung, P.J., Briddick, H., Briddick, W.C., & Rehfuss,
M.C. (2011). Career style interview: A contextualized approach to
career counseling. The Career Development Quarterly, 59(3), pp.
274-287. doi: 10.1002/j.2161-0045.2011.tb00069.x
Taljaard-Plaut, J., & Strauss, R. (1998). A generation in transition:
Understanding stress-related disorders in African black university
students. Psychoanalytic Psychotherapy in South Africa, 6(2), pp.
56−64.
Tanzer, N.K. (2005) Developing tests for use in multiple cultures and
languages: A plea for simultaneous development. In Hambleton,
R.K., Merenda, P.F., & Spielberger, C.D. (Eds.) (2005). Adapting
educational and psychological tests for cross-cultural assessment
(Chapter 10, pp. 235–244). Mahwah, NJ: Lawrence Erlbaum
Associates, Inc.
Taylor, C., Jamieson, J., Eignor, D. & Kirsch, I. (1998). The
relationship between computer familiarity and performance on
computer-based TOEFL test tasks. TOEFL Research, 61.
Princeton, NJ: Educational Testing Service.
Taylor, I.A. (2000). The construct comparability of the NEO PI-R
Questionnaire for black and white employees. Unpublished
doctoral thesis, University of the Free State.
Taylor, J.M., & Radford, E.J. (1986). Psychometric testing as an unfair
labour practice. SA Journal of Psychology, 16, pp. 79–96.
Taylor, N. (2004). The construction of a South African five-factor
personality questionnaire. Unpublished master’s dissertation, Rand
Afrikaans University, South Africa.
Taylor, N. (2008). Construct, item, and response bias across cultures
in personality measurement. Unpublished doctoral thesis.
Johannesburg, South Africa: University of Johannesburg.
Taylor, N. (2016). Generally representative is generally representative:
Comment on Shuttleworth-Edwards. The Clinical
Neuropsychologist, 30(7), pp. 1017−1022. Doi:
http://dx.doi.org/10.1080/13854046.2016.1213884.
Taylor, N., & De Beer, J. (2010). The relationship between emotional
intelligence and derailment across generations. Johannesburg,
South Africa: JvR Psychometrics.
Taylor, N. & De Bruin, G.P. (2003). Development of the South African
Five Factor Personality Inventory: Preliminary findings. Paper
presented at the Ninth Congress of the Psychological Society of
South Africa (PsySSA), Johannesburg.
Taylor, N., & De Bruin, G.P. (2006). Basic Traits Inventory: Technical
manual. Johannesburg, South Africa: Jopie van Rooyen &
Partners.
Taylor, N., & De Bruin, G.P. (2017). Basic Traits Inventory: Technical
manual (5th Version/Print). Johannesburg, South Africa: JvR
Psychometrics.
Taylor, N., & Vorster, P. (2015). Minnesota Multiphasic Personality
Inventory™-2: Nuclear power facility psychometric properties.
Johannesburg, South Africa: JvR Psychometrics.
Taylor, T.R. (1987). Test bias: The roles and responsibilities of test
users and test publishers. NIPR Special Report Pers – 424.
Pretoria: HSRC.
Taylor, T.R. (1994). A review of three approaches to cognitive
assesssment, and a proposed integrated approach based on a
unifying theoretical framework. SA Journal of Psychology, 24(4),
pp. 184–193.
Taylor, T.R. & Boeyens, J.C.A. (1991). The comparability of the
scores of blacks and whites on the South African Personality
Questionnaire: an exploratory study. SA Journal of Psychology,
21(2), pp. 1–10.
Tchanturia, K., Anderluh, M., Morris, R., Rabe-Hesketh, S., Collier, D.,
Sanchez, P., & Treasure, J. (2004). Cognitive flexibility in anorexia
nervosa and bulimia nervosa. Journal of the International
Neuropsychological Society, 10(4), pp. 513–520.
Teig, N., & Scherer, R. (2016, July 19). Bringing formal and informal
reasoning together: A new era of assessment? Frontiers in
Psychology, 7. Doi: https://doi.org/10.3389/fpsyg.2016. 01097
Temane, Q.M. & Wissing, M.P. (2006). The role of subjective
perception of health in the dynamics of context and psychological
well-being. SA Journal of Psychology, 36(3), pp. 564–581.
Terre’Blanche, M. & Durrheim, K. (1999). Research in practice. Cape
Town: UCT Press.
Theron, C.C. (2007). Confessions, scapegoats and flying pigs:
Psychometric testing and the law. SA Journal of Industrial
Psychology, 33(1), pp. 102–117.
Theron, C. (2009). The diversity-validity dilemma: In search of
minimum adverse impact and maximum utility. SA Journal of
Industrial Psychology, 35(1), pp. 183–195, Art. #765, 13 pages.
doi:10.4102/sajip.v35i1.765. Retrieved on 28 November 2012 from
http://www.sajip.co.za
Theron, D. & Roodt, G. (1999). Variability in multi-rater competency
assessment. Journal of Industrial Psychology, 25(2), pp. 21–27.
Theron, D. & Roodt, G. (2000). Mental models as moderating variable
in competency assessments. Journal of Industrial Psychology,
26(2), pp. 14–19.
Theron, L.C. (2013). Assessing school readiness using the Junior
South African Individual Scales: A pathway to resilience. In S.
Laher & K. Cockcroft (Eds.), Psychological assessment in South
Africa: Research and applications (pp. 60−73). Johannesburg:
Wits University Press.
Thissen, D. & Steinberg, L. (1988). Data analysis using Item
Response Theory. Psychological Bulletin, 104, pp. 385–395
Thomson, L-A. (2007). The relationship between personality traits and
life balance: A quantitative study in the South African corporate
sector. Unpublished master’s dissertation. Johannesburg, South
Africa: University of Johannesburg.
Thorndike.E.L. (1918). The nature, purposes, and general methods of
measurements of educational products. Chapter II in G.M. Whipple
(Ed.), The seventeenth yearbook of the National Society for Study
of Education. Part II. The Measurement of Educational Products.
Bloomington, IL: Public School Publishing Co.
Thornton, G. C., & Gibbons, A.M. (2009). Validity of assessment
centers for personnel selection. Human Resource Management
Review, 19(3), pp. 169−187. Doi:
https://doi.org/10.1016/j.hrmr.2009.02.002.
Thorndike, R.L. & Hagen, E.P. (1977). Measurement and evaluation in
psychology and education (4th ed.). New York: John Wiley &
Sons.
Thornton, G. III & Rupp, D.E. (2006). Assessment Centers in Human
Resource management: Strategies for prediction, diagnosis, and
development. Mahwah, New Jersey: Lawrence Erlbaum.
Thuynsma, C. & de Beer, L.T. (2017). Burnout, depressive symptoms,
job demands and satisfaction with life: Discriminant validity and
explained variance. South African Journal of Psychology, 47(1),
pp. 46−59. doi: https://doi.org/10.1177%2F0081246316638564
Tomita, A., Labys, C. A., & Burns, J.K. (2014). The relationship
between immigration and depression in South Africa: Evidence
from the first South African National Income Dynamics Study.
Journal of Immigrant and Minority Health, 16(6), pp. 1062−1068.
Torgerson, W.S. (1958). Theory and methods of scaling. New York:
John Wiley.
Tracey, T.J.G. (2007). Moderators of the interest congruence-
occupational outcome relation. International Journal of Educational
and Vocational Guidance, 7, pp. 37–45.
Triadó, C., Villar, F., Solé, C. & Celdrán, M. (2007). Construct validity
of Ryff’s scale of psychological well-being in Spanish older adults.
Psychological Reports, 100, pp. 1151−1164.
doi:10.2466/PR0.100.4.1151-1164
Tuerlinckx, F., De Boeck, P., & Lens, W. (2002). Measuring needs
with the Thematic Apperception Test: A psychometric study.
Journal of Personality and Social Psychology, 82(3), pp. 448–461.
doi:10.1037/0022-3514.82.3.448
Ungerer, L.M. & Strasheim, A. (2011). The moderating effect of living
standards on the relationship between individual-level culture and
life satisfaction: A value segmentation perspective. Management
Dynamics, 20(3), pp. 25–50.
Unity technologies (2017). Home page: Imagine, build and succeed
with Unity. Retrieved on 25 September 2018 from Unity:
https://unity3d.com
U.S. Department of Labor (2017, November 2017). O*Net resource
Center. Retrieved on 13 September 2018 from:
http://www.onetcenter.org
Valchev, V.H., Nel, J.A., van de Vijver, F.J.R., Meiring, D., De Bruin,
G.P., & Rothman, S. (2013). Similarities and differences in implicit
personality concepts across enthno-cultural groups in South Africa.
Journal of Cross-Cultural Psychology, 44, pp. 365−388. doi:
10.1177/0022022112443856.
Valchev, V. H., Van de Vijver, F.J.R., Nel, J. A., Rothmann, S., &
Meiring, D. (2013). The use of traits and contextual information in
free personality descriptions across ethnocultural groups in South
Africa. Journal of Personality and Social Psychology, 104, pp.
1077−1091.
Van Dam, N.T. & Earlywine, M. (2011). Validation of the Center for
Epidemiologic Studies Depression Scale-Revised (CESD-R):
Pragmatic depression assessment in the general population.
Psychiatry Research, 186, pp. 128−132. doi: 0.1016/j.psychres.
2010.08.018
Van den Berg, A.R. (1996). Intelligence tests. In K. Owen & J.J.
Taljaard (Eds). Handbook for the use of psychological and
scholastic tests of the HSRC, pp. 157–190. Pretoria: HSRC.
Vandenberg, R.J. (2002). Toward a further understanding of and
improvement in measurement invariance methods and
procedures. Organizational Research Methods, 5(2), pp. 139–158.
doi: 10.1177/1094428102005002001
Vandenberg, R. J., & Lance, C.E. (2000). A review and synthesis of
the measurement invariance literature: Suggestions, practices, and
recommendations for organizational research. Organizational
Research Methods, 3(1), pp. 4–70. doi:
10.1177/109442810031002
Van der Merwe, R.P. (2002). Psychometric testing and human
resource management. SA Journal of Industrial Psychology, 28(2),
pp. 77–86.
Van der Merwe, K., Ntinda, K & Mpofu, E. (2016). African
perspectives of personality psychology. In L.J. Nicholas (Ed.),
Personality psychology (pp. 32−50). Cape Town, South Africa:
Oxford University Press Southern Africa (Pty) Ltd.
Van der Post, W.Z., De Coning, T.J. & v.d.M. Smit, E. (1997). An
instrument to measure organisational culture. South African
Journal for Business Management, 4, pp. 147–165.
Van de Vijver, F.J.R. (2002). Cross-cultural assessment: Value for
money? Applied Psychology: An International Review, 51(4), pp.
545–566.
Van de Vijver, F.J.R. (2016). Test adaptations. In F.T. Leong, D.
Bartram, F.M. Cheung, K.F. Geisinger, & D Iliescu (Eds.)., The ITC
international handbook of testing and assessment (pp. 364−376).
New York, NY: Oxford University Press.
Van de Vijver, F.J.R. & Leung, K. (1997) Methods and data analysis
for cross-cultural research. Sage: New York
Van de Vijver, F.J.R., & Leung, K. (2011). Equivalence and bias: A
review of concepts, models, and data-analytic procedures. In D.
Matsumoto & F.J.R. van de Vijver, Cross-cultural Research
Methods in Psychology (pp. 17–45). Cambridge, NY: Cambridge
University Press.
Van de Vijver, F.J. R., & Matsumoto, D. (2011). Introduction to the
methodological issues associated with cross-cultural research. In
D. Matsumoto & F.J.R. van de Vijver, Cross-cultural Research
Methods in Psychology (pp. 1–16). Cambridge, NY: Cambridge
University Press.
Van de Vijver, F.J.R., & Tanzer, N.K. (2004). Bias and equivalence in
cross-cultural assessment: an overview. European Review of
Applied Psychology, 54(2), pp. 119-135.
https://doi.org/10.1016/j.erap.2003.12.004.
Van Dierendonck, D. (2004). The construct validity of Ryff’s scales of
psychological well-being and its extension with spiritual well-being.
Personality and Individual Differences, 36, pp. 629−643.
doi:10.1016/S0191-8869(03)00122-3
Van Eeden, R. (1991). Manual for the Senior South African Individual
Scale – Revised (SSAIS–R). Part I: Background and
standardization. Pretoria: HSRC.
Van Eeden, R. & Prinsloo, C.H. (1997). Using the South African
version of the 16PF in a multicultural context. SA Journal of
Psychology, 27, pp. 151–159.
Van Eeden, R. & Mantsha, T.R. (2007). Theoretical and
methodological considerations in the translation of the 16PF5 into
an African language. SA Journal of Psychology, 37(1), pp. 62–81.
Van Eeden, R. & Van Vuuren, J. (2017). The cognitive processing
potential of infants: Exploring the impact of an early childhood
development programme. The South African Journal of Childhood
Education 7(1), a499. doi: https://doi.org/10.4102/sajce.v7i1.499.
Van Eeden, R. & Visser, D. (1992). The validity of the Senior South
African Individual Scale – Revised (SSAIS–R) for different
population groups. SA Journal of Psychology, 22, pp. 163–171.
Van Heerden, R. (2007). Exploring Normal South African and British
Children: A Comparative Study utilizing the Griffiths Mental
Development Scales – Extended Revised. Unpublished master’s
treatise. Nelson Mandela Metropolitan University.
Van Lingen, J.M., Douman, D.L., & Wannenburg, J.M. (2011). A
cross-sectional exploration of undergraduate nursing student
wellness and academic outcomes at a South African higher
education institution. SA Journal of Psychology, 41(3), pp. 396–
408.
Van Ommen, C. (2005). Putting the PC in IQ: Images in the Wechsler-
Adult Intelligence Scale – third edition (WAIS III). SA Journal of
Psychology, 35(3), pp. 532–554.
Van Rooyen, K. (2005). The Performance of South African and British
Children on the Griffiths Mental Development Scales – Extended
Revised: A Comparative Study. Unpublished master’s treatise.
Nelson Mandela Metropolitan University.
Van Rooyen, K. & Nqweni, Z.C. (2012). Culture and Posttraumatic
Stress Disorder (PTSD): a proposed conceptual framework. SA
Journal of Psychology, 42(1), pp. 51–60.
Van Rooyen, S.W. (2000). The psychosocial adjustment and stress
levels of shift workers in the South African Police Service, Flying
Squad. Unpublished master’s treatise, University of Port Elizabeth.
Van Tonder, C. & Roodt, G. (2008). Organisation development:
Theory and practice. Pretoria: Van Schaik.
Van Wijk, C.H. (2012). Self-reported generalised anxiety and
psychomotor test performance In healthy South Africans. SA
Journal of Psychology, 42(1), pp. 7–14.
Van Wijk, C.H. & Meintjies, W.A.J. (2015a). Grooved pegboard for
adult employed South Africans: Normative data and Human
Immunodeficiency Virus associations. South African Journal of
Psychology, 45(4), pp. 521–535. doi:
https://doi.org/10.1177/0081246315587692
Van Wijk, C.H. & Meintjies, W.A.J. (2015b). Grooved pegboard for
adult employed South Africans: Socio-demographic influences.
South African Journal of Psychology, 45(4), pp. 536–550. doi:
https://doi.org/10.1177/0081246315587693
Van Zyl, C.J.J., & Becker, J. (2012). Investigating the ethnic
equivalence of the 16PF5-SA. Paper presented at the XXX
International Congress of Psychology, CTICC, Cape Town, South
Africa.
Van Zyl, C.J.J., & De Bruin, G.P. (2016). Work-related Risk and
Integrity Scale technical manual (2016 edition). Johannesburg,
South Africa: JvR Psychometrics.
Van Zyl, C.J.J., & Taylor, N. (2012). Evaluating the MBTI® Form M in
a South African context. SA Journal of Industrial Psychology/SA
Tydskrif vir Bedryfsielkunde, 38(1), Art. #977, 15 pages.
http://dx.doi.org/10.4102/sajip.v38i1.977 Access at
http://www.sajip.co.za/index.php/sajip/article/viewFile/977/1249
Van Zyl, E.S. (2004). Die effek can biografiese veranderlikes op die
vlakke en oorsake van stress by ‘n groep finalejaar-mediese
student. South African Family Practice, 46(4), pp. 26–29.
Van Zyl, E. & Visser, D. (1988). Differential item functioning in the
Figure Classification test. SA Journal of Industrial Psychology,
24(2), pp. 25–33.
Van Zyl, E.S. & Van der Walt, H.S. (1991). Manual for the Experience
of Work and Life Circumstances Questionaire. Pretoria: HSRC.
Van Zyl, L. E., & Rothmann, S. (2012). Flourishing of students in a
tertiary education institution in South Africa. Journal of Psychology
in Africa, 22(4), pp. 593−599.
Van Zyl, P.M., Joubert, G., Bowen, E., Du Plooy, F., Francis, C.,
Jadhunancan, S., Fredericks, F. & Metz, L. (2017). Depression,
anxiety, stress and substance use in medical students in a five-
year curriculum. South African Journal of Health Professions
Education, 9(2), pp. 67−72. doi:10.7196/AJHPE.2017.v9i2.705
Veenhoven, R. (1991). Is happiness relative? Social Indicators
Research, 24(1), pp. 1–34.
Venter, A. (2006). The relationship between the big five personality traits
and emotional intelligence competencies among employees in a
Durban based medical aid administrator. Unpublished master’s
thesis. Durban, South Africa: University of KwaZulu-Natal.
Vernon, P.A. (1990). An overview of chronometric measures of
intelligence. School Psychology Review, 19(4), pp. 399–410.
Vincent, K.R. (1991). Black/White IQ differences: Does age make the
difference? Journal of Clinical Psychology, 47(2), pp. 266–270.
Visser, D., & Du Toit, J.M. (2004). Using the Occupational Personality
Questionnaire (OPQ) for measuring broad traits. SA Journal of
Industrial Psychology, 30(4), pp. 65–77.
Visser, D., & Viviers, R. (2010). Construct equivalence of the OPQ32n
for Black and White people in South Africa. SA Journal of Industrial
Psychology/SA Tydskrif vir Bedryfsielkunde, 36(1), pp. 1–11. Art.
#748, 11 pages. DOI: 10.4102/sajip.v36i1.748
Vogt, L. & Laher, S. (2008). The five-factor model of personality and
individualism/collectivism in South Africa: An exploratory study.
Psychology in Society, 37, pp: 39−54.
Vogt, L. & Laher, S. (2009). The relationship between
individualism/collectivism and the Five Factor Model of personality:
An exploratory study. Psychology in Society, 37, pp. 39–54.
Von Stumm, S. (2012). You are what you eat? Meal type, socio-
economic status and cognitive ability in childhood. Intelligence, 40,
pp. 576–583.
Vrana, S. R., & Vrana, D.T. (2017). Can a computer administer a
Wechsler Intelligence Test? Professional Psychology: Research
and Practice, 48(3), pp. 181−198.
Vygotsky, L.S. (1978). Mind in society: The development of higher-
order psychological processes. Cambridge, MA: Harvard
University Press.
Wang, P.S., Beck, A.L., Berglund, P., McKenas, D.K., Pronk, P.N.,
Simon, G.E. & Kessler, R.S. (2004). Effects of major depression on
moment-in-time work performance. The American Journal of
Psychiatry, 161(10), pp. 1885–1892.
Ward, C.L., Lombard, C.J., & Gwebushe, N. (2006). Critical incident
exposure in South African emergency services personnel:
Prevalence and associated mental health issues. Emergency
Medical Journal, 23, pp. 226–231. doi: 10.1136/emj.2005.025908
Watkins, C.E., Jr., Campbell, V.L., Nieberding, R., & Hallmark, R.
(1995). Contemporary practice of psychological assessment by
clinical psychologists. Professional Psychology: Research and
Practice, 26, pp. 54–60.
Watson, D., Clark, L.A., & Tellegen, A. (1988). Development and
validation of brief measures of positive and negative affect: The
PANAS scales. Journal of Personality and Social Psychology,
54(6), pp. 1063–1070.
Watson, M., & Kuit, W. (2007). Postmodern career counselling and
beyond. In K. Maree (ed.), Shaping the story: A guide to facilitating
narrative counselling (pp. 73–86). Pretoria, South Africa: Van
Schaik,
Wechsler, D. (1944). The measurement of adult intelligence (3rd ed.).
Baltimore: Williams & Wilkins.
Wechsler, D. (1958). The measurement and appraisal of adult
intelligence (4th edition). Baltimore: Williams and Wilkins.
Wechsler, D. (2014). Wechsler Adult Intelligence Scale-Fourth Edition.
Bloomington, MN: NSC Pearson.
Weiss, L. (2003). WISC-IV: Clinical reality and the future of the four-
factor structure. Paper presented at the 8th European Congress of
Psychology, Vienna, Austria, 6–11 July 2003.
Weiss, D.J., Dawis, R.V., Engand, G.W. & Lofquist, L.H. (1967).
Manual for the Minnesota Satisfaction Questionnaire. Minneapolis:
University of Minnesota.
Wessels, S.M. (2011). The Children’s Auditory Performance Scale
(CHAPS): An Exploration of its Usefulness in Identifying Central
Auditory Processing Disorder in a South African Clinical Setting.
Unpublished Doctoral thesis, Nelson Mandela Metropolitan
University, Port Elizabeth.
Westaway, M. S., & Maluka, C.S. (2005). Are life satisfaction and self-
esteem distinct constructs? Black South African perspectives.
Psychological Reports, 97, pp. 567–575.
Whetzel, D. L., & McDaniel, M.A. (2009). Situational judgment tests:
An overview of current research. Human Resource Management
Review, 19, pp. 188–202. doi:10.1016/j.hrmr.2009.03.007
Whipple, G.M. Whipple (Ed.), The seventeenth yearbook of the
National Society for Study of Education. Part II: The Measurement
of Educational Products. Bloomington, IL: Public School Publishing
Co.
WHO Global
Status
Report
on
Alcohol
(2004).
Retriev
ed 27
September. 2008, from
http://www.who.int/substance_abuse/publications/en/south_africa.p
WHO Online Question and Answer: What is Mental Health? (2008).
Retrieved 27 September, 2008, from
http://www.who.int/features/qa/62/en/index.html
Wicken, A.W. (2000). Wellness in industry. Work, 15, pp. 95–99.
Wiggett, C. (2006). Guide dog ownership and psychological well-
being. Unpublished doctoral dissertation, University of
Stellenbosch, South Africa.
Wilcocks, R.W. (1931). The South African Group Test of Intelligence.
Stellenbosch: University of Stellenbosch.
Williamson, D.M., Bauer, M., Steinberg, L.S., Mislevy, R.J., Behrens,
J.T. and DeMark, S.F. (2004). Design rationale for complex
performance assessment. International Journal of Testing, 4(4),
pp. 303–332.
Willig, A.C. (1988). A case of blaming the victim: The Dunn
monograph on bilingual Hispanic children on the U.S. mainland.
Hispanic Journal of Behavioral Sciences, 10, pp. 219–236.
Wilson, M (2004). Developing measures: An item response modeling
approach. Mahwah, NJ: Lawrence Erlbaum Associates.
Winter, D.G., & Stewart, A.J. (1977). Power motive reliability as a
function of retest instructions. Journal of Consulting and Clinical
Psychology, 45, pp. 436–440.
Winter, D.G., Stewart, A.J., John, O.P., Klohnen, E.C., & Duncan, L.E.
(1998). Traits and motives: Toward an integration of two traditions
in personality research. Psychological Review, 105, pp. 230–250.
Wise, S.L., Barnes, L.B., Harvey, A.L., & Plake, B.S. (1989). Effects of
computer anxiety and computer experience on computer-based
achievement test performance of college students. Applied
Measurement in Education, 2, pp. 235–242.
Wissing, J.A.B., Du Toit, M.M., Wissing, M.P., & Temane, Q.M.
(2008). Psychometric properties of various scales measuring
psychological well-being in a South African context: The FORT 1
project. Journal of Psychology in Africa, 18(4), pp. 511–520.
Wissing et al. (2010). Validation of three Setswana measures for
psychological wellbeing. SA Journal of Industrial Psychology,
36(2), pp. 1–8.
Worley, J. A., Vassar, M. Wheeler, D.L. & Barnes, L.B. (2008). Factor
structure of scores from the Maslach Burnout Inventory: A review
and meta-analysis of 45 exploratory and confirmatory factor-
analytic studies. Educational & Psychological Measurement, 68,
pp. 797–823.
Wyse, A.E. & Mapuranga, R. (2009). Differential Item Functioning
analysis using Rasch item information functions. International
Journal of Testing, 9, pp. 333–357.
Xu, G., Wang, C., & Shang, Z. (2016). On initial item selection in
cognitive diagnostic computerized adaptive testing. British Journal
of Mathematical and Statistical Psychology, 69, pp. 291−315.
Yang, W., Lance, C.E. & Hui, H.C. (2006). Psychometric properties of
the Chinese Self-Directed Search (1994 edition). Journal of
Vocational Behavior, 68(3), pp. 560–576.
Yeo, R.A., Ryman, S.G., Pommy, J., Thoma, R.J., & Jung, R.E.
(2016). General cognitive ability and fluctuating asymmetry of brain
surface area. Intelligence, 56, pp. 93−98.
Zenisky, A.L. & Sireci, S.G. (2002). Technological innovations in
large-scale assessment. Applied Measurement in Education, 15,
pp. 337–362.
Zhou, J. (2009). A review of assessment engineering principles with
select applications to the Certified Public Accountant Examination.
Technical report prepared for The American Institute of Certified
Public Accountants. Retrieved 30 November 2012 from
http://www.aicpa.org/becomeacpa/cpaexam/psychometricsandscori
a_review_of_assessment_engineering.pdf
Zhu, J. & Tulsky, D.S. (2000). Co-norming the WAIS-III and WMS-III:
Is there a test-order effect on IQ and memory scores? Clinical
Neuropsychology, 14(4), pp. 461–467.
Zickar, M.J., Cortina, J., & Carter, N.T. (2010). Evaluation of
measures: Sources of sufficiency, error, and contamination. In J.L.
Farr & N.T. Tippins, Handbook of Employee Selection (pp. 399–
416). New York: Psychology Press (Taylor & Francis Group).
Zikmund, W.G., Babin, B.J., Carr, J.C., & Griffin, M. (2010). Business
research methods (8th edition). Mason, OH: South-Western
Publishing.
Zinn, S. & Langdown, N. (2011). E-book usage amongst academic
librarians in South Africa. South African Journal of Libraries and
Information Science, 77(1), pp. 104−115.
Zuilkowski, S.S., McCoy, D.C., Serpell, R., Matafwali, B. & Fink, G.
(2016). Dimensionality and the development of cognitive
assessments for children in sub-Saharan Africa. Journal of Cross-
Cultural Psychology, 47(3), pp. 341–354. doi
http://dx.doi.org/10.1177/0022022115624155).
Zunker, V.G. (2016). Career counselling: A holistic approach. Boston,
MA: Cengage Learning.
Index
Page numbers in italics indicate information contained in
illustrations or tables.
A
ability 183–184
cognitive 292
groups 189–190
ability level and reliability 65–66
Ability, Processing of Information, and Learning Battery 27,
273 absenteeism 206
and HIV/Aids 207
and ill health 206
and mental illness 208
academic achievement 75–76, 290–
291 academic counselling 239
accountability 308
educational 290
acculturation 322
acquiescence bias 65
administering measures 144–145, 145, 328
administering measures, practitioner’s duties 161
assessment anxiety 158–159
control over groups 156
establishing rapport 156–158
managing irregularities 160–161
mental disabilities 163–164
motivating test-takers 156
physical disabilities 163
providing instructions 159–160
recording behaviour 161
relationship between practitioner, test-taker 155–158
scientific attitude 155–156
time limits 160
young children 162–163
administering measures, practitioner’s duties after assessment
166 collecting, securing materials 164–165
recording, scoring, interpreting 165
administering measures, preparation 146,
154
checking conditions 148–150
checking materials, equipment 147–
148 ethical issues 147
familiarity with measures, instructions
148 informed consent 153–154 linguistic
factors 152–153
personal circumstances, timing 150–
151 planning sequence, length 151–152
and test sophistication 153
administrative error and reliability 66
Adolescent Personality Questionnaire
227 adult assessment 158
Adult Neuropsychological Questionnaire 147
adverse impact 273
Affectometer-2 Scale 214
affect, positive and negative 203
African-centred assessment 8–9
African-centred item development 342–343
African-centred measures 337, 337–338, 339
African-centred non-verbal cognitive reasoning stimuli 91
African-centred, Western-oriented approaches combined 335, 336,
336–337, 341–342
age-related changes 315–316
alcohol abuse 208
alternate-form reliability 62
Annual National Assessments 290
answer mechanisms, protocols 92–93
anti-test lobby 34, 35
anxiety 205
assessment 158–159
and test-taking 158–159, 330
apartheid 25–28
apperception 236–237
aptitude
and career counselling 244
school-age learners 291
see also intelligence measures, aptitude
Aptitude Test for School Beginners 193, 194,
284 artificial intelligence 257, 262–263
assessment 3–4
African-centred 8–9
definition 4
shift in role of professionals 4–5 and
technological developments 4–5
and tests, testing 4–5 tools 4
B
Basic Traits Inventory 227–228, 228,
337 and careers 246–247
Bayley Scales of Infant Development 185,
285 Beck Depression Inventory 215
Beck Depression Scale 146
behavioural cluster analysis 305–306
Bender-Gestalt Test 195 bias
construct 118
and construct validity 332–333
definition 108
extremity 65
prediction 116–118
see also equivalence, statistical approaches
to Big Five model of personality traits 227
bilingual assessment 104 bilingual test-taker
approach 106
Bill of Rights, test-taker’s 309
Binet-Simon Scale 16–17, 170
biological context 315–317
biological intelligence 172
biological measures of intelligence 173–
174 bipolar disorder 205
Broad Band Competency Assessment Battery
274 burnout 209, 210, 218–219
C
Campbell Interest and Skills Survey 274
career, academic counselling 239
Career Construction Interview 252
career construction theory and life-design counselling 251–
253 career-counselling assessment 242–243
and aptitude 244
in a changing environment 253
developmental approach 248–249
and intelligence 244
and personality 246–247
person-environment fit approach 243–248
systems approach 249–251
and values 247
and vocational interests 245–246
Career Interest Profile 252, 274
career maturity 248
Career Path Finder 274
category scales 42, 42
Cattell-Horn-Carroll model 174
causal interpretation 303
Caveon test security 331
central tendency, measures of 47–48, 48–49
Centre for Epidemiologic Studies Depression Scale 215–216,
340 cheating 331–332
minimising 149–150
children
administering assessments 162–163
conveying test results 309–310
and rapport 157
research 282–283
Children’s Personality Questionnaire 226–
227 Chinese Personality Inventory 22
chirology 14
classification of psychological measures 132–134, 135, 135
Clerical Test Battery 273
clinical interpretation 306–307
clinical practice 34–35
closed questions 296
coaching 139
coefficient
of determination 78
of equivalence 62
of internal consistency 62
of stability 62
Coefficient Alpha 63, 63
cognitive ability vs cognitive functions
292 see also under intelligence
cognitive diagnostic assessment 171
cognitive development 174
cognitive functioning 169–170, 292
Cognitive Process Profile 273
comparative rating formats 46
competency assessment approach 272–273
Competency Assessment Series 274
competency-based assessment 23–24
competing values 140, 140–141, 141–142, 296–297
composite scale 46
computer-based and Internet-delivered assessment 255–256,
263 artificial intelligence 262–263
and cloud storage 259
ethical and legal use 264–267
gamified assessments 260–261
important terms 256–257
and simulations 257, 260–262
in South Africa 258
in the twenty-first century 258–263
virtual content 261
computer-based tests 19–21, 29, 32,
87 administration 256
interpretation 257
scoring 256
computerised adaptive testing 170, 256
computerised simulation assessments
257 concept-driven adaptations 102
conceptual intelligence 174 concurrent
validity 73–74
confidentiality 307–308
configural invariance 114
confirmatory factor analysis 112–113
co-norming of measures 52
consistency of measurement 60–61
constant-sum scales 44, 44
construct bias 118
construct-identification procedures 72–73
constructional performance 196
constructivism 243, 249, 251
construct maps 86–87
content-description procedures 71–72
content validity 71–72
contextual intelligence 172, 174
continuous data 40
convergent validity 73
conveying test results 307
children’s 309–310
ethical considerations 307–308
methods of 308–309
corrected reliability coefficient 62, 62
counsellors, registered 129
counterproductive behaviour 238–239
criterion
contamination 76, 78
deficiency, relevance 76
keying 86, 231
measures 75–77
criterion-prediction
procedures 73–75
validity 73
criterion-referenced achievement tests 291
criterion-referenced assessment 67
criterion-referenced measures 85
interpretation 307
Critical Reasoning Test Battery 273 cross-
cultural test adaptation 21, 101–102
challenges in South Africa 103–104
equivalence of adapted versions of measures 104–
107 equivalence of measures 104–105
linguistic equivalence 105–107
multicultural test development 120–122
reasons for 101
statistical approaches to equivalence 107–
119 steps for maximising success 120 types
of 102, 102–103, 103
cross-cultural use of personality assessment measures 232–233
cross-validation 75
crystallised intelligence 173
culturally appropriate measures 341, 341
African-centred item development 342–343
capacity-building 345
delineating norm groups 343–344
ethical practices 345–347
qualitative measures 344
repositories of research studies 344–345
culturally sensitive Rorschach Comprehensive System 35
cultural meaningfulness 334–335
combining African-centred, Western-oriented approaches 335,
336, 336–337, 341–342
indigenous, African-centred measures 337, 337–338, 339
roadmap for next phase 341–347
Western-oriented measures for SA 339, 340, 340–
341 culture and results 321–323
culture-free, culture reduced, culture-common tests
21 Curriculum Assessment Policy Statements 291
Customer Service Aptitude Profile
273 cut-off scores 56–57, 302
D
data
continuous 40
displaying 47, 47
generating nominal 44, 44
generating ordinal 45, 45
Denver Developmental Screening Test 282
depersonalisation 218 depression 205
E
education 28–29
and aptitude 193–194
higher 34, 293–294
scholastic tests 197
educational achievement measures 52–53, 290–
291 educationally disadvantaged 157
educationally focused screening 283–
284 education, school-age learners 288
achievement measures 290–291
aptitude measures 291
cognitive measures 291–292
dynamic assessment 292–293
and educational accountability 290
personality-related measures 292
effort and test-taking 330
emic approach 233
emotional intelligence 86, 89, 92, 171, 176
emotionality 159
Employee Assistance Programme
211 Employment Equity Act
and assessment, modern: South Africa 30–
32 and computer-based tests 266 and
prediction bias 117–118
and reliability 61
environmental mastery 203
equivalence, statistical approaches to 107
of adapted versions of measures 104–
107 and bias 107–109
and bias at item level 109–112
and bias in factor structures between groups 112–
115 construct bias and construct equivalence 118
and invariance 108
and measurement bias 108
and measurement equivalence 108–109
method bias 119
nomological bias 119
and prediction bias 116–118
errors 41–42
error variance invariance 115
essay questions 291
ethical considerations
administering measures, preparation 147 computer-based
and Internet-delivered assessment 264–267 conveying test
results 307–308
see also fair and ethical practices
ethical dilemmas 139
ethical practices
culturally appropriate measures 345–347
statutory control of assessment measures, SA 136–
139 see also fair and ethical practices
etic approach 232–233
eudaimonic perspective 202, 203
evaluation of psychological measures 132–134, 135,
135 evaluative interpretation 303–304
even-numbered vs odd-numbered rating 46–47
evidence 5
expectancy
tables 56, 56
of tester 156
Experience of Work and Life Circumstances Questionnaire 216–
217 experimental psychologists 16
experimental version of measure 84, 84
expert witnesses 298–299
exploratory factor analysis 112
extraverts, introverts 230
extremity bias 65
F
face validity 71–72
factor analysis 72–73, 224
factor-analytic strategy 226
factorial validity 72–73
factor variance-covariance invariance
115 factual-driven adaptations 102
factual vs inferred information 296
fair and ethical practices 136
and assessment practitioners 136–137
competing values 140, 140–141, 141–
142 defining 136
ethical dilemmas 139
in industry 140, 140–141, 141–142
multiple constituents 139–140, 140–141, 141–142
preparing test-takers 138–139
professional practices 137–138
responsibilities of organisations 142–
143 rights, responsibilities of test-takers
138
faking bad/good 231, 330–331
familiarity/recognisability-driven adaptations
103 Family Environment Scale 301 feeling
individuals 230
15 Factor Questionnaire 227
15 Factor Questionnaire Plus 274
fluid intelligence 173
Folstein’s Mini Mental Status Examination 146
forced-choice items 88
forced-choice scales 44
generating nominal data 44, 44
generating ordinal data 45, 45
forensic psychologists and test use 131
format-driven adaptations 103 Full
Scale IQ 180
full-score equivalence 113–114
functional equivalence 113
G
gamified assessments 257, 260–261, 278
General Ability Measure for Adults 273
General Adaptability Battery 26
General Health Questionnaire 214–215
Generalised Anxiety Disorder scale 216
General Reasoning Test Battery 273
General Scholastic Aptitude Test 188, 189, 190
Global Intelligence Scale 180–181
Goodenough-Harris Drawing Test 283
grade equivalents 52–53
Graduate Talent Simulations 274,
275 graphic rating scales 44, 44
graphology 14–15
Griffiths III 181, 182–183, 185
Griffiths Mental Development Scales – Extended Revised 181
group characteristics 279
group processes 278
group structure 279
Grover-Counter Scale of Cognitive Development 181, 183, 185
Guttman scales 45, 45
H
halo effect 65
Hamilton Depression Scale 146
happiness 202
Health Professions Act 28, 32, 127–
128 classifying a measure 134–135
and linguistic factors 152–153
hedonistic perspective 202
higher education 34
well-being of students 211–213
High School Personality Questionnaire
227 HIV/Aids
test battery 146–147
in the workplace 206–207
Hogan assessment series 274
Hogan Development Survey 239
Hogan Motives, Values, and Preferences Inventory
274 Hogan Personality Inventory 229 home
environment 323–324
Home Screening Questionnaire
301 honesty contract 332
Human Sciences Research Council 31, 34
humorology 13–14
I
identity 222
in-basket test 278
incremental validity 74
Individual Scale for General Scholastic Aptitude
180 industrial psychologists and test use 130
industry 34, 271–272
and aptitude 194
assessment of individuals 272–278 assessment
of organisations 279–280 assessment of
workgroups, work teams 278–279
ethical assessment practices in 140, 140–141, 141–142
mental illness in 207–208
personnel selection 272–275
infant intelligence, predictability of 315
infant, preschool developmental assessment
280 developmental assessment, shortcomings
285 developmental screening 282–283
diagnostic measures 284–285 educationally
focused screening 283–284
functions assessed 281
info. from other observations 285–286
why we assess development 280–281
inferred information 296 influence of tester
330 information-processing and
intelligence 174 informed consent 153–154
informed opinions 300
J
Job Demands-Resources model 208–209,
209 job satisfaction 210
Job Specific Solutions 274
judging vs perceiving individuals 230
Jung Personality Questionnaire 230–
231
Junior South African Individual Scales 180–181, 183,
185 juvenile delinquents 157
K
K-ABC 185
Kant, Immanuel 38
known-groups validity 74
L
language
bias 89
and results 319–321
see also under linguistic
language and test adaptation 102, 103–104
back-translation 106
forward-translation 105–106
Learning Orientation Index 274
learning potential 319
and intelligence 175
Learning Potential Computerised Adaptive Test 89, 171, 188–189,
190, 273, 337
leniency bias 65
life-design counselling 251–253
Life Orientation Test –Revised 216
Life Role Inventory 248–249
life satisfaction 203
Likert-type scales 42, 43, 43
linearly transformed standard scores
53 linguistic equivalence 105–107
empirical investigations 106–107
judgemental approaches 105–106
linguistic factors 152–153
administering measures, preparation 152–
153 see also under language
linguistics-driven adaptations 102
Locus of Control Inventory 217
M
malingering 330–331
Mantel-Haenszel procedure 110
Maree Career Matrix 246
Maslach Burnout Inventory 218–219
mastery testing 67
maximal and typical performance
273 MB-10 245
McCall’s T-score 54, 54
McCarthy Scales of Mental Abilities 284–285
McMaster Health Index Questionnaire 217
mean 48, 48
measurement 38
absolute zero 39
categories of levels 39–41
definition 38
equal intervals 38–39, 39
errors 41–42
levels of 38–39
magnitude 38
properties of scales 38–39
psychology professionals who may use 128–
131 measurement and statistical concepts 37–38
definition of measurement 38
measurement invariance approaches 114–115
measurement scales 42
deciding on format 46–47
different scaling options 42–45
measures, developing 82
administration 96
assemble, pre-test experimental version 92–93
classification 97
item-analysis phase 93–95
item development, creation, sourcing 90–
92 planning phase 83–90
publish, refine continuously 96–98
revise, standardise final version 95–96
steps in 82–83, 84, 84–98
technical evaluation, establishing norms
96 measures, evaluating 98–99 Mechanical
Insight test 192
mechanical, non-mechanical interpretation 304–306
median 48, 48
mental age 315–316
mental-age scales 52
mental disabilities 163–164
practitioner’s duties 163–164
mental disorders, deficiencies 16
mental illness 204–205
in the workplace 207–208
mental status examination 296
meta-analysis 75
method bias 89
methodological considerations 328
influence of tester 330
patterns in scores 328–330
standardised procedures 328
status of test-taker 330–332
metric equivalence 113
metric invariance 114
Minnesota Multiphasic Personality Inventory 18, 231–
232 Minnesota Satisfaction Questionnaire 219 mode 48
N
negative past experiences 157
negative self-orientated thoughts 159
neuropsychologists and test use 131
nominal measurement levels 40
non-mechanical interpretation 304–
306 non-response errors 64–65
normalised standard score 54 normally
distributed test scores 47, 47 norm-
driven adaptations 102
norm groups, establishing 51–52 norm-
referenced tests, interpretation 306–307
norms 51–57
definition 51
norm scores, scales 52–55
interrelationships of 55
issues with use of 57
setting standards, cut-off scores 55–57
O
objective, subjective formats 88 Occupational
Interest Profile 274 Occupational Personality
Profile 274 Occupational Personality
Questionnaire 229, 274 open and closed
questions 296 open-ended items 88
opinions 8, 300
ordinal measurement levels 40
organisational characteristics 279
organisational processes 279
organisational research 280
organisational structures 279
organisations
assessment of 279–280
ethical responsibilities 142–143
output-based approaches 272–273,
275 overcrowding 323–324
P
paired-comparison scales 44, 44
palmistry 14
parameter estimates 113
patterns in scores 328–330
Pearson product-moment correlation 50, 50
Pearson product-moment correlation coefficient 50, 50
perceiving individuals 230
percentages and percentiles 53
percentile rank score 53
perceptual measures 195–196
Perceptual Reasoning Index 180, 181–182
performance
constructional 196
job 76
maximal and typical 273
motor 196
ratings 275–276
in specialised training 76
performance-based items 88
personal growth 204
personality
and career-counselling 246–247
and counterproductive behaviour 238–
239 early views on 13–14
personality assessment 221 conceptual
scheme 222–223 cross-cultural use of
measures 232–233 relatively stable
personality traits 223–235 structured
measures 18
uses and applications 239–240
personality assessment, motives and concerns 235–
236 explicit motives 236
implicit motives 236–237
projective methods 236–238
Personality Inventory for Children 232
person-environment fit 243–248
personnel selection 239, 272–275
phrenology 14
physical disabilities 163
physical impairments 316–317
physical well-being 150
physiognomy 13
portfolio assessment 291
Positive Affect Negative Affect Scales
214 poverty 324
power measures 6 practice effects
332 practice examples, tests 139
pragmatics-driven adaptations 103
prediction bias and equivalence
116
and Employment Equity Act 117–118
procedures to establish 116, 117, 117
predictive validity 74
preparation prior to session 146
selecting measures for test battery 146–147
preschoolers 157, 162–163 prisoners,
juvenile delinquents 157
Processing Speed Index 180, 182
Professional Board for Psychology 32, 36, 132–136
proficiency tests 197
profile analysis 304
Progressive International Reading Literacy Study
290 projective assessment techniques 222 projective
tests 88
proximal factors 323
psychodiagnostic assessment 294–295
interviews 295–297
limitations of 299–300
psychological knowledge, expertise 298–
299 psycholegal assessment 298–299
psycholegal contexts 295 psychological
assessment 24
psychological measurement see
measurement psychologists and test use 130
psychometric intelligence 172
psychometrics 4, 15–16
Psychometrics Committee 32, 132–135
psychometrists and test use 129
psychopathology 318
psychotechnicians and test use 128
psychotherapists and test use 129
psychotherapy 239
public protection 135–136
purpose in life 204
purposive falsification 65
Q
quality of life 202
Quanti-Skill Accounting Assessment 273
question- vs statement-item formats 46
R
random errors 41
range 49
range restriction and reliability 65,
66 rapport 156–158, 330 Rasch
model 110
ratings 76
ratio 40
rational method of measure 86
Raven Progressive Matrices 21, 191
Raven’s Advanced Progressive Matrices 178, 339
recognisability-driven adaptations 103 regression
analysis 79, 79
Regulations Defining the Scope of the Profession of
Psychology 127–128, 128–131, 131
reliability coefficient 61, 62–66
inter-scorer 63
intra-scorer 64
magnitude of 67
reliability of a measure 59–60
approaches to estimating 64
defining 60–61, 61
factors affecting 64–66
interpretation of 66–67
types of 61–64
reporting results, written form 310–311
research
cognitive functioning 170–171, 184–187, 190–191, 194–
197 computer-based assessment 260–262
cross-cultural measures 339–341, 344–
345 organisational 280
personality assessment 227, 232, 233,
239 well-being 203
research psychologists and test use
130 respondent error 64–66, 66
response bias 64–65, 89 response set
222
results, factors affecting 313–314
results in context 314
age-related changes 315–316
biological context 315–317
culture 321–323
environmental factors 323–325
home environment 323–324
intrapsychic context 317–318
language 319–321
physical impairments 316–317
psychopathology 318
schooling 318–319
socio-economic status 324–325
test-wiseness 325–327
transient conditions 318
urbanisation 325
Rey-Osterrieth Complex Figure Test 146
rights, responsibilities of test-takers 138
role saliency 248
Rorschach Comprehensive System 35
Rorschach Inkblot Test 88, 160, 237–238,
340 Rosenthal effect 156
Ryff’s Psychological Well-being Scales 217–218
S
sample homogeneity 78
San people 27
Satisfaction with Life Scale 216
scalar invariance 114–115
scatter 329
scholastic tests 197
school-age children 157, 291–292
School-entry Group Screening Measure 288–289
School-readiness Evaluation by Trained Testers 284
scientific attitude 155–156
self-acceptance 203
Self-directed Search 245
self-selection bias 64–65
semantic differential scales 43, 43
Senior South African Individual Scale 181, 183,
186 Sense of Coherence 216
sensing vs intuitive individuals 230
sentence-completion tests 238
serious games 260–261, 278
SHL Verify ability tests 192, 192, 194
simple regression equation 50, 50
simulations 257, 260–262, 277
simultaneous multicultural test development 121
simultaneous multilingual test development 22
single attribute vs comparative rating formats 46
single dimensional vs composite scale 46
single general factor 173
Situational Judgment Tests 277
situational tests 276–278
16 Personality Factor
Questionnaire and careers 246
and cultural groups 340
personality assessment 224–225, 225, 225–
227 and personnel selection 274
skewed test scores 47, 47
social-desirability bias 65
social intelligence 172
socially desirable responses 222–223
socio-economic status 324–325
Sources of Work Stress Inventory 218
South African Burnout Scale 219, 338
South African Career Interest Inventory 245, 338
South African Personality Inventory 121, 227, 233, 234–235, 337
speech impairments 317
speeded measure, reliability of 65
speed measures 6
speed tests 88–89
split-half reliability 62, 62
standard deviation 49–50, 50
standard error
of estimation 78, 78–79, 79
of measurement 329
standardised conditions 16
standardised procedures 328
standardised tests 18
Standard Measurement Error 67,
67 standard normal distribution 51
standard scores 53–54 stanine
scale 54, 54
State-Trait Anxiety Inventory 214
statistical concepts 47–51
status of test-taker 330–332
statutory control of assessment measures,
SA exercise of control 127–136
fair and ethical practices 136–139
reasons for 126
sten scale 54, 54
stringency bias 65
Strong Interest Inventory 274
structural equivalence 113 structured
methods 222 subgroup norming 57
subjective formats 88 subjective well-
being 202–203 substance use
disorders 208 successive test
development 120–121 systematic
errors 41–42
T
teacher-developed measures 291 Technical
Test Battery 273 terminological/factual-
driven adaptations 102 test adaptations 102,
102–103, 103 test bias 6
U
unconscious misrepresentation 65
unitary validity 77
urbanisation 325
Utrecht Work Engagement Scale 219
V
validity
concurrent 303
construct 72, 303, 332–333
content 71–72, 303
convergent 73
criterion-prediction 73
differential 74
and interpretation 301–304
validity coefficient 77
factors affecting 77–78
magnitude of 77
validity generalisation 75
validity of a measure 69–70
defining 70–71
indices, interpretation of 77–79
types of 71–77
validity triangle 71, 71
Value Orientations 274
values 247
and career-counselling assessment 247
competing 140, 140–141, 141–142, 296–297
variability, measure of 49–50
variance 49, 49
variances, predictor/criterion variables 78,
78 variations
in assessment conditions and reliability
66 in instructions and reliability 66
in interpretation of instructions and reliability
66 in scoring or ratings and reliability 66
vendor-neutral assessment portals
257 Verbal Comprehension Index 180,
181 Verify Range of Ability Tests 274
see also SHL Verify ability tests
vignettes 278
Vineland Adaptive Behaviour Scales 282–
283 virtual content 261
virtual reality/augmented reality 257
vocational interests 245–246
W
Wechsler Adult Intelligence Scale 180, 181–182. 183, 186,
340 Wechsler Intelligence Scales 17–18, 21
for children 21
well-being 200–201
defining 201
domains of 202–205
and positive psychology 201–202
social, economic determinants 203
subjective 202–203
well-being in the workplace 205, 208–
210 cost of ill health 206–208
wellness programmes 210–211
why it matters 205–206
well-being, measures of 213
in diverse contexts 213–218
in the work context 218–219
well-being, university students 211–
213 domains of holistic view of 212
Wellness Questionnaire for Higher Education 217, 338
workaholism 210
work engagement 209–210
Working Memory Index 180, 182
Work Locus of Control Scale 219
Work-related Risk and Integrity Scales 238–239
Y
young children 162–163
Z
z-scores 53, 55