Professional Documents
Culture Documents
A Mixed Methods Model of Scale Development and Validation Analysis
A Mixed Methods Model of Scale Development and Validation Analysis
A Mixed Methods Model of Scale Development and Validation Analysis
net/publication/331295116
CITATIONS READS
12 2,745
1 author:
Yuchun Zhou
Ohio University
15 PUBLICATIONS 221 CITATIONS
SEE PROFILE
All content following this page was uploaded by Yuchun Zhou on 14 May 2019.
Yuchun Zhou
To cite this article: Yuchun Zhou (2019) A Mixed Methods Model of Scale Development and
Validation Analysis, Measurement: Interdisciplinary Research and Perspectives, 17:1, 38-47, DOI:
10.1080/15366367.2018.1479088
ABSTRACT KEYWORDS
Using mixed methods to develop new scales is not a new idea since the Scale development;
2000s. However, there exists inadequate literature that discusses scale validation analysis; mixed
development using mixed methods, with steps including how to design methods
the study, how to implement the process, and how to conduct validation.
This study proposes a hands-on model of using mixed methods to develop
new scales and using multiple approaches to conduct validation analysis.
The proposed model consists of five steps that highlight both mixed
methods’ integration techniques and psychometric methods. The model
of scale development and validation analysis is practical and useful for
researchers who desire to develop a reliable scale.
Introduction
In social sciences and human behaviors, scholars usually found the needs to develop new scales due
to lack of existing instruments in their fields. Traditionally, researchers referred to psychometric
literature on reliability and validity for item generation (Rowan & Wulff, 2007) and survey meth-
odology for questionnaire design (May, 2001) when developing new scales. In recent years, research-
ers realized that scale development was not merely a procedure within a research project but a
systematic study from a research design level to implementation, and accordingly, mixed methods
has been employed to scale development (Bryman, 2006; Collins, Onwuegbuzie, & Sutton, 2006;
Greene, Caracelli, & Graham, 1989).
The assumption of mixed methods research is that mixing both qualitative and quantitative
methods could enhance the understanding of a research phenomenon than either method by itself
(Creswell & Plano Clark, 2011; Johnson, Onwuegbuzie, & Turner, 2007). The development and
evolution of mixed methods has experienced 6 stages in the last 30 years: the formative stage (1980s
and before), the paradigm stage (1980s to 1990s), the procedural stage (1980s to present), the
advocacy stage (early 2000s to present), the reflective stage (2000s to present), and the expansion
stage (2010s to present) (Creswell & Plano Clark, 2011; Tashakkori & Teddlie, 1998; Teddlie &
Tashakkori, 2009). In more recent years, research on mixed methods’ expansion focused on its
applications in and adaptations to specific disciplines or topics, such as experiment intervention,
program evaluation, longitudinal research, and instrument development (Glogowska, 2011; Karasz &
Singelis, 2009; Kettles, Creswell, & Zhang, 2011; Tewksbury, 2009; Plano Clark et al., 2015; Plano
Clark & Wang, 2010; Shaw, Connelly, & Zecevic, 2010).
Using mixed methods to develop new scales is not a new idea since the 2000s. A few methodol-
ogists have discussed the rationales of using mixed methods in scale development since 2006 (Collins
et al., 2006; Creswell & Plano Clark, 2011; Onwuegbuzie, Bustamante, & Nelson, 2010; Smolleck,
Zembal-Saul, & Yoder, 2006). For instance, Collins et al. (2006) explicated the necessity and
CONTACT Yuchun Zhou zhouy@ohio.edu 302G McCracken Hall, Ohio University, Athens, OH 45701.
Color versions of one or more of the figures in the article can be found online at www.tandfonline.com/hmes.
© 2019 Taylor & Francis Group, LLC
MEASUREMENT: INTERDISCIPLINARY RESEARCH AND PERSPECTIVES 39
appropriateness of using mixed methods to “assess the appropriateness and/or utility of existing
instrument(s); create new instrument(s) and assess its appropriateness and/or utility” (p. 76).
Although researchers argued the advantages of using mixed methods to develop new scales,
literature providing systematic instructions on how to do it has been scarce and incomplete. For
instance, Onwuegbuzie et al. (2010) published a 10-phase IDCV instrument development framework
for developing and assessing the fidelity of a quantitative instrument. Their framework contained the
idea of mixing approaches; however, it remained at the implementation level and did not give
instructions on the research design level. Therefore, this study integrated mixed methods research
into psychometrics and proposed a hands-on model of using a sequential mixed methods design to
develop scales.
By reading this study, researchers could improve their understanding of using mixed methods for
scale development and use it appropriately. Hopefully, mixed methods would be adopted in practice
by researchers from a wide range of disciplines.
researchers aspiring to design a mixed methods study for instrument development, this study
proposed a mixed methods model that provided clear guidance on research design, integration
phases, validation techniques, and psychometric consideration involving scale development and
validation.
Furthermore, previous literature has also addressed how to properly use mixed methods for
validation in specific fields, including health sciences (Hitchocock et al, Hitchcock et al., 2006),
psychology (Luyt, 2012; Ungar & Liebenberg, 2011), education (Burton & Mazerolle, 2011; Crede &
Borrego, 2013; Nassar-McMillan, Wyer, Oliver-Hoyo, & Ryder-Burge, 2010; Smolleck et al., 2006),
childhood trauma (Boeije, Slagt, & Wesel, 2013), and language test (Lee & Greene, 2007).
In addition to the multiple methods deployed in the validation analysis of an instrument,
Onwuegbbuzie and Johnson (2006) also paid attention to the legitimation in the validation analysis
via using mixed methods. The legitimation included sampling integration, inside-outside, weakness
minimization, sequential, conversion, paradigmatic mixing, commensurability, multiple validities,
and political legitimation. Following Onwuegbbuzie and Johnson (2006), Dellinger and Leech (2007)
and Leech, Dellinger, Brannagan, and Tanaka (2010) further discussed how to evaluate a mixed
methods study for validation purpose. The framework emphasized the quality of mixed methods
research design, legitimation considerations, interpretive rigor, inferential consistency, as well as the
utilization and consequential elements. In sum, mixed methods is an appropriate methodology that
should be used to collect various evidence of validity for a new instrument; however, mixed methods
would not guarantee the quality of validation analysis. Only when researchers used mixed methods
with rigor, could they properly test their instrument. Therefore, the discussion in this study on how
to integrate a specific mixed methods research design into psychometrics is extremely important to
researchers who need to develop and analyze new scales.
the factors, reduce the items, and examine the factor structure (Burton & Mazerolle, 2011;
Dahodwala et al., 2012; Martin & Sass, 2010; Melka et al., 2011; Miller et al., 2012). Confirmatory
factor analysis was another commonly used strategy to cross-validate the factor structure following
the exploratory factor analysis (Burton & Mazerolle, 2011; Dahodwala et al., 2012; Melka et al., 2011;
Miller et al., 2012; Weiss & Smith, 1999; Worthingon & Whittaker, 2011). When the model fit
indices were acceptable, validity evidence was provided for construct validity, which indicated the
items could measure the scale construct.
To sum up, the existing literature indicated the needs and appropriateness of using mixed
methods for scale development and validation. However, no study provided a comprehensive
model that encompasses mixed methods designs and psychometric consideration in scale develop-
ment. Therefore, this study proposed a mixed methods model with five steps to construct new scales
and concurrently examine multiple evidence for greater validity. The recommended steps not only
emphasize mixed methods design but also focus on specific mixing and validation techniques to
achieve appropriate psychometric properties.
phases. As Figure 1 indicates, ovals are used to distinguish the mixing steps from other square boxes
with single method.
At the research design level, the mixed methods model of scale development and validation analysis
(MSDVA) adapts Creswell and Plano Clark (2011)’s exploratory instrument design and embraces
validation phases. It consists of five steps: (1) qualitatively investigating the scale construct, which is
also a qualitative validation process to collect content-validity evidence, (2) converting qualitative
findings to scale items, which is an integration strategy in mixed methods research, (3) conducting
mixing validation to review items’ content-based validity, (4) administering test items and collect item
responses, and (5) conducting quantitative validation to analyze item properties and to examine
construct-validity evidence. For items with poor psychometric properties, they should be revised and
sent back to step (3) for another run of validation and pilot testing until they get through item analysis
in step (5). More details and suggestions for each step in the model are presented as followed.
stage, the evidence is mainly about content-based validity, namely, whether the items comprehen-
sively present the scale construct to be measured.
The main qualitative approaches include reflection, debriefing, and panel review. The common
quantitative methods include sorting and calculation. Reflection refers to defining the scale construct
to be measured. The researcher should attempt to answer questions such as, what the scale is used to
measure and how to define the construct. This practice of reflection could provide evidence for
content validity of the new scale (Burton & Mazerolle, 2011; Dahodwala et al., 2012; Onwuegbuzie
et al., 2010; Smolleck et al., 2006).
Debriefing is a discussion on the relationship between items and the construct. By discussing the
structure of the new scale, the hypothesized model between items and the latent construct is
generated (Burton & Mazerolle, 2011; Dahodwala et al., 2012; Melka et al., 2011; Miller et al.,
2012; Weiss & Smith, 1999; Worthingon & Whittaker, 2011). The hypothesized model will be tested
via factor analysis in the later steps.
Panel review is to have a panel of experts review the items the representativeness and complete-
ness. Experts’ comments will provide evidence for content validity of the new scale (Dahodwala
et al., 2012; Durham et al., 2011; Morell & Tan, 2009; Smolleck et al., 2006). Afterwards, revision of
the items according to the panel’s feedback should be completed.
Sorting is a quantitative method to calculate experts’ agreement on the content of items. A
group of experts are asked to review the items and determine what items should be grouped
together. They sort items to different categories and then define the categories. The sorting
procedure will provide the evidence of content-based and construct validity because it discusses
why certain items should be categorized in one construct (Agarwa, 2011; Agarwa et al., 2011;
Moore & Benbasat, 1991, Moore & Benbasat, 2001). All in all, multiple approaches could be used
to examine the new items’ accuracy and comprehensiveness, and the results should be integrated
to provide evidence of content-based validity.
After combining the results from different analyses, the researchers should revise and/or delete
poor items and send the revised items back to Step (3) for iterative runs of mixing validation and
pilot testing until the statistical validation results are acceptable through Step (5). The whole process
of the model is represented in Figure 1.
Conclusion
In conclusion, this study summarized the existing literature on using mixed methods to develop
instrument. The review of literature indicated why mixed methods is the most appropriate research
design for scale development. Adapting Creswell and Plano Clark (2011)’s exploratory instrument
design, the author of this study proposed a five-step mixed methods model of scale development
(MSDVA) that comprises mixed methods research design and psychometric issues. As the figure of
the model showed, the scale development procedures and validation procedures are integrated
during the whole process. Mixing occurs at research design level and research implementation
level. For instance, at some steps, development and validation are conducted concurrently. At certain
steps, specific mixing strategies are highlighted based on mixed methods research design considera-
tion. In brief, the model provides comprehensive and detailed instructions for researchers to develop
new scales using a sequential mixed methods design.
Besides the discussion on scale development, the paper also aims to improve readers’ under-
standing of mixed methods as well as its applications. In the future, more mixed methods studies are
needed to present empirical examples of scale development in social sciences. Hopefully, mixed
methods would be well recognized and applied in a wide range of disciplines in practice.
References
Agarwa, N. (2011). Verifying survey items for construct validity: A two-stage sorting procedure for questionnaire
design in information behavior research. ASIST, 10, 9–13.
Agarwa, N., Xu, Y., & Poo, D. (2011). A context-based investigation into source use by information seekers. Journal of
American Society for Information Science and Technology, 62(6), 1087–1104.
American Educational Research Association, American Psychological Association, & National Council on
Measurement in Education. (AERA/APA/NCME). (1999). Standards for educational and psychological testing.
Washington, DC: American Educational Research Association.
Boeije, H., Slagt, M., & Wesel, F. (2013). The contribution of mixed methods research to the field of childhood trauma:
A narrative review focused on data integration. Journal of Mixed Methods Research, 7(4), 347–369.
Borsboom, D., Mellenbergh, G., & Heerden, J. (2004). The concept of validity. Psychological Review, 111(4), 1061–
1071.
Brod, M., Tesler, L., & Christensen, T. (2009). Qualitative research and content validity: Developing best practices
based on science and experience. Quality of Life Research, 18, 1263–1278.
Bryman, A. (2006). Integrating quantitative and qualitative research: How is it done. Qualitative Research, 6, 97–113.
Burton, L., & Mazerolle, S. (2011). Survey instrument validity part I: Principles of survey instrument development and
validation in athletic training education research. Athletic Training Education Journal, 6(1), 27–35.
Collins, K., Onwuegbuzie, A., & Sutton, I. (2006). A model incorporating the rationale and purpose for conducting
mixed-methods research in special education and beyond. Learning Disabilities: A Contemporary Journal, 4(1), 67–
100.
Crede, E., & Borrego, M. (2013). From ethnography to items: A mixed methods approach to developing a survey to
examine graduate engineering student retention. Journal of Mixed Methods Research, 7(1), 62–80.
Creswell, J. W., & Plano Clark, V. L. (2011). Designing and conducting mixed methods research (2nd ed.). Thousand
Oaks, CA: Sage.
Dahodwala, N., Karlawish, J., Shea, J., Zubritsky, C., Stern, M., & Mandell, D. (2012). Validation of an instrument to
measure older adults’ expectations regarding movement. PLoS One, 7(8), e43854.
Dellinger, A., & Leech, N. (2007). Toward a unified validation framework in mixed methods research. Journal of Mixed
Methods Research, 1(4), 309–332.
Durham, J., Tan, B., & White, R. (2011). Utilizing mixed research methods to develop a quantitative assessment tool:
An example from explosive remnants of a war clearance program. Journal of Mixed Methods, 5(3), 212–226.
Glogowska, M. (2011). Paradigms, pragmatism and possibilities: Mixed-methods research in speech and language
therapy. International Journal of Language Communication Disorder, 46(3), 251–260.
46 Y. ZHOU
Greene, J. C., Caracelli, V. J., & Graham, W. F. (1989). Toward a conceptual framework for mixed-method evaluation
designs. Educational Evaluation and Policy Analysis, 11, 255–274.
Hitchcock, J., Sarkar, S., Nastasi, B., Burkholder, G., Varjas, K., & Jayasena, A. (2006). Validating culture- and gender-
specific constructs: A mixed-method approach to advance assessment procedures in cross-cultural settings. Journal
of Applied School Psychology, 22(2), 13–33.
Holsapple, M., Finelli, C., Carpenter, D., Harding, T., & Sutkus, J. (2009). Work in progress – A mixed-methods
approach to developing an instrument measuring engineering students’ positive ethical behavior. 39th ASEE/IEEE
Frontiers in Education Conference, Session T3E-1, San Antonio, TX.
Hubley, A., & Zumbo, B. (2011). Validity and the consequences of test interpretation and use. Social Indicators
Research, 103, 219–230.
Johnson, R. B., Onwuegbuzie, A. J., & Turner, L. (2007). Toward a definition of mixed methods research. Journal of
Mixed Methods Research, 1(2), 112–133.
Karasz, A., & Singelis, T. (2009). Qualitative and mixed methods research in cross-cultural psychology: Introduction to
the special issue. Journal of Cross Cultural Psychology, 40(6), 909–916.
Kettles, A., Creswell, J., & Zhang, W. (2011). Mixed methods research in mental health nursing. Journal of Psychiatric
and Mental Health Nursing, 18, 535–542.
Kline, R. B. (2011). Principles and practice of structural equation modeling (3rd ed.). New York: Guilford Press.
Lee, Y., & Greene, J. (2007). The predictive validity of an ESL placement test: A mixed methods approach. Journal of
Mixed Methods Research, 1(4), 366–389.
Leech, N., Dellinger, A., Brannagan, K., & Tanaka, H. (2010). Evaluating mixed research studies: A mixed methods
approach. Journal of Mixed Methods Research, 4(1), 17–31.
Luyt, R. (2012). A framework for mixing methods in quantitative measurement development, validation, and revision:
A case study. Journal of Mixed Methods Research, 6(4), 294–316.
Martin, N., & Sass, D. (2010). Construct validation of the behavior and instructional management scale. Teaching and
Teacher Education, 26(5), 1124–1135.
May, T. (2001). Social research: Issues, methods and process. Berkshire, UK: Open University Press.
Melka, S., Lancaster, S., Bryant, A., Rodriguez, B., & Weston, R. (2011). An exploratory and confirmatory factor
analysis of the affective control scale in an undergraduate sample. Journal of Psychopathology Behavior Assessment,
33, 501–513.
Messick, S. (1989). Meaning and values in test validation: The science and ethics of assessment. Educational
Researcher, 18, 5–11.
Meurer, S., Rubio, D., Counte, M., & Burroughs, T. (2002). Development of a healthcare quality improvement
measurement tool: Results of a content validity study. Hospital Topics: Research and Perspectives on Healthcare,
80(2), 7–13.
Miller, M., Kim, J., Chen, G., & Alvarez, A. (2012). Exploratory and confirmatory factor analyses of the Asian
American racism-related stress inventory. Assessment, 19(1), 53–64.
Moore, G., & Benbasat, I. (1991). Development of an instrument to measure the perceptions of adopting an
information technology innovation. Information Systems Research, 2(3), 192–222.
Moore, G., & Benbasat, I. (2001). Development of an instrument to measure the perceptions of adopting an
information technology innovation. Information Systems Research, 2(3), 192–222. [Copyright, 1991, The Institute
of Management Sciences].
Morell, L., & Tan, R. (2009). Validating for use and interpretation: A mixed methods contribution illustrated. Journal
of Mixed Methods Research, 3(3), 242–264.
Nassar-McMillan, S., Wyer, M., Oliver-Hoyo, M., & Ryder-Burge, A. (2010). Using focus groups in preliminary
instrument development: Expected and unexpected lessons learned. The Qualitative Report, 15(6), 1621–1634.
Nastasi, B., Hitchcock, J., Sarkar, S., Burkholder, G., Varjas, K., & Jayasena, A. (2007). Mixed methods in intervention
research: Theory to adaptation. Journal of Mixed Methods Research, 1(2), 164–182.
Newman, I., Lim, J., & Pineda, F. (2013). Content validity using a mixed methods approach: Its application and
development through the use of a table of specifications methodology. Journal of Mixed Methods, 7(3), 243–260.
Onwuegbbuzie, A., & Johnson, R. (2006). The validity issue in mixed research. Research in the Schools, 13(1), 48–63.
Onwuegbuzie, A., Bustamante, R., & Nelson, J. (2010). Mixed research as a tool for developing quantitative instru-
ments. Journal of Mixed Methods, 4(1), 56–78.
Plano Clark, V., Anderson, N., Wertz, J., Zhou, Y., Schumacher, K., & Miaskowski, C. (2015). Conceptualizing
longitudinal mixed methods designs: A methodological review of health sciences research. Journal of Mixed
Methods Research, 9(4), 297–319.
Plano Clark, V. L., & Wang, S. C. (2010). Adapting mixed methods research to multicultural counseling. In J. G.
Ponterotto, J. M. Casas, L. A. Suzuki, & C. M. Alexander (Eds.), Handbook of multicultural counseling (3rd ed., pp.
427–438). Thousand Oaks, CA: Sage.
Rowan, N., & Wulff, D. (2007). Using qualitative methods to inform scale development. The Qualitative Report, 12(3),
450–466.
MEASUREMENT: INTERDISCIPLINARY RESEARCH AND PERSPECTIVES 47
Shaw, J., Connelly, D., & Zecevic, A. (2010). Pragmatism in practice: Mixed methods research for physiotherapy.
Physiotherapy Theory and Practice, 26(8), 510–518. doi:10.3109/09593981003660222
Sireci, S. (2009). Packing and unpacking sources of validity evidence: History repeats itself again. In R. Lissitz (Ed.), The
concept of validity: Revisions, new directions, and applications (pp. 19–37). Charlotte, NC: Information Age
Publishing, Inc.
Sireci, S. G., & Sukin, T. (2013). Test validity. In K. Geisinger et al. (Eds.), APA handbook of testing and assessment in
psychology, 1 (pp. 61–84). Washington DC: American Psychology Association.
Smolleck, L., Zembal-Saul, C., & Yoder, E. (2006). The development and validation of an instrument to measure
preservice teachers’ self-efficacy in regard to the teaching of science as inquiry. Journal of Science Teacher
Education, 17, 137–163.
Tashakkori, A., & Teddlie, C. (1998). Mixed methodology: Combining qualitative and quantitative approaches.
Thousand Oaks, CA: Sage.
Teddlie, C., & Tashakkori, A. (2009). Foundations of mixed methods research. Thousand Oaks, CA: Sage.
Tewksbury, R. (2009). Qualitative versus quantitative methods: Understanding why qualitative methods are superior
for criminology and criminal justice. Journal of Theoretical and Philosophical Criminology, 1(1), 38–58.
Ungar, M., & Liebenberg, L. (2011). Assessing resilience across cultures using mixed methods: Construction of the
child and youth resilience measure. Journal of Mixed Methods, 5(2), 126–149.
Vogt, D., King, D., & King, L. (2004). Focus groups in psychological assessment: Enhancing content validity by
consulting members of the target population. Psychological Assessment, 16(3), 231–243.
Weiss, M., & Smith, A. (1999). Quality of youth sport friendships: Measurement development and validation. Journal
of Sport & Exercise Psychology, 21, 145–166.
Worthingon, R., & Whittaker, T. (2011). Scale development research: A content analysis and recommendations for
best practice. The Counseling Psychologist, 34(6), 806–838.