Professional Documents
Culture Documents
Safety Science: Awatef Ergai, Tara Cohen, Julia Sharp, Doug Wiegmann, Anand Gramopadhye, Scott Shappell
Safety Science: Awatef Ergai, Tara Cohen, Julia Sharp, Doug Wiegmann, Anand Gramopadhye, Scott Shappell
Safety Science
journal homepage: www.elsevier.com/locate/ssci
a r t i c l e i n f o a b s t r a c t
Article history: The Human Factors Analysis and Classification System (HFACS) is a framework for classifying and
Received 5 March 2015 analyzing human factors associated with accidents and incidents. The purpose of the present study
Received in revised form 11 July 2015 was to examine the inter- and intra-rater reliability of the HFACS data classification process.
Accepted 21 September 2015
Methods: A total 125 safety professionals from a variety of industries were recruited from a series of
Available online 11 November 2015
two-day HFACS training workshops. Participants classified 95 real-world causal factors (five causal
factors for each of the 19 HFACS categories) extracted from a variety of industrial accidents. Inter-rater
Keywords:
reliability of the HFACS coding process was evaluated by comparing performance across participants
HFACS
Inter-rater reliability
immediately following training and intra-rater reliability was evaluated by having the same participants
Intra-rater reliability repeat the coding process following a two-week delay.
Classification Results: Krippendorff’s Alpha was used to evaluate the reliability of the coding process across the various
Human-error HFACS levels and categories. Results revealed the HFACS taxonomy to be reliable in terms of inter- and
intra-rater reliability, with the latter producing slightly higher Alpha values.
Conclusion: Results support the inter- and intra-rater reliability of the HFACS framework but also reveal
additional opportunities for improving HFACS training and implementation.
Ó 2015 Elsevier Ltd. All rights reserved.
http://dx.doi.org/10.1016/j.ssci.2015.09.028
0925-7535/Ó 2015 Elsevier Ltd. All rights reserved.
394 A. Ergai et al. / Safety Science 82 (2016) 393–398
ORGANIZATIONAL
INFLUENCES
PRECONDITIONS
FOR
UNSAFE ACTS
Errors
Errors Violations
results become suspect. Decisions based on such analyses may Among the six studies that were designed to assess inter- and/
therefore lead to onerous comparisons across industries and or intra-rater reliability, some demonstrated acceptable levels of
produce ineffective mitigation/prevention plans that have little reliability (i.e., Li and Harris, 2006; O’Connor, 2008; O’Connor
meaningful impact on reducing risk or improving system safety. and Walliser, 2010). However, this was not always the case as
In general, reliability refers to the extent to which a framework, others (Olson and Shorrock, 2010; O’Connor and Walker, 2011;
experiment, test or measuring instrument yields the same result Olsen, 2011) showed less reliability. Accounting for these seem-
over repeated trials (Carmines and Zeller, 1979). Evaluating relia- ingly variable results, Cohen et al., 2015 suggested that lower
bility is a primary concern for many fields including the behavioral, levels of HFACS reliability were attributable to a variety of factors,
psychological, medical, and social sciences, particularly as new including: inadequate training of coders, the use of a small number
methods, tests, devices, and instruments are developed. There are of accident cases/causal factors, unnecessary modifications made
two types of reliability important in scientific research that to the HFACS framework, and an inconsistency in the methods
requires data classification: inter-rater and intra-rater reliability. used to assess both inter- and intra-rater reliabilities.
Inter-rater reliability refers to the framework’s ability to produce Unfortunately, many of these conclusions are speculative given
the same results irrespective of who conducts the analysis, that many of the articles reviewed by the authors lacked sufficient
whereas intra-rater reliability refers to the consistency of each detail to discern the exact nature of the conditions under which the
individual rater. Minor variations may occur in both cases; how- coding was conducted. To further address these issues, this study
ever, in general, the more stable the results, the more confident assessed the reliability of the HFACS framework as a general acci-
one can be that the results are reproducible and trustworthy. dent analysis tool using a large number of trained coders and mul-
A number of papers have previously examined the reliability of tiple real-world accident causal factors from a variety of industries.
the HFACS methodology. A recent review of these studies indicates
that the framework produces generally reliable results (Cohen,
et al., 2015). The Cohen review examined 111 HFACS manuscripts 2. Methods
in detail, of which 14 peer-reviewed articles reported inter- and/or
intra-rater reliability of HFACS. Notably however; only six of these 2.1. Participants
14 articles (Li and Harris, 2006; O’Connor, 2008; O’Connor and
Walliser, 2010; Olson and Shorrock, 2010; O’Connor and Walker, One hundred and twenty five safety professionals from a variety
2011; Olsen, 2011) were specifically designed to test reliability. of industries (e.g. aviation, mining, medicine and manufacturing)
The others merely reported reliabilities as part of quality check were recruited from a series of two-day training workshops
of data used in a larger study. designed and provided by the developers of HFACS (S.S. & D.W.).
A. Ergai et al. / Safety Science 82 (2016) 393–398 395
Table 1
Description of the HFACS categories.
Organizational influences
Organizational climate: Prevailing atmosphere/vision within the organization including such things as policies, command structure, and culture
Operational process: Formal process by which the vision of an organization is carried out including operations, procedures, and oversight among others
Resource management: How human, monetary, and equipment resources necessary to carry out the vision are managed
Unsafe supervision
Inadequate supervision: Oversight and management of personnel and resources including training, professional guidance, and operational leadership among other
aspects.
Planned inappropriate operations: Management and assignment of work including aspects of risk management, crew pairing, operational tempo, etc.
Failed to correct known problems: Those instances when deficiencies among individuals, equipment, training, or other related safety areas are ‘‘known” to the supervisor,
yet are allowed to continue uncorrected
Supervisory violations: The willful disregard for existing rules, regulations, instructions, or standard operating procedures by management during the course of their
duties
Unsafe acts
Errors
Decision errors: These ‘‘thinking” errors represent conscious, goal-intended behavior that proceeds as designed, yet the plan proves inadequate or inappropriate for the
situation. These errors typically manifest as poorly executed procedures, improper choices, or simply the misinterpretation and/or misuse of relevant information
Skill-based errors: Highly practiced behavior that occurs with little or no conscious thought. These ‘‘doing” errors frequently appear as breakdown in visual scan
patterns, inadvertent activation/deactivation of switches, forgotten intentions, and omitted items in checklists often appear. Even the manner or technique with
which one performs a task is included
Perceptual errors: These errors arise when sensory input is degraded, as is often the case when flying at night, in poor weather, or in otherwise visually impoverished
environments. Faced with acting on imperfect or incomplete information, aircrew run the risk of misjudging distances, altitude, and decent rates, as well as
responding incorrectly to a variety of visual/vestibular illusions
Violations
Routine violations: Often referred to as ‘‘bending the rules” this type of violation tends to be habitual by nature and is often enabled by a system of supervision and
management that tolerates such departures from the rules.
Exceptional violations: Isolated departures from authority, neither typical of the individual nor condoned by management
2.2. Training from the National Transportation Safety Board (NTSB), Occupa-
tional Safety and Health Administration (OSHA), and other HFACS
Trainers with expertise in both human factors and the HFACS accident databases such as aviation, maintenance, food service,
methodology provided participants two days of intensive HFACS healthcare, mining and lodging. Each causal factor was selected
training. The first day included didactic training on HFACS includ- for clarity and was presented to participants in the format in which
ing a brief overview of human error. This was followed by an it was reported in the accident database. Some causal factors were
extensive discussion of each causal category within HFACS includ- modified for grammatical purposes only, and in a manner that did
ing detailed descriptions and examples. Day one culminated with not affect the content of the statements. This approach was
an in-class case study that illustrated the human factors associated adopted to enhance content and external validity of the causal fac-
with a real-world accident using HFACS. tor statements. Given the original accident reports were not coded
Day two started with a comprehensive review of the HFACS sys- using HFACS, a scoring key of correct answers was created for each
tem including a question and answer period. This was followed by of the 95 causal statements. Correct answers were determined a
several hands-on group exercises coding human causal factors priori by consensus among the HFACS expert research team.
within the HFACS framework. Also included were additional case An online survey platform was used as the primary data-
studies to further clarify the HFACS framework and its use in acci- gathering instrument in this study so that the evaluation could
dent/incident investigation. The training concluded with a large be conducted away from the original training facility for purposes
group discussion of any issues or areas in need of clarification. of the two-week follow-up assessment. In the form, each of the 95
causal factors was presented one at a time. The causal factor was
2.3. Coding assessment formatted as a statement, followed by the 19 possible HFACS cau-
sal categories. The participants were required to read and identify
Five causal factors associated with each of the 19 HFACS causal each causal factor, and then attribute it to the causal category that
categories (n = 95) were extracted from actual accident reports best described it by checking the appropriate box. Participants
396 A. Ergai et al. / Safety Science 82 (2016) 393–398
were not provided feedback on their performance nor were they Table 2
provided a copy of their exam. Overall a for each HFACS tier and category; inter-rater reliability.
completing this coding process and participants were allowed to Unsafe supervision 0.73
use notes and reference materials to help complete the task. This Inadequate supervision 0.51
Planned inappropriate operations 0.49
process was used to ensure that the assessment was not simply a
Failed to correct a known problem 0.82
memory test but rather mimicked the coding process commonly Supervisory violation 0.53
utilized in the real-world.
Organizational influences 0.80
Resource management 0.62
Organizational climate 0.80
3. Results
Organizational process 0.69
Previous studies related to coding and classifying causal factors prevention plans and strategies. Once a company has classified its
have recruited participants from industries that match those of the accident and near miss cases using the HFACS taxonomy, it can
causal factors (e.g. using pilots to classify causal factors from avia- analyze these data searching for trends that point to weaknesses
tion accidents). For this study, participants from a variety of indus- in certain areas of the system. In addition, conducting association
tries were used to code causal factors from multiple industries; analysis among HFACS categories can help identify additional areas
therefore the background of the coders did not necessarily match for improvement. Information of this nature not only provides the
the industrial accidents from which the causal factors were safety professional with supplementary knowledge to guide lim-
extracted. Although the authors were careful to re-write any state- ited resources toward a more focused intervention, but also offers
ments that may have been grammatically incorrect, it is possible benefits to employee health through lowering frequency and
that the nomenclature used in the description of the causal factors, severity of work accidents, all of which have a positive impact on
may have caused confusion for those individuals outside of the cost and well-being at work.
industry in which the causal factor came. This may explain why
some of the reliability levels may not have been as high as References
expected. Future research should focus on using a large number
of coders and casual factors so that participants can be matched Baysari, M.T., Caponecchia, C., McIntosh, A.S., Wilson, J.R., 2009. Classification of
errors contributing to rail incidents and accidents: a comparison of two human
to their individual domain of expertise. error identification techniques. Saf. Sci. 47, 948–957.
Carmines, E., Zeller, R., 1979. Reliability and Validity Assessment. Sage Publications,
Thousands Oaks, CA.
5. Limitations Chen, S., Wall, A., Davies, P., Yang, Z., Wang, J., Chou, Y., 2013. A human and
organizational factors (HOFs) analysis method for marine casualties using
HFACS-Maritime Accidents (HFACS-MA). Saf. Sci. 60, 105–114.
One limitation of this study involves the number of participants
Cohen, T., Wiegmann, D., Shappell, S., 2015. Evaluating the reliability of the human
who failed to complete the second assessment. Of the 125 partici- factors analysis and classification system. Aviat. Space Environ. Med. 86 (8),
pants who completed the first coding activity, only 59 (47%) com- 728–735.
ElBardissi, A.W., Wiegmann, D.A., Dearani, J.A., Daley, R.C., Sundt, T.M., 2007.
pleted the second assessment. Because of this drop-out rate, while
Application of the human factors analysis and classification system
the sample size for the analysis of the intra-rater reliability was methodology to the cardiovascular surgery operating room. Soc. Thoracic
still large, it was significantly smaller than that of the inter-rater Surg. 83, 1412–1419.
reliability. Because participation was voluntary, unfortunately Gwent, Kilem, 2010. Handbook of Inter-rater Reliability, (2) Advanced Analytics LLC.
Krippendorff, K., 2004. Content Analysis: An Introduction to Its Methodology. Sage
those individuals who chose to participate in the second coding Publications, Incorporated.
activity may have been more inherently interested in the HFACS. Li, W.C., Harris, D., 2006. Pilot error and its relationship with higher organizational
These individuals also may have been more attentive and consci- levels: HFACS analysis of 523 accidents. Aviat. Space Environ. Med. 77 (10),
1056–1061.
entious during both the training and the coding process. The nat- O’Connor, P., 2008. HFACS with an additional layer of granularity: validity and
ure of this study also precludes knowing about the activities that utility in accident analysis. Aviat. Space Environ. Med. 79 (6), 599–606.
these participants engaged in during the two weeks between the O’Connor, P., Walliser, Philips E., 2010. Evaluation of a human factors analysis and
classification system used by trained raters. Aviat. Space Environ. Med. 81 (10),
first and second assessment. Some participants may have intermit- 957–960.
tently reviewed what they learned during training or actually O’Connor, P., Walker, P., 2011. Evaluation of a human factors analysis and
applied HFACS as part of their work assignments. All of these fac- classification system as used by simulated mishap boards. Aviat. Space
Environ. Med. 82 (1), 44–48.
tors could account for the slightly higher levels of intra-rater reli- Olson, N.S., Shorrock, S.T., 2010. Evaluation of the HFACS-ADF safety classification
ability levels compared to inter-rater reliability levels. In any system: inter-coder consensus and intra-coder consistency. Accid. Anal. Prev.
case, however, trends in reliability levels between intra- and 42, 437–444.
Olsen, N.S., 2011. Coding ATC incident data using HFACS: inter-coder consensus. Saf.
inter-reliability were generally similar.
Sci. 49 (10), 1365–1370.
Patterson, J.M., Shappell, S.A., 2010. Operator error and system efficiencies: analysis
of 508 mining incidents and accidents from Queensland, Australia using HFACS.
6. Conclusions Accid. Anal. Prev. 42, 1379–1385.
Reason, James, 1990. Human Error. Cambridge University Press, New York.
This study furthers the research on the utility of HFACS as an Reinach, S., Viale, A., 2006. Application of a human error framework to conduct train
accident/incident investigations. Accid. Anal. Prev. 38, 396–406.
accident casual factors classification tool with the goal of investi- Shappell, S., Detwiler, C., Holcomb, K., Hackworth, C., Boquet, A., Wiegmann, D.,
gating both inter-rater and intra-rater reliability of the HFACS cod- 2007. Human error and commercial aviation accidents: a comprehensive, fine-
ing process. This study supplements past research by using a large grained analysis using HFACS. Hum. Factors 49, 227–242.
Shohamy, E., Gordon, C., Kraemer, R., 1992. The effect of rater’s background and
sample size of coders and causal factors from actual incidents/ training on the reliability of direct writing tests. Modern Language J. 76 (1),
accidents from a variety of industries. Results support the inter- 27–33.
and intra-rater reliability of the HFACS framework but also suggest Weigle, S.C., 1998. Using FACETS to model rater training effects. Language Test. 15
(2), 263–287.
additional avenues of continued improvement.
Wiegmann, D.A., Shappell, S.A., 2003. A Human Error Approach to Aviation Accident
Reliable HFACS data is essential for empirical research on safety Analysis: the Human Factors Analysis and Classification System. Ashgate,
systems and on the effectiveness of any mitigation and/or accident Burlington, VT.