Professional Documents
Culture Documents
The Intersection of Artificial
The Intersection of Artificial
by
DISSERTATION
Presented to the Graduate Faculty of
The University of Texas at San Antonio
in Partial Fulfillment
of the Requirements
for the Degree of
COMMITTEE MEMBERS:
Rohit Valecha, Ph.D., Chair
H.Raghav Rao, Ph.D., Co-Chair
Paul Rad, Ph.D.
Kathryn Keeton, Ph.D.
This dissertation is dedicated to the many hands which have guided and pushed me through
unforeseen challenges…knowing I could make it. Namely, I dedicate this to my husband, Joe.
You’ve been my foundation for when my own footing has failed me, never failing to pick me up
and keep me in the fight. I am not a poet, but if I was, I’d write you the greatest love song in the
world….let this just be a Tribute. I also dedicate this to the many mentors who have dedicated
many hours in my development. Had it not been for your words of encouragement and honesty, I
would not be the Airman I am today. I salute you and thank God I’ve had such wonderful
influences who have consistently embraced my quirks. More importantly, y’all have imparted the
trust of brotherhood/sisterhood which I value to my very soul. Lastly, I dedicate this to my late
mother, Sherry. I miss you so much and know you’ve watched over me during this endeavor.
ACKNOWLEDGEMENTS
Writing this dissertation was one of the best experiences I could have had because of my
outstanding committee members. The dedication and time spent helping me grow as an
academic and officer within the United States Air Force is unfathomably appreciated. I charged
into this program with a time limit. Understanding this, each of my committee members and the
entire department stepped up to the challenge and did not waiver their expectations. To Dr.
Rohit Valecha and Dr. H. Raghav Rao, my Chair and Co-Chair, respectively. From the start, I
knew we would be a great team. People with Buffalo, NY lineage tend to find each other in the
craziest places, Texas being one of them (Go BILLS!). Words cannot express how much of a
positive influence these two men have on my academic development at UTSA. Although the
work was hard, Dr. Valecha’s optimism and patience allowed me to see the fun in research and
made it unquestionably enjoyable. I have learned how to better value patience within the
scientific process and I credit this to Dr. Valecha’s mentorship. Moreover, Dr. Rao’s expertise
and insight guided our research endeavors to ensure the work we dedicated many hours to was
impactful. Through his challenging seminars and genuine feedback, I have grown better as
scholar. My committee members, Dr. Paul Rad and Dr. Kathryn Keeton have also been essential
to my development within this program. Before I met Dr. Rad, Python would have been another
snake in the jungle, and “AI” would have just been Skynet from The Terminator…or HAL 9000.
He was my first introduction to coding in Python and general Artificial Intelligence. I thank him
sincerely for helping me see the forest through the trees and connecting me with my other
committee members. Furthermore, my many conversations with Dr. Keeton gave me a sense of
applicability and relevance. Her reassurance and experience with military veterans encouraged
iii
me that the research we worked on together would make some waves and possibly help people. I
thank her deeply for her unwavering support throughout this process.
Dr. Nicole Beebe is another faculty member I’d like to highlight and acknowledge. I
know there was a bit of a gamble accepting me into the folds, but I am grateful for her support.
As an Air Force veteran herself, I know she understands the responsibility that comes with
wearing the uniform and I appreciate her putting her faith in my success. I hope to proudly
represent Nicole, our fantastic department, and UTSA as I rush back to the greatest job in the
world. I will showcase the education and hopefully make things better for our Airmen. Lastly,
I’ll echo an acknowledgment from another one of her student’s dissertation; Nicole has always
been a positive force of nature, setting a gold standing in my book for women in science and
technology. Maybe one day when I grow up, I’ll be just as successful as her. I thank her for
supporting me throughout this crazy endeavor and for allowing me to join the Ph.D. ranks.
I’d be remorse not to acknowledge my fantastic cohort. We joined this program at a very
challenging time in our history given the global pandemic. Social distancing and high infection
rates forced us to interact virtually which made socializing difficult. We didn’t “meet” in person
until Dr. Bachura’s teaching seminar in the Summer of 2021, which was the end of our first year.
I cherish the friendships we have nurtured together through a global pandemic, challenging
seminars, and research endeavors. That said, you all have made this experience unforgettable,
and I am truly indebted to you all for helping me navigate this program. I look forward to seeing
you all graduate. You can trust I’ll be in the crowd of many other people who believe in you,
ready to hoist a beer into the air. Keep up the great work and continue to charge forward!
Another acknowledgement I’d like to take a paragraph to acknowledge Lee Ring and my
CrossFit Lobo family. Lee, thank you for continuously being a persistent, respectful, intuitive,
iv
challenging, and knowledgeable influence on my coaching career. You’re a wonderful mentor
and I appreciate your brutal honesty. Words cannot express how much I appreciated our offline
banter on my research and the refreshing insights you could provide as a prior veteran yourself.
To my Lobo Family, you all have provided me with a community I can rely on for advice,
counsel, and the occasional competitive edge. I cherish the special memories we’ve built
together, whether suffering through one of Lee’s workouts or enduring one of my “train-with-
the-trainer” sessions. Your friendship gave me a great sense of community and belonging while
enduring the academic rigor at UTSA. I look forward to the future memories to come!
Last and most certainly not least, I’d like to personally thank all the research assistants,
colleagues, and teams which have helped me collect, clean, and process the data utilized in this
manuscript. Without your attention to detail and willingness to help, I would not be here. Thank
v
“This Master’s Thesis/Recital Document or Doctoral Dissertation was produced in accordance
with guidelines which permit the inclusion as part of the Master’s Thesis/Recital Document or
Doctoral Dissertation the text of an original paper, or papers, submitted for publication. The
Master’s Thesis/Recital Document or Doctoral Dissertation must still conform to all other
requirements explained in the “Guide for the Preparation of a Master’s Thesis/Recital
Document 6 or Doctoral Dissertation at The University of Texas at San Antonio.” It must
include a comprehensive abstract, a full introduction and literature review, and a final overall
conclusion. Additional material (procedural and design data as well as descriptions of
equipment) must be provided in sufficient detail to allow a clear and precise judgment to be
made of the importance and originality of the research reported.
August 2023
vi
THE INTERSECTION OF ARTIFICIAL INTELLIGENCE IN HUMAN RESOURCE
Within this research, we sought to explore and understand where Artificial Intelligence
(AI) intersects with human resource information systems. We further investigated the AI
phenomenon by leveraging machine learning tools to understand the following :1) where AI
exists within human resource management (HRM) via a systematic literature review *; 2) how to
propose a research framework to investigate data science competencies within private and
federal sector job postings, and 3) how military veterans with data science skills communicate
their experience concerning employability. The following sections discuss these papers further.
The first essay is a systematic literature review investigating where AI exists within
published HRM literature. Within this literature review, we leveraged a 2-phased methodology to
navigate 315,053 articles from various multidisciplinary publication sources. The findings of this
review indicate research opportunities to grow within HRM's competency management systems
(pay and benefits). Furthermore, it highlighted a gap within AI research exploring qualitative
components of recruitment information systems, which spurred the motivation for the rest of this
dissertation.
*
Published: International Journal of Information Management Data Insights, 1(2), 2021
vii
Paper 2: JC-Compass: A Framework for Conducting Competency-Based Job Posting
description research. This paper leverages a design science approach to propose and explore a
new competency-based job description research method. The results indicate that the federal
sector potentially places a decisive influence on problem-solving terms rather than artificial
intelligence terms. We also discovered that the private sector strongly influenced statistics and
ethics terminology, whereas the federal focused on problem solving and ethics.
The third and final paper explores how military veterans seeking jobs within the data
science community communicate their skills and how their employability is affected by their
choice of words within resumes. Understanding that drawdowns are posturing some military
veterans to leave the service and pursue other endeavors, we sought to evaluate how Gulf War II
("post 9/11") veterans advertise their skills relative to how many times they have been
unemployed throughout their resume. In conducting this research, we aspire to provide insight
into how veterans seeking data science jobs could better posture themselves to be more
marketable and provide employers insight into current trends and expectations.
†
Submitted and presented at 2022 Pre-ICIS SIGDSA Symposium, award recipient of “Best Paper”
viii
TABLE OF CONTENTS
Chapter 7: Conclusion........................................................................................................44
Appendix ...........................................................................................................................47
References ..........................................................................................................................49
Essay II: JC-Compass: A Framework for Conducting Competency-Based Job Posting Research
ix
Chapter 5: General Overview of Text Analysis Methods For Competency Research ......72
Chapter 9: Conclusion......................................................................................................113
Appendices .......................................................................................................................115
References ........................................................................................................................124
Essay III: Veteran Talent Within Data Science: An Exploratory Resume Analysis on the
Chapter 8: Conclusion......................................................................................................176
Appendices .......................................................................................................................177
References ........................................................................................................................180
Vita
x
LIST OF TABLES
Literature Review
xi
Table 9b Results of Preliminary Validation (Federal) ..........................................................99
Essay 3: Veteran Talent Within Data Science: An Exploratory Resume Analysis on the
xii
Table 13 Top Programming Languages Related to Resumes .............................................166
xiii
LIST OF FIGURES
Literature Review
Essay 3: Veteran Talent Within Data Science: An Exploratory Resume Analysis on the
Figure 2 Employability Construct: Modified from Fugate et al. (2004) and Ngoma and
xiv
Figure 5 Data Science Programming Languages vs Unemployment ................................167
xv
ESSAY I. ARTIFICIAL INTELLIGENCE IN TACTICAL HUMAN RESOURCE
1
CHAPTER 1: ABSTRACT
(HRMS) and HR Information Systems (HRIS). The tactical procedures of recruitment, employee
performance evaluation and satisfaction, compensation and benefit analysis, best practice
analysis, discipline management, and employee training and development systems have seen a
growth in the incorporation of AI. To better understand this evolution, we seek to explore
publication sources and literature that feature the application of AI within HRM. By utilizing a
systematic literature review methodology, this paper identifies which tactical HRIS (T-HRIS)
components are featured in literature and how each T-HRIS component is represented. This
paper gives insight to which component of tactical HRM/HRIS receives attention and identifies
2
CHAPTER 2: INTRODUCTION
as digitization infiltrates the tedious processes which exist within its respective operations. From
earlier inventions like the computer and the internet, HRM has found a way to navigate these
(Hmoud and Várallyai, 2020). Like a trebuchet, advanced technology launched the rapid
evolution of Human Resource Information Systems (HRIS) as newer capabilities like Artificial
tactical HRIS (T-HRIS). The amount of organizational, personnel, and task-orientated data HR is
enhances sustainable business models (Di Vaio et al., 2020). However, this evolution and growth
in capabilities comes with a responsibility of understanding the current state of AI within tactical
HR processes, requiring both HR professionals and academics to dive into existing literature
which highlights AI-enhanced HR capabilities and areas of growth within the HR discipline.
where Artificial Intelligence exists within HRM and T-HRIS (Di Vaoi et al., 2020; arg et al.,
2021a; Qamar et al., 2021; Vrontis et al., 2021). However, these extensive reviews have failed to
consider how AI applications are utilized in a managerial and technical standpoint, providing
little insight to which components of HRIS are underrepresented with AI capabilities. Identifying
this deficiency, we seek to explore this consideration to provide the academic community and
professional sector insight to where AI is potentially lacking and where it is thriving. The aim of
this research is to explore AI within the HRM and tactical HRIS discipline. Thus, we conduct a
systematic literature review (SLR) to provide a baseline to understand the status of T-HRIS
3
components within literature and how it is represented. This SLR, has the objective of
identifying and understanding the components of tactical HRIS that are represented most in
literature. Within this manuscript, we attempt to bridge this understanding by answering the
following research questions: RQ1) What are the tactical HRIS components which exist in
published literature? RQ2) How have the components of tactical HRIS been represented in
literature?
sector with insights to where AI is potentially lacking and where it is thriving. This work
scientifically reviews existing literature, identifying what has been accomplished by organizing
T-HRIS components on a technical and managerial spectrum. Given the everevolving state of
technology and newer applications coming to fruition within HRM, the implications of this
research are imperative for academics and industry professionals to historically understand the
direction research has taken with regards to AI and T-HRIS. This historical understanding will
provide insight to potential deficiencies and boons which may exist with regards to AI
applications.
The arrangement of this paper is as follows. It first provides a background and framework
to tactical HRIS and HRM, the evolution of AI within HR, and the AI methods acknowledged in
this paper. Secondly, it provides an explanation of the SLR methodology used. Next, it provides
the results that include insight to the which components of T-HRIS exist in literature, how they
are represented, which publication sources described the research, and future research
considerations. Finally, the paper concludes with a discussion of future research opportunities.
4
CHAPTER 3: BACKGROUND AND RELATED WORK OF T-HRIS AND AI
Within this section, we define what tactical HRIS is comprised of, propose a framework
which outlines the structure of tactical HRIS, identify previous literature reviews which have
been conducted in this field, and define common AI methodologies pertaining to this literature
review.
Tactical HRIS (T-HRIS) refers to the Human Resource personnel and technology
responsible for performing specific tasks inherent to the profession of Human Resource
Management to achieve organizational goals and objectives. T-HRIS equips HR leaders with the
information systems, employee training and development systems, and employee discipline
5
intellectual assets: managerial HRIS (otherwise known as “soft”, “collaborative” and “people-
centric”) and technical HRIS (otherwise known as “hard”, “calculative” and “data-centric”)
(Collings et al., 2018; Mayfield et al., 2003; Stewart, 2007). Technical HRIS reflects logic,
reasoning, data, and understanding. In contrast, the managerial side, is dedicated to relationship
building, workplace synthesis and creativity, and the care of the employees who work for the
organization (Cregan et al., 2021; Laker and Powell, 2011). To expand upon the structure of T-
HRIS, Figure 1 provides a framework outlining the fundamental components which make up T-
expertise, culture, and commitment (Mayfield et al., 2003; Stewart, 2007). The term
“managerial” is about soft skills, which are actions and interactions with others and how we
capabilities that enable HRM professionals to better connect with employees and effectively
make decisions within their respective companies. At the tactical level, enhancing the
streamlining time-consuming tasks and giving time back to employees and customers alike.
practices to enhance workplace relationships between the organization and its respective
employees (Cregan et al., 2021). Two components distinguish the collaborative nature of
managerial HRIS. As seen in Figure 1, the first branch of managerial HRIS is the Human
Relations component.
6
Human Relations
The human relations element of managerial HRIS relates to technologies that enhance an
organization’s ability to generate and maintain professional and effective interpersonal and
to connect and exchange information and ideas between two or more individuals by way of any
channel (in person, online, or written) and build relationships (Collinson, 1996; Laker and
organization’s ability to effectively self-reflect on successes and failures and grow. These
interactions tend to be very self-reflective and perceptive to the environment (Laker and Powell,
organization’s ability to manage its employees effectively. The first component is Employee
Performance and Satisfaction (EPS), which refers to how an organization and its managers
connect with its employees, understanding employee diagnostics, and retains talent. This pillar
evaluations and feedback sessions for managers to vector employee behavior and provide them a
human connection to discuss successes and concerns within the work environment.
employee behavioral rehabilitation and termination processes (Tariq et al., 2016). Regarding
company policy. However, it also requires strong inter/intrapersonal skills to navigate delicate
conversations and potential terminations carefully. Furthermore, connecting with employees and
7
understanding their family demographics to ensure they are correctly being compensated and are
receiving the correct benefits is important to the CBA pillar of T-HRIS. Lastly, providing a
technological infrastructure that effectively and professionally manages the EPS processes
provides both managers and employees more time to communicate and grow within the
organization. We provide Table 1 to further define and explore specific task descriptions of these
responsibilities.
8
“Calculative” Technical HRIS
reputation for reflecting the data-driven and technical skills and capabilities within an
organization (Eubanks, 2018; Laker and Powell, 2011). Managerial HRIS centers itself around
organization. Technical HRIS revolves around technologies and information systems that
facilitate data analysis, technical understanding, and efficient workflow. This side of the HRIS
coin contains two branches that capture the responsibilities within an organization: 1) Resource-
challenges within an organization, technical HRIS is seen as “calculative” and direct, using data
to influence decisions and navigate the challenge of maintaining the competitive edge. Specific
categories within technical T-HRIS include customer engagement and workflow Best Practices
within the organization, Recruiting Information Systems for talent management, Employee
Training and Development systems, and Compensation and Benefits Information Systems which
manage an employee’s benefits and pay. To further expand on the responsibilities of each
category, Table 2 summarizes task descriptions of these “technical” categories within T-HRIS.
Resource Driven
The phrase “resource” alludes to a variety of different ingredients that help propel an
organization into success and notoriety. Some examples include allocated budgets, time,
technological equipment, and infrastructure. In HRM lens, acquiring and retaining premier talent
creates a demand for streamlined HRIS processes to assist in hiring decisions and prevent a
9
company from falling short due to the inability to screen, evaluate, interview, and onboard
Furthermore, the pressure of reducing discrimination within the recruiting process and
industry the organization is a part of (Rathi, 2018). These modern challenges have created a
processes when acquiring new talent to grow the company. From the technical HRIS perspective,
acquiring the employee and retaining them comes first. The managerial practices come later after
10
Policies and Standards of Practice (SOPs)
policies, and logical SOPs create two sub-branches within the second branch of technical HRIS.
The first branch sets its attention to Best Practices that exist within an organization. Specifically,
this data revolves around employee output relative to the input, customer engagement and
satisfaction statistics, policies to reduce administrative burdens, tracking and analyzing employee
performance evaluations, and feedback interviews (Eubanks, 2018; Kumar, 2012; Rathi, 2018).
The second branch of technical HRIS sets its focus on analyzing and interpreting
employee training and development data. This branch signifies an organization’s ability to invest
in human capital and bolster assets that already exist within to strengthen retention and further
develop their skills. Providing employees with training and education to perform new or hone
existing skills can be seen as a return on investment, which this component of HRIS evaluates.
Furthermore, this branch statistically identifies deficiencies and knowledge gaps within an
organization, prompting training managers to engage more effectively and secure training for
The historical review (Bhuiyan et al., 2014) provides an in-depth insight into the
evolution of HRIS. The drastic change in information technology during the 1990s
revolutionized HR professionals’ roles. Many of the hands-on and analog practices within HRM
experienced a great deal of technological evolution (Bhuiyan et al., 2014). Punch cards that log
an employee’s time became digital interfaces that allow employees to use their biometric
information to “clock in” to work. Engaging with one’s manager or supervisor to request paid-
time-off became a simple email rather than a conversation and logging a hand-written calendar.
11
Furthermore, digital job advertisements and applications have become more prevalent in the job
With the rise of big data and computing capabilities, HRM and HRIS constantly need to
cope with the amount of data they receive more effectively. The next frontier of digitization
evolution has required incorporating AI within these HRIS functions. HR personnel screening of
prospective applicants and their respective resumes can be done by AI applications that utilize
fuzzy and neuro-fuzzy-based agent approaches. AI can also create a short list of qualified
candidates based on what the HR department is looking for in prospective hires and candidates
There has been recent academic interest in how organizations implement AI within HRM
business practices (Abdeldayem and Aldulaimi, 2020; Collins, Dennehy, Conboy, & Mikalef,
2021; Di Vaio, Palladino, Hassan, & Escobar, 2020; Garg, Kiwelekar, Netak, & Ghodake, 2021;
Vrontis, Christofi, Tarba, Makrides, & Trichina, 2021). We refer to three systematic literature
reviews that have previously investigated AI within HRM (Di Vaio, Palladino, Hassan, &
Escobar, 2020; Garg, Sinha, Kar, & Mani, 2021a; Qamar, Agrawal, Samad, & Chiappetta
Jabbour, 2021; Vrontis, Christofi, Tarba, Makrides, & Trichina, 2021) and highlight how this
The previous literature reviews that have been conducted regarding AI and HRM
practices, although similar in nature, offer very different methodologies and perspectives to
consider. For instance, the bibliometric analysis conducted by Di Vaio et al. (2020) reviewed 73
academic articles that met their specified criteria. Specifically, Di Vaio et al. (2020) focused their
SLR toward understanding the state of the art of AI and small business models (which included
12
HRM practices) and providing a sense of direction for future research. Whereas, the SLR
conducted by Vrontis et al. (2021) analyzed 45 articles and mapped out research of AI within
learning opportunities, recruiting, training, job performance, and job replacement). The SLR
conducted by Qamar et al. (2021) provided insight into the state-of-the-art applications of AI
within the HRM domain and sets the stage for a future research agenda. They uniquely
articles. Lastly, (Garg, Sinha, Kar, & Mani, 2021a) provided a semi-systematic review of 105
management functions within the HRM spectrum and highlighted the need for HR experts and
practices.
Garg et al. Semi-systematic literature review; understand current state of ML 2002-2018 105
integration within HRM; showcase relationship between HR
experts and ML specialists
This Study Explore Tactical HRIS literature and come to understand which 2014-2020 33
components are exist in literature and how they are further
represented.
13
Complementing these three SLRs, this paper studies how AI exists within tactical HRIS
(T-HRIS) and HRM practices. We propose a framework that expands the difference between
technical T-HRIS (resource and data-driven components) and managerial T-HRIS (human-
centric components). Our analysis surveys the last six years of this development (2014-2020),
picking up where Bhuiyan et al.’s (2014) survey on the evolution of HRIS left off. Furthermore,
it uniquely focuses on AI’s application and relationships within tactical components of HRIS and
HRM.
AI uses many analytical methods ( Maettig & Foot, 2020). In this paper, we focus on four
methods that frequent AI and HRM/HRIS literature. Table 4 provides definitions of the four AI
categories.
Machine Learning
“Machine learning” refers to a machine’s ability to learn and perform a process given a
goal and defined steps (tasks) to train off to reach said goal. A generic example of machine
learning within the context of THRIS would be training a machine to recommend an employee
for promotion. The goal in this example would be the promotion recommendation (the task), and
the steps to get there would be the qualifications and criteria weighted against candidates in this
decision (the performance metric). Ideally, the training data would be a list of previously
recommended employees for promotion and their respective records, contributing to the
decision. The machine would then study those tens, hundreds, or thousands of records and
identify patterns to recommend given a set of new data (the stimulus/experience) (Mitchell,
2006).
14
Machine learning also utilizes deep learning techniques, which utilize neural networks
(NN) to analyze further and accomplish these complicated recommendation tasks. NN can
analyze copious amounts of data quickly and assist humans in recommending business decisions
based on analysis completed via NNs (Khan et al., 2020). As an example, such NNs can find
correlations between real-life events and sentiment changes via thousands of tweets (from
Twitter) and identify topics which influenced the sentiment of the public toward events
communicate with humans in their native tongue. Such capability has enabled machines to better
understand both speech and text and generate relevant responses to the stimulus it receives from
the human.
Such AI practices have become more prevalent when interacting with customers and
assisting employees as chatbots and language analysis algorithms have started introducing
themselves and streamlining various T-HRIS processes like employee onboarding, recruitment,
training, and leave requests (Garg et al., 2021) Majumder and Mondal, 2021). A chatbot can
interact with the individual and answer their questions autonomously (Eubanks, 2018; Majumder
Machine Vision
Machine vision uses cameras and other sensory devices to provide machines with the
ability to process visual data to better understand the world around them and make autonomous
decisions based on the received data. Examples of such capabilities are evident in the
15
cancerous abnormalities at the cellular level, facial recognition systems, and the US Post Office’s
ability to sort letters containing handwritten addresses (Mitchell, 2006; Rudin, 2019).
Within the context of T-HRIS, machine vision capabilities have started to integrate
algorithms that review video data to assist in automating the recruitment process by providing
HR professions a first impression, personality analysis, and recommending the pursuit of the
Recommendation Engines
Recommendation engines are AI tools that have increasingly infiltrated everyday life in
activities ranging from online shopping to find the right playlist of music or podcast to engage.
These engines allow companies like Amazon, Facebook, and Netflix to offer personalized
experiences to their customers (Xiao and Benbasat, 2007). These unique AI engines highlight
content or products that are more likely to appeal to the customers or employees of the company
based on behavioral trending data collected on them. The engines also analyze demographic
information on the individuals to help connect people via social media websites (“you may also
know” functions).
T-HRIS has also seen a growth in these capabilities via candidate recommendations,
not only streamline what once hours of individualized analysis were, but they have empowered
HRM professionals with the ability to process copious amounts of data and identify trends that
are outside of human capabilities to the computer. These recommendation engines are in action
in a variety of publicly accessible applications. For instance, LinkedIn, Instagram, and Facebook
16
are well known for connecting colleagues with one another. LinkedIn serves as a more
“professional” social media outlet (think virtual curriculum vitae), Instagram and Facebook cater
to the more personal aspects and hobby-based activities in one’s life (Papacharissi, 2009;
In summary, this section has provided insight and definitions to tactical HRIS
components and showcased a framework which highlights their placement within the structure.
We have further explored previous literature reviews which have been accomplished regarding
AI applications within HRM and highlighted their purposes, primary citations, and date-range
which they cover. Lastly, we have defined AI methodologies with HRM and applicable uses of
Table 4. AI Methods
Name AI Descriptions
Machine Learning The goal is defined and the steps to reach said goal are learned by
a machine (training). Uses algorithms and/or deep learning
techniques (neural networks) to learn autonomously from data
provided and assists by making decisions and providing insights
(Khan et al, 2020; Kumar 2019; Ridhwan and Hargreaves, 2021)
Natural Language Processing Automatic manipulation of natural language such as speech and
text. Aims to understand input from human speech as well as
generating responses in human languages (Kumar 2019)
17
CHAPTER 4: METHODOLOGY
This chapter provides a description of the systematic literature review (SLR) process
utilized within this study ((Collins, Dennehy, Conboy, & Mikalef, 2021; Di Vaio, Palladino,
Hassan, & Escobar, 2020; Vrontis, Christofi, Tarba, Makrides, & Trichina, 2021). Using the
SLR methodology we: 1) generate copious amounts of literature to analyze; 2) seek to answer
specific research questions; 3) seek to extract relevant pieces of academic literature relating
Review Process
Following Collins et al. (2020) and Okoli (2015), we conduct this review in 2 phases:
The first phase of this process filters literature, while the second phase extensively focuses on the
content of each article. The goal of this SLR is to understand the representation of T-HRIS in
published literature and how the representation of specific components differs. RQ1 seeks to gain
seeks to understand how the components of T-HRIS have been represented in literature by
identifying the analysis of the paper (empirical analysis, qualitative analysis, conceptual analysis,
or literature review). Through this SLR, we seek to provide the research community with a
review that investigates AI within HRM and T-HRIS. Considering AI and HRM are
1) Business Source Complete (EBSCO) (Nolan and Garavan, 2016; Vrontis et al., 2021)
3) Web of Science (WOS) (Collins, Dennehy, Conboy, & Mikalef, 2021; Di Vaio,
4) ABI-INFORM (ABI) (Baskerville & Myers, 2009; Nolan & Garavan, 2016).
18
Upon identifying our databases, we established our inclusion and exclusion criteria for
Phase 1 (Journal Demographic Filtration) and Phase 2 (Content Filtration). For Phase 1, there are
two filtration steps: 1) the initial search and 2) the journal verification. Our inclusion and
the initial search (step 1, Phase 1). The article must be: 1) peer-reviewed; 2) written in English;
2021; Dhamija and Bag, 2020; Di Vaio et al., 2020; Nolan and Garavan, 2016; Qamar et al.,
2021; Scandura and Williams, 2000). The exclusion of literature occurred if it: 1) was not written
piece.
For the journal verification (step 2, Phase 1), we identified 90 academic journals
(Appendix A from a wide range of disciplines that have historically published academic articles
and conference proceedings relating to HRM and AI (Collins, Dennehy, Conboy, & Mikalef,
2021; Di Vaio, Palladino, Hassan, & Escobar, 2020; Pisani, 2009; Qamar, Agrawal, Samad, &
Chiappetta Jabbour, 2021; Vrontis, Christofi, Tarba, Makrides, & Trichina, 2021). These
publication sources serve as inclusion criteria #5 and exclusion criteria #4. We included the
article if it belonged to the 90 journals or excluded it if it did not belong to any of those journals.
Once the journal verification state is complete, Phase 2 of the methodology beings: The
Content Filtration phase. The inclusion and exclusion criteria of this phase consist of two steps.
The first step is a title, abstract, and keywords (if provided) review, where we reviewed the title,
abstract and associated keywords of the article to see if it matched our pre-established definitions
of T-HRIS (Table 1 and Table 2) and AI Methods (Table 4). We also deleted duplicate articles
19
within step 1 of Phase 2. The second step of Phase 2 consists of reviewing the article in its
entirety to verify it met the requirement to be an article whose focus is on AI and T-HRIS
components of HRM.
Topic String
Artificial Intelligence ("AI" OR “artificial intelligence” OR "Natural Language Processing"
OR "NLP" OR "Chatbot" OR "Machine Vision" OR "Machine
Learning" OR "ML" OR “Recommendation Engine“ OR “Deep
Learning” OR “Neural Networks”)
Human Resource Management (“Human Resource Management” OR “HRM” OR “Human Resource
Information Systems” OR “HRIS” OR “Human Resources” OR “HR”
OR “Human Resource Management Systems” OR “HRMS”)
Full String
("AI" OR “artificial intelligence” OR "Natural Language Processing"
OR "NLP" OR "Chatbot" OR "Machine Vision" OR "Machine
Learning" OR "ML" OR “Recommendation Engine“ OR “Deep
Learning” OR “Neural Networks”) AND (“Human Resource
Management” OR “HRM” OR “Human Resource Information
Systems” OR “HRIS” OR “Human Resources” OR “HR” OR “Human
Resource Management Systems” OR “HRMS”)
Once the inclusion and exclusions criteria became concrete for Phase 1 and Phase 2, we
then sought to standardize our search string. One general search string strategy is to base the
string on specified research questions and a list of synonyms (Collins et al., 2021; Kitchenham,
2012). We pulled from our research questions and pre-defined AI methods (see Table 4). Our
search string consists of two pieces. The first piece is our key phrases regarding AI, while our
second piece comprised keywords relating to HRM, HRIS, and HRMS. We used a Boolean
practice when developing our search string. The "OR" operator exists between different words.
compartmentalized the AI-specific keywords from the HRspecific terms. Table 5 outlines the
20
Methodology in Action
Figure 2 demonstrates the selection process of our literature. Plugging the standardized
string into each database, cumulatively yielded 315,053 articles. Isolating the articles which met
the date, language, and peer-review requirements, achieved a total of 36,856 articles.
Subsequently we filtered based on the inclusion and exclusion criteria previously discussed.
Following this step, we further sorted through these articles based on the academic journals to
which they belonged, yielding a tertiary selection of 697 articles from the four publication
The tertiary selections from Phase 1 feed Phase 2 of our methodology. From this number
(697), we look at the content of each article selected. For step 1 of phase 2, we reviewed each
article’s title, abstract, and keywords based on definitions established in section 2 (Table 1, Table
2, and Table 4). Phase 2 was our initial screening to validate which of the 697 articles truly
connected AI with T-HRIS within HRM. We also took this time to delete duplicate articles from
the databases to secure a list of unique articles. Upon completing step 1 of Phase 2, we yielded a
secondary selection of 74 unique articles. These 74 articles fed the last and final step of our
method, which required us to read through the entirety of each article and validate that they were
met the criteria to be the primary studies selected based on our SLR.
21
Figure 2. SLR Methodology
22
CHAPTER 5: FINDINGS
The findings and analysis of literature within AI and HRIS are presented in this chapter.
showcases three unique branches which fall within this managerial T-HRIS literature umbrella.
The first branch of this consideration is employee performance and satisfaction, and the second
understanding where weaknesses exist. Given that the nature of delivering feedback is sensitive
professionals with more time to understand and interpret the analysis provided by the machine
and carefully deliver the feedback and findings to the employee in question. Furthermore, the
tool to ensure their employees are engaged and receiving fair and correct entitlements based on
autonomously engage with employees and streamline processes via simulating human-to-human
interactions. Their proposed model provides insight into an employee’s satisfaction and
performance concerning cultural and motivational factors, communication techniques, and the
work descriptions of each employee. The presented model demonstrates essential factors to
consider in AI-based communication tools and how they interact with one another. Where
23
faceto-face techniques excelled indirectly in motivating employees and increasing their
motivating and directly influencing an employee’s satisfaction. Thus, the tactical act of
communicating with employees and making that connection presented a challenge for AI-
Similarly, (Tong et al., 2021) research explores AI’s theoretical application in employee
performance feedback. It empirically evaluates the quality of feedback (which may increase
employee productivity) and the employee perception of the AI from a field experiment. Given
AI’s superior standardized and analytic nature, evaluating employee input and output, providing
areas of improvement, and predicting future performance has become a streamlined process for
HR professionals and T-HRIS applications. HR teams would manually review and scrub an
can do the analytics for them and provide recommendations directly to the employee. Despite the
brutally honest system, AI feedback can negatively affect the employees by disclosing their
weaknesses.
Consequently, direct honesty discourages them from furthering (otherwise known as the
"adverse disclosure effect"). This effect creates a unique paradox when considering improving an
directly and allow them to achieve complacency or disclose the employee’s areas of
improvement, potentially discouraging them further to seek improvement and achieve the same
complacency. Ultimately, (Tong et al., 2021) experiment and research conclude that these
24
Both (Pratt et al., 2021) and (Tong et al., 2021) highlight one of the most significant
weaknesses AI has within the managerial T-HRIS applications, lacking emotional intelligence.
effective and meaningful communication with employees. Managers and HR professionals need
to have the emotional wherewithal to cope with difficult emotional hurdles that may present
themselves. However, they also need to have the ability to communicate effectively while
respecting their employees’ boundaries. Prentice et al. (2020) explore this notion by empirically
studying the effects emotional intelligence and AI has on employee retention, satisfaction, and
performance within the hotel industry. The findings within their study conclude that emotional
intelligence is a valid predictor of employee attitudes and behaviors, and that AI does well to
focus on the technical and functional efficiency an employee has on organizational performance.
Despite having a significant impact on employee performance, AI holds little significance about
Finally, (Garg et al., 2021) provide insight to employee satisfaction within the logistics
feedback and satisfaction using AI algorithms. This further empowers an employee to answer a
climate survey on their experiences, analyzes the respondent’s input, and provides the
organization actionable insight which allows organization stakeholders more insight to how to
improve employee engagement, retention, and efficiency. Through this analysis, (Garg et al.,
2021) provide both the professional and academic community insight on how the use of AI has a
profound effect on the organization and its respective employees. These capabilities benefit the
25
employee’s voice and provides an avenue to communicate concerns and boons directly to the
organization.
The SLR we undertook yielded 4 total articles which featured Employee Performance
and Satisfaction. These 4 articles were consistent of 2 qualitative paper and 2 literature reviews.
The qualitative analyses conducted gives insight to individualization of HRM practices and
AImediated social exchanges and how AI impacts HRM practices and the attitudes and
behaviors of employees. Their collective findings via interviews suggest generational differences
in the adoption of AI and that AI-enabled bots and digital personal assistants are utilized in
analytical and routine tasks involving employees (Kaminska and Borzillo, 2018; Malik et al.,
2020). Within Malik et al.’s (2020) analysis, is suggested that these AI-enabled programs bolster
employee commitment and satisfaction within an organization. While, Kaminska and Borzillo
(2018) suggest fundamental differences between Genration X and Y employees and the adoption
of enterprise social networking systems driven by AI. The 2 literature reviews discovered within
this SLR provide insight to how AI technologies have infiltrated organizational settings by
natural language processing AI technologies and how it affects the feelings and thinking of
employees within the work environment. While Kock et al. (2020) systematically review AI
capabilities within employee performance reviews and how cognitive factors relating to rating
26
AI Literature within Discipline Management Systems
The research conducted by Dressel and Farid (2018) explores a new light of discipline
management by evaluating the predictive analysis of recidivism within the criminal justice
system. With the exponential growth of data and the development of AI-empowered machines to
cope with this growth, commercial software runs the risk of developing inaccurate predictions
that could affect human beings’ lives. Specifically, AI-algorithmic developments have fueled
tools like the Correctional Offender Management Profiling for Alternative Sanctions
two years of assessment from 137 different features about the individual and their respective
criminal history. (Escalante et al., 2017)’s analysis compared the overall accuracy and bias in
human assessments (individuals with no criminal justice background) with the algorithmic
assessment of COMPAS. Through (Dressel and Farid, 2018) ’s empirical analysis, the prediction
of the COMPAS algorithm was no more accurate or fair than the predictions of humans with
little to no criminal justice experience. The results were near indistinguishable, thus showcasing
the dangers of blindly implementing AI practices in high-risk decisions where human lives are
on the line.
with a discrete avenue to communicate and report abhorrent behaviors within an organization,
minimizing the fear of retribution and feelings of shame. Olson (2018) highlights Spot’s online
chatbot, which engages with the harassment victim on their experiences. This chatbot utilizes
NLP techniques to provide a "cognitive" interview to the employee and gain insight discretely on
details about their harassment claim. As Spot has grown within the company, it has developed
27
organization’s corporate cultures. This development has provided insight into where reports are
deriving from and identifying patterns that HR professionals cannot humanly conceive through
pattern recognition. Although it may not solve the problems directly, (Escalante et al., 2017)
concludes Spot gives insight into more significant issues that employees are not willing to speak
about aloud.
The qualitative analysis conducted by Bhattacharyya and Nair (2019) discusses how
future of work applications (robotics, artificial intelligence, internet of things) have multiple
facets via semistructured open-ended interviews with 26 respondents. Their analysis gives
insight to how organizations will have data dependencies and that employee will be expected to
synthesize data for sense making and decision processes which could affect work performance
AI applications within Technical T-HRIS have enabled once timeconsuming and task-
saturated processes to be automated and streamlined for both the employee and customer. These
tasks are data-driven to ensure the organization is optimized and can maintain or obtain a
competitive edge. The first branch of Technical T-HRIS we will review concerning AI is the
workflow and customer engagement within the Best Practices branch of T-HRIS. The second
branch explored will be the Recruitment Information Systems component which contains four
separate pieces of literature which review the effects and consequences of implementing AI
within these processes. The third branch is the Employee and Training Development Systems
branch. This branch contains three pieces of literature that explore the effects of AI within
28
employee education systems. Compensation and Benefits Analysis is the last branch, which
and benefits systems provides HR professionals with a tool to ensure their employees receive the
correct entitlements based on their unique circumstances while effectively budgeting company
expenses.
Cain et al. (2019) provide a compilation and systematic review of current studies and
journals which examine the impact of AI and robotics have within the hospitality and tourism
industry. Their review proposes emphasizing the importance of exploring research within the
human-robot interaction as these technologies grow to help streamline processes to give time
back to employees. Furthermore, they insist that research should focus on understanding how
customers interact with automated AIs capable of performing an employee’s tasks quickly.
Rahmani and Kamberaj (2021) review how introducing AIenhanced chatbots that interact
with customers and employees provides organizations with machines capable of learning and
automatically adapting to the environment based on data received and processed. Their analysis
provided insight into how the automation of administrative tasks within employee onboarding,
process improvement, and recruitment provide humans with the necessary information, time, and
psychological energy to make well-informed decisions for their organization. Their paper
assesses how an organization can use chatbots to streamline administration and explores what
technology they can run. It discusses how AI-enhanced chatbots can simulate comparably
“human” conversations using NLP and machine learning processes to enhance the customer
experience.
29
Further expanding on the use of Chatbots, Majumder and Mondal (2021) provides an in-
depth overview of the use of chatbots within HRM and the streamlined processes. The influence
chatbots have within decision-making processes is prevalent, but they also facilitate a better
example, Majumder and Mondal (2021) highlight the innovative application of an AI chatbot
within employee training modules, making them more interactive and engaging directly with the
employee, compared to the mandatory videos these same employees once had to watch.
Lastly, Chakraborty and Kar (2021) provide an example on how AI can be used to mine
data to gain insight on the wellbeing of employees during the unprecedented times following the
COVID-19 pandemic. Specifically, they conducted a mixed methodology analysis which utilized
COVID-19 created for employees across the work spectrum. Specifically, they addressed how
professionals across industries are affected by the pandemic, how that impact translates into
academia, and the nature of the impact on the social welfare of faculty members. Ultimately,
their analysis concluded there are systemic challenges revolving around infrastructure and digital
These innovations spotlight the benefits of having a machine engage directly with the
employee to answer questions they may have regarding pay, benefits, or any other HR-related
topic. Not only does this free up time for the HR professional, but it also provides a near-human
experience for the employee to directly engage with an HR entity to get the information they
30
AI Literature from SLR regarding Organizational Best Practices
Through our SLR, organizational best practices saw an influx of research within AI
capabilities as it was the most represented T-HRIS component with 11 empirical articles, 2
conceptual research articles, and 1 qualitative research article. Regarding the empirical analyses,
organizational best practices (Bani-Hani and Khasawneh, 2019; Chang and Jung., 2017; Cheng-
Kui et al., 2020; Fehrenbacher, 2017; Jabr and Zheng, 2014; Saha et al., 2016). Furthermore,
these empirical analyses provide insight to how AI technologies can replicate and influence HR
decision-processes within organizations, highlighting both the challenges and benefits that come
with the responsibility of leveraging these autonomous systems (Bani-Hani and Khasawneh,
2019; Castillo et al., 2018; Chang and Jung., 2017; Cheng-Kui et al., 2020; Fehrenbacher, 2017;
Jabr and Zheng, 2014; Kretzer and Maedche, 2018; Lankton et al., 2015; Lee and Ahn, 2020;
The conceptual research articles identified within this SLR regarding best practices
provide insight to how to cope and navigate computing systems which take on human-like
abilities to make decisions within its environment while providing transparency and a sense of
intuitive intelligence (Huang and Rust, 2018; Scheutz and Venkastesh, 2020). These conceptual
articles provide insight to how an organization and its employees can better interface with AI
technologies to enhance workplace procedures, services, and interactions (Huang and Rust,
Finally, one qualitative research article regarding best practices was identified within our
SLR. Yorks et al. (2020) interview 7 doctors and nurses from a variety of medical fields to gain
insight to AI adoption within the workplace and how it has transformed processes. Their
31
interview findings provided insight to how certain professional demographics (healthcare
workers) are reacting to the technological evolution within their respective workplaces.
Johnson et al. (2020) explore implementing electronic HRM and AI to help recruit highly
qualified employees, increase individual retention rates, and decrease the amount of time it takes
to onboard/replace new employees within the hospitality and tourism industry. Their research
proposes that AI facilitates two cognitive elements that promote decision-making within an
organization.
The first introduced is a cognitive insight which implies the algorithms and ML
techniques AI utilizes assists in interpreting the copious amount of data received and discovers
patterns not previously identified by the organization. Such advancements have enabled
organizations to engage in activities more effectively like predicting potential candidates for
hiring and potential internal hires within the organization. The second cognitive element Noone
and Coulter (2012) address is cognitive engagement, which implies using a chatbot to simulate a
human-like and near-natural social interaction with the employee and customer through NLP
technologies. These intelligent agents help with customer decisions and employee questions
within the “tech-support” realm giving time back to HR professionals at the tactical level (Noone
Given the recent global pandemic, Koch et al. (2021) utilize advanced web mining
robotics to evaluate how companies in Germany adjusted to the new processes caused by
COVID-19. Given the increased emphasis on public safety, HR professionals heavily relied on
Specifically, these authors explored how the German public sector job market reacted to the
32
pandemic and how this pandemic affected IT professionals with regards to electronic recruitment
systems (job listings, interview processes, telework capabilities). By utilizing intelligent robots to
scrape data necessary for their research, Koch et al. (2021) were capable of conjuring results
which indicate the importance of advertising work from home availability, increased planning
uncertainty (longer job postings), and that job advertisements with a healthcare background grew
To expand on the utilization of intelligent agents and chatbots, Rahmani and Kamberaj
(2021) further develop the research by assessing how a company can effectively use chatbots, the
technology on which chatbots run, and evaluating how well a chatbot can simulate a
types of chatbots and their respective architectures. This in-depth exploration provides HRM
professionals with insight into what chatbots exist. However, it also provides insight into how
interviewing a director from an IT company and an online survey of 210 respondents, (Stewart,
2007) could determine the following: 1) The implementation of AI in automated tasks has
proven to be beneficial for employees and customers. 2) AI should not be responsible for the
entire procedure of hiring new employees. 3) Empathy and emotional bias should play a role in
the employee recruitment process, where AI can help with eliminating bias the human-factor is
still important.
Within the SLR we conducted, 8 empirical, 3 conceptual, and 1 qualitative articles were
33
influence recruitment systems (Li et al., 2014) and how well AI systems can enhance the
candidate selection process within company recruitment initiatives (Karatop et al., 2015;
Martinez-Gil et al., 2020; Pessach et al., 2020; Sajjadiani et al., 2019; van Esch et al., 2019). The
3 conceptual papers expanded on how AI can impact how a company brands and markets itself
to potential recruits (Dabirian et al., 2017). Additionally, these conceptual papers provide insight
to when recruiters should incorporate AI within practices (Black and van Esch, 2020) and how to
use it to develop models which can predict company turnover via case-based reasoning (Wang et
al., 2017). Finally, the qualitative research article discovered within this SLR regarding
services connect organizational learning and recruitment via interviewing 12 soft-ware firms
located within Norway. These interviews highlighted a tension associated with the need to create
stable individual knowledge systems for employees and dependencies on external software
is crucial in securing various organizational goals. For one, ensuring employee participation in
annual security compliance and company-specific module training is a responsibility that are
levied upon HR personnel and departments writ large. Mandatory training compliance aside,
grooming employees by presenting educational opportunities eligible for postures for internal
hires and retention benefits. Understanding that the number of employees within an organization
varies, T-HRIS processes are ripe opportunities for AI to grow and streamline educational
based on an analysis of their aptitude, interests, and success potential is neigh. These capabilities
34
empower HR professionals to explore an employee’s potential further and not be burdened with
the hours of analysis it would take to run this analysis on each employee.
To further expand, Noone and Coulter (2012) evaluate Zaxby’s, a popular fast-food
restaurant originating from Athens, GA, and how they utilize autonomous robotic applications to
reduce service times and food waste substantially. They further explore how these same AI
applications can increase labor training efficiencies and opportunities for enhanced process
applications. Where information had been passed manually from person to person, quick service
restaurants, like Zaxby’s, have successfully applied AI-enhanced tools which store production
knowledge to enhance continuity of best practices to train future employees upon arrival.
Maettig and Foot (2020) further explore improving technical THRIS concerning
human-in-the-loop strategies. Like Rahmani and Kamberaj, 2021, Maettig and Foot (2020)
acknowledge that digital training assistants are great for storing best practices from older
employees to train newer hires. They further acknowledge the value in having digital assistants
analyze performance trends; however, they address that AI-enhanced digital assistants lack the
detailed experience knowledge that older employees may have. Particularly, the implicit
Similarly, Xu and Xiao (2020) introduce the concept of using AIenhanced virtual reality
Through their empirical analysis, these authors highlighted the application of virtual reality
technology within enterprise HRM, which can improve efficiency, reduce costs, and enhance the
35
Thus, these AI-enhanced training assistants are best used as augmentation devices to
recommending, and analyzing training, allowing humans to play a prominent role in filling in the
gaps, and adding a tailored and personalized approach to the onboarding process.
Regarding Employee Training and Development Systems, our SLR provided two articles
which expanded upon the incorporation of AI. The first article is empirical and conceptual in
nature. Through interviews, Maity (2019) discovered 33.33% of respondents believed intuitive e-
learning interfaces would be beneficial within their workplace. Maity (2019) further concludes
that 92.6% of HR training professionals believe that AI software within the digital learning
environment should interact with employees to further their engagement and development within
the company. The second article is conceptual in nature, as Lima (2020) explores the smart
organizations and how companies can start to build smarter learning platforms to increase
performance levels and help innovative organizations develop talented, creative, and diverse
employees.
The advancements of AI within resource-driven T-HRIS are also within the scope of
employee compensation and benefits. In particular, Robert et al. (2020) explore how
organizations have rapidly deployed AI systems to manage their employees and provide three
organizations to move beyond vague statements of what is considered fair to more specified
36
metrics to secure equal practices and train the AI. Should an AI be used to determine an
employees work similar hours in the same position and receive equal pay regardless of specific
can permeate and streamline. Data analytics relating to pay equity is crucial in securing fair and
ethical compensation and benefits practices within an HR department. Not only does an HR
department need to balance the needs of the organization’s employees, but it also must secure
crunching and inferring the unfathomable amount of information collected on its employees
enables HR professionals with the time to review the analysis and validate fair practices.
(Ahmed, 2018) provides insight into how crucial interpretability and transparency of the AI
mechanism is, as it provides HR professionals with an insight into the justification the AI
produces (Bourhis et al., 2019). Thus, this provides HR departments a more trustworthy process
(Hmoud and Várallyai, 2020). The notion of using black-boxes within high-stakes decision-
in unethical practices within an organization (Ahmed, 2018; Amer-Yahia et al., 2020; Rudin,
2019).
Within this SLR, there were no articles regarding Employee Compensation and Benefits
Systems discovered within the 33 articles that met our selection criteria.
37
How have the components of AI and T-HRIS been represented in the literature?
This research aimed to identify how the components of T-HRIS are represented within
the literature. Upon reviewing the methodologies of these articles, we identified that 19 were
empirical analysis, 5 were qualitative analysis, 7 were conceptual analysis, and 2 were literature
reviews. Based on this SLR, empirical studies dominated the results of our SLR, as our technical
Through our extensive review and using the definitions outlined in Table 2, we identified
which AI methodologies and applications are more represented within HRM and T-HRIS
literature (Table 6). Regarding the AI-type (y-axis), the variable titled “General AI Reference”
represents an article that referenced AI in a general sense and did not specify an AI method
within their paper. Using the definitions our analysis concluded a high representation of machine
Recommendation Engines
Machine Vision
Machine Learning
General AI Reference
0 2 4 6 8 10 12 14
38
Component Representation within each Analysis-Type
literature reviews identified, Best Practices and Recruitment Information Systems and Employee
Training and Development were the most represented component of T-HRIS regarding AI
application research. The managerial components, Employee Performance and Satisfaction and
Discipline Management, did not have representation with regards to empirical analysis.
was only referenced passively and in general terms of AI applications in 8 articles within this
SLR; there were no articles identified within this SLR which primarily focused on Compensation
and Benefit Information Systems and AI applications (see Figure 3). Given that technical T-
HRIS components are resource-driven and based on policies and SOPs, we have identified
through our SLR that empirical studies are currently more represented by technical T-HRIS
12
10
8
6
4
2
0
Best Practices Recruitment Employee Training Compensation and Employee Discipline
(Technical) Information and Development Benefit Information Performance and Management
Systems (Technical) Systems Satisfaction Systems
(Technical) (Technical) (Managerial) (Managerial)
39
Specifically, the Policies and Standards of Practice components of technical T-HRIS are
most represented by empirical research. Within the empirical analyses from our SLR, we further
identified common AI themes which these articles highlighted (see Figure 4). This provides
insight to the lack of empirical studies and analyses within managerial T-HRIS components. For
the components which were represented (Best Practices, Recruitment Information Systems, and
Employee Training and Development), our SLR identified the underrepresentation of machine
vision applications within HRM and T-HRIS, while the other methodologies (Recommendation
Engines, Natural Language Processing, Machine Learning, and AI Writ Large) are better
9
8
7
6
5
4
3
2
1
0
General AI Machine Learning Machine Vision Natural Language Recommendation Not Reprsented
Reference Processing Engines
Given the context of this SLR, we have identified a potential avenue to expand upon
empirical academic research within T-HRIS. Within our rigorous SLR, no managerial T-HRIS
components were represented within the group of empirical articles, giving reason to believe
future research could be done in this field to expand the T-HRIS research agenda.
40
Of the four qualitative studies, our SLR identified, four T-HRIS components were
Satisfaction, and Discipline Management Systems). Each component contained one piece of
the second least represented article type within this SLR. Nevertheless, it was the only article
Of the seven conceptual analyses, our SLR identified, there was representation in both
technical T-HRIS and managerial T-HRIS components. Like the empirical representation, the
conceptual analysis favors technical T-HRIS components considering six of the seven articles
related to Best Practices (three articles), Recruitment Information Systems (two articles), and
Employee Training and Development (one article). The only component of managerial T-HRIS
represented was Employee Performance and Satisfaction, leaving room for future research
opportunities.
Finally, of the two literature reviews, our SLR identified, only one component out of all
six possible (four technical, two managerial) was applicable. The Employee Performance and
Satisfaction component was represented in both reviews, leaving room for future literature
41
CHAPTER 6: DISCUSSION
Based on the results analysis of our research, the application of AI within tactical HRIS
practices needs further exploration within academia in a variety of areas. First, we consider the
contributions this SLR has provided to the existing literature and surrounding communities.
Secondly, we discuss future research opportunities this SLR has shined a light upon to further
develop this field and agenda. Lastly, we discuss the limitations of our research and how future
Through our research, we seek to shine a light on the importance of AI and HRM
practices through the exploration of existing literature relating to tactical HRIS components. The
prominent growth of digitalization within the work force and the evolution of AI within HRM
has helped improve areas like employee experiences and performance (Garg et al., 2021). Thus,
tactical HRIS components and where academics writ large are focusing their attention. We
sought to provide inductive insights to the current state of AI within HRM based on a literature-
driven and systematic methodology that is repeatable for future research opportunities to come
HRIS components, we discovered several gaps which may help expand the agenda of AI within
HRM. First, there is a glaring gap in published literature (relative to our search) between
managerial and technical HRIS practices. Where task-oriented and datadriving HRIS
components see more representation in literature, the more intuitive and emotionally intelligent
practices lack attention. Comparatively, this provides an opportunity to dive into why this gap in
42
research exists and further exploration as to why organizations veer away from incorporating AI
heavily in academic literature, given the parameters we set. Although we cast a large net of 90
journals to consider during our analysis, no articles of the 33 identified celebrated direct AI
applications within Employee Compensation and Benefits programs. This discovery invites
directly affect benefits and compensation programs within an organization and make that tactical
Moving forward, the implications of this research can equally guide the academic
community and industry professionals on understanding where AI exists within T-HRIS. This
offers an opportunity for synergy between the academics and industry professionals, as this work
provides insight to potential research areas regarding T-HRIS components and where AI is
processes. Moreover, the effects of this research provide the academic community with
awareness to the state of which AI is explored and the lack of research within managerial T-
43
CHAPTER 7: CONCLUSION
The purpose of this research was to review the integration of AI within HRM and HRIS
components; admittedly, there were limitations we encountered during our research that merit
exploring. First, we used a broad range of keywords like “artificial intelligence” and “human
resource management” within our search string to provide us with a wide coverage of topics in
HR and AI. This may have inhibited our ability to gather insights from other authors which may
have emerged outside of the selected keywords used in our string. Furthermore, we limited our
search to a specified number of research journals, which may have impacted the number we were
able to generate and analyze within our search. Lastly, we were limited in the access to
databases. To enhance this search, incorporating other reputable databases like SCOPUS may
and managerial (human-centric) T-HRIS. The reviewed literature presents a research gap
between technical and managerial T-HRIS, which is evident upon reviewing 33 articles across
four different databases, six T-HRIS components, four different AI methodologies, and four
article types. This analysis discovered that the more data-driven and task-orientated T-HRIS
applications are saturated in literature, and the managerial practices are less represented. The
review of existing research provides a foundation and direction for future research to address the
Going into this analysis, we set a goal to identify which T-HRIS components existed in
academic literature and how they were represented and sought to answer the research questions
derived from this goal. Based on our SLR, we could sift through 350,053 articles based on
specified criteria and narrow the list down to 33 articles that met the requirements and
44
parameters we sought to investigate. Upon completing our meticulous analysis, we have decided
to address some opportunities for future research based on the SLR conducted: First, technical T-
HRIS (resource and task-driven) components are overrepresented. There is potential for future
exploration on the managerial T-HRIS components, which are more human-centric. Second,
there are significantly more empirical studies when analyzing the relationship between AI and T-
HRIS, leaving room to expand the research agenda by providing more qualitative analysis,
conceptual analysis, and literature review research. In addition, out of all the literature analyzed
within this SLR, to include Phase 1 and Phase 2 of our methodology, this research solely takes
technical and managerial HRM and HRIS practices into consideration and individually analyzes
them. This review also paves the way for future exploration regarding technical and managerial
HRM and HRIS and its relationship with AI. Finally, based on our SLR, the Information
Systems and Management publication community claims 15 of the 33 total articles, which gives
us reason to believe these sources may be more receptive to research regarding AI and HRM
practices. This highlights potential blind spots for other publication venues seeking analysis on
Through this research, we seek to enhance research developments within the HR, IS, and
AI communities by bringing to light gaps within existing literature through a systematic review.
Although we faced limitations regarding our search string, journal analysis, and database usage,
we reinforce the future research agenda by providing new avenues to explore. We further
reinforce the motion that there is room for future academics and professionals to explore AI
for Employee Compensation and Benefits systems research and AI applications. To conclude, we
celebrate the academic and literature advancements of AI within HRIS and HRM and are excited
45
highlight the research potential which exists within tactical HRIS components. This review
provides the necessary groundwork to further grow this research agenda and gain a deeper
understanding of why the identified gaps exist with in literature and where professional and
academic stakeholders can focus their attention to further grow the field.
46
APPENDIX
47
Journal of Vocational Education and Training Strategic Management Journal
Journal of Workplace Learning Work, Employment and Society
Human Resource Development Journals Continued
Small Business Journal (7):
(12):
Learning Organization Entrepreneurship Theory and Practice
International Entrepreneurship and Management
Computer Science and Engineering Journals (4):
Journal
International Journal of Entrepreneurial
Computers and Industrial Engineering
Behavior and Research
International Journal of Entrepreneurship and
Computers in Human Behavior
Innovation
International Journal of Human-Computer Studies International Small Business Journal
Journal of Small Business and Enterprise
Journal of Computer Mediated Communication
Development
Journal of Small Business Management
48
REFERENCES
Ahmed, O. (2018). Artificial intelligence in HR. International Journal of Research and Analytical
Reviews, 5(4), 971-978.
Amer-Yahia, S., Basu Roy, S., Chen, L., Morishima, A., Abello Monedero, J., Bourhis, P., ... &
Yoshida, K. (2020). Making ai machines work for humans in fow. ACM SIGMOD Record, 49(2), 30-
35.
Badri, H., Bashiri, M., & Hejazi, T. H. (2013). Integrated strategic and tactical planning in a supply
chain network design with a heuristic solution method. Computers & Operations Research, 40(4),
1143–1154. https://doi.org/10.1016/j.cor.2012.11.005
Bani-Hani, D., & Khasawneh, M. (2019). A Recursive General Regression Neural Network (R-
GRNN) Oracle for classification problems. EXPERT SYSTEMS WITH APPLICATIONS, 135, 273–
286. https://doi.org/10.1016/j.eswa.2019.06.018
Bhattacharyya, S. S., & Nair, S. (2019). Explicating the future of work: Perspectives from India. The
Journal of Management Development, 38(3), 175–194.http://dx.doi.org.libweb.lib.utsa.edu/10.1108/
JMD-01-2019-0032
Black, J. S., & van Esch, P. (2020). AI-enabled recruiting: What is it and how should a manager use
it? Business Horizons, 63(2), 215–226. https://doi.org/10.1016/j.bushor.2019.12.001
Bourhis, P., Demartini, G., Elbassuoni, S., Hoareau, E., & Rao, H. R. (2019). Ethical Challenges in
the Future of Work. IEEE Data Eng. Bull., 42(4), 55-64.
Bhuiyan, F., Chowdhury, M. M., & Ferdous, F. (2014). Historical evolution of human resource
information system (HRIS): An interface between HR and computer technology. Human Resource
Management Research, 4(4), 75-80.
Cain, L. N., Thomas, J. H., & Alonso, M. (2019). From sci-fi to sci-fact: The state of robotics and AI
in the hospitality industry. Journal of Hospitality and Tourism Technology, 10(4), 624–650. http://
dx.doi.org.libweb.lib.utsa.edu/10.1108/JHTT-07-2018-0066
Castillo, A., Vander Meer, D., & Castellanos, A. (2018). ExUP recommendations: Inferring user’s
product metadata preferences from single-criterion rating systems. Decision Support Systems, 108,
69–78. https://doi.org/10.1016/j.dss.2018.02.006
Chakraborty, A., & Kar, A. K. (2021). How did COVID-19 impact working professionals – a
typology of impacts focused on education sector. The International Journal of Information and
Learning Technology, 38(3), 273–282. https://doi.org/10.1108/IJILT-06-2020-0125
49
Chang, W., & Jung, C. (2017). A hybrid approach for personalized service staff recommendation.
Information Systems Frontiers, 19(1), 149–163. ProQuest One Academic; SciTech Premium
Collection. https://doi.org/10.1007/s10796-015-9597-7
Cheng-Kui, H., Wang Tawei, & Huang Tzu-Yen. (2020). Initial Evidence on the Impact of Big Data
Implementation on Firm Performance. Information Systems Frontiers, 22(2), 475–487. ProQuest One
Academic; SciTech Premium Collection. https://doi.org/10.1007/s10796-018-9872-5
Collings, D. G., Wood, G. T., & Szamosi, L. T. (Eds.). (2018). Human Resource Management: A
Critical Approach (2nd ed.). Routledge. https://doi.org/10.4324/9781315299556
Cregan, C., Kulik, C. T., Johnston, S., & Bartram, T. (2021). The influence of calculative (“hard”)
and collaborative (“soft”) HRM on the layoff-performance relationship in high performance
workplaces. Human Resource Management Journal, 31(1), 202–224. https://doi.org/10.1111/1748-
8583.12291
Dabirian, A., Kietzmann, J., & Diba, H. (2017). A great place to work!? Understanding crowdsourced
employer branding. Business Horizons, 60(2), 197–205. https://doi.org/10.1016/j.bushor.2016.11.005
Dahlbom, P., Siikanen, N., Sajasalo, P., & Jarvenpää, M. (2020). Big data and HR analytics in the
digital era. Baltic Journal of Management, 15(1), 120–138. https://doi.org/10.1108/BJM-11-2018-
0393
De Kock, F. S., Lievens, F., & Born, M. Ph. (2020). The profile of the ‘Good Judge’ in HRM: A
systematic review and agenda for future research. Human Resource Management Review, 30(2),
100667. https://doi.org/10.1016/j.hrmr.2018.09.003
Dhamija, P., & Bag, S. (2020). Role of artificial intelligence in operations environment: A review and
bibliometric analysis. The TQM Journal, 32(4), 869–896. https://doi.org/10.1108/TQM-10-2019-
0243
Doctor, F., Hagras, H., Roberts, D., & Callaghan, V. (2009a). A fuzzy based agent for group decision
support of applicants ranking within recruitment systems. 2009 IEEE Symposium on Intelligent
Agents, 8–15. https://doi.org/10.1109/IA.2009.4927494
Doctor, F., Hagras, H., Roberts, D., & Callaghan, V. (2009b). A neuro-fuzzy based agent for group
decision support in applicant ranking within human resources systems. 2009 IEEE International
Conference on Fuzzy Systems, 744–750. https://doi.org/10.1109/FUZZY.2009.5277379
Dressel, J., & Farid, H. (2018). The accuracy, fairness, and limits of predicting recidivism. Science
Advances, 4(1), eaao5580. https://doi.org/10.1126/sciadv.aao5580
50
Escalante, H. J., Guyon, I., Escalera, S., Jacques, J., Madadi, M., Baró, X., Ayache, S., Viegas, E.,
Güçlütürk, Y., Güçlü, U., van Gerven, M. A. J., & van Lier, R. (2017). Design of an explainable
machine learning challenge for video interviews. 2017 International Joint Conference on Neural
Networks (IJCNN), 3688–3695. https://doi.org/10.1109/IJCNN.2017.7966320
Eubanks, B. (2018). Artificial Intelligence for HR: Use AI to Support and Develop a Successful
Workforce. Kogan Page Publishers.
Garg, S., Sinha, S., Kar, A. K., & Mani, M. (2021). A review of machine learning applications in
human resource management. International Journal of Productivity and Performance Management,
ahead-of-print(ahead-of-print). https://doi.org/10.1108/IJPPM-08-2020-0427
Garg, R., Kiwelekar, A. W., Netak, L. D., & Ghodake, A. (2021). i-Pulse: A NLP based novel
approach for employee engagement in logistics organization. International Journal of Information
Management Data Insights, 1(1), 100011. https://doi.org/10.1016/j.jjimei.2021.100011
Foot, H. (2020). Approach to improving training of human workers in industrial applications through
the use of Intelligence Augmentation and Human-in-the-Loop. 2020 15th International Conference on
Computer Science Education (ICCSE), 283–288. https://doi.org/10.1109/ICCSE49874.2020.9201867
Hmoud, B. I., & Várallyai, L. (2020). Artificial Intelligence in Human Resources Information
Systems: Investigating its Trust and Adoption Determinants. International Journal of Engineering and
Management Sciences, 5(1), 749–765.
Hmoud, B., & Laszlo, V. (2019). Will artificial intelligence take over human resources recruitment
and selection?. Network Intelligence Studies, 7(13), 21-30.
Huang, M.-H., & Rust, R. T. (2018). Artificial Intelligence in Service. Journal of Service Research,
21(2), 155–172. https://doi.org/10.1177/1094670517752459
Hughes, C., Robert, L., Frady, K., Arroyos, A., Hughes, C., Robert, L., Frady, K., & Arroyos, A.
(2019). Artificial Intelligence, Employee Engagement, Fairness, and Job Outcomes. In Managing
Technology and Middle- and Low-skilled Employees (pp. 61–68). Emerald Publishing Limited.
https://doi.org/10.1108/978-1-78973-077-720191005
Jabr, W., & Zheng, Z. (Eric). (2014). Know Yourself and Know Your Enemy: An Analysis of Firm
Recommendations and Consumer Reviews in a Competitive Environment. MIS Quarterly, 38(3),
635-A10.
Johnson, R. D., Stone, D. L., & Lukaszewski, K. M. (2020). The benefits of eHRM and AI for talent
acquisition. Journal of Tourism Futures, 7(1), 40–52. https://doi.org/10.1108/JTF-02-2020-0013
51
Jones, C., & Saundry, R. (2012). The practice of discipline: Evaluating the roles and relationship
between managers and HR professionals. Human Resource Management Journal, 22(3), 252–266.
https://doi.org/10.1111/j.1748-8583.2011.00175.x
Jøranli, I. (2018). Managing organisational knowledge through recruitment: Searching and selecting
embodied competencies. Journal of Knowledge Management, 22(1), 183–200. Entrepreneurship
Database; ProQuest One Academic; SciTech Premium Collection. https://doi.org/10.1108/JKM-12-
2016-0541
Kaminska, R., & Borzillo, S. (2018). Challenges to the learning organization in the context of
generational diversity and social networks. Learning Organization, 25(2), 92–101. https://doi
.org/10.1108/TLO-03-2017-0033
Karatop, B., Kubat, C., & Uygun, Ö. (2015). Talent management in manufacturing system using
fuzzy logic approach. Computers & Industrial Engineering, 86, 127–136. https://doi.org/10.1016/
j.cie.2014.09.015
Karim, A. J. (2011). The significance of management information systems for enhancing strategic
and tactical planning. JISTEM - Journal of Information Systems and Technology Management, 8,
459–470. https://doi.org/10.4301/S1807-17752011000200011
Khan, W. A., Chung, S. H., Awan, M. U., & Wen, X. (2020). Machine learning facilitated business
intelligence (Part I). Industrial Management & Data Systems, 120(1), 164–195. https://doi.org/10.1
108/watIMDS-07-2019-0361
Kitchenham, B. A. (2012, September). Systematic review in software engineering: where we are and
where we should be going. In Proceedings of the 2nd international workshop on Evidential
assessment of software technologies. 1-2.
Koch, J., Plattfaut, R., & Kregel, I. (2021). Looking for Talent in Times of Crisis – The Impact of the
Covid-19 Pandemic on Public Sector Job Openings. International Journal of Information
Management Data Insights, 1(2), 100014. https://doi.org/10.1016/j.jjimei.2021.100014
Kotera, Y., Sheffield, D., & Van Gordon, W. (2019). The applications of neuro-linguistic
programming in organizational settings: A systematic review of psychological outcomes. Human
Resource Development Quarterly, 30(1), 101–116. https://doi.org/10.1002/hrdq.21334
Kretzer, M., & Maedche, A. (2018). Designing Social Nudges for Enterprise Recommendation
Agents: An Investigation in the Business Intelligence Systems Context. Journal of the Association for
Information Systems, 19(12), 1145–1186. http://dx.doi.org.libweb.lib.utsa.edu/10.17705/1jais.00523
Kumar, R. (2012). Human resource information system: An innovative strategy for human resource
management. Gian Jyoti E-Journal, 1(2), 1-12.
52
Laker, D. R., & Powell, J. L. (2011). The differences between hard and soft skills and their relative
impact on training transfer. Human Resource Development Quarterly, 22(1), 111–122. https://doi
.org/10.1002/hrdq.20063
Lankton, N. K., McKnight, D. H., & Tripp, J. (2015). Technology, Humanness, and Trust: Rethinking
Trust in Technology. Journal of the Association for Information Systems, 16(10), 880–918.
Lee, D., & Ahn, C. (2020). Industrial human resource management optimization based on skills and
characteristics. Computers & Industrial Engineering, 144, N.PAG-N.PAG. https://doi.org/10.1016
/j.cie.2020.106463
Li, Z., Xu, W., Zhang, L., & Lau, R. (2014). An ontology-based Web mining method for
unemployment rate prediction. DECISION SUPPORT SYSTEMS, 66, 114–122. https://doi.org/
10.1016/j.dss.2014.06.007
Lima, M. (2020). Smarter organizations: Insights from a smart city hybrid framework. International
Entrepreneurship and Management Journal, 16(4), 1281–1300. Entrepreneurship Database; ProQuest
One Academic. https://doi.org/10.1007/s11365-020-00690-x
Liu, Y., Pant, G., & Sheng, O. R. L. (2020). Predicting Labor Market Competition: Leveraging
Interfirm Network and Employee Skills. Information Systems Research, 31(4), 1443–1466. https://
doi.org/10.1287/isre.2020.0954
Maity, S. (2019). Identifying opportunities for artificial intelligence in the evolution of training and
development practices. Journal of Management Development, 38(8), 651–663. https://doi.org/10.1
108/JMD-03-2019-0069
Majumder, S., & Mondal, A. (2021). Are chatbots really useful for human resource management?
International Journal of Speech Technology. https://doi.org/10.1007/s10772-021-09834-y
Malik, A., Budhwar, P., Patel, C., & Srikanth, N. R. (2020). May the bots be with you! Delivering
HR cost-effectiveness and individualised employee experiences in an MNE. The International
Journal of Human Resource Management, 0(0), 1–31. https://doi.org/10.1080/09585192.202
0.1859582
Martinez-Gil, J., Paoletti, A. L., & Pichler, M. (2020). A Novel Approach for Learning How to
Automatically Match Job Offers and Candidate Profiles. Information Systems Frontiers, 22(6),
1265–1274. ProQuest One Academic; SciTech Premium Collection. https://doi.org/10.1007/s10796-
019-09929-7
Mayfield, M., Mayfield, J., & Lunce, S. (2003). Human resource information systems: A review and
model development. Advances in Competitiveness Research, 11(1), 139–151.
McCarthy, J., Minsky, M., Rochester, N., & Shannon, C. (2006). A Proposal for the Dartmouth
Summer Research Project on Artificial Intelligence: August 31, 1955—ProQuest. AI Magazine,
27(4), 12–14.
53
Mitchell, T. M. (2006). The discipline of machine learning (Vol. 9). Pittsburgh: Carnegie Mellon
University, School of Computer Science, Machine Learning Department.
Moor, J. (2006). The Dartmouth College Artificial Intelligence Conference: The Next Fifty Years. AI
Magazine, 27(4), 87–87. https://doi.org/10.1609/aimag.v27i4.1911
Muhammad, A. U., Shah, Z. A., & Azhar, K. A. (2021). The Increasing Role of Hris in Facilitating
Hr Functions in Pakistan’s Banking Sector. International Journal of Information, Business and
Management, 13(1), 24–34.
Noone, B. M., & Coulter, R. C. (2012). Applying Modern Robotics Technologies to Demand
Prediction and Production Management in the Quick-Service Restaurant Sector. Cornell Hospitality
Quarterly, 53(2), 122–133. https://doi.org/10.1177/1938965511434112
Obeidat, B. Y. (2012). The Relationship between Human Resource Information System (HRIS)
Functions and Human Resource Management (HRM) Functionalities. Journal of Management
Research, 4(4). https://doi.org/10.5296/jmr.v4i4.2262
Olson, P. (2018, March 2). This Chatbot Is Helping People Track Harassment At Work. Forbes.
https://www.forbes.com/sites/parmyolson/2018/03/02/chatbot-spot-sexual-harassment-ai/
Pessach, D., Singer, G., Avrahami, D., Chalutz Ben-Gal, H., Shmueli, E., & Ben-Gal, I. (2020).
Employees recruitment: A prescriptive analytics approach via machine learning and mathematical
programming. Decision Support Systems, 134, 113290. https://doi.org/10.1016/j.dss.2020.113290
Pratt, M., Mohcine, B., Taskin, N., & Cakula, S. (2021). Use of AI for Improving Employee
Motivation and Satisfaction (pp. 289–299). https://doi.org/10.1007/978-3-030-68201-9_30
Prentice, C., Lopes, S. D., & Wang, X. (2020). Emotional intelligence or artificial intelligence– an
employee perspective. Journal of Hospitality Marketing & Management, 29(4), 377–403. https://doi
.org/10.1080/19368623.2019.1647124
Rahmani, D., & Kamberaj, H. (2021). Implementation and Usage of Artificial Intelligence Powered
Chatbots in Human Resources Management Systems.
54
Rathi, D. R. (2018). Artificial intelligence and the future of hr practices. International Journal of
Applied Research, 4(6), 113-116.
Ridhwan, K. M., & Hargreaves, C. A. (2021). Leveraging Twitter Data to Understand Public
Sentiment for the COVID‐19 Outbreak in Singapore. International Journal of Information
Management Data Insights, 100021.
Robert, L. P., Pierce, C., Marquis, L., Kim, S., & Alahmad, R. (2020). Designing fair AI for
managing employees in organizations: A review, critique, and design agenda. Human–Computer
Interaction, 35(5–6), 545–575. https://doi.org/10.1080/07370024.2020.1735391
Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and
use interpretable models instead. Nature Machine Intelligence, 1(5), 206–215.
https://doi.org/10.1038/s42256-019-0048-x
Rybinski, K., & Tsay, V. (2018). The Application of Machine Learning in Faculty Assessment: A
Case Study of Narxoz University. HUMAN RESOURCE MANAGEMENT, 122/123(3/4), 145–170.
Saha, P., Bose, I., & Mahanti, A. (2016). A knowledge based scheme for risk assessment in loan
processing by banks. DECISION SUPPORT SYSTEMS, 84, 78–88. https://doi.org/10.1016/j.dss.
2016.02.002
Schuetz, S., & Venkatesh, V. (2020). Research Perspectives: The Rise of Human Machines: How
Cognitive Computing Systems Challenge Assumptions of User-System Interaction. Journal of the
Association for Information Systems, 21(2), 460–482. http://dx.doi.org.libweb.lib.utsa.edu/10.177
05/1jais.00608
Shum, H., He, X., & Li, D. (2018). From Eliza to XiaoIce: Challenges and opportunities with social
chatbots. Frontiers of Information Technology & Electronic Engineering, 19(1), 10–26. https://doi.
org/10.1631/FITEE.1700826
Stapleton, P., Luiz, G., & Chatwin, H. (2017). Generation Validation: The Role of Social Comparison
in Use of Instagram Among Emerging Adults. Cyberpsychology, Behavior, and Social Networking,
20(3), 142–149. https://doi.org/10.1089/cyber.2016.0444
Stewart, T. A. (2007). The wealth of knowledge: Intellectual capital and the twenty-first century
organization. Currency.
Tariq, O., Sang, J., & Gulzar, K. (2016). Design and Implementation of Human Resource Information
Systems Based on MVC a Case Study Vocational Education in Iraq. International Journal of U- and
e- Service, Science and Technology, 9(11), 15–26. https://doi.org/10.14257/ijunesst.2016.9.11.02
55
Tong, S., Jia, N., Luo, X., & Fang, Z. (2021). The Janus Face of Artificial Intelligence Feedback:
Deployment Versus Disclosure Effects on Employee Performance. Strategic Management Journal.
University of Maryland. (2021). University Libraries Research Guides: Steps of a Systematic Review.
University of Maryland Libraries. https://lib.guides.umd.edu/SR/steps
Using machine learning to translate applicant work history into predictors of performance and
turnover. - PsycNET. (n.d.). Retrieved August 25, 2021, from https://psycnet.apa.org/doiLan
ding?doi=10.1037%2Fapl0000405
van Esch, P., Black, J., & Ferolie, J. (2019). Marketing AI recruitment: The next phase in job
application and selection. COMPUTERS IN HUMAN BEHAVIOR, 90, 215–222. https://doi.org/
10.1016/j.chb.2018.09.009
van Esch, P., & Black, J. S. (2019). Factors that influence new generation candidates to engage with
and complete digital, AI-enabled recruiting. Business Horizons, 62(6), 729–739. https://doi.org/10.1
016/j.bushor.2019.07.004
Wang, X., Wang, L., Zhang, L., Xu, X., Zhang, W., & Xu, Y. (2017). Developing an employee
turnover risk evaluation model using case-based reasoning. Information Systems Frontiers, 19(3),
569–576. ProQuest One Academic; SciTech Premium Collection. https://doi.org/10.1007/s10796-
015-9615-9
Watson, H. (2017). Preparing for the cognitive generation of decision support. MIS Quarterly
Executive, 16, 153–169.
Wohlin, C., Runeson, P., Höst, M., Ohlsson, M. C., Regnell, B., & Wesslén, A. (2012).
Experimentation in software engineering. Springer Science & Business Media.
Xiao, B., & Benbasat, I. (2007). E-Commerce Product Recommendation Agents: Use,
Characteristics, and Impact. MIS Quarterly, 31(1), 137–209. https://doi.org/10.2307/25148784
Xu, D., & Xiao, X. (2020). Influence of the Development of VR Technology on Enterprise Human
Resource Management in the Era of Artificial Intelligence. IEEE Access, 1–1. https://doi.org
/10.1109/ACCESS.2020.3020622
Yorks, L., Rotatori, D., Sung, S., & Justice, S. (2020). Workplace Reflection in the Age of AI:
Materiality, Technology, and Machines. Advances in Developing Human Resources, 22(3), 308–319.
https://doi.org/10.1177/1523422320927299
Zehir, C., Karaboğa, T., & Başar, D. (2020). The Transformation of Human Resource Management
and Its Impact on Overall Business Performance: Big Data Analytics and AI Technologies in
Strategic HRM. In U. Hacioglu (Ed.), Digital Business Strategies in Blockchain Ecosystems:
Transformational Design and Future of Global Business. Springer International Publishing, 265-279.
https://doi.org/10.1007/978-3-030-29739-8_12
56
ESSAY II. JC-COMPASS: A FRAMEWORK FOR CONDUCTING COMPETENCY-
57
CHAPTER 1: ABSTRACT
As Artificial Intelligence (AI) and Machine Learning (ML) capabilities expand across the
data science career field, Human Resource (HR) professionals face the challenge of acquiring top
talent to fill necessary vacancies within their respective organization. Thus, developing strategies
to orient and communicate the needs of the organization have become paramount in maintaining
a competitive edge in recruiting employees for job vacancies. Accordingly, there has been
considerable emphasis on understanding the specific skills and experience demanded of data
scientists within job postings. However, most methods have yielded lists of specific skills
demanded of data scientists and very little insight on systematic approaches to thematically
Thus, we build and introduce a framework to guide such research called Job Competency
Compass: Initiate, Establish, Analyze, and Socialize (JC-Compass). The JC-Compass framework
provides a focused and systematic approach with detailed guidelines for industry and academic
entities interested in understanding and researching thematic trends within job postings by
building dictionaries of word similarity . The four steps to conducting job description research
competencies, and 4) socialize the findings. To illustrate how the JC-Compass can be utilized, an
example empirical study leveraging the TF-IDF method is reported. This example analysis
illustrates how JC-Compass can guide industry professionals and scholars in analyzing thematic
58
CHAPTER 2: INTRODUCTION
With the evolution of Artificial Intelligence (AI) and Machine Learning (ML) saturating
the workforce, companies face the challenge of hiring and internally securing the right talent to
manage, develop, and operate these assets ethically (Enholm et al., 2022). This evolution has
consequently affected how human resource professionals manage the necessary skill demands
that come with the evolution of technology and work environments. The same human resource
professionals are responsible for effectively communicating the company's needs and tailoring
this demand to attract prospective applicants and train current employees. Thus, competency-
based approaches have become a popular technique to oversee the portfolio of skills the
organization owns, as there is an added pressure to ensure those competencies capture the
Big data, AI, and ML have accelerated business intelligence capabilities for the United
States federal government and private sector. As a result, research endeavors exploring job
targeting accuracy and skills needed to manage challenges within jobs have grown considerably
in recent years. Social capabilities (soft skills), team experience (both leader and followership),
and technical understanding have become key points of interest for academic and industry
researchers investigating job postings (Anton et al., 2020; Huang, 2021). However, one major
research gap that IS scholars have currently left unexplored is the investigation of leveraging
Today, there has been little work which has reported on large-scale identification,
collection and analysis of competency themes which exist within job postings. Within current job
posting research, many authors and investigators explore specific skills needed for the job; albeit
59
we understand skill identification is imperative to grasp current workforce demands given a
specific career field. Modeling our research after Kim et al.’s (2005) design science approach,
we propose a job posting research framework which steps away from analyzing specific skills
and encourages NLP capabilities to investigate thematic trends based within job postings. Such a
systematic approach to explore job posting themes as well as enhance acquisition strategies.
The arrangement of this paper is organized into the following sections. First, we discuss
the background on job posting research which leverages NLP capabilities. This section compares
job posting research with other tactical human resource management (T-HRM) research that has
specifically used NLP capabilities. Furthermore, this section highlights relevant and recent job
posting research and existing frameworks. Next, details are provided concerning how to
operationalize a job posting research project. Within this section, we highlight methods in
identifying appropriate data sources (job posting sites), as well as techniques for collecting
content from these sources. We further provide insight to potential analytical directions once the
job postings have been collected. Following this section, we conduct a post-hoc analysis to
further explore and identify how the groups may differ from each other by year. Finally, we
discuss the contributions and implications of this work. To summarize, this paper provides a
60
CHAPTER 3: BACKGROUND
While human resource analytics is a rich topic within business literature, there has been a
drastic growth in interest surrounding the importance of business intelligence and streamlining
decision processes via the incorporation of machine learning and AI. That said, HRM is not a
catch-all term. There are many spectrums, perspectives, and unique responsibilities of human
resource management that exist. For instance, there is a strategic lens of HRM which considers
the overall effects of activities which enable an organization to achieve its goals. This lens
considers how HRM -as a whole- affects performance-related outcomes at the organizational
level (Boon et al, 2018; Perlman, 2016). Unethical practices (biases), market competition, and
reduction of resources are challenges that exist at this level which affect an organization’s health,
reputation, and performance (Perlman, 2016). Beneath the strategic level, there is a tactical lens
which refers to the managerial and technical actions and responsibilities which fall under the
HRM scope of responsibility (Votto et al., 2021). Managerial components are centered around
employee performance, expertise, culture, and discipline. This subsection of T-HRM consists of
employee performance, satisfaction, and discipline management. The technical component refers
Given the rich amount of data T-HRM encompasses, it is no stranger to machine learning
and AI methodologies to help cope with analyzing trends. Specifically, NLP capabilities have
become a useful tool when analyzing text, speech, and semantics. In essence, NLP capabilities
seek to understand input from human speech and/or text and generate responses understandable
to humans (Kumar, 2019). Such technologies include chatbots which that interact with humans
61
and autonomously answer questions related to recruitment qualifications, training opportunities,
leave requests, and onboarding procedures (Garg et al., 2021; Majumder and Mondal, 2021).
Other T-HRM components have leveraged NLP capabilities to analyze copious amounts of data
and extract impactful information from it. Such practices have been seen screening prospective
resumes (Chuang et al., 2009; Cole et al., 2003), reviewing employee climate surveys (Layug
2018), and detecting fake job postings (Amaar et al., 2022). The ability for NLP to help HRM
contribute to an extremely impactful body of work. Thus, the exploration and development of
new capabilities for NLP could support organizations and businesses in generating stronger
Specifically, recent research within the recruitment T-HRM component has investigated
the ability to leverage job postings from across the internet to make inferences on trending skill
demands of employers. There are online communities which foster a space for businesses to post
job requisitions such as LinkedIn, Indeed, GlassDoor, Google, Monster, and many others. Within
these prominent websites are millions of job postings, usually tied to keywords related to the
vacant position. This enables prospective employees to type keywords into a search engine and
receive job postings catered to their search criteria. From these sites, employers and applicants
are connected. Information, application submissions, company hyperlinks, and other valuable
information are shared between the two parties. Given the amount of text information embedded
within each job posting, NLP has become an invaluable tool for analyzing thousands of job
postings to generate an understanding of trends and patterns. Although business literature is rich
with job posting research, there lacks a structured framework to guide job posting research. That
said, the following subsection dives deeper into recent literature surrounding the “recruitment”
62
T-HRM domain. In particular, job posting research and the use of NLP to conduct analyses. We
With the recent explosion of academic and industry interest in NLP capabilities within
recruitment interests, HRM literature has yielded a variety of analyses that have built a
foundation for our research. Highlighting these impacts, we seek to convey how the contribution
of our research is unique and differs from existing initiatives. To comprehensively review
previous literature, we refer to eight peer-reviewed articles involving job description analysis and
NLP capabilities which are summarized in Table 1. We also review three published frameworks
which discuss job posting research efforts which is summarized in Table 2. We extracted these
articles from Google Scholar, leveraging the “most relevant” search option. Regarding Table 1,
we leveraged the following strings to generate a list of articles to review: “Job Posting Research
NLP”, “Job Description Research NLP”, “Job Posting Analysis NLP” and “Job Description
Analysis NLP”. We manually reviewed the first page of results (most relevant) and extracted
analyses which were directly related to job postings and NLP capabilities. Regarding Table 2, we
leveraged the same technique with different search strings. The search strings we utilized for
Table 2 were “Job Posting NLP Research Framework” and “Job Description NLP Research
Framework”.
Regarding Table 1, each article is unique and provides a thorough analysis of job
postings, we identify a similar structure in how these analyses are conducted, such that a
potential framework could be derived and built from them to enhance future research efforts. For
instance, each article identifies a source to collect their job postings from, be it Indeed or general
web crawling efforts. Furthermore, each article determines a collection methodology such as
63
scraping websites or manual extraction. Lastly, each article identifies an analysis technique such
The articles summarized in Table 2 showcase existing frameworks within the job posting
domain. Two of the four articles promote frameworks for extracting information out of resumes
to match up against job descriptions. The remaining articles discuss systematic approaches to
reviewing job descriptions via temporal, spatial, and curriculum considerations. By analyzing
could systematically lead future research endeavors in analyzing job descriptions with NLP
capabilities. Rather than specialize in resume extraction and specific considerations against job
postings, we identified a prospect to add a new framework to the existing literature which
framework promotes the use of Word2Vec modeling to build dictionaries based on established
job competencies within the organization. Our proposed framework differs from existing
organization deems relevant to the career field. The following subsection dives deeper into the
64
Table 1. Previous Job Posting NLP Literature Summary Continued
Kortum et Computer 291 NLP Job Leverage text Evaluated data via Generated
al. 2022 Vision/NLP Postings; 591 mining similarity test and collection of
Job Computer techniques to extracted skills needed skills related to
Descriptions Vision Job compare from both categories of AI and ML,
(Indeed) Postings Computer Vision jobs showcased NLP
to NLP job and Computer
descriptions Vision Jobs
skills differ
significantly
Varelas et Online Job Not Disclosed Leverage NLP to Evaluated Support Proves simpler
al. 2022 Portals from classify jobs Vector Model, Random methods can be
Greek according to Forest, K-Nearest equal or more
Portals ISCO Neighbors, Stochastic powerful than
Occupation Gradient Descent, and deep learning
Codes for Labor Neural Network techniques if
Market classifier performance they are
Intelligence via accuracy and F1- combined
efforts scores properly
Vo et al. Public Job 2,907 Job Leverage NLP to Benchmarked Developed and
2022 Postings Descriptions; recommend Computer validated a tool
(GlassDoor); 1,886 Course personalized Science/Information which
Course Descriptions course/job Technology universities can
Descriptions recommendation Recommendation use to help
(university s for Computer System (CSIT-CRS) students find
websites) Science/Informat against other jobs
ion Technology established/validated
Students models & user survey
Zhu et al. Postdoctoral 823 Leveraged web Utilized NLP toolkit Demonstrated
2021 Job Postings Engineering scraping tools comprehensively ability of NLP
(Indeed) and Computer and NLP to review Knowledge techniques to
Science extract and Skill Assessments, generate insights
Postdoc Job analyze word frequency, and to postdoctoral
Postings Postdoctoral job domain disciplines experiences
postings within the computer within the
science and engineering computer science
career field – conducted and engineering
Kruskal Wallis test to career field
demonstrate
significance of
differences
Lunn et US 3,824 Job Demonstrate Python Libraries: Programming
al. 2020 Computer Postings (770- NLP capabilities WordCloud Bigram and testing are
Science Job New York; of extrapolating Analysis, Pandas (data important skills
Postings 774- San job data from manipulation), for employment
(Indeed) Francisco; online resources Matplotlib (generate within the
745- San Jose; diagrams), NumPy Computer
752- Seattle; (mathematics extension Science
783- of Matplotlib). community.
Washington Validated by randomly Specifically,
DC) inspecting 50 postings Python and
manually for accuracy machine learning
knowledge is
preferred.
65
Table 1. Previous Job Posting NLP Literature Summary Continued
Meyer US 198 Job Conduct content Leverage R- Showcased
2019 Healthcare Postings analysis which programming NLP to uptick in demand
Data categorizes core code job postings, of data scientists
Science Job information from develop skill matrices, within healthcare
Postings job postings and visualize content organizations
(Indeed) within healthcare analysis and need for
industry professional
development,
training and
education to
grow career field
Sibarani UK Job 872 Job Analyze job Ontology-Based Automated
et al. 2017 Postings Advertisement postings based Information Extraction efforts proved
(Indeed) s on the context of and Paris/Keele method feasible in
skills and for cluster analysis extracting job
competencies compared against postings when
needed to fill a manual (human) compared to
role equivalent for accuracy of
extracting skills from human
job posts equivalent
66
CHAPTER 4: INTRODUCING COMPETENCY-BASED HUMAN RESOURCE
investigating job descriptions have focused interests on specific skill trends and mechanisms to
providing resources to match candidates with listed postings, as further discussed in Table 2.
Although previous research initiatives have been skill-based and extensive, limitations
surrounding the generalizability of the results have been identified (Lunn et al. 2020; Meyer,
2019). In other words, the specificity of the targeted job demographic has encouraged future
research efforts to explore different career fields and themes to enhance job description analysis
techniques (Lunn et al., 2020). Prior research surrounding frameworks have indicated a paucity
Srivastava, 2014; Laber and O’Connor, 2000). Considering these deficiencies, cross-disciplinary
approaches for analyzing job descriptions are needed (Yu et al., 2021).
To address this call, we consider a design science paradigm leveraged by Kim et al.
science, artifacts take many forms. For instance, an IT artifact can be a construct, model,
technological rule, or method applied in the development and use of information systems and
science (Gregor and Hevner, 2013; Hevner et al., 2004). Hevner et al. (2004) define methods as
a mechanism which provides guidance on how to solve or explore problems within a solution
space. To further expand, methods can range from formal equations and algorithms to textual
descriptions of “best practices” (Abbasi et al., 2012; Hevner et al., 2004). Studies prior have
67
leveraged the design science paradigm to develop frameworks to guide research endeavors
(Abbassi et al., 2012; Benjamin et al., 2019; Kim et al., 2005). Accordingly, we were inspired to
In the absence of sufficient design guidelines, studies have placed a strong emphasis on
the need for design theories to govern the development process of an IT artifact (Abbasi et al.,
2012; Abbasi and Chen, 2008; Markus et al., 2002). Thus, to achieve developing a cross-
HRM) as a kernel theory to guide the development of our proposed job competency research
modeling methods which can be used to address the research gaps mentioned previously. This
results in a proposal for an enhanced job description analysis technique. In subsequent sections, a
job description analysis framework based on C-HRM is presented and later evaluated with an
empirical example.
Competency-based HRM has become an integral practice within organizations over the
last thirty years, as it has enabled businesses to better structure the needs from its human capital.
Competency models have historically been used to help organizations match employees to jobs
(Carretta, 1992), design improvement to job redesign efforts, recruitment strategies, and career
2007).
68
Table 3. Various Definitions of Competency
Taking these definitions into consideration, we derive that competencies refer to themes
that are directly related to an employee’s work performance. To expand, competencies act as an
umbrella which house specific skills, knowledge, and characteristics related to the employee’s
work responsibility, such that enhanced skills lead to better competency comprehension which
then leads to better work performance (Ratnawat, 2015). Within literature, competencies have
been used for a wide range of HRM function such as employee selection, retention initiatives,
and organizational planning (Ratnawat, 2015; Rodriguez, 2002). Thus, leveraging a competency
69
Human Resource Competency Models
is often associated with frameworks and guidelines to help organizations develop systematic best
practices. Within this research, we first observe Kandula’s (2013) macro competency modeling
technique. Specifically, this approach outlines the need to initiate strategic oversight and
competency framework should have a strategic tie to the organization. In other words, when
developing macro competencies, it is imperative to take the organization’s vision, mission, and
goals into consideration when determining competencies for its employees. Furthermore, once
the competencies are developed, socializing the results, and gaining insight to future directions
is determined that competency frameworks that have no linkage or weak linkage to company
vision , mission, and goals will contribute very little to the desired results.
Mansfied (1996) and Chouhan and Srivastava (2014). There are two additionally approaches
however, this is considered the fastest and broadest approach to establishing competencies within
an organization (Chouhan and Srivastava, 2014). Industry and HRM professionals seeking a fast
encourages defining a set of competencies for a broad range of related jobs (e.g., all data science
jobs) (Chouhan and Srivastava, 2014; Mansfield, 1996). Where the former approach engages
employees directly, this second approach charges a research team to develop competencies using
concepts from books, job descriptions, articles, and existing frameworks within leadership,
70
business, organizational development, and human resource outlets. Where the “Single-Job”
technique is tailored to a specific job within an organization and the “Multiple-Job” focus more
There are two primary considerations we take away from the “One-Size-Fits-All”
approach. The first is identifying the population for whom the model will apply (e.g., data
science jobs). Secondly, developing an HR application to assist in establishing and analyzing the
proposed competencies is also a common practice within this competency model (Mansfield,
developing competencies.
Chouhan and Srivastava (2014), and Mansfield (1996) to create our research model. Kandula’s
(2013) perspective is represented at the beginning and end of the model, “Initiate” and
“Socialize”, respectively. The “initiate” component encourages the study of company goals,
mission, and vision, whereas the “socialize” component encourages reflection and
communication of results. Furthermore, the research insights of Chouhan and Srivastava (2014)
and Mansfield (1996) are represented within the second and third section in the figure. The
“establish” component encourages the identification of job population and review of literature,
while the analyze phase reinforces the need to establish analysis guidelines. Subsequent sections
dive deeper into specified methods extrapolated from each of these theories to further develop
71
CHAPTER 5: GENERAL OVERVIEW OF TEXT ANALYSIS METHODS FOR
COMPETENCY RESEARCH
research model was understanding recent text analysis methods used within HRM research
to C-HRM, it is important to review text analysis techniques which could potentially tie into
research frameworks. Where previous sections of this paper have provided insight to published
NLP literature and driving C-HRM theory leading the development of this proposed research
embedding and text frequency analysis methods to competency-based job description research
initiatives.
evident Word2Vec was published before the others observed. Subsequent models identified
Word2Vec limitations and further optimized its capabilities to accomplish new tasks such as text
sentiment classification (Shen and Liu, 2021). For instance, Shen and Liu (2021) highlight
significant differences between Word2Vec and the Bidirectional Encoder Representation from
Transformers (BERT) model. Where Word2Vec generates word embeddings that are context-
independent, BERT generates embeddings which allow a word to have multiple vectors because
it takes context into consideration when Word2Vec does not (Shen and Liu, 2021).
However, these newer models developed limitations of their own, despite their enriching
abilities. Although the first of its kind and its limitations, Word2Vec is still considered an
efficient word embedding strategy to investigate similarity relations between specified domain
72
terminologies (competencies) (Goldberg and Levy, 2014; Quan et al. 2018). It continues to grow
in popularity and assist in semantic research endeavors given its minimal memory requirement
and word similarity analysis capabilities (Kai et al., 2019; Gao et al., 2022).
In a historical context, Word2Vec was developed by Tomas Mikolov and his colleagues
distribution and composition of words within a corpus (Goldberg and Levy, 2014; Mikolov et al.,
2013). Having identified that NLP systems historically atomized words, Mikolov et al. (2013)
sought to explore the notion of word similarity by developing a tool which vectorized words.
Word vectorization refers to an NLP method which maps words from a corpus to a
corresponding vector of real numbers, which is then used for word similarity or prediction
analysis. Vector representations which are close to one another indicate word similarity
Like most NLP methods, Word2Vec heavily relies upon three major steps when
conducting similarity analyses: 1) data preparation, 2) model training, and 3) method analysis.
Data preparation refers to the identification of a data source, collection and cleaning techniques,
and tokenization the words for the model. The second step refers to how the model is trained.
Within this step, model parameters are reviewed and determined based on the needs of the
research objective. Lastly, method analysis refers to the use of a method unique to the model’s
library to conduct an analysis. For the sake of the later mentioned framework, the most_similar()
method is referenced. This tool finds the top-N most similar words that contribute positively
toward similarity. Essentially, the user trains the Word2Vec model on a corpus of words model
from which the user calls upon the similarity method to identify which N words were most
73
similar to a designated word (or competency). Figure 1 summarizes the word embedding steps
competency exceeds the interests of this research endeavor. The proposed framework seeks to
create dictionaries of words that are similar to established competencies. Specific context of the
phrase and/or competency is irrelevant and exceeds the interests of this endeavor. That said, it
could be a future direction of this research endeavor. Given the prominent nature of Word2Vec
for NLP tasks (Adewumi et al., 2021), we proceeded with this being our primary model within
74
Text Frequency Analyses
job description research. Where word embeddings help identify prominent words throughout a
corpus, text frequency analyses provide an additional empirical metric for measuring importance.
When studying competencies within job descriptions, it is crucial to measure the degree of
importance surrounding each established competency (Tang et al., 2022). Job descriptions
developed by HRM professionals reflect the needs of the organization and associated
competency qualities that are in great demand through the enterprise, which may be repeatedly
Per Qaiser and Ali (2018), the total length of documents can vary, which leads to the
possibility that any term may occur more frequently in longer word-count documents when
compared to smaller documents. Thus, the occurrence of any term in a document is divided by
the total terms present in that document. This division calculates the actual term frequency.
Regarding inverse document frequency (IDF), this component acknowledges that not all
words can be treated equally, given that each keyword has a different importance level (Qaiser
and Ali, 2018). Thus, IDF assigns lower weights to frequent words and assigns greater weights to
words that are not as frequent. TF-IDF score is simply the multiplication of term frequency (TF)
75
CHAPTER 6: THE JOB COMPETENCY RESEARCH FRAMEWORK
DEVELOPMENT
With the help of C-HRM perspectives of Kandula (2013), Chouhan and Srivastava
(2014), and Mansfield (1996), we have identified three requirements for our research model,
socializing the findings. In this section, we describe the components of our proposed framework
The first section of this model involves initiating job description research. If a research
project wishes to succeed, it must be directly connected to a goal established by the research
party, such that there is a clear link in research intentions and problem-solving initiatives
(Probert et al., 2003). Within competency development, a weak linkage will not contribute to the
desired results (Kandula, 2013). Therefore, the first section of this framework encourages the
initiation of research by establishing research goals to guide analysis. This can be achieved by
explore. Once researchers have developed their research goals, establishing, and analyzing the
Establish Competencies
Keeping the research goal in mind, we consider Mansfield’s (1996) and Chouhan and
did not include the first or third modeling technique can be found in the next paragraphs.
76
As defined in previous sections, the “Single-Job Competency Model” is an approach
which leverages employees and management input to derive competencies. Although engaging,
this approach is considered costly with regards to time and money, especially during the data
leadership), especially if participation is incentivized, costs the organization time and money
when taking the initial steps to developing competencies. Furthermore, once the data has been
collected, it must be analyzed and derive common themes throughout the stakeholder feedback.
Regarding the “Multiple-Job Model”, this method is traditionally used when the jobs in
question have nothing in common. For example, if an organization has data science and human
resource management vacancies. Therefore, this this method relies on the construction of general
“Building block” competencies tend to reflect wholistic non-technical skills that are universal,
each with a unique definition and ascribed behaviors detailing how an employee can demonstrate
proficiency. Where the other two approaches focus either on a specific job or category of jobs,
the Multiple-Job Approach proposes the development of universal competencies with job-
block” competency, the job-specific expectations to demonstrate proficiency may differ for data
scientists and human resource employees. Understanding this method is job agnostic, it falls out
of our scope of framework research interest. The intent of the proposed framework is to help
academic and industry professionals research and better understand a career field.
77
As mentioned in section 3.2., Chouhan and Srivastava’s (2014) and Mansfield’s (1996)
objective of the proposed research framework is to arm both academic and industry researchers
with a systematic approach to thematically exploring job postings for a specified career field
(e.g. data scientists), we found the “Once-Size-Fits-All” methodology to be the most cost
effective and efficient way to derive competencies. That said, there are limitations to choosing
this model. For instance, it is heuristic in nature. Competencies may fluctuate or change based
on the literature, job descriptions, and resources the research team reviews may limit consistency
throughout literature. However, the inconsistencies could provide a boon to the academic and
industrial community because it opens an opportunity for rich discussions and perspective to be
published sources. Thus, creating a rich collection of scholarly insight on a designated job
category. Thus, the initiation component of the model adopts the second “One-Size-Fits-All”
model.
Analyze Competencies
Once the competencies have been established, analyzing the proposed competencies
follows. These three-phases consist of data preparation, model training, and similarity analysis
(see Figure 1). The following subsections provide deeper insight into each component.
Data Preparation
The preparation phase is the backbone of the analysis. Within this phase it is imperative
to establish where the job description data is going to come from, how the researchers plan to
extract the data, and what measures the research team will take to clean and prepare the data for
tokenization.
78
With regards to where the data will come from, scholars have identified a wealth of
sources to acquire job information. Namely, websites that allow applicants to search keywords
and set personalized search parameters such as LinkedIn, Indeed, and GlassDoor have grown in
popularity when conducting job description research (Christensen et al., 2005; Gupta, 2021;
Hosain and Liu, 2020; Landers et al., 2019; Marinescu et al., 2020; McCabe, 2017; Muthyala et
al., 2017). Thus, identifying a source to conduct research is dependent on how the research team
plans to extract the needed data. There are open-source data depots like GitHub or Kaggle which
allow scholars to share data for other research endeavors. Additionally, there is also the option
of scraping data directly from the desired websites. Although, both techniques come with
benefits and concerns. With regards to open-source data, the research team is trusting the source
to have not tampered with or defame the data. Although it is the easiest option to acquire job
description information, researchers are faced with the assumption that it is an honest
representation of what was pulled. Regarding web scraping, many websites have policies and
measures which prevent machine learning techniques from extracting information from them.
Furthermore, memberships and money-walls are also a factor as some companies monetize the
data they are responsible for. There is the option of reaching out directly to a desired company
and acquiring a developer API token; however, not all job posting websites offer this option.
Although web scraping may yield the most data, it faces many challenges to acquire it.
The last consideration of this phase is the preprocessing and tokenization component.
Assuming the data source has been selected and job postings have been collected, it is imperative
to prepare the data for the Word2Vec model. This requires data cleansing such as eliminating
null data and duplicate entries. Furthermore, leveraging the Natural Language Took Kit (NLTK)
python library to efficiently prepare text data for analysis has proven to be an effective tool for
79
NLP analysis (Millstein, 2020; Mohan, 2015). The NLTK library is a robust and renowned tool
within the NLP community that contains a variety of functions to help clean. This research
model proposes the use of regular expressions to conduct our initial preprocessing and the
Wisdom and Gupta, 2016). The objective of cleaning this information is to remove noisy data
from the job postings to generate key meaningful words (tokens). By doing this, the Word2Vec
model used in phase 2 can better vectorize clear word associations, given our established
hosts free tutorials, online courses, coding competitions, and webinars by industry experts to
Table 4 summarizes our preprocessing steps and associated literature and sources. To
effectively preprocess the information, we leveraged published articles and books to guide the
80
Table 4. Data Pre-Processing Steps Continued
Stem Words (strip end of “comput”, “exampl” Anandarajan et al., 2018
word)
Lemmatize Words “computer”, “example” Anandarajan et al., 2018
(identify root word)
Model Training
The once the tokenized words have been established, training Word2Vec model within
the Genism library (Ozkur et al., 2022) becomes essential in developing dictionaries of similar
words based on the established competencies. Tuning the parameters of the model become
imperative within this component, as the wrong combination of hyper-parameters are known to
produce poor quality vectors (Adewumi et al., 2021). Thus, reviewing the number of vector
directions (size), context word window size (window), minimum word count (min_count), and
training algorithm (sg) are a few of many optional parameters to take into consideration when
Model Analysis
inputted, and the function will return the 10 closest words from the target word (Kurnia, 2020).
This function establishes the competency dictionaries required for thematic analysis.
Understanding the tokens have been stemmed and lemmatized, the input words must reflect a
similar format. Once the dictionaries have been organized and established, the analysis phase can
prominent and commonly used term weighing methodology which provides indication to word
importance (Chen et al., 2016; Jalilifard et al., 2021). TF-IDF scores range between zero and
one. If the word or dictionary receives a score that is closer to zero, this indicates it is not
81
relevant within the document. On the other hand, if the word or dictionary receives a score that is
Within this framework, each established competency dictionary will have a score
between zero and one within each job posting. Once all TF-IDF scores have been established for
each dictionary against each job posting, an average can be taken to better understand which
competencies are more prevalent. However, Post-Hoc statistical testing is imperative to confirm
differences which may appear, Phase 4 provides a space to interpret the results of the TF-IDF
Socialization
The last phase of this proposed research model enables researchers to interpret and
communicate the results of the TF-IDF competency analysis. By investigating the calculated
averages of TF-IDF scores, researchers can determine which competency dictionaries were more
prevalent. Furthermore, statistical testing and means comparisons can be conducted to confirm
and validate findings. For instance, ANOVA testing can be used for comparing group means and
confirm differences which may exist (Park, 2009). Additionally, if comparing two data sources, a
two-tailed test can be leveraged in determining significance as it challenges the null hypotheses
that the datasets are similar (UCLA: Statistical Consulting Group, n.d.). Once all statistical
considerations have been considered, researcher can then review the results of their analysis and
report their findings to enrich the existing pool of literature. Although, the socialize component
goes beyond communicating the results of the analysis. Reflecting on limitations and future
82
CHAPTER 7: THE JC-COMPASS FRAMEWORK FOR JOB DESCRIPTION
RESEARCH
The purpose of this section is to tie theory and analysis components together to establish
a research framework. Due to the variety of ways to analyze job descriptions, a standardized and
would be of great value for industry and academic professionals. The Job Competency Initiate,
Establish, Analyze, and Socialize (JC-Compass) framework described here serves to guide
researchers through their own exploration of job descriptions given a research objective. This
framework leverages the research theory which combines C-HRM and NLP practices and is
intended to help professionals see a project through its entire life cycle when using Word2Vec
There are four primary sections of this cross-disciplinary model. The first is identified as
initiation, which refers to the creation of research purpose. This can be achieved via creating
research goals, identifying a research problem, and/or developing research questions regarding a
83
designated phenomenon. The second section of this research framework is the establishing of
managers, data scientists, etc.) and reviewing existing frameworks, literature, and guidance
surrounding the targeted demographic to establish competencies. Once the competencies are
created, the framework shifts to the analysis phase where the researchers are directed extensively
analyze the competencies via reviewing job descriptions. Within this phase, data preparation
commences. This refers to the identification job description data sources, collection, cleaning,
and tokenization of words. Additionally, the Word2Vec model is trained on the preprocessed
words (tokens) to enable the last component, similarity analysis. The method analysis enables the
component is comprised of communicating the results of the analyses and statistical comparisons
encourages the while the second is methodology. Given this framework marries C-HRM theory
with NLP practice, we observe the NLP components are embedded within the competency
dictionary development and job description analysis. The initiation section ignites the research
project, placing the responsibility of establishing and reviewing research goals and developing
job competencies on the scholar. The methodology section is broken into four phases: 1)
preparation, organization, analysis, and socialization (Meyer, 2019). This section is used to guide
competency-based job description analysis via Python libraries. Subsequent sections discuss
84
CHAPTER 8: EMPIRICAL DEMONSTRATION OF JC-COMPASS
framework and dictionaries. This chapter is broken down into four sections. The first section
discusses the initiation phase of this example. The second section covers the establishment of
competencies to guide the research. The third section discusses the analysis methodology.
Lastly, the final section discuses the results and socializes the findings.
To evaluate our job description analysis framework, we demonstrate the value of the
description data from the private sector and United States Federal government. The premise of
this example serves two objectives. The first objective is to demonstrate how this framework can
be used when working with archival data. The second is to explore how the United States
Federal Government thematically differs from the private sector when communicating job
requirements for data science vacancies. The United States Bureau of Labor Statistics projects a
36% growth of the data science career field between 2021 and 2031, which is much faster than
the average for all other occupations within the federal and private sector. This evolution has
consequently affected how human resource professionals manage the necessary skill demands
that come with the evolution of technology and work environments. Historically, the federal
sector relies heavily on contracted data science personnel rather than full-time staff, which has
stemmed the Office of Personnel Management (OPM) to derive a new data science job series to
better recruit more fulltime data science personnel. This shift in interest consequently creates a
competitive environment for recruiting top-tier talent within the data science field. Thus, we seek
to fortify recent research initiatives by diverting from the trend of analyzing specific skills
85
desired of data scientists and investigate how the federal and private sector communicate their
respective desires via natural language processing techniques to establish nomenclature and
lexicons related to data science competencies derived from theory and industry practice. In doing
so, this example seeks to showcase the proposed research model by educating the industrial and
academic community on how the private and federal sector thematically communicate their
needs from data scientists rather than just analyze desire skills. By utilizing our proposed
research model, this example draws attention to thematic differences within job descriptions,
which in turn could bolster recruitment strategies such that HRM professionals can remain agile
Establish Competencies
As newer AI and ML technologies penetrate industries and data science jobs worldwide,
securing the right talent to manage and develop these tools has become a premium for
organizations within the private and federal sectors. Renowned organizations like the National
Security Agency, Google, IBM, and the Department of Homeland Security leverage data
scientists to conduct predictive analysis on various complex problems that challenge the
organization's success and national security. That said, establishing job competencies has
become a necessary tool for human resource managers across the federal and private sectors
(Gangani et al., 2006; Hattingh et al., 2019; Rodriguez et al., 2002). As the data science career
field has evolved, newer technologies and requirements have presented new responsibilities for
data scientists. This has created a multitude of data science frameworks and varying
competencies across the industry. To guide this research example, we leverage the practices of
86
competency frameworks published by Deloitte and IBM to propose 5 core competencies related
develop ethical safeguards across six key dimensions (Ammanath, 2022; Mökander and Floridi,
2021). The six key dimensions of this framework are as follows: 1) transparent and explainable
AI, 2) fair and impartial, 3) robust and reliable, 4) privacy, 5) safe and secure, and 6) responsible
and accountable AI. Understanding ethical practices of data science are imperative to building a
successful organization. We extracted two core competencies from this Deloitte framework to
gauge how job descriptions solicit skills. Specifically, this framework addresses the unique
advancements from committing unfair or unsafe practices. Thus, the first competency we
extracted for our data competency framework is ethics. This competency encompasses the social
responsibilities and skills necessary to uphold within the work environment. The framework
explicitly addresses the unique capabilities of these smarter technologies and proposes ways to
the unique capabilities of these intelligent technologies and proposes ways to safeguard
technology advancements from committing unfair or unsafe practices. Artificial Intelligence (AI)
recommendation engines, and deep learning techniques (to name a few) (Votto et al., 2021).
Incorporating it as a data science competency within our model ensures that sub-interests related
87
Figure 3. Adopted Data Science Competencies
The second core framework we leverage to build our competency network stems from
IBM's Data Science Skills Competency Model. Within this framework, there are 28 critical
competencies broken into seven groups. These competencies hold data scientists accountable
within the work environment. These key competencies range from problem-solving to
understanding analytical cycles (Gottipati et al., 2021; IBM, 2020). We extracted three additional
primary competencies from this framework to guide our dictionary development. The first
competency extracted from this framework is "machine learning," which relates to employees
components. The second competency extracted from this framework is "problem-solving," which
refers to an employee's ability to identify and characterize business problems and develop
solutions. The last component we extracted from this framework was an employee's statistical
capabilities, otherwise denoted as "statistics" within our model. The statistics component refers
AI is in the narrative leading into skill development. For the sake of this research
example, IBM's framework alone is limited. Although it addresses the rapid growth of AI
88
capabilities and required responsibilities, it does not mention "ethics" within the document, nor
does it include artificial intelligence within the established competencies. Therefore, we leverage
the Deloitte framework to fill in these gaps and propose five primary competencies which guide
this research: data, ethics, problem-solving, and machine learning (Figure 3). These
competencies will later be the foundation for building term dictionaries to run term-frequency
AI is in the narrative leading into skill development. For the sake of this research
example, IBM's framework alone is limited. Although it addresses the rapid growth of AI
capabilities and required responsibilities, it does not mention "ethics" within the document, nor
does it include artificial intelligence within the established competencies. Therefore, we leverage
the Deloitte framework to fill in these gaps and propose five primary competencies which guide
this research: data, ethics, problem-solving, and machine learning (Figure 3). These
competencies will later be the foundation for building term dictionaries to run term-frequency
Data Science Competencies within the Federal Government and Private Sector
Technological capabilities have allowed for developing new and innovative data science
tools to streamline once tedious processes to simple key-stroke actions. The US federal
government and private sector have experienced growth spurts within new data science
capabilities, requiring new and innovative talent to manage these tools. Regarding the federal
government, institutional barriers regarding training and awareness limit the implementation and
acceptance of these newer technologies (Kreuger et al., 2019). By reinforcing systematic training
requirements and hiring standards, the US federal government has successfully developed
reskilling opportunities related to machine learning, artificial intelligence, and data science to
89
adopt AI (Tarraf et al., 2019). Furthermore, literature surrounding smarter IT and the federal
government reinforces the need for strong-minded Chief Information Officers to enhance the
strategic role of IT within government initiatives, especially as near-peer adversaries and peer
competitors tirelessly strive toward a competitive advantage (Khallaf and Majadalawieh, 2012;
Peppard, 2007; Wagemann Jr., 2020). Although ambitious, there is a delicate balance the federal
government must adhere to the allocated budget. ‘Doing more with less’ initiatives promote
internal data and statistical upskilling opportunities to respond to the growing need for data
scientists. However, these initiatives have also created struggles within decision-making
processes, given the complex political, managerial, and democratic changes which fluctuate
throughout the government (Janessen and Estevez, 2013; Tarraf et al., 2019). Thus, the Office of
Personnel Management (OPM) renovated its data science occupational series, which outlines the
need for multifaceted data scientists with an interdisciplinary background. Through these
guidelines, the federal government identifies a broad and general range of necessary skills that
need identification when considering talent, emphasizing skills directly related to machine
learning and statistics (OPM, 2021). Through this renovated definition of data science
requirements, upskilling initiatives to grow data scientists have ignited, allowing existing
employees to apply for these opportunities and return to their work centers with a deeper
In contrast, the private (for-profit) sector has much smoother decision-making processes,
given the general objective of selling products or services to consumers to generate stakeholder
wealth (Nutt, 2005). With the increasing challenges of collecting, transforming, and analyzing
big data, the private sector seeks to recruit flexible, agile, and digitally skilled individuals with
strong AI and ML backgrounds. They are needed to oversee capabilities (Sart and Yildiz, 2022).
90
Thus, investing in employees with experience in machine learning and artificial intelligence
enhances the development of emerging markets within the private sector (Strusani and
Houngbonon, 2019).
Lastly, the federal and private sectors expect ethical practices and a fundamental
understanding of data. However, ethical issues surrounding privacy, data validity, fairness, bias,
and transparency are not the only issue for the federal or private sector (Ammanath, 2022; Egger
et al., 2022). Both sectors share a responsibility to maintain ethical practices, if not to do the
right thing but also to save face from the public eye, given the media’s capabilities (Ouchchy et
al., 2020). Furthermore, new mining and data collection capabilities emphasize that applicants
understand they need to have experience handling pertinent data within the organization
ethically. Regarding this research, ethical competency refers to the sector’s ability to
Analyze Competencies
We analyze the content of job postings from both the United States federal government
and the private sector using our proposed research model. The first phase is the preparation phase
which encompasses data collection methodologies and data processing. The second phase of this
approach consists of organizing the data. Within this phase, we leverage the results from the
preparation phase to build term-similarity dictionaries to guide our analysis. The third phase is
the analysis phase which encompasses the conduction of a term-frequency inverse document
frequency (TF-IDF) analysis. Within this step, we analyzed the job postings' TF-IDF score and
91
Data Collection
Regarding data collection, we leveraged two data sources to acquire our job posting data
(Kaggle and USAJobs.gov). We collected two archival data sets to train our model and build our
dictionaries. Regarding our private sector training data, we leveraged a public and archival
dataset from Kaggle (Nasar et al., 2021) which yielded 3,567 unique data science job
descriptions from 2021 and 2022, upon cleaning the data set (removing duplicates and null data).
For our federal job postings, we acquired permission and tokens from USAJobs.gov to
pull data science job postings directly from the website. The privileges bestowed upon us
allowed us to pull archival job postings listed in 2021 and 2022. We omitted any positions
currently open to ensure both datasets were archival. With this query, we received 3,086 job
postings and descriptions directly related to data science. A total of 6,653 job descriptions were
used within this analysis of federal and private sector data science job postings. Table 5
Understanding the date quantity differences between the federal and private sector
datasets, we reviewed analyzed the uniqueness of tokens for each dataset and discuss the results
within the Cleaning and Tokenization section of this paper (see Table 7).
Descriptive Statistics
Before diving into the various phases of our methodology, we showcase our descriptive
statistics for the federal and private sector datasets. Within our descriptive statistics, we
showcase the number of job descriptions collected for each year summarized within Table 5.
Given the interests of this research, we specifically focused on the job description variable, as the
scraped from GlassDoor did not provide location or usable variables for further analysis.
92
Table 5. Token Comparison
2022
Once our data was acquired, we then preprocessed and tokenized the information in
preparation to train and test the natural language processing (NLP) neural network.
Upon tokenizing the job postings, we reviewed how similar the datasets were to gauge
relevance of comparison. Given the larger sample size, the private sector identified 5,842 tokens
with regard to the job descriptions. The federal sector yielded 1,916 total. Although the private
sector had more unique tokens, the federal sector shared 1,768 of the tokens. Thus, the federal
sector was approximately 93% similar to the private sector data. Table 6 summarizes the numeric
similarity and dissimilarity of each data set. Given the strong similarity score of the federal
Model Training
Upon preprocessing, tokenizing, and validating the similarity of the data, we trained a
Word2Vec model for the private and federal sectors to identify similar words in a text. This
unsupervised neural network is widely used in NLP for generating word embeddings (Htait et al.,
2018). Within these two models, the similarity is the distance between keywords, otherwise
93
known as the cosine similarity score. Following Valecha et al.’s (2021) misinformation analysis,
we utilized the skip-gram algorithm, which predictively learned the word embeddings; this is
numeric vector representation. We also trained the algorithm on a window of two, such that the
context spans over two words to the left and two to the right of the target word during training
(Valecha et al., 2021). To avoid overfitting our model, we used the default model setting for the
number of times a word should appear for it to be considered during the training process (five
times). Once our models are trained, the preparation phase of our analysis ends.
Method Analysis
We proceeded to organize our word embeddings to posture for a TF-IDF analysis. Within
the organization phase, one primary step is establishing our dictionaries. To accomplish this, we
continue to use the NLTK library to call upon the most_similar() function to extract similar
words related to our core competencies: ethics, artificial intelligence, machine learning, problem-
solving, and statistics (Poomagal et al., 2022). Our search queries needed to contain stemmed
and lemmatized words to accomplish this. Thus Table 7 summarizes our search queries within
Upon running the most_similar() function, we generated the top 10 most similar words
related to our established data science competencies (Martinez Soriano et al., 2019). Once we
acquired all similar words for each competency, we built two comprehensive dictionaries, one
94
for the federal sector and the other for the private sector. To provide further insight, Figures 4a-b
showcase separate graphs which summarize the most similar words for the designated data
science competency. Each grouping is unique and does not share a similar word. The left graph
showcases the top 10 words, most like “ethic,” for both the federal and private sectors.
Furthermore, the data points on each graph represent the cosine similarity score assigned
to each word from the Word2Vec model. However, the graph on the right showcases the top 10
words, most like “machine learning,” of which the private and federal sectors share “artificial”.
This is the only competency in which both datasets and models share similar words. To provide
insight into each dictionary, we have included figures within Appendix B and C which
summarize the comprehensive dictionary for the federal and the private sector. Duplicate words
95
Artificial Intelligence Dictionary Problem Solving Dictionary
Comparison Comparison
Cosine Similarity Cosine Similarity
Private Sector Cosine Similarity
Private Sector Cosine Similarity Score Score
Federal Sector Cosine
Federal Sector Cosine Similarity Score Similarity Score
cognit pragmat
nuclear1 augment valid1 solver
spatial ai overse break
vehicl 0.5 fusion concept 0.5 diagnos
creat chip qualit difficult
visual 0 multimod substant 0 conceptu
dollar exploit trial frame
access signal propos thinker
imageri tree experiment dig
compens audio diver intuit
threat dissemin
Statistics Dictionary
Comparison
Cosine Similarity
Private Sector Cosine
Similarity Score
econometr
content1 theori
semest inferenti
econom 0.5 causal
thesi exploratori
differenti 0 actuari
substanti mathemat
comput theoret
logic
probabl algebra
correl
content analysis. MAXQDA is a powerful software package for analyzing qualitative data and
conducting content analysis (Oliveira et al., 2015). Because this suite interfaces with Windows-
based software and programs, we were capable of uploading our .CSV files for each of our
datasets once we converted them to .XLSX (excel worksheet). Utilizing the dictionary function
96
within MAXQDA, we uploaded our two dictionaries into the software to run two separate
content analyses, one for our federal data and the other for our private sector data. The results
from the MAXQDA content analysis yielded an excel document with the word count of each job
description and the number of words that appeared within the job description relative to the
dictionary for each job posting. Upon receiving this excel document from MAXQDA, we added
the two columns from the master document which portray the location of the job posting as well
as the number of days the position was open for application. To minimize error, we organized the
job postings by year posted and assigned dummy variables to each job description. From here,
we numerically ordered each description for the MAXQDA excel document and the master
document to ensure the locations were with the correct job posting. We then conducted our TF-
IDF analysis. By examining the TF-IDF analysis, we can compare how the federal and private
sectors differ regarding leveraging the established data science competencies (Figure 3) within
We further coded the job postings to indicate the presence of each competency. For
instance, if the job posting contained terms related to the AI Dictionary that the Word2Vec
model produced, we coded it with “1”, otherwise it was coded “0”. The next section discusses
We based our machine coding validation on strategies and systematic methods used by
Valecha et al. (2021a), Larsen et al. (2020), Wang et al. (2018), and Gonzalez and Sol (2012).
the data science competencies within the job posting. This process was performed in two steps:
1) having two information technology (IT) experts manually code randomly selected job postings
97
for both the private and federal sector, and (2) correlate the human coding with the machine’s
We set up a preliminary briefing to manage expectations. The research team met with
both IT experts simultaneously to discuss the rules of engagement for coding these job postings.
It was communicated that the work was to be completed independently with no collaboration.
We then discussed definitions for each competency to confirm an understanding and allow the
expert to ask any questions. Once the meeting concluded, both experts were separated and
proceeded to code. Regarding the experts, they confirmed to have over 9 years’ experience
within the IT field, have had exposure to data science, and have both held managerial/hiring
positions.
Regarding step 1, we sought to have the IT experts code job postings in 3 batches. The
job postings within the first 3 batches were identical. Given that this is a systematic process, our
goal was to repeat the coding process until we achieved a total kappa value greater than the .7
threshold (Valecha et al., 2021a; McHugh, 2012; Landis and Koch, 1977). For each batch, we
provided 30 random and identical job postings to each expert. 15 of the thirty were private sector
job postings, while the remaining were federal. Each batch provided a new set of random job
postings which were shared to the coders. The experts reviewed the job postings and looked for
cues relating to ethics, artificial intelligence, machine learning, problem solving, and statistics.
The experts coded the following: “0” in the absence of information about that competency, and
‘1’ otherwise. We compared the coding mechanisms of the validators against one another and
their performance against the machine. The kappa results of step 1 are summarized in tables 8a-
b and 9a-b.
98
Table 8a. Results of Inter-Coder Reliability (Private Sector)
PRIVATE Batch 1 (N=15) Batch 2 Batch 3 Total
SECTOR (N=15) (N=15) (N=45)
99
Table 9b. Results of Preliminary Validation (Federal Sector) Continued
Total 58% 73% 81% 71%
Ethics 60% 73% 80% 71%
Machine Learning 67% 73% 73% 71%
Artificial Intelligence 53% 73% 87% 71%
Problem Solving 60% 67% 87% 71%
Statistics 50% 80% 80% 70%
fourth batch of job postings where were random and unique to the validator. This step is unique
from the others as we sought to compare the machine’s coding to that of the human experts. The
research team administered 30 random job postings to each expert (15 federal and 15 private
sector). This step did ensure that the job postings were not identical. This means each expert got
a different set of randomly selected job postings to review. Within Batch 4 we considered
percent agreement and Cohen’s Kappa. Higher kappa scores indicate an agreed understanding
between validators (Valecha et al, 2021a; McHugh, 2012; Landis and Koch, 1977). Table 10a-b
100
Table 10a. Results of Batch 4 (Validator 1)
Prob Solv 87% 70%
Stat 87% 66%
Following validation, we then conducted our TF-IDF analysis. By examining the TF-IDF
analysis, we can compare how the federal and private sectors differ regarding leveraging the
established data science competencies (Figure 3) within the job postings. To create a robust
analysis, we incorporated synonyms for each term the machine output (Bafna et al. 2016; Akay
et al., 2014). The dictionary of each competency and the associated synonyms can be located in
Appendix D and E.
Descriptive Statistics
Tables 11a-b demonstrates a Spearman correlation test which investigates the presence of
each competency for each sector. Each statistically significant correlation had a value of less
101
than .7, indicating no significant multi-collinearity problems exist between the presence of each
102
Table 11b. Spearman Correlation (Federal Sector)
Socialization
Within this section, we convey the results of our analysis and discuss the theoretical and
significant. In order to do so, we conducted a TF-IDF analysis for each job posting relative to
103
each data science competency dictionary for the federal and private sectors to measure
comparisons. In other words, we analyzed each job within the federal (n=649) and private
(n=1,492) sector and computed TF-IDF scores for the data science competencies identified in
Figure 3. By doing so, we could measure the relevance and significance of the specific dictionary
within each sector. Upon computing the TF-IDF score for each job posting, we calculated the
mean of the dictionary’s TF-IDF score for each job postings to run our comparison analysis
between the federal and private sectors. Furthermore, we conducted an ANOVA analysis to
investigate the difference in means between the federal and private sectors and their relative
significance.
The federal sector places a stronger emphasis on competencies like statistics and ethics
within job postings, whereas the private sector favors problem solving and ethics. Within federal
job postings, statistics ranked first place as the most prominent competency represented within
the job postings. The private sector ranks problem solving as its top competency. The ethics
component shares the second-most prominent position within the ranking. Table 12 summarizes
the rankings of TF-IDF scores for both the federal and private sector.
104
Comparative Mapping
We compared the TF-IDF scores for the private and federal sectors to further analyze our
results (Figure 5a-b). Despite the comparatively low numbers of TF-IDF scores within the
private sector, the results indicate a more substantial presence of the statistics competency within
the federal sector when compared to the private. Comparatively, the private sector may place a
stronger emphasis within the job description on communicating the desired ability problem solve
when compared to the federal sector. Lastly, our research identified that the private sector has
significantly lower TF-IDF scores than the federal sector. The keywords most like ethics,
problem-solving, artificial intelligence, machine learning, and statistics within the private sector
were not used frequently throughout job descriptions, despite having strong cosine similarity
scores. The lower score could indicate a lack of representation throughout the job postings
despite being identified as a word that is closely related to the identified competencies.
Statistics Statistics
0.0005 0.008
0.0004 0.006
0.0003
0.004
Artificial 0.0002 Artificial
Ethics Ethics
Intelligence 0.0001 Intelligence 0.002
0 0
105
Means Difference Analysis
To test the mean differences among the data science competencies of each group (federal
and private), we conducted an ANOVA analysis via IBM’s SPSS software (Kim et al., 2005).
We also conducted a post-hoc t-test analysis comparing the sectors. The significance level for the
According to the TF-IDF ANOVA results there is sufficient evidence that supports most
of the mean TF-IDF differs among the five competencies for federal and private job sectors at a
.05 level of significance. Regarding all competencies, the value is less than the established
significance level, the differences within the means are statistically significant. Therefore, the
ANOVA test results show that the TF-IDF mean values of the five competencies obtained from
106
the federal and private sectors are not identical, implying that the different competencies among
Ethics TFIDF
676.29 <.001 -34.05 2,139 <.001 <.001
Through our post-hoc TF-IDF t-test analysis, we identified and validated the rejection of
the null such that the variances of the majority of the competencies differ from one another. This
analysis validated all competencies have a difference in means which are statistically
significantly different from zero. The difference in means for the Statistics competency is not
statistically significantly different from zero, which implies the sectors use this dictionary
Understanding our primary dataset consisted of insights from 2022 job postings, we
sought to conduct a post-hoc analysis regarding job titles to identify any potential differences
between the groups of data leveraging the TF-IDF metrics. This section explores this post-hoc
analysis.
107
Post-Hoc Private and Federal Sector Job Title Comparison
Understanding job data science titles may differ for both the private and federal sector,
we leveraged Mbah et al. (2017) and Saltz and Grady’s (2017) strategy in categorizing each job
posting based on synonyms and words contained within each job title. Saltz and Grady (2017)
propose five data science job title categories which consist of: 1) Data Scientist, 2) Data Science
Regarding data scientist, this reflects jobs that seek out individuals who can find and interpret
rich and specialized data sources and manage large amounts of data. This practice also includes
being able to visualize the data to aid in understanding complex data. Data science researcher
refers to jobs that seek out researchers/academic interested in exploring actionable knowledge
related to specific scientific problems, business processes, or hidden relationships. With regards
to the data science architect, these jobs refer to hiring an individual capable of designing and
maintaining the architecture of data science applications and/or facilities. Tangential, data
science programmer refers to jobs which seek out individuals who can code and develop
software for analytic applications and support enterprise-level processes. Where an architect
designs and creates blueprints, the programmer acts as the construction unit responsible for
building the desired construct/blueprint. Lastly, data/business analyst refers to jobs which seek
out individuals who can analyze large varieties of data and extract information about systems and
management/leadership positions. Keywords within duty titles include: 1) director, head, lead,
108
Table 15. Private and Federal Job Titles
With regards to the job titles, the highest data science job title within the private sector is
reflected by data science architect. Contrarily, data/business analyst is reflected for the federal
sector. This implies the federal sector seems more business-related data science skills while the
private sector communicates needs for specialized data scientists. Furthermore, both the private
and federal sector share their lowest numbers being in data science programmers and data
science researchers. This indicates that both sectors have fewer demands for academic and
Once we coded the job titles and garnered an understanding of the demographics, we
conducted a comparative analysis of the TF-IDF scores related to each job title category. Within
this comparison, we the average TF-IDF scores calculated within our primary analysis and
ranked them to see how prominent each competency is related to each job category. Regarding
the private sector, problem solving reigned the most prominent within each duty title category.
Comparatively, the statistics competency was the most prominent throughout the federal sector.
When comparing both sectors, this analysis identified that although the statistics competency
was the most represented within the federal, it was the least represented within the private sector
109
job title categories. This could imply that a stronger emphasis is placed on the desire for
applicants to have stronger problem solving skills within the private sector, whereas the federal
sector favors more technical skills related to the statistics field. Figure 6a-b summarizes these
findings.
0.03
0.025
0.02
0.015
0.01
0.005
0
Data Science Data Science Data Science Data Scientist Data/Business Management
Architect Programmer Researcher
0.0005
0.0004
0.0003
0.0002
0.0001
0
Data Science Data Science Data Science Data Scientist Data/Business Management
Architect Programmer Researcher
By leveraging the competencies derived from IBM and Deloitte, we conducted a content
analysis on job postings from the United States federal government and private sector to gain
insight into the emphasis of each data science competency. Through this analysis, we identified
three themes for future research. The first theme we identified is that there is a lack of research
investigating the differences in job postings between the federal government and private sectors.
Future research endeavors could dive deeper into specific skill requirements (R-programming,
analytics, python) and how they differ from the federal and private sectors. Secondly, statistics
within the private sector represented the lowest competency in job postings. Our analysis
encourages future research endeavors to investigate the relationship between the statistical
expectations of employees and the semantics within job descriptions. Lastly, our research
identified that the private and federal sector’s TF-IDF scores fluctuate in presence which could
indicate shifts in demand for prospective employees. Given the global COVID pandemic, future
research could continue the temporal exploration of labor force demands within the data science
community.
the limitations of this research. Although systematic, the methodology for data collection limited
us in the variables we could explore. If we had access to temporal considerations (how long the
posting was open) or attractiveness of the posting (number of applications received), we could
have dived deeper into how the dictionaries could affect the recruitment efforts of the federal and
private sectors. Furthermore, the sample sizes of the job postings were limited, despite each
being unique and from an archival source. The research team eliminated any duplicate job
postings to reduce redundancy, which caused the federal job postings to drop drastically. This is
111
due to the federal sector recycling job descriptions. Other job posting portals could be leveraged
to expand this research endeavor; however, we chose to keep the sources the same for the sake of
symmetry and standardization. Other venues to acquire job posting data may produce a larger
sample size such as stronger web scraping programs and other coding mechanisms. Moreover,
future research could investigate how data science job descriptions and associated data
competency dictionaries have evolved for the federal and private sectors. Given the datasets
leveraged for this analysis, we did not have the data to conduct that in the depth of research.
Additionally, our data source reflects a snapshot in time from when the data was pulled.
Lastly, there are other pre-trained topic modeling models within Genism that could be
leveraged instead of Word2Vec. Namely, Doc2Vec, Latent Dirichlet Allocation, Author Topic
Modeling, and Hierarchical Dirichlet Processing are a few modeling techniques which fall under
the Genism library. Moreover, the post-hoc statistical models can also be enhanced or exchanged
for more effective techniques as newer developments occur. Through these limitations, we seek
112
CHAPTER 9: CONCLUSION
Understanding big data, AI, and ML have accelerated business intelligence capabilities
for the United States federal government and private sector, research endeavors have also grown
in this interest. However, there is a lack of IS research frameworks within the HRM domain
which leverage NLP tools. This paper seeks to address the IS research gap by proposing the JC-
research. Figure 7 summarizes our proposed method to guide future research competency-based
research endeavors.
We acknowledge there are limitations to this research and framework. For instance, there
are other models which are optimized for context analysis such as the models addressed in Table
4. This could provide new considerations for contextually understanding competencies within
job postings rather than relying on word similarities. This framework serves as a foundation, in
hopes that new frameworks come to light and enrich the IS and HRM research domain.
Moreover, we intend for this framework to be a starting point for scholars from academic and
industry backgrounds alike. As NLP capabilities become more sophisticated, the need for tools
and models to guide researchers in using these capabilities becomes imperative to exploring this
research domain.
113
Where previous literature has homed explicitly in on skill requirements, we wanted to
divert from this trend to focus on a competency approach to analyzing job descriptions. Through
this framework proposal, empirical example, we demonstrated how this framework can benefit
business experts ranging from a variety of different career fields given its systematic and
transparent foundation. We hope the proposed competency model helps HR and IS professionals
114
APPENDICES
terms which are obtained from a neural-network model (Bengio et al., 2000; Fernández-Reyes
and Shinde, 2019). There are two main categories of embedding text: 1) word embedding and 2)
sentence embedding. Word embedding is the practice of encoding of each word present in the
text into a numerical vector, while sentence embedding is the encoding of each sentence present
in the text. Given the interest of exploring specified competencies within job descriptions, word
embeddings empower researchers with the ability to look at specific words and identify
Thus, this research focuses its attention on word embedding capabilities. Given the recency of
NLP expansion and development, new word embedding models have come into existence. For
instance, Word2Vec (Mikolov et al., 2013; Fernández-Reyes and Shinde, 2019), FastText
(Bojanowski et al., 2017), GloVe (Pennington et al., 2014), ELMo (Peters et al., 2018) and
BERT (Delvin et al., 2018). Table 4 summarizes each word embedding model, its general
115
Table 1. Word Embedding Model Summaries Continued
ELMo Word Context Tasks Not designed for transfer Peters et al.
learning, needs to be trained (2018)
for specific tasks
BERT Word Context Tasks High memory cost; slow to Delvin et al.
train; not stand alone system (2018)
(designed to be input into
other systems)
116
Appendix B: Private Sector Dictionary (Word2Vec Output)
117
Appendix C: Federal Sector Dictionary (Word2Vec Output)
118
Appendix D: FULL Private Sector Dictionary (TF-IDF Analysis)
119
Appendix D: FULL Private Sector Dictionary (TF-IDF Analysis) Continued
Subject Neural Visionary Noticeable Numerical
Fair Infer Frame Salient Seminal
Model Derive Body Tree Unconventional
Performance Decide Cage Sapling Unusual
Show Understand Philosopher Forest Exploratory
Advocate Theoretic/Theory Fabric Seedling Preparatory
Confirm Thesis Framework Shrub Prior
Encourage Vector Physique Timber Actuary
Endorse Machine Learn Scaffolding Wood Analyst
Hold To Robotics Structure Auditor
Quantitative Thinker Bookkeeper
Promote Dig Clerk
Side With Gibe Mathematic
Stand By Intuit Theoretic
Vindicate Hunch Scientific
Ethic Instinct Abstract
Integrity Problem Academic
Solv/E/Ing
Moral Calcul/us/ate/ation
Principle
Virtue
120
Appendix E: FULL Federal Sector Dictionary (TF-IDF Analysis)
121
Appendix E: FULL Federal Sector Dictionary (TF-IDF Analysis) Continued
Mater Clear Urge Visual Contrasting
Actual Elegant Preliminary Greenback Distant
Appreciable Fresh Substant Create Distinct
Earthly Graceful Big Invent Divergent
Perceptible Hygienic Consequential Build Diverse
Physical Immaculate Considerable Conceive Offbeat
Substantial Neat Extraordinary Constitute Other
Screen Orderly Hefty Construct Economy
Cover Pure Generous Make Particular
Curtain Squeaky Clean Large Design Peculiar
Net Spotless Massive Devise Various
Stringent Simple Meaningful Discover Thesis
Binding Tidy Serious Establish Contention
Demanding Unblemished Significant Forge Hypothesis
Draconian Washed Solid Organize Budgetary
Exacting Farm Sizable Form Opinion
Forceful Acreage Steady Found Premise
Harsh Estate Strong Generate Proposition
Inflexible Field Valuable Initiate Supposition
Ironclad Garden Vast Plan Theory
Rigorous Grassland Quality Produce Commercial
Severe Homestead Concept Set Up Fiscal
Stiff Lawn Approach Shape Industrial
Strict Meadow Conception Spawn Monetary
Tough Nursery Image Start Semester
Cognit Orchard Notion Vehicle Content
Pasture Perception Automobile Contented
Ethic Plantation Theory Bicycle Comfortable
Administrate Ranch Thought Bus Fulfilled
Carry Out Ai View Boat Gratified
Conduct Opportunity Oversee Cab Pleased
Direct Load Manage Car Happy
Execute Anomaly Command Jeep Satisfied
Govern Detect Supervise Taxi Willing
Oversee Multidimension Survey Truck Mathematic/s/al
Supervise Query Watch Wagon Algebra
Valid Van Calculation
Accurate Spatial Calculus
Authentic Nuclear Geometry
Binding Basic Math
Compelling Essential Statistic
Conclusive Fundamental
Credible Important
Good Key
Lawful Paramount
Legal Pivotal
122
Appendix E: FULL Federal Sector Dictionary (TF-IDF Analysis) Continued
Legitimate Significant
Logical
Original
Persuasive
True
Well-Founded
Problem
Solv/e/ing
123
REFERENCES
Abbasi, A., & Chen, H. (2008). CyberGate: A Design Framework and System for Text Analysis of
Computer-Mediated Communication. MIS Quarterly, 32(4), 811–837. https://doi.org/10.2307/2514
8873
Abbasi, Albrecht, Vance, & Hansen. (2012). MetaFraud: A Meta-Learning Framework for Detecting
Financial Fraud. MIS Quarterly, 36(4), 1293. https://doi.org/10.2307/41703508
Adewumi, T. P., Liwicki, F., & Liwicki, M. (2021). Word2Vec: Optimal Hyper-Parameters and Their
Impact on NLP Downstream Tasks (arXiv:2003.11645). arXiv. http://arxiv.org/abs/2003.11645
Akay, A., Dragomir, A., & Erlandsson, B.-E. (2015). Network-Based Modeling and Intelligent Data
Mining of Social Media for Improving Care. IEEE Journal of Biomedical and Health Informatics,
19(1), 210–218. https://doi.org/10.1109/JBHI.2014.2336251
Alsudais, A., Aldumaykhi, A., & Otai, S. (2022). Comparison of Job Titles for Specific Terms:
Investigating “Data Science.” In E. Pardede, P. Delir Haghighi, I. Khalil, & G. Kotsis (Eds.),
Information Integration and Web Intelligence (pp. 98–103). Springer Nature Switzerland. https:
//doi.org/10.1007/978-3-031-21047-1_
Amaar, A., Aljedaani, W., Rustam, F., Ullah, S., Rupapara, V., & Ludi, S. (2022). Detection of Fake
Job Postings by Utilizing Machine Learning and Natural Language Processing Approaches. Neural
Processing Letters, 54(3), 2219–2247. https://doi.org/10.1007/s11063-021-10727-z
Ammanath, B. (2022). Trustworthy AI: A Business Guide for Navigating Trust and Ethics in AI.
John Wiley & Sons.
Anandarajan, M., Hill, C., & Nolan, T. (2019). Text Preprocessing. In M. Anandarajan, C. Hill, & T.
Nolan (Eds.), Practical Text Analytics: Maximizing the Value of Text Data (pp. 45–59). Springer
International Publishing. https://doi.org/10.1007/978-3-319-95663-3_4
Anton, E., Behne, A., & Teuteberg, F. (2020). The Humans Behind Artificial Intelligence–An
Operationalisation of AI Competencies.
Bafna, P., Pramod, D., & Vaidya, A. (2016). Document clustering: TF-IDF approach. 2016
International Conference on Electrical, Electronics, and Optimization Techniques (ICEEOT), 61–66.
https://doi.org/10.1109/ICEEOT.2016.7754750
Barducci, A., Iannaccone, S., La Gatta, V., Moscato, V., Sperlì, G., & Zavota, S. (2022). An end-to-
end framework for information extraction from Italian resumes. Expert Systems with Applications,
210, 118487. https://doi.org/10.1016/j.eswa.2022.118487
Bengio, Y., Ducharme, R., & Vincent, P. (2000). A Neural Probabilistic Language Model. Advances
in Neural Information Processing Systems, 13. https://proceedings.neurips.cc/paper/2000/hash/728f2
06c2a01bf572b5940d7d9a8fa4c-Abstract.html
124
Benjamin, V., Valacich, J. S., & Chen, H. (2019). DICE-E: A Framework for Conducting Darknet
Identification, Collection, Evaluation with Ethics. MIS Quarterly, 43(1), 1–22. https://doi.org/10.
25300/MISQ/2019/13808
Bojanowski, P., Grave, E., Joulin, A., & Mikolov, T. (2017). Enriching Word Vectors with Subword
Information. Transactions of the Association for Computational Linguistics, 5, 135–146. https://doi
.org/10.1162/tacl_a_00051
Boon, C., Eckardt, R., Lepak, D. P., & Boselie, P. (2018). Integrating strategic human capital and
strategic human resource management. The International Journal of Human Resource Management,
29(1), 34–67. https://doi.org/10.1080/09585192.2017.1380063
Brumer, Y., Shapira, B., Rokach, L., & Barkan, O. (2017). Predicting Relevance Scores for Triples
from Type-Like Relations using Neural Embedding.
Carretta, A. (1992). Career and succession planning. In Competency Based Human Resource
Management. HayGroup.
Chen, J., Chen, C., & Liang, Y. (2016). Optimized TF-IDF Algorithm with the Adaptive Weight of
Position of Word. 114–117. https://doi.org/10.2991/aiie-16.2016.28
Chouhan, V. S., & Srivastava, S. (2014). Understanding Competencies and Competency Modeling ―
A Literature Survey. IOSR Journal of Business and Management, 16(1), 14–22. https://doi.org/10.
9790/487X-16111422
Christensen, B. J., Lentz, R., Mortensen, D. T., Neumann, G. R., & Werwatz, A. (2005). On‐the‐Job
Search and the Wage Distribution. Journal of Labor Economics, 23(1), 31–58. https://doi.org/10.1
086/425432
Chuang, Z., Ming, W., Guang, L. C., Bo, X., & Zhi-qing, L. (2009). Resume Parser: Semi-structured
Chinese Document Analysis. 2009 WRI World Congress on Computer Science and Information
Engineering, 5, 12–16. https://doi.org/10.1109/CSIE.2009.562
Cole, M. S., Feild, H. S., & Giles, W. F. (2003). Using Recruiter Assessments of Applicants’ Resume
Content to Predict Applicant Mental Ability and Big Five Personality Dimensions. International
Journal of Selection and Assessment, 11(1), 78–88. https://doi.org/10.1111/1468-2389.00228
Deguchi, T., Seo, S., & Ishii, N. (2022). Meaning of the Clusters on Dimensionality Reduction by
Word Clustering. 2022 12th International Congress on Advanced Applied Informatics (IIAI-AAI),
325–330. https://doi.org/10.1109/IIAIAAI55812.2022.00072
125
Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2018). BERT: Pre-training of Deep
Bidirectional Transformers for Language Understanding. https://doi.org/10.48550/arXiv.1810.04805
Embarak, Dr. O. (2018). Data Analysis and Visualization Using Python: Analyze Data to Create
Visualizations for BI Systems. Apress. https://doi.org/10.1007/978-1-4842-4109-7
Enholm, I. M., Papagiannidis, E., Mikalef, P., & Krogstie, J. (2022). Artificial Intelligence and
Business Value: A Literature Review. Information Systems Frontiers, 24(5), 1709–1734.
https://doi.org/10.1007/s10796-021-10186-w
Falatah, R. (2021). The Impact of the Coronavirus Disease (COVID-19) Pandemic on Nurses’
Turnover Intention: An Integrative Review. Nursing Reports, 11(4), Article 4. https://doi.org/1
0.3390/nursrep11040075
Fernández-Reyes, F. C., & Shinde, S. (2019). CV Retrieval System based on job description
matching using hybrid word embeddings. Computer Speech & Language, 56, 73–79. https://doi.or
g/10.1016/j.csl.2019.01.003
Gao, Q., Huang, X., Dong, K., Liang, Z., & Wu, J. (2022). Semantic-enhanced topic evolution
analysis: A combination of the dynamic topic model and word2vec. Scientometrics, 127(3), 1543–
1563. https://doi.org/10.1007/s11192-022-04275-z
Gardiner, A., Aasheim, C., Rutner, P., & Williams, S. (2018). Skill Requirements in Big Data: A
Content Analysis of Job Advertisements. Journal of Computer Information Systems, 58(4), 374–384.
https://doi.org/10.1080/08874417.2017.1289354
GeeksforGeeks. (2018, February 2). Part of Speech Tagging with Stop words using NLTK in python.
GeeksforGeeks. https://www.geeksforgeeks.org/part-speech-tagging-stop-words-using-nltk-python/
Goldberg, Y., & Levy, O. (2014). word2vec Explained: Deriving Mikolov et al.’s negative-sampling
word-embedding method (arXiv:1402.3722). arXiv. http://arxiv.org/abs/1402.3722
Goldfarb, A., Taska, B., & Teodoridis, F. (2023). Could machine learning be a general purpose
technology? A comparison of emerging technologies using data from online job postings. Research
Policy, 52(1), 104653. https://doi.org/10.1016/j.respol.2022.104653
Gonzalez, R., & Sol, H. (2012). Validation and Design Science Research in Information Systems. In
Research Methodologies, Innovations and Philosophies in Software Systems Engineering and
Information Systems (pp. 403–426). https://doi.org/10.4018/978-1-4666-0179-6.ch021
Gottipati, S., Shim, K. J., & Sahoo, S. (2021). Glassdoor Job Description Analytics – Analyzing Data
Science Professional Roles and Skills. 2021 IEEE Global Engineering Education Conference
(EDUCON), 1329–1336. https://doi.org/10.1109/EDUCON46332.2021.9453931
126
Gregor, S., & Hevner, A. R. (2013). Positioning and Presenting Design Science Research for
Maximum Impact. MIS Quarterly, 37(2), 337–355. https://doi.org/10.25300/MISQ/2013/37.2.01
Gupta, A. (2022). Company, Job, and Perceived Advantage Influence on Job Hunting: Evaluating
Glassdoor Job Hunting platform. Journal of Digitovation and Information System, 2(1), Article 1.
https://doi.org/10.54433/JDIIS.2022100011
Ha, T., Lee, M., Yun, B., & Coh, B.-Y. (2022). Job Forecasting Based on the Patent Information: A
Word Embedding-Based Approach. IEEE Access, 10, 7223–7233.
https://doi.org/10.1109/ACCESS.2022.3141910
Hardeniya, N., Perkins, J., Chopra, D., Joshi, N., & Mathur, I. (2016). Natural Language Processing:
Python and NLTK. Packt Publishing Ltd.
Heckman, J. (2021, May 14). First cohort of federal employees graduate from OMB’s data science
upskilling program. Federal News Network. https://federalnewsnetwork.com/workforce/2021/05
/first-cohort-of-federal-employees-graduate-from-ombs-data-science-upskilling-program/
Heidarysafa, M., Kowsari, K., Bashiri, M., & Brown, D. E. (2021). Toward a Knowledge Discovery
Framework for Data Science Job Market in the United States (arXiv:2106.11077). arXiv. http://ar
xiv.org/abs/2106.11077
Hevner, A. R., March, S. T., Park, J., & Ram, S. (2004). Design Science in Information Systems
Research. MIS Quarterly, 28(1), 75–105. https://doi.org/10.2307/25148625
Hosain, S., & Liu, P. (2020). LinkedIn for Searching Better Job Opportunity: Passive Jobseekers’
Perceived Experience. The Qualitative Report, 25(10), 3719–3732.
Htait, A., Fournier, S., Bellot, P., Htait, A., Fournier, S., & Bellot, P. (2018). Unsupervised Creation
of Normalization Dictionaries for Micro-Blogs in Arabic, French and English. Computación y
Sistemas, 22(3), 729–737. https://doi.org/10.13053/cys-22-3-3034
Hu, K., Luo, Q., Qi, K., Yang, S., Mao, J., Fu, X., Zheng, J., Wu, H., Guo, Y., & Zhu, Q. (2019).
Understanding the topic evolution of scientific literatures like an evolving city: Using Google
Word2Vec model and spatial autocorrelation analysis. Information Processing & Management, 56(4),
1185–1203. https://doi.org/10.1016/j.ipm.2019.02.014
IBM. (2020). The Data Science Skills Competency Model. IBM Corporation, 12.
Jalilifard, A., Caridá, V. F., Mansano, A. F., Cristo, R. S., & da Fonseca, F. P. C. (2021). Semantic
Sensitive TF-IDF to Determine Word Relevance in Documents. In S. M. Thampi, E. Gelenbe, M.
Atiquzzaman, V. Chaudhary, & K.-C. Li (Eds.), Advances in Computing and Network
Communications (pp. 327–337). Springer. https://doi.org/10.1007/978-981-33-6987-0_27
127
Janssen, M., & Estevez, E. (2013). Lean government and platform-based governance—Doing more
with less. Government Information Quarterly, 30, S1–S8. https://doi.org/10.1016/j.giq.2012.11.003
Khallaf, A., & Majdalawieh, M. (2012). Investigating the Impact of CIO Competencies on IT
Security Performance of the U.S. Federal Government Agencies. Information Systems Management,
29(1), 55–78. https://doi.org/10.1080/10580530.2012.634298
Kim, D. J., Song, Y. I., Braynov, S. B., & Rao, H. R. (2005). A multidimensional trust formation
model in B-to-C e-commerce: A conceptual framework and content analyses of academia/practitioner
perspectives. Decision Support Systems, 40(2), 143–165. https://doi.org/10.1016/j.dss.2004.01.006
Kortum, H., Rebstadt, J., & Thomas, O. (2022). Dissection of AI Job Advertisements: A Text
Mining-based Analysis of Employee Skills in the Disciplines Computer Vision and Natural Language
Processing. Hawaii International Conference on System Sciences. https://doi.org/10.24251/HICSS.
2022.635
Kreuter, F., Ghani, R., & Lane, J. (2019). Change Through Data: A Data Analytics Training Program
for Government Employees. Harvard Data Science Review. https://doi.org/10.1162/99608f92ed35
3ae3
Kurnia, R. I. (2020). Classification of User Comment Using Word2vec and SVM Classifier.
International Journal of Advanced Trends in Computer Science and Engineering, 9(1), 643–648.
https://doi.org/10.30534/ijatcse/2020/90912020
Laber, M. E., & O’Connor, J. K. (2000). Competency modeling: Ready, set, research. The Industrial-
Organizational Psychologist, 37(4), 91–96.
Landers, R., Brusso, R., & Auer, E. (2019). Crowdsourcing Job Satisfaction Data: Examining the
Construct Validity of Glassdoor.com Ratings. Personnel Assessment and Decisions, 5(3). https://doi.
org/10.25035/pad.2019.03.006
Landis, J. R., & Koch, G. G. (1977). The Measurement of Observer Agreement for Categorical Data.
Biometrics, 33(1), 159–174. https://doi.org/10.2307/2529310
Larsen, K., Lukyanenko, R., Mueller, R., Storey, V., Vander Meer, D., Parsons, J., & Hovorka, D.
(2020, June 1). Validity in Design Science Research.
Layug, C. (2018). Extracting Major Topics From Survey Text Responses Using Natural Language
Processing. Naval Post Graduate School.
Le, Q., & Mikolov, T. (2014). Distributed Representations of Sentences and Documents. International
Conference on Machine Learning.
128
López, F., & Romero, V. (2014). Mastering Python Regular Expressions. Packt Publishing Ltd.
Lunn, S., Zhu, J., & Ross, M. (2020). Utilizing Web Scraping and Natural Language Processing to
Better Inform Pedagogical Practice. 2020 IEEE Frontiers in Education Conference (FIE), 1–9.
https://doi.org/10.1109/FIE44824.2020.9274270
Marinescu, I. E., Skandalis, D., & Zhao, D. (2020). Job Search, Job Posting and Unemployment
Insurance During the COVID-19 Crisis (SSRN Scholarly Paper No. 3664265). https://doi.org/
10.2139/ssrn.3664265
Markus, M. L., Majchrzak, A., & Gasser, L. (2002). A Design Theory for Systems That Support
Emergent Knowledge Processes. MIS Quarterly, 26(3), 179–212.
Martinez Soriano, I., Castro Peña, J. L., Fernandez Breis, J. T., San Román, I., Alonso Barriuso, A.,
& Guevara Baraza, D. (2019). Snomed2Vec: Representation of SNOMED CT Terms with
Word2Vec. 2019 IEEE 32nd International Symposium on Computer-Based Medical Systems
(CBMS), 678–683. https://doi.org/10.1109/CBMS.2019.00138
Mbah, R. B., Rege, M., & Misra, B. (2017). Discovering Job Market Trends with Text Analytics.
2017 International Conference on Information Technology (ICIT), 137–142. https://doi.org/10.11
09/ICIT.2017.29
McHugh, M. L. (2012). Interrater reliability: The kappa statistic. Biochemia Medica, 22(3), 276–282.
Meyer, M. A. (2019). Healthcare data scientist qualifications, skills, and job focus: A content analysis
of job postings. Journal of the American Medical Informatics Association, 26(5), 383–391. https://do
i.org/10.1093/jamia/ocy181
Mikolov, T., Chen, K., Corrado, G., & Dean, J. (2013). Efficient Estimation of Word Representations
in Vector Space (arXiv:1301.3781). arXiv. http://arxiv.org/abs/1301.3781
Mökander, J., & Floridi, L. (2021). Ethics-Based Auditing to Develop Trustworthy AI. Minds and
Machines, 31(2), 323–327. https://doi.org/10.1007/s11023-021-09557-8
Muthyala, R., Wood, S., Jin, Y., Qin, Y., Gao, H., & Rai, A. (2017). Data-Driven Job Search Engine
Using Skills and Company Attribute Filters. 2017 IEEE International Conference on Data Mining
Workshops (ICDMW), 199–206. https://doi.org/10.1109/ICDMW.2017.33
129
Nash, M. (2018). ‘Let’s work on your weaknesses’: Australian CrossFit coaching, masculinity and
neoliberal framings of ‘health’ and ‘fitness.’ Sport in Society, 21(9), 1432–1453. https://doi.org/
10.1080/17430437.2017.1390565
Nutt, P. C. (2006). Comparing Public and Private Sector Decision-Making Practices. Journal of
Public Administration Research and Theory, 16(2), 289–318. https://doi.org/10.1093/jopart/mui041
Oliveira, M., Bitencourt, C. C., Santos, A. C. M. Z. dos, & Teixeira, E. K. (2015). Thematic Content
Analysis: Is There a Difference Between the Support Provided by the MAXQDA® and NVivo®
Software Packages? Revista de Administração Da UFSM, 9(1), 72–82. https://doi.org/10.5902
/1983465911213
Ouchchy, L., Coin, A., & Dubljević, V. (2020). AI in the headlines: The portrayal of the ethical
issues of artificial intelligence in the media. AI & SOCIETY, 35(4), 927–936. https://doi.org/10
.1007/s00146-020-00965-5
Ozkur, M., Benlier, N., Takan, I., Vasileiou, C., Georgakilas, A. G., Pavlopoulou, A., Cetin, Z., &
Saygili, E. I. (2022). Ginger for Healthy Ageing: A Systematic Review on Current Evidence of Its
Antioxidant, Anti-Inflammatory, and Anticancer Properties. Oxidative Medicine and Cellular
Longevity, 2022, e4748447. https://doi.org/10.1155/2022/4748447
Park, H. M. (2009). Comparing Group Means: T-tests and One-way ANOVA Using Stata, SAS, R,
and SPSS. https://scholarworks.iu.edu/dspace/handle/2022/19735
Pennington, J., Socher, R., & Manning, C. (2014). Glove: Global Vectors for Word Representation.
Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing
(EMNLP), 1532–1543. https://doi.org/10.3115/v1/D14-1162
Perlman, B. J. (2016). Human Resource Management at the Local Level: Strategic Thinking and
Tactical Action. State and Local Government Review, 48(2), 114–120. https://doi.org/10.1177/016
0323X16659115
Peters, M. E., Neumann, M., Iyyer, M., Gardner, M., Clark, C., Lee, K., & Zettlemoyer, L. (2018).
Deep contextualized word representations (arXiv:1802.05365). arXiv. http://arxiv.org/abs/1802.0
5365
Poomagal, S., Malar, B., Ranganayaki, E. M., Deepika, K., & Dheepak, G. (2022). Sentiment
Thesaurus, Synset and Word2Vec Based Improvement in Bigram Model for Classifying Product
Reviews. Sn Computer Science, 3(6), 422. https://doi.org/10.1007/s42979-022-01305-8
130
Mechanical Engineers, Part B: Journal of Engineering Manufacture, 217(9), 1183–1195. https://doi.o
rg/10.1243/095440503322420115
Qaiser, S., & Ali, R. (2018). Text Mining: Use of TF-IDF to Examine the Relevance of Words to
Documents. International Journal of Computer Applications, 181. https://doi.org/10.5120/ijca20
18917395
Quan, P., Shi, Y., Niu, L., Liu, Y., & Zhang, T. (2018). Automatic Chinese Multiple-Choice Question
Generation for Human Resource Performance Appraisal. Procedia Computer Science, 139, 165–172.
https://doi.org/10.1016/j.procs.2018.10.235
Rajagopalan, G. (2021). Regular Expressions and Math with Python. In G. Rajagopalan (Ed.), A
Python Data Analyst’s Toolkit: Learn Python and Python-based Libraries with Applications in Data
Analysis and Statistics (pp. 77–99). Apress. https://doi.org/10.1007/978-1-4842-6399-0_3
Rodriguez, D., Patel, R., Bright, A., Gregory, D., & Gowing, M. (2002). Developing competency
models to promote integrated human resource practices. Human Resource Management: Published in
Cooperation with the School of Business Administration, The University of Michigan and in Alliance
with the Society of Human Resources Management, 41(3), 309–324.
Saltz, J. S., & Grady, N. W. (2017). The ambiguity of data science team roles and the need for a data
science workforce framework. 2017 IEEE International Conference on Big Data (Big Data), 2355–
2361. https://doi.org/10.1109/BigData.2017.8258190
Sart, G., & Yildiz, O. (2022). Digitalism and Jobs of the Future [Chapter]. Digital Transformation
and Internationalization Strategies in Organizations; IGI Global. https://doi.org/10.4018/978-1-7998-
8169-8.ch001
Shaikh, R., Phulkar, N., Bhute, H., Shaikh, S. K., & Bhapkar, P. (2021). An Intelligent framework for
E-Recruitment System Based on Text Categorization and Semantic Analysis. 2021 Third
International Conference on Inventive Research in Computing Applications (ICIRCA), 1076–1080.
https://doi.org/10.1109/ICIRCA51532.2021.9544102
Shen, Y., & Liu, J. (2021). Comparison of Text Sentiment Analysis based on Bert and Word2vec.
2021 IEEE 3rd International Conference on Frontiers Technology of Information and Computer
(ICFTIC), 144–147. https://doi.org/10.1109/ICFTIC54370.2021.9647258
Sibarani, E. M., Scerri, S., Morales, C., Auer, S., & Collarana, D. (2017). Ontology-guided Job
Market Demand Analysis: A Cross-Sectional Study for the Data Science field. Proceedings of the
13th International Conference on Semantic Systems, 25–32.
https://doi.org/10.1145/3132218.3132228
131
Strusani, D., & Houngbonon, G. V. (2019). The Role of Artificial Intelligence in Supporting
Development in Emerging Markets [Brief]. World Bank. https://doi.org/10.1596/32365
Tang, Y., Bai, S., & Cui, L. (2022). An Empirical Study on the Deficiencies and Optimization of the
Management System of Tourist Attractions Based on Human Resource Management. Advances in
Multimedia, 2022, e2133830. https://doi.org/10.1155/2022/2133830
Tarraf, D., Shelton, W., Parker, E., Alkire, B., Carew, D., Grana, J., Levedahl, A., Leveille, J.,
Mondschein, J., Ryseff, J., Wyne, A., Elinoff, D., Geist, E., Harris, B., Hui, E., Kenney, C.,
Newberry, S., Sachs, C., Schirmer, P., Warren, K. (2019). The Department of Defense Posture for
Artificial Intelligence: Assessment and Recommendations. RAND Corporation. https://doi.org/
10.7249/RR4229
UCLA: Statistical Consulting Group. (n.d.). FAQ: What are the differences between one-tailed and
two-tailed tests? Retrieved January 28, 2023, from https://stats.oarc.ucla.edu/other/mult-
pkg/faq/general/faq-what-are-the-differences-between-one-tailed-and-two-tailed-tests/
Valecha, R., Sharman, R., Rao, H. R., & Upadhyaya, S. (2013). A Dispatch-Mediated
Communication Model for Emergency Response Systems. ACM Transactions on Management
Information Systems, 4(1), 2:1-2:25. https://doi.org/10.1145/2445560.2445562
Valecha, R., Srinivasan, S. K., Volety, T., Kwon, K. H., Agrawal, M., & Rao, H. R. (2021). Fake
News Sharing: An Investigation of Threat and Coping Cues in the Context of the Zika Virus. Digital
Threats: Research and Practice, 2(2), 1–16. https://doi.org/10.1145/3410025
Valecha, R., Volety, T., Rao, H. R., & Kwon, K. H. (2021). Misinformation Sharing on Twitter
During Zika: An Investigation of the Effect of Threat and Distance. IEEE Internet Computing, 25(1),
31–39. https://doi.org/10.1109/MIC.2020.3044543
Varelas, G., Lagios, D., Ntouroukis, S., Zervas, P., Parsons, K., & Tzimas, G. (2022). Employing
Natural Language Processing Techniques for Online Job Vacancies Classification. In Artificial
Intelligence Applications and Innovations. AIAI 2022 IFIP WG 12.5 International Workshops, 333–
344. Springer International Publishing. https://doi.org/10.1007/978-3-031-08341-9_27
Vathanophas, V. (2007). Competency Requirements for Effective Job Performance in Thai Public
Sector. Contemporary Management Research, 3(1), Article 1. https://doi.org/10.7903/cmr.49
Vijayarani, S., & Janani, R. (2016). Text Mining: Open-Source Tokenization Tools: An Analysis. 3,
37–47.
132
Vo, N. N. Y., Vu, Q. T., Vu, N. H., Vu, T. A., Mach, B. D., & Xu, G. (2022). Domain-specific NLP
system to support learning path and curriculum design at tech universities. Computers and Education:
Artificial Intelligence, 3, 100042. https://doi.org/10.1016/j.caeai.2021.100042
Votto, A. M., Valecha, R., Najafirad, P., & Rao, H. R. (2021). Artificial Intelligence in Tactical
Human Resource Management: A Systematic Literature Review. International Journal of Information
Management Data Insights, 1(2), 100047. https://doi.org/10.1016/j.jjimei.2021.100047
Wagemann Jr, J. (2020). The United States Air Force and Artificial Intelligence: Moving Forward by
Learning from Past Technology Implementation. Air Force Fellows Program Maxwell AFB.
Wang, X., Parameswaran, S., Bagul, D. M., & Kishore, R. (2018). Can online social support be
detrimental in stigmatized chronic diseases? A quadratic model of the effects of informational and
emotional support on self-care behavior of HIV patients. Journal of the American Medical
Informatics Association, 25(8), 931–944. https://doi.org/10.1093/jamia/ocy012
Yu, R., Das, S., Gurajada, S., Varshney, K., Raghavan, H., & Lastra-Anadon, C. (2021). A Research
Framework for Understanding Education-Occupation Alignment with NLP Techniques. Proceedings
of the 1st Workshop on NLP for Positive Impact, 100–106. https://doi.org/10.18653/v1/2021.nlp4p
osimpact-1.11
Zhu, J., Zerbe, E., Ross, M. S., & Berdanier, C. G. P. (n.d.). The Stated and Hidden Expectations:
Applying Natural Language Processing Techniques to Understand Postdoctoral Job Postings.
133
ESSAY III. VETERAN TALENT WITHIN DATA SCIENCE: AN EXPLORATORY
OF ACTIVE-DUTY VETERANS
134
CHAPTER 1: ABSTRACT
United States military veterans face employment challenges as they transition from
military service to the civilian sector. For one, communicating one’s skills and worth has
challenged both applicants and recruiters in understanding prior military experience, especially
for those who served within a combat capacity. Coincidentally, Artificial Intelligence (AI) and
Machine Learning (ML) capabilities have grown across the human resource career field to assist
in talent management endeavors. This has enabled Human Resource (HR) professionals and
presented within resumes. Accordingly, newer approaches to reviewing veteran resumes have
come to fruition. Understanding these capabilities are relatively new, this paper seeks to add to
the growing literature interest in leveraging hybrid approaches to analyze military experience
within veteran resumes claiming to have skills within the data science career field. Specifically,
this paper leverages LIWC-22, human-coding, and Word2Vec analyses to investigate veteran
135
CHAPTER 2: INTRODUCTION
Decades of war and conflict within Iraq, Afghanistan, and other forward-deployable
countries has brought renewed attention to employment challenges of United States veterans.
Historical troop-level drawdowns yielding the lowest number of service members in Afghanistan
have vectored many veterans from a variety of career fields homeward toward the United States,
some pursuing endeavors outside of the military (Garamone, 2021). In the wake of a of troop-
level drawdown in Afghanistan and Iraq, it is likely the number of veterans leaving the service
unemployment for veterans, Keeling et al. (2019) observe that these statistics may be misleading.
Statistics reported by the Bureau of Labor Statistics have indicated a record 2.5% low-rate of
veteran unemployment (January 2023) and marginal difference in unemployment rates between
veterans and nonveterans. The survey methodology investigates 60,000 households per month
and provides a cross-sectional comparison which could represent a small portion of veterans
(Loughran, 2014; Keeling et al., 2019). Furthermore, when observing the different populations of
veterans, the percentage of unemployment varies from generation to generation. For instance, by
averaging the 2022 “Employment Situation” data from the Bureau of Labor Statistics, Gulf War
II Era veterans (any who have served after September 11th, 2001) represent the largest
percentage within the veteran population making up 44.49% (see Figure 1).
136
2022 Average Unemployment Population Percentage
18.86
44.49
10.63
26.15
Gulf War II Era Gulf War I Era WWII, Korean War, Vietnam Era Other Service Periods
The total unemployment percentage averaged to about 2.83% in 2022, Gulf War II Era
may have had higher unemployment rates, the focus of this research seeks to investigate the Gulf
War II Era veterans seeking employment within the civilian sector given its population size and
relevancy toward veterans entering the civilian labor force. Thus, within this exploratory
research, we seek to further understand factors which may contribute to this population’s
unemployment by investigating how Gulf War II Era veterans market their experience.
Commonly cited reasons for high unemployment durations for these veterans, relative to non-
conditions, physical disabilities, and inept job search strategies (Loughran, 2014).
Upon entering the civilian labor force, veterans are faced with the challenge of
communicating the unique skills, experiences, and achievements earned throughout their military
career in such a way that is marketable to a civilian employer within their resume. Although the
Department of Defense provides a Transition Assistance Program and Veteran Affair's benefits
137
to equip veterans with tools for communicating their skills via resume modules and curriculum,
veterans have claimed that these programs have done little to assist in truly preparing them for
drafting effective resumes and translating military experience (Harrod et al., 2023).
Coincidentally, the unemployment rate of Gulf War II Era veterans continues to remain above
the average veteran unemployment, and there have been claims that there is very little research
surrounding United States military veterans and employability within the civilian sector and that
there is potential for this research interest to grow (Goss, 2020; Stone, 2016).
With the Bureau of Labor Statistics projecting a 36% growth of the data science career
field between 2021 and 2031, veterans departing the military may see opportunity to start
marketing themselves to this career field. This growth within the data science career field is
considered significantly faster than the average for all other occupations listed throughout the
United States within O*Net. Understanding military veterans departing Active Duty may pursue
a variety of different career fields, we seek to explore Gulf War II Era resumes that have
deliberately communicated data science skills. In doing so, this paper investigates the
employability of military veterans that have communicated data science skills by analyzing the
style. We further incorporate a veteran’s career identity by observing data science competencies
(Votto et al., 2023). Lastly, we factor in control variables related to the veteran with regards to
their military rank, years served on active duty, holding a security clearance, college education,
managerial experience, data science experience, and total number of civilian jobs held. Once
these variables are calculated, we proceed to explore any relationships that may exist between
138
the psychometric variables, data science competencies, controls, and civilian unemployment
durations.
By exploring these relationships, we aspire for this paper to produce three primary
contributions. First, we add to the growing pool of research surrounding employability and
United States military veteran. We also demonstrate the capabilities of leveraging an academic
IT artifact to conduct research on a unique and important demographic within the United States
labor force. Secondly, this research fortifies academic interests in exploring psychometric factors
and data science competencies which may contribute to employability of veterans. Lastly, this
paper provides insight to veterans and employers within the data science community on potential
The arrangement of this paper is organized into the following sections. Section 2
discusses previous literature which has investigated veteran unemployment. We also highlight
literature that has leveraged LIWC-22 to conduct language analysis. Additionally, we introduce
the theoretical background and research model which guides the hypotheses development of this
research. Section 3 discusses the methodology of our analysis. Within this section we discuss
data collection and cleaning techniques. We further discuss analysis techniques leveraged to
explore the relationships between unemployment durations and psychometric variables within
resumes. The next section examines the results the analyses and further discusses post-hoc
limitations, and future research directions of this research. Lastly, the final section provides a
conclusion which summarizes this work and encourages further exploration of veteran
unemployment.
139
CHAPTER 3: LITERATURE REVIEW AND THEORIETICAL BACKGROUND
Within this section, we introduce the definition of a military veteran and published
research that has contributed to the exploration of veteran employability. We further Fugate et
al.’s (2004) proposed construct of employability which guides our research model.
individual who served within the reserve, guard, or federal active military and were discharged
or released therefrom under conditions other than dishonorable (Department of Veteran Affairs,
2019). Understanding these individuals are equipped with unique skills based on their military
experiences, veteran employability has been a research interest given concerns surrounding
unemployment and challenges to assimilate back into the civilian workforce. Therefore, we
employability to understand how this has already been explored and how our research could
Within Table 1, we have summarized nine pieces of literature that have thoroughly
investigated veteran employability. This table highlights the authors and year of publication. It
further provides insight to the sample analyzed, research methods, and findings. When
Votto et al. (2021). We first established the database from which the literature would be pulled
from. For this research, we leveraged Google Scholar given its multidisciplinary publication
sources and open-source nature. We then determined search parameters consisting of a date
range and search string. Our scope of interest was publications within the Gulf War II era
timeframe. Therefore, the established date range was any peer-reviewed publication between
140
September 11th, 2001, and March 2023. The search string we used further narrowed the scope by
identifying peer-reviewed literature with the following keywords: “United States”, “Military
Veteran”, and “Employability”. We used the Boolean operator, “and”, to connect the keywords
together and ensure the results contained each of the established keywords. The third step was
extracting the most relevant pieces of literature from Google Scholar. Understanding hundreds of
results are revealed after searching, we filtered the results by “most relevant” and chose the top-
ten pieces of literature which were perceived as “most-relevant” based on the established
keywords. The last step of our literature review consisted of reading each article to confirm it
Of the identified literature reviewed, four of ten were doctoral dissertations published
between 2016 and 2020. Additionally, three journal articles and two book chapters were
identified. We disqualified one piece of literature, as it was outside the scope of our interest due
to focusing on the Canadian Armed Forces. It was noted that of all the publications within the
Google Scholar database, the top ten most relevant results yielded more dissertations than
141
Table 1. Previous Veteran Employability Research Continued
Stone 2016 Dissertation 511 Survey analysis Perceptions of hiring
MTurk to gauge how managers play a large role
Participant human resource in applicant selection—
s managers military stereotypes were
perceive military supported to affect how
veterans veterans are perceived
James 2017 Dissertation 10 Female Interview 5 prevalent themes
Military Analysis on identified: 1) importance of
Officers military networking, 2) workplace
transition to culture, 3) role adjustment,
civilian jobs 4) military stereotypes, and
5) lessons learned/shared
Fossey et al. 2019 Book Chapter Military- Systematic Developed set of principles
to-Civilian policy review of to assist in developing
policies published NATO effective policies for
developed guidance military transitioning to
by NATO surrounding civilian life
military
transition and
benefits
Goss 2020 Dissertation 15 Post Qualitative Identified four themes:
9/11 Army Descriptive 1)feeling of uncertainty
Officers Study (interviews shifting roles, 2) reliance on
and focus inner self and veteran
groups) on identity to build confidence,
military 3) establish support
transition to network, and 4) remain
civilian mission-focused to prepare
employment, for transition
one-on-one
interviews &
focus groups
Rutter 2020 Dissertation 10-15 Qualitative Identified seven themes:
prior Descriptive 1) planning and preparation,
military Analysis (online 2) work recommendations,
non- questionnaire, 3) cultural change, 4)
commissio semi-structured competencies, 5)
ned officer interview) on networking, 6) role in the
veterans motivation workplace, and 7) support
influencing systems
military-to-
civilian transition
Lampka and 2017 Journal Article 5 Veteran Mixed Methods 3 Qualitative findings which
Kowaleski Resumes (Qualitative hindered successful
analysis for employment within civilian
resumes, workforce: 1) military
quantitative for linguistics, 2)skills were
unemployment embedded, deemphasized
durations) importance, 3) no
management experience
communicated.
142
Table 1. Previous Veteran Employability Research Continued
3 Quantitative finding:
regression analysis showed
a strong positive
relationship between
civilian and unemployment
rates
McLennan 2021 Book Chapter Bureau of N/A Overview of military
Labor occupations and pay grades
Statistics and their effect on
military employability
occupation
listing
Perkins et al. 2022 Journal Article 854 Survey Analysis Two key findings: 1) Career
Veterans to gain insight on planning/exploration and
participati components that resume writing were
ng in predict significantly important to a
employme employment and veteran in search of a job
nt help veterans after the military, 2)
programs find jobs veterans leveraging a
mentor or coach were 2.2
times more likely to finda
job 6- to 9-months after
separation
Theoretical Background
construct (MDC). Law et al. (1998) define MDCs as directionality of the relationships between a
construct and its dimensions. In this instance, this MDC refers to one’s employability (the
construct) and its relationship to associated dimensions. Fugate et al. (2004) view this
relationship as aggregate, such that the dimensions cause or create the construct (one’s
human capital (PHC) and career identity (CI) for each resume (Fugate et al., 2004; Ngoma and
Ntale, 2016). The employability construct and its dimensions are summarized in Figure 1.
Prior to discussing the individual dimension, we first seek to define the concept of
employability and the control variables associated with each resume. The construct of
143
employability embodies individual characteristics which affect how individuals’ interface within
foster adaptive behavior and cognition with regards to communicating skills and performing
tasks (Fugate et al., 2004). To measure employability, unemployment durations have been
For instance, Lampka and Kowaleski (2017) leverage a mixed methods approach to gain better
understanding of the relationships which exist between veteran unemployment and skills
Figure 2. Employability Construct: Modified from Fugate et al (2004) and Ngoma and
Ntale (2016)
individual which related to their career and job affiliation. Essentially, these are the traits which
qualitative elements such as tonality, analytical thinking, and authenticity when communicating
experience (McEvoy et al., 2005; Nisha and Rajasekaran, 2018). We also consider two
communication styles observed within a resume derived from Osgood et al. (1957). The first is
144
active orientation which refers to diction that imply active engagement and initiative. The second
is week diction which refers to the use of language related to submission, vulnerability and
withdrawl (Lim 2002; Stone and Hunt, 1963; Osgood et al., 1957). Therefore, PHC is broken
into two branches: 1) psychometric variables and 2) communication styles. Within this
subsection, we will dive into literature surrounding each element of PHC and derive hypotheses
Regarding the PHC element of analytical thinking, we observe two pieces of literature
which highlight its relationship to employability and derive our first hypothesis. Nisha and
Rajaeskaran (2018) conduct a literature review which highlights a variety of traits that are
considered desirable skills for potential employees seeking new employment opportunities.
Among the skills listed, analytical thinking ranked within the most marketable that employers
seek within prospective employees. Furthermore, Kulkarni et al. (2017) designs an employability
skill matrix which studies engineering graduates and their ability to problem solve. Through this
matrix, increased analytical thinking and expertise in technical knowledge yielded improved
employability within the student sample. Thus, we develop the following hypothesis regarding
The next observed element of PHC is authenticity. This element refers to the perceived
honesty and genuineness of input (Boyd, 2022). Within literature, effective self-branding
requires the applicant and employer to be authentic and clearly understand the “why”
surrounding both sides of the hiring process (Ainspan and Saboe, 2021). Specifically, an
employee must be true and honest with the skills they bring to the labor pool and communicate
145
with them effectively. Contrarily, the employer must effectively communicate desired skills and
expectations of an employee.
Tonality is the third element within PHC. This element refers to the social and
psychological responses to emotional (positive or negative) stimuli (Cohn et al., 2004). In the
context of employability, Fugate and Kinicki (2008) refer to tonality as the cognitive reappraisals
which indicate initiative and proactive behavior, such that positive tones reflect those who are
optimistic and take initiative. Considering the desirability of employees who take initiative and
bring a positive outlook to the organization, we posit the following for further analysis:
We further consider weak and active orientation with regards to the communication style
of each resume. Active orientation refers to the individual’s ability to react to intellectual
challenges within a space and take initiative (Memarovic et al., 2012). Contrarily, weak
orientation refers to an individual’s inability to seize initiative and lean toward submissive
tendencies (Stone and Hunt, 1963). Given these characteristics, we posit the following regarding
communication styles:
Fugate et al. (2004) introduce career identity as one’s self-definition within a career
context. This construct is referred to as a cognitive compass which motivates one to adapt and
create opportunities to match work aspirations. This construct provides a strong foundation for
employability through an applicant’s ability to self-reflect on goals and communicate their skills
146
needed to achieve them. Thus, we consider data science competencies addressed by Votto et al.’s
(2023). These competencies were derived from literature and guidance published by IBM and
Deloitte and investigate how veterans communicate their identities within competencies and how
technologies (Ammanath, 2022; Mökander and Floridi, 2021). The development of ethical
expectations goes beyond that of technology. The desire for employees to have a grounded sense
of ethics, especially when handling sensitive data, has become a desired competency of
employees. Thus, we extract our first competency, ethics, from this framework. This competency
reflects how employees communicate their ability to enforce equitable and fair practices within
the data science work environment. Additionally, we extract another competency, Artificial
Intelligence, from this framework to address how employees communicate their aptitude to
leverage these intelligent tools. That said, Artificial Intelligence encompasses a variety of topics
related to machine learning and deep learning (Votto et al., 2021). one core competency from
this Deloitte framework to gauge how military veterans communicate their experience and skills
to make themselves employable within the data science career field. Thus, we posit the
following:
The second framework we leveraged to derive competencies was developed by IBM and
is titled “Data Science Skill’s Competency Model”. Although this framework contains twenty-
eight data science competencies across seven groups, we extracted three primary considerations
147
for this analysis. The first competency we adopted was problem solving. This refers to a
identify business problems, and develop solutions. The second competency we extracted from
IBM’s model was “machine learning”. This competency refers to an applicant’s ability to
statistical components. Lastly, we extracted “statistics” from the IBM model as our last
understand probability theory, sampling, and other vital statistical components (IBM, 2020).
derived. The ability to reflect on past experiences and communicate skills related to this career
field bolsters career identity within this career field and promotes potential to enhance
employability. Thus, through the posited hypotheses, we hope to further explore how veterans
communicate their skills and understand how these competencies are related to employability.
Figure 2 summarizes the hypotheses and research model driving this research initiative.
148
CHAPTER 4: METHODOLOGY
Within this section, we describe the process of data acquisition and cleansing, descriptive
statistics, and analytic techniques. We utilize Meyer ‘s (2019) three-phased methodology when
conducting our analysis. The first phase consists of data collection and preparation. The second
phase discusses how the dimensions are organized. The last phase is the analysis and reporting
component. The following subsections provide deeper insight into each of the phases.
With regards to the source of our resumes, we manually extracted resumes from Indeed
(Muthyala et al., 2017). Understanding APIs exist to assist and streamline resume extraction
endeavors, we pursued the manual route given the specific demographic (Gulf War II Era
We acknowledge our desired data may contain sensitive information relating to military
veterans. Resumes may contain personal emails, telephone numbers, and addresses related to the
individual. Therefore, we pursued an Information Review Board to review our proposed research
interests and confirm confidentiality measures would ensure ethical research practices were
maintained and respected. Upon receiving IRB approval, we continued to analyze resumes from
Indeed.
Following similar strategies within systematic literature reviews (Votto et al., 2021), we
conducted a nine-step filtering process. This was a nine-step process (summarized in Table 2)
ultimately yielded 395 resumes for our analysis. Within the first step, we typed “data science”
into Indeed’s search field yielding 1,467,289 results. We then limited the search to “Job Title”
and “Skills” within the United States. These parameters yielded 130,094 resumes. We further
narrowed our parameters down to resumes that had been updated within the last six months,
149
excluding any candidates that had been contacted within the last 30 days. This filtration yielded
50,105 resumes. For the fifth step, we observed resumes that were “read to work now”, which
yielded 27,725 resumes. We then selected candidates that were “active or former military” which
yielded 1,125 resumes. We then manually reviewed each resume and extracted prior active-duty
Gulf War II Era veterans and excluding any members who were still serving on active duty. This
yielded 504 resumes. For the eighth step, we continued to manually review each resume and
eliminate any National Guard or Reserve members, yielding 430 resumes. For the last step, we
eliminated any personally identifiable information from each resume to ensure anonymity and
recruited an active-duty military member to help audit the resumes to ensure they veterans served
on active duty and were not currently in the reserves or national guard. This yielded our final
Data Operationalization
To compute the PHC variables, we first leveraged LIWC-22 (Parameswaran et al., 2022)
to quantify the presence of analytic thinking, authenticity, and tone within a resume. This
whose constructs have both been tested for validity and reliability (Boyd, 2022; Parameswaran et
al., 2022; Boyd and Pennebaker, 2015). Additionally, we used the General Inquirer to find the
psychological indexes of Osgood et al.’s (1957) active and weak lexicon. In doing so, we
constructed a machine algorithm to observe the presence of Osgood et al.’s (1957) active and
week orientation lexicons and derive a percentage, on the same scale as the LIWC-22 variables.
The reliability and validity of Osgood et al.’s (1957) has been tested throughout literature
(Stoklasa et al. 2017; Stone and Hunt, 1963). All PHC variables are represented as a percentage,
150
relative to the word count of each resume. Table 3 summarizes all continuous variables observed
within this analysis. The last two variables located within Table 3 are later discussed.
Regarding the CI variables, we consider the private sector data science competency
dictionaries established and validated by Votto et al. (2023). Although, this research leverages
these dictionaries to evaluate resumes instead of job descriptions. These dictionaries were
derived from data science job descriptions scraped from Glassdoor. Votto et al. (2023) utilized a
151
Word2Vec model to extract the top-10 most similar related to each data science competency
(ethics, machine learning, artificial intelligence, problem solving, and statistics). We leveraged
MAXQDA to run our quantitative content analysis to identify which resumes had a presence of
each data science competency (Oliveira et al., 2015). If the resume contained words from the
calculate the required variables. Regarding the dependent variable, the research assistant
calculated how many weeks each individual resume indicated an employment gap between
civilian jobs. Upon completion, each resume had a number to reflect the number of weeks the
Additionally, the research assistant extracted the control variables for our model. There
were seven controls total we considered with our model. The resumes did not include
considerations involving education and work experience (Cole et al. 2004). The first control we
considered was whether the veteran indicated earning a college degree (1 = degree
communicated, 0 = otherwise). The third control variable military rank of the veteran (1 =
officer, 0 = enlisted). The fourth control consisted of quantifying how many years the individual
served on active duty (1 = over 10 years, 0 = less than 10 years). The fifth and sixth control
variables examined the most recent job the individual held. The fifth confirmed if their most
recent job was managerial in nature (1 = managerial, 0 = not managerial), while the sixth
confirmed the veteran’s most recent job was related to data science (1 = data science related, 0 =
152
otherwise). The last control variable we considered was how many civilian jobs the veteran held.
This variable is expanded upon in Table 3, while the remaining categorical variables are
summarized in Table 4.
Descriptive Statistics
Regarding the descriptive statistics, we leveraged SPSS to analyze our frequency tables
and correlation matrix ( Kim et al., 2005). Table 5 and Table 6 report the descriptive statistics of
the variables of our interest. We also conducted a Spearman Correlation Test to gauge the
correlations between the variables of our interest. There were no variables that shared a high
correlation as the highest was .392 (Parameswaran et al., 2022; Valecha et al., 2021; Venkatesan
153
Table 4. Descriptive Statistics: Continuous Variables
Descriptive Statistics
Minimum Maximum Mean Variance
Analytic 83.53 98.76 94.2468 6.507
Authentic 4.14 94.67 31.6974 201.691
Tone 3.61 98.56 55.8252 394.867
Weak .00 1.662 .25 .064
Active .58 9.04 3.12 1.54
Total_Num_Civ_Jobs 0 22 3.63 6.12
Intelligence
Clearance
Job (Data
Graduate
Artificial
Learning
Statistics
Previous
Previous
Machine
Problem
Security
Solving
College
Ethics
Years
Rank
Job
Mean .81 .08 .44 .15 .45 .34 .29 .50 .40 .30 .45
Variance .15 .07 .25 .13 .25 .23 .21 .25 .24 .21 .25
Analytic Techniques
Due to the violation of normality within our variables, negative binomial regression has been
suggested as a method to analyze our variety of variables given the robust nature of this analysis
(Valecha et al., 2021; Osgood, 2000). Therefore, each of the hypotheses were tested using
negative binomial regression. The equation and results of this analysis are as follows:
154
Table 6. Main Model Results
155
CHAPTER 5: RESULTS
The results of the negative binomial regression analysis are summarized in Table 7.
These results indicate significant negative effects on civilian unemployment for several key
variables. Regarding PHC, tone (p-value < .001) and active orientation (p-value < .001) were
statistically significant and had a negative effect on civilian unemployment. This indicates the
more positive and optimistic the member’s tone is throughout their resume, the more employable
they could be. Similarly, leveraging an active disposition when communicating one’s worth
positively affects employability. Regarding CI, three of the five data science competencies
negatively impacted unemployment durations. These competencies are ethics (p-value < .05),
machine learning (p-value < .001), and artificial intelligence (p-value < .001). This indicates that
the incorporation of ethical behavior and skills related to artificial intelligence and machine
learning bolster one’s employability within their resume when pursuing data science jobs.
Robustness Checks
To expand upon our initial findings, we conducted three robustness checks. This
subsection provides deeper insight into these robustness checks. The first robustness check
considered was an alternate measure test. This investigates the consistency of results (Balance et
al., 1987) within an analysis by considering a different dependent variable. Within this test, we
observed the number of weeks unemployed since their last job listed within the resume instead of
the total number of weeks unemployed as a civilian. Through this test, we confirmed a member’s
tone (p-value < .05), and mention of ethics (p-value < .001), machine learning (p-value < .001),
and artificial intelligence (p-value < .001) continued to negatively affect unemployment duration.
This implies consistency between data science competencies and the psychometric variable of
156
Table 7. Alternate Measures Analysis
Psychological Human Capital (Psychometric Variables)
Analytic Thinking (-.070)
~10.0846
Authenticity (-.047)
~.0193
Tone (-.025)*
~.0061
Psychological Human Capital (Communication Style)
Weak Disposition (-13.370)***
~.2.5266
Active Lexicon (-.154)
~.1083
Career Identity (Data Science Competencies)
Ethics (-.1.647)***
(Glassdoor) ~.5033
Machine Learning (-6.160)***
(Glassdoor) ~1.0844
Artificial Intelligence (-5.026)***
(Glassdoor) ~.9975
Problem Solving (-.560)
(Glassdoor) ~.5089
Statistics (6.603)***
(Glassdoor) ~.7967
Control Variables
College Graduate (3.627)***
~.7541
Military Rank (4.446)***
~1.1200
Security Clearance (-4.501)***
~.6536
Years Active Duty (-37.498)
~ --
Previous Job (2.069)**
(Managerial) ~.7176
Previous Job (Data (-3.849)***
Science) ~.5671
Total Number of (-1.054)***
Civilian Jobs ~.1804
Note: p-value <0.10; *p-value <0.05; **p-value <0.01; ***p-value <.001
Note: () covariate; ~standard error
science competency dictionaries from a different private-sector job description source, LinkedIn.
Within this test, we sought to explore how using a different source for our dictionaries could
affect the significance of our established variables from the main model. We followed the same
Word2Vec method mentioned in Votto et al. (2023) and generated a new list of the top-10 most
157
similar words (Deguhci et al., 2021; Brumer et al., 2017). After eliminating duplicate job
descriptions and tokenizing the words, we trained the model on 1,000 data science job
descriptions from LinkedIn. The results of the Word2Vec “most_similar” method can be found
in the Appendices B and C the Glassdoor and LinkedIn datasets. Once the dictionaries were
created, we coded each resume based on the presence of each dictionary (1 = present, 0 =
otherwise). Upon successfully coding the resumes, we conducted another Spearman Correlation
test to ensure the new variables were not multicollinear with the established controls and other
independent variables. There were no variables that shared a high correlation as the highest was
.282 (Parameswaran et al., 2022; Valecha et al., 2021; Venkatesan et al., 2021). We then reran
our original model, substituting the original private sector dictionaries from Glassdoor with the
dictionaries created from LinkedIn. The results of this secondary analysis yielded considerable
and consistent results. To summarize, both tone (p-value < .001) and active orientation (p-value
< .001) stayed true to negatively impacting unemployment durations. Furthermore, the CI
variables of machine learning (p-value < .05) and artificial intelligence (p-value < .001)
remained the same by negatively impacting unemployment durations. Thus, supporting our
original findings within the main model. Table 9 summarizes these findings.
158
Table 8. Alternate Dictionary Analysis Continued
Machine Learning (-.562)*
(LinkedIn) ~.2474
Artificial Intelligence (-1.413)***
(LinkedIn) ~.2534
Problem Solving (.693)*
(LinkedIn) ~.2742
Statistics (2.113)***
(LinkedIn) ~.2401
Control Variables
College Graduate (2.917)***
~.3229
Military Rank (-2.263)***
~.4606
Security Clearance (.770)***
~.2170
Years Active Duty (-2.481)***
~.3595
Previous Job (-.905)**
(Managerial) ~.2893
Previous Job (Data (.094)
Science) ~.719
Total Number of (.386)***
Civilian Jobs ~.0659
Note: p-value <0.10; *p-value <0.05; **p-value <0.01; ***p-value <.001
Note: () covariate; ~standard error
The last robustness check we conducted was an endogeneity (Hausman) test. This test is
designed to investigate the any correlations that may exist between our exploratory independent
variables and error terms (Lu et al., 2018). To examine the endogeneity issue, we acknowledge
it is important to choose an instrumental variable for each model (Venkatesan et al., 2021).
Within this analysis, we consider the entire word count of each resume as the instrumental
variable and consider the Glassdoor data science competencies from Votto et al. (2023).
Leveraging SPSS, we first generated our residuals (Sargan-Hansen J-Statistic) for each model.
This was accomplished by establishing our dependent variable as the data science competency
and word count as our independent variable. We then ran a regression analysis to generate the
error residual for each competency. Once the residuals were generated, we plugged them
residual into our original model, one at a time, to better understand some of the relationships
which may exist. This resulted in five different models which are summarized in Table 10.
159
Models 1 (ethics) and 4 (problem solving) confirmed the variables were not interdependent with
one another by the statistically insignificant results. This confirms that there are no omitted
variable biases caused by the effect word count has on each data science competency. Models 2
(machine learning), 3 (artificial intelligence), and 5 (statistics) did not meet the endogeneity
criteria due to resulting in statistically significant results. That said, all five models shared the
same recurring statistically significant results from the original model of this analysis. Tone,
active orientation, ethics, machine learning, and artificial intelligence continued to negatively
impact unemployment durations, despite the residuals being added to the model. Despite several
of the models not meeting the endogeneity criteria, there was consistency demonstrated within
160
Table 9. Results of Endogeneity Test Continued
College Graduate (.3.072)*** (2.970)*** (3.083)*** (3.032)*** (3.088)***
~.3386 ~.3388 ~.3390 ~.3369 ~.3394
Military Rank (-2.066)*** (-2.173)*** (-2.052)*** (-2.118)*** (-2.042)***
~.4623 ~.4619 ~.4612 ~.4648 ~.4618
Security Clearance (.956)*** (.756)*** (.969)*** (.907)*** (.973)***
~.2393 ~2237 ~.2383 ~.2685 ~.2389
Years Active Duty (-2.482)*** (-2.596)*** (-2.473)*** (-2.509)*** (-2.476)***
~.3695 ~3793 ~.3695 ~.3691 ~.3697
Previous Job (Managerial) (-1.711)*** (-1.411)*** (-1.731)*** (-1.673)*** (-1.723)***
~.2639 ~.2571 ~.2649 ~.2625 ~.2640
Previous Job (Data Science) (.363) (.305) (.347) (.352) (.368)
~2450 ~.2417 ~.2437 ~.2454 ~.2451
Total Number of Civilian Jobs (.617)*** (.610)*** (.624)*** (.600)*** (6.25)***
~.0715 ~.0685 ~.0720 ~.0705 ~.0720
Residual on Model
Competency Residual (-2.655) (-2.583)*** (-2.898)* (-1.017) (-4.826)*
~1.3592 ~.5569 ~1.3513 ~.7098 ~2.2313
Note: p-value <0.10; *p-value <0.05; **p-value <0.01; ***p-value <.001
Note: () covariate; ~standard error
Hypotheses Results
As mentioned previously, PHC was represented by the veteran’s tone and active
orientation when communicating their worth within the resume. Simultaneously, CI was
represented by the presence of ethics, machine learning, and artificial intelligence. Each of these
variables negatively impacted unemployment durations within the resume, thus indicating the
potential for enhancing a veteran’s employability. Regarding the control variables, there were
three which negatively impacted unemployment durations. The first was military rank which
implied that if the veteran was an officer, their unemployment duration was less than that of an
enlisted member. The next was the number of years served on active duty. Based on this
analysis, if the veteran served over 10 years in the military, the unemployment duration within
the resume was negatively affected. Lastly, if the military veteran’s most recent job was
managerial in nature, then the unemployment duration was also negatively affected. There is one
control variable worth mentioning that positively affected unemployment durations. The number
of civilian jobs held by the veteran positively affected unemployment durations. This implies that
it hindered the employability of the veteran. This can be explained by the veteran holding
161
numerous jobs and bouncing around from opportunity to opportunity, thus creating more gaps
within their experience pool. Such behavior could have employers shy away from employing this
individual as they’ve demonstrated an inability to remain at a job for a desired duration of time
insignificant results. This implies that these variables may not affect a veteran’s employability.
Furthermore, we had one data science competency which positively affected unemployment
durations. This went against our initial hypothesis that the presence of statistics would positively
impact employability. We speculate that the cause of this is related to the increase of AI and
machine learning capabilities within the data science career field. Although statistics may be a
core component of machine learning practices, it may not be as marketable on its own as the
demand for machine learning engineers and scientists grows. To summarize our findings, Table
162
Table 10. Hypotheses Results Continued
Career Data Science Ethics H3a .003 Hypothesis Supported
Identity Competency
163
CHAPTER 6: POST-HOC ANALYSIS
Within this chapter, we discuss the two post-hoc analysis we performed to open the AI
black box within military veteran resumes. The first analysis explores the Top-25 data science
technologies listed on O*Net and investigates their presence within veteran resumes. Regarding
the second post-hoc analysis, we conduct a topic modeling analysis to gain insight to potential
of Labor to showcase occupations and communicate desired skills related to said occupation
(Peterson et al., 2001). This website seeks to digitize information regarding professions that exist
within the United States to help prospective applicants understand the rising demands (Lennon et
al., 2023). Considering the data science career field is expected to grow 36% between 2021 and
2031 (Bureau of Labor Statistics, 2023), we sought to leverage the Top-25 data science skills
posted within the O*Net database. Upon collecting these variables, we separated them into two
164
We then created dictionaries for each category and calculated the existence of each
technological skill within each resume. We quantified how many resumes contained each
technological skill and averaged the number of months each veteran was unemployed, as seen in
Table 13 and Figure 3. To avoid redundancy, we only counted the skill once per resume to
165
O*Net Data Science Software vs Average Unemployment
(Months)
60.00
50.00
40.00
30.00
20.00
10.00
0.00
Apache Amazon Tableau Apache SAS Scala Microsoft Apache Git MATLAB Docker Microsoft
Spark Web Hadoop Azure Hive Excel
Services
Regarding software, Amazon Web Services, Tableau, Git, and Microsoft Excel were the
most prevalent within veteran resumes, each variable demonstrating a lower unemployment
duration when compared to some of the other software. This indicates veterans have more
experience with these technologies and do well to advertise these skills within resumes,
potentially reducing unemployment durations. Of the four mentioned, Microsoft Excel was the
strongest represented with over 50% of veteran resumes communicating experience with this
suite.
166
Table 13. Top Programming Languages Related to Resumes Continued
Scikit-Learn 4.56 18 9.96
PyTorch 1.77 7 7.74
C++ 21.27 84 12.53
Pandas 11.65 46 7.22
Keras 4.05 16 3.26
NumPy 9.37 37 8.08
NoSQL 3.29 13 10.40
PySpark 1.27 5 1.41
50.00
40.00
30.00
20.00
10.00
0.00
The most prominent programming languages that exist within veteran resumes include
Python, SQL, and Oracle Java. There is an inverse relationship between these languages and
within the percentage of resumes and lower unemployment durations. This would imply that
veterans expressing experience with these skills are more marketable with regards to securing a
observing 3 software and programming languages to investigate the effects they would have on
167
the main model of our analysis (Shah et al., 2022; Li et al., 2021; Kumar et al., 2018). The top-
three software identified (Li et al., 2021), in order of resume presence, are as follows: 1)
Microsoft Excel (data management software), 2) Git (system software), and 3) Tableau (data
visualization software). In addition, the top-three programming languages (in order of presence)
are as follows: 1) SQL, 2) Python, and 3) Oracle Java. The following subsections detail the
This subsection provides a deeper insight into the empirical investigation we conducted
to understand the potential effects the top-three software and programming languages may have
on veteran employability.
As seen in Table 15, Microsoft Excel and Git are statistically significant and have a
negative effect on unemployment durations. This implies that there is potential for the inclusion
of these skills within a resume to enhance a veteran’s employability. Contrarily, Tableau was
statistically significant; although, its inclusion within the model positively impacted
unemployment durations. This indicates that the inclusion of this software within a resume may
168
Table 14. Data Science Software Model Results Continued
Active Lexicon (-.859)***
~.1379
Career Identity (Data Science Competencies)
Ethics (-.813)**
~.2762
Machine Learning (-.457)
~.3106
Artificial Intelligence (-1.248)***
~.3247
Problem Solving (.670)*
~.2776
Statistics (3.533)***
~.3332
Control Variables
College Graduate (4.155)***
~.4310
Military Rank (-2.215)***
~.5253
Security Clearance (.1.909)***
~.3022
Years Active Duty (-3.303)***
~.4365
Previous Job (Managerial) (-2.398)***
~.2893
Previous Job (Data (.006)
Science) ~.2938
Total Number of Civilian (.568)***
Jobs ~.0756
Software
Microsoft Excel (-1.284)***
~.3308
Tableau (.652)*
~.3175
Git (-2.972)***
~.4254
Note: p-value <0.10; *p-value <0.05; **p-value <0.01; ***p-value <.001
Note: () covariate; ~standard error
We also considered the effects of programming languages within our main model. As
indicated in Table 16, the inclusion of Python and Oracle Java negatively affects unemployment
169
durations within our model. This result indicates that veterans who communicate skills related to
these languages may make themselves more employable, given the negative effects on
unemployment durations. Although SQL was the most prevalent within resumes, it was not
considered statistically significant. This indicates that there is potential the inclusion of this
programming language may not be as marketable than the former-mentioned languages. The
170
Table 15. Data Science Programming Language Model Results Continued
Military Rank (-2.359)***
~.4802
Security Clearance (.746)***
~.2236
Years Active Duty (-3.343)***
~.4348
Previous Job (Managerial) (-1.565)***
~.2589
Previous Job (Data (.686)**
Science) ~.2471
Total Number of Civilian (.494)***
Jobs ~.0726
Programming Languages
Python (-1.556)***
~.2627
SQL (.565)
~.3109
171
CHAPTER 7: DISCUSSION
This study makes three two contributions to literature. First, we contribute to the broad
texts surrounding data science, artificial intelligence, and human resource management by taking
leverages natural language processing techniques using Word2Vec to generate data science
competency dictionaries, which are then used to investigate veteran resumes. This approach
shines a light on potential capabilities that exist which could be used to better understand how
AI and machine learning techniques could be used within the human resource management
domain. To better understand how Gulf War II veterans communicate their data science skills,
we hypothesized that psychological human capital and career identity affect their employability
(Ngoma and Ntale, 2016; Fugate et al., 2004). By leveraging technologies like LIWC-22,
Word2Vec, and MAXQDA we successfully investigated potential variables which could affect a
veteran’s employability. This demonstrates the potential for future research to continue the
exploration of leveraging human resource data within AI and machine learning initiatives.
The second contribution of this research shines a light on factors which may affect a
veteran’s employability when seeking a job within the data science career field. Specifically, our
data and analysis suggest the positive tone of one’s resume decreases unemployment durations
throughout resumes. This implies that if a veteran uses optimistic diction or words that convey a
positive outlook on their experience, then the potential for lower unemployment durations exists.
Furthermore, if the veteran’s communication style is active, implying they take initiative, this
also has a positive influence on employability. Regarding the data science competencies, ethics,
artificial intelligence, and machine learning were also positive influences on a veteran’s
172
component of conducting analyses, especially with organizations that manage big data
artificial intelligence technologies could improve a veterans employability, as these skills are
Through this analysis, we hope to inspire new literature to explore the various
components of human resource management and data science, especially by highlighting some
of the limitations which exist within this research. First, human resource professionals are
charged with managing a rich source of data relating to employees and their respective
organizations. At times, this data is qualitative in nature, which creates potential opportunities for
however, there are more word embedding algorithms which exist that could provide insight to
contextual factors within resumes and job descriptions. This research was limited by only
leveraging Word2Vec. Future research initiatives could explore how other word embedding
algorithms like FastText (Bojanowski et al., 2017), GloVe (Pennington et al., 2014), ELMo
(Peters et al., 2018) and BERT (Delvin et al., 2018) analyze resumes and job descriptions.
Additionally, we acknowledge that this research is limited by the small sample size of our
resume pool. Understanding we leveraged a manual approach to cleaning and auditing these
resumes to ensure the demographic we desired was captured, there are sophisticated web
scraping capabilities which exist that may streamline the process and produce a larger sample-
173
size for analysis. Future research endeavors could leverage these newer technologies to acquire a
To expand upon this research initiative, other communication styles could be explored
when considering resumes. This research only considered Osgood et al.’s (1957) active and
weak lexicon, given the intent and nature of resumes. Other potential communication styles
could be explored such as Laswell’s Value dictionaries or Harvard’s IV-4 dictionaries (Lim,
2002).
differences which may exist between veterans seeking jobs within data science. This research
investigated any veteran that was considered a Gulf War II veteran. This could be broken down
into tertiary components. For example, exploring how well a veteran who was recently
discharged from the military communicates their skills versus veterans who have been out for
several years. This could potentially give insight to how well veterans are postured to write their
resumes upon leaving the military and provide an avenue for NLP capabilities to step in and help
Lastly, future research initiatives could further investigate the various transition
assistance programs offered to military veterans. As discussed in Harrod et al.’s (2023) interview
research, veterans have claimed that the current curriculum has not prepared them well to
translate their skills or properly structure their resume. Through this research on veteran
employability, newer research directions could explore resumes of veterans who have recently
attended a transition assistance program and investigate the structure, content, and marketability
of their resumes. In doing so, this may provide deeper insight to potential curriculum
174
enhancements for the resume-building component of transition assistance programs offered to
175
CHAPTER 8: CONCLUSION
Within this paper, we investigated Gulf War II veteran resumes that communicated skills
within data science. We leveraged a multidisciplinary approach, allowing Fugate et al. (2004)
and Ngoma and Ntale’s (2016) theory on employability to guide our NLP investigations.
Through this analysis, we demonstrated factors which may contribute to positively affecting a
veteran’s employability when pursuing a career within the data science community. Through the
limitations addressed in the previous section, we clearly denote that there are rich avenues worth
exploring with regards to human resource management and the data science community. Outside
of the mentioned contributions, this research hopes to provide two additional perspectives. First,
we hope this research shines a light on how multidisciplinary approaches could benefit research
initiatives. Secondly, we hope this research inspires future scholars to continue the conversation
of where machine learning capabilities could exist within human resource management function.
We acknowledge there are considerable risks involving biases and transparency when
incorporating these newer technologies. With that, we hope this research invites more
conversation and dialogue on how to appropriately leverage and manage these newer
176
APPENDICES
1
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 7
1 --
2 -0.106* --
-
0.165* .238*
3 ** ** --
-
.130*
4 .063 ** .028 --
5 .072 -.086 -.064 .116* --
.144* .214*
6 -0.97 ** ** -.045 -.050 --
.167** .202*
7 * -.50 -.121* .108* ** -.113* --
.288*
8 .083 -.121* .049 .022 .015 .015 ** --
.180* .209*
9 -.038 .035 .035 .045 .015 ** .071 ** --
1 .163* .170* .392* .204* .188*
0 .068 .076 .041 ** ** -.060 ** ** ** --
1 .163* .193*
1 .117* .009 .013 ** .066 -.037 .113* .117* -.027 ** --
1 .231** .185* .176* .136* .138*
2 * -.100* -.041 ** .074 -.047 ** ** .058 .098 ** --
-
1 .136* .138*
3 .50 ** -.008 ** .036 .040 .099* .122* .104* .044 .016 .057 --
-
1 .151* .188* .131* .172* .212*
4 .079 ** -.114* ** .099* -.041 .022 .044 .001 -.029 ** ** ** --
1 .282** .241* .148* .230* .217*
5 * .020 -.013 ** .090 -.054 ** .125* .101* .032 .115* ** .085 ** --
-
1 .219** .219* .159* .151* .291* .192* .216* .174* .259* .154*
6 * -.048 -.020 ** ** ** ** ** .104* ** ** ** .097 .014 ** --
1 .111** .214* .154* .146* .160* .05
7 .107* * .108* -.053 .047 * .004 ** ** .082 .123* .040 -.056 -.069 ** 6 --
Note: 1= Analytic; 2=Authentic; 3=Tone; 4=Acitve_Lexicon; 5=Weak_Lexicon; 6=Ethics; 7=Machine
Learning; 8=Artificial Intelligence; 9=Problem Solving; 10=Statistics; 11=College Graduate; 12=Rank;
13=Clearance; 14=Years Active Duty; 15=Managerial; 16=DataSciJob; 17=TotalNumJobs
Note: p-value <0.10; *p-value <0.05; **p-value <0.01; ***p-value <.001
177
Appendix B: Glassdoor Dictionary (Votto et al., 2023)
178
Appendix C: LinkedIn Dictionary
179
REFERENCES
Ainspan, N. D., & Saboe, K. N. (2021). Military Veteran Employment: A Guide for the Data-Driven
Leader. Oxford University Press.
Ammanath, B. (2022). Trustworthy AI: A Business Guide for Navigating Trust and Ethics in AI. John
Wiley & Sons.
Bakhshi, H., Downing, J. M., Osborne, M. A., & Schneider, P. (2017). THE FUTURE OF SKILLS.
Pearson.
Ballance, R. H., Forstner, H., & Murray, T. (1987). Consistency Tests of Alternative Measures of
Comparative Advantage. The Review of Economics and Statistics, 69(1), 157–161. https://doi.or
g/10.2307/1937915
Bartlett, J. E., Kotrlik, J. W., & Higgins, C. C. (n.d.). Organizational Research: Determining
Appropriate Sample Size in Survey Research.
Bergman, M., & Herd, A. (2017). Proven Leadership = College Credit: Enhancing Employability of
Transitioning Military Members Through Prior Learning Assessment. Advances in Developing
Human Resources, 19(1), 78–87. https://doi.org/10.1177/1523422316682949
Berntson, E., Näswall, K., & Sverke, M. (2008). Investigating the relationship between employability
and self-efficacy: A cross-lagged analysis. European Journal of Work and Organizational
Psychology, 17(4), 413–425. https://doi.org/10.1080/13594320801969699
Bojanowski, P., Grave, E., Joulin, A., & Mikolov, T. (2017). Enriching Word Vectors with Subword
Information. Transactions of the Association for Computational Linguistics, 5, 135–146. https://doi
.org/10.1162/tacl_a_00051
Boser, S. (2007). Power, Ethics, and the IRB: Dissonance Over Human Participant Review of
Participatory Research. Qualitative Inquiry, 13(8), 1060–1074. https://doi.org/10.1177/107780
0407308220
Boyd, R. L. (2022). The Development and Psychometric Properties of LIWC-22. Austin, TX:
University of Texas at Austin.
180
Boyd, R. L., & Pennebaker, J. W. (2015). Did Shakespeare Write Double Falsehood ? Identifying
Individuals by Creating Psychological Signatures With Text Analysis. Psychological Science, 26(5),
570–582. https://doi.org/10.1177/0956797614566658
Brumer, Y., Shapira, B., Rokach, L., & Barkan, O. (2017). Predicting Relevance Scores for Triples
from Type-Like Relations using Neural Embedding.
Bureau of Labor Statstics. (2023). Data Scientists: Occupational Outlook Handbook: : U.S. Bureau of
Labor Statistics. https://www.bls.gov/ooh/math/data-scientists.htm
Cohn, M. A., Mehl, M. R., & Pennebaker, J. W. (2004). Linguistic Markers of Psychological Change
Surrounding September 11, 2001. Psychological Science, 15(10), 687–693. https://doi.org/10.11
11/j.0956-7976.2004.00741.x
Cole, M. S., Feild, H. S., Giles, W. F., & Harris, S. G. (2004). Job Type and Recruiters’ Inferences of
Applicant Personality Drawn from Resume Biodata: Their Relationships with Hiring
Recommendations. International Journal of Selection and Assessment, 12(4), 363–367. https://doi.
org/10.1111/j.0965-075X.2004.00291.x
Cole, M. S., Rubin, R. S., Feild, H. S., & Giles, W. F. (2007). Recruiters’ Perceptions and Use of
Applicant Résumé Information: Screening the Recent Graduate. Applied Psychology, 56(2), 319–343.
https://doi.org/10.1111/j.1464-0597.2007.00288.x
Curry Hall, K., Harrell, M., Bicksler, B., Stewart, R., & Fisher, M. (2014). Veteran Employment:
Lessons from the 100,000 Jobs Mission. RAND Corporation. https://doi.org/10.7249/RR836
Deguchi, T., Seo, S., & Ishii, N. (2022). Meaning of the Clusters on Dimensionality Reduction by
Word Clustering. 2022 12th International Congress on Advanced Applied Informatics (IIAI-AAI),
325–330. https://doi.org/10.1109/IIAIAAI55812.2022.00072
Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2018). BERT: Pre-training of Deep
Bidirectional Transformers for Language Understanding. https://doi.org/10.48550/arXiv.1810.04805
Drouin, M., Boyd, R. L., Hancock, J. T., & James, A. (2017). Linguistic analysis of chat transcripts
from child predator undercover sex stings. The Journal of Forensic Psychiatry & Psychology, 28(4),
437–457. https://doi.org/10.1080/14789949.2017.1291707
Enholm, I. M., Papagiannidis, E., Mikalef, P., & Krogstie, J. (2022). Artificial Intelligence and
Business Value: A Literature Review. Information Systems Frontiers, 24(5), 1709–1734. https://
doi.org/10.1007/s10796-021-10186-w
Fossey, M., Lazier, R., Neil Lewis, M., Williamson, N., & Caddick, N. (2019). Chapter 4—Military-
to-civilian transition policies, processes, and program efforts. In C. A. Castro & S. Dursun (Eds.),
181
Military Veteran Reintegration (pp. 51–74). Academic Press. https://doi.org/10.1016/B978-0-12-
815312-3.00004-8
Fox, A. K., & Royne Stafford, M. B. (2021). Olympians on Twitter: A Linguistic Perspective of the
Role of Authenticity, Clout, and Expertise in Social Media Advertising. Journal of Current Issues &
Research in Advertising, 42(3), 294–309. https://doi.org/10.1080/10641734.2020.1763521
Fugate, M., Kinicki, A. J., & Ashforth, B. E. (2004). Employability: A psycho-social construct, its
dimensions, and applications. Journal of Vocational Behavior, 65(1), 14–38. https://doi.org/10.
1016/j.jvb.2003.10.005
Goss, K. K. (n.d.). How Post-9/11 Army Officers Describe Their Successful Transition from the
Military to Civilian Employment in the United States [Ed.D., Grand Canyon University]. Retrieved
March 23, 2023, from https://www.proquest.com/docview/2465715130/abstract/A77EF185065
5400EPQ/1.
Graham, B., & Paul, C. (2010). Does higher education really lead to higher employability and wages
in the RMI? US Census Bureau Report.
Habbat, N., Anoun, H., & Hassouni, L. (2022). Exploration, Sentiment Analysis, Topic Modeling,
and Visualization of Moroccan Twitter Data. In J. Kacprzyk, V. E. Balas, & M. Ezziyyani (Eds.),
Advanced Intelligent Systems for Sustainable Development (AI2SD’2020) (pp. 1067–1083). Springer
International Publishing. https://doi.org/10.1007/978-3-030-90639-9_87
IBM. (2020). The Data Science Skills Competency Model. IBM Corporation, 12.
Harrod, M., Miller, E. M., Henry, J., & Zivin, K. (2017). “I’ve never been able to stay in a job”: A
qualitative study of Veterans’ experiences of maintaining employment. Work, 57(2), 259–268.
https://doi.org/10.3233/WOR-172551
182
Jarrahi, M. H. (2018). Artificial intelligence and the future of work: Human-AI symbiosis in
organizational decision making. Business Horizons, 61(4), 577–586. https://doi.org/10.101
6/j.bushor.2018.03.007
Jennings, R. E., Lanaj, K., Koopman, J., & McNamara, G. (2022). Reflecting on one’s best possible
self as a leader: Implications for professional employees at work. Personnel Psychology, 75(1), 69–
90. https://doi.org/10.1111/peps.12447
Jordan, K. N., Sterling, J., Pennebaker, J. W., & Boyd, R. L. (2019). Examining long-term trends in
politics and culture through language of political leaders and cultural institutions. Proceedings of the
National Academy of Sciences, 116(9), 3476–3481. https://doi.org/10.1073/pnas.1811987116
Kacewicz, E., Pennebaker, J. W., Davis, M., Jeon, M., & Graesser, A. C. (2014). Pronoun Use
Reflects Standings in Social Hierarchies. Journal of Language and Social Psychology, 33(2), 125–
143. https://doi.org/10.1177/0261927X13502654
Kalichman, S. C., Link to external site, this link will open in a new window, & Smyth, J. M. (2021).
“And you don’t like, don’t like the way I talk”: Authenticity in the language of bruce springsteen.
Psychology of Aesthetics, Creativity, and the Arts. https://doi.org/10.1037/aca0000402
Kaur, R., & Singh, H. (2018). A Literature Review on the Employability and the Effects of Ex-
Military Personnel in Corporate Boardrooms. International Affairs and Global Strategy, 61, 50–58.
Keeling, M. E., Ozuna, S. M., Kintzle, S., & Castro, C. A. (2019). Veterans’ Civilian Employment
Experiences: Lessons Learnt From Focus Groups. Journal of Career Development, 46(6), 692–705.
https://doi.org/10.1177/0894845318776785
Kim, D. J., Song, Y. I., Braynov, S. B., & Rao, H. R. (2005). A multidimensional trust formation
model in B-to-C e-commerce: A conceptual framework and content analyses of academia/practitioner
perspectives. Decision Support Systems, 40(2), 143–165. https://doi.org/10.1016/j.dss.2004.01.006
Kulkarni, V., Bewoor, A. K., Malathi, P., & Balapgol, B. S. (2017). Employability Skill Matrix for
Engineering Graduates of Tier-II Institutes. Journal of Engineering Education Transformations,
30(3).
Kumar, P., Kumar, P., Zaidi, N., & Rathore, V. S. (2018). Analysis and Comparative Exploration of
Elastic Search, MongoDB and Hadoop Big Data Processing. In M. Pant, K. Ray, T. K. Sharma, S.
Rawat, & A. Bandyopadhyay (Eds.), Soft Computing: Theories and Applications (pp. 605–615).
Springer. https://doi.org/10.1007/978-981-10-5699-4_57
Lampka, E., & Kowalewski, S. J. (2017). Veterans in the Workplace: An Analysis of Military Veteran
Underutilization in the Civilian Workforce. 2(1).
Lehner, O. M., Ittonen, K., Silvola, H., Ström, E., & Wührleitner, A. (2022). Artificial intelligence
based decision-making in accounting and auditing: Ethical challenges and normative thinking.
183
Accounting, Auditing & Accountability Journal, 35(9), 109–135. https://doi.org/10.1108/AAAJ-09-
2020-4934
Lewis, P. V. (1985). Defining ‘business ethics’: Like nailing jello to a wall. Journal of Business
Ethics, 4(5), 377–383. https://doi.org/10.1007/BF02388590
Li, G., Yuan, C., Kamarthi, S., Moghaddam, M., & Jin, X. (2021). Data science skills and domain
knowledge requirements in the manufacturing industry: A gap analysis. Journal of Manufacturing
Systems, 60, 692–706. https://doi.org/10.1016/j.jmsy.2021.07.007
Lim, E. T. (2002). Five Trends in Presidential Rhetoric: An Analysis of Rhetoric from George
Washington to Bill Clinton. Presidential Studies Quarterly, 32(2), 328–366.
Liu, W. (2018). High-involvement human resource practices, employee learning and employability.
Career Development International, 23(3), 312–326. https://doi.org/10.1108/CDI-10-2017-0177
Lu, G., Ding, X. (David), Peng, D. X., & Hao-Chun Chuang, H. (2018). Addressing endogeneity in
operations management research: Recent developments, common problems, and directions for future
research. Journal of Operations Management, 64, 53–64. https://doi.org/10.1016/j.jom.2018.10.001
Margaret Little, B. (2011). Employability for the workers – what does this mean? Education +
Training, 53(1), 57–66. https://doi.org/10.1108/00400911111102360
Markowitz, D., Kouchaki, M., Gino, F., Hancock, J., & Boyd, R. (2022). Authentic First Impressions
Relate to Interpersonal, Social, and Entrepreneurial Success. Social Psychological & Personality
Science. https://journals-sagepub-com.libweb.lib.utsa.edu/doi/full/10.1177/19485506221086138
McAllister, C. P., Mackey, J. D., Hackney, K. J., & Perrewé, P. L. (2015). From Combat to Khakis:
An Exploratory Examination of Job Stress With Veterans. Military Psychology, 27(2), 93–107.
https://doi.org/10.1037/mil0000068
McEvoy, G. M., Hayton, J. C., Warnick, A. P., Mumford, T. V., Hanks, S. H., & Blahna, M. J.
(2005). A Competency-Based Model for Developing Human Resource Professionals. Journal of
Management Education, 29(3), 383–402. https://doi.org/10.1177/1052562904267538
McKenny, A. F., Short, J. C., & Payne, G. T. (2013). Using Computer-Aided Text Analysis to Elevate
Constructs: An Illustration Using Psychological Capital. Organizational Research Methods, 16(1),
152–184. https://doi.org/10.1177/1094428112459910
184
McLennan, W. (2021). Understanding Veteran’s Resumes and Conducting Veteran Interviews. In N.
D. Ainspan & K. N. Saboe, Military Veteran Employment: A Guide for the Data-Driven Leader.
Oxford University Press.
Memarovic, N., Langheinrich, M., Alt, F., Elhart, I., Hosio, S., & Rubegni, E. (2012). Using public
displays to stimulate passive engagement, active engagement, and discovery in public spaces.
Proceedings of the 4th Media Architecture Biennale Conference: Participation, 55–64.
https://doi.org/10.1145/2421076.2421086
Minerley, S. (2018). Mind The Gap: Unemployment Discrimination. Government Law Review, 1(11).
Minnis, S. E. (n.d.). A phenomenological exploration of combat veterans’ experiences as they
transition to civilian employment using higher education as career development [Ph.D., Texas A&M
University]. Retrieved March 6, 2023, from https://www.proquest.com/docview/1650237914/
abstract/7E5B68B3D714235PQ/1
Mökander, J., & Floridi, L. (2021). Ethics-Based Auditing to Develop Trustworthy AI. Minds and
Machines, 31(2), 323–327. https://doi.org/10.1007/s11023-021-09557-8
Monzani, D., Vergani, L., Pizzoli, S. F. M., Marton, G., & Pravettoni, G. (2021). Emotional Tone,
Analytical Thinking, and Somatosensory Processes of a Sample of Italian Tweets During the First
Phases of the COVID-19 Pandemic: Observational Study. Journal of Medical Internet Research,
23(10), e29820. https://doi.org/10.2196/29820
Muthyala, R., Wood, S., Jin, Y., Qin, Y., Gao, H., & Rai, A. (2017). Data-Driven Job Search Engine
Using Skills and Company Attribute Filters. 2017 IEEE International Conference on Data Mining
Workshops (ICDMW), 199–206. https://doi.org/10.1109/ICDMW.2017.33
Newman, M., Pennebaker, J. W., Berry, D. S., & Richards, J. M. (2003). Lying Words: Predicting
Deception from Linguistic Styles. Personality and Social Psychology Bulletin, 25(5), 665–675.
Ngoma, M., & Ntale, P. (2016). Psychological capital, career identity and graduate employability in
Uganda: The mediating role of social capital. International Journal of Training and Development,
20(2), 124–139. https://doi.org/10.1111/ijtd.12073
Nisha, S. M., & Rajasekaran, V. (2018). Employability Skills: A Review. IUP Journal of Soft Skills,
12(1), 29–37.
Niu, Y., Zhu, Y., Xu, X., & Hunter-Johnson, Y. (2022). Exploring Self-Perceived Employability and
Ambition of Student Veterans in a Higher Education Institution in the United States. Journal of
Interdisciplinary Studies in Education, 11(2), Article 2.
Norlander, P., Ho, G. C., Shih, M., Walters, D. J., & Pittinsky, T. L. (2020). The Role of
Psychological Stigmatization in Unemployment Discrimination. Basic and Applied Social
Psychology, 42(1), 29–49. https://doi.org/10.1080/01973533.2019.1689363
185
Oliveira, M., Bitencourt, C. C., Santos, A. C. M. Z. dos, & Teixeira, E. K. (2015). Thematic Content
Analysis: Is There a Difference Between the Support Provided by the MAXQDA® and NVivo®
Software Packages? Revista de Administração Da UFSM, 9(1), 72–82. https://doi.org/10.5902
/1983465911213
O*Net. (n.d.). National Center for O*Net Development: Data Scientists 15-2051.000. O*Net Online.
Retrieved June 13, 2023, from https://www.onetonline.org/link/summary/15-2051.00
Oosthuizen, R. M., Coetzee, M., & Mntonintshi, F. (2014). Investigating the relationship between
employees’ career anchors and their psychosocial employability attributes in a financial company. SA
Journal of Human Resource Management, 12(1), Article 1. https://doi.org/10.4102/sajhrm.v12i1.650
Osgood, C. E., Suci, G. J., & Tannenbaum, P. H. (1957). The Measurement of Meaning. University of
Illinois Press.
Parameswaran, S., Mukherjee, P., & Valecha, R. (2022). I Like My Anonymity: An Empirical
Investigation of the Effect of Multidimensional Review Text and Role Anonymity on Helpfulness of
Employer Reviews. Information Systems Frontiers. https://doi.org/10.1007/s10796-022-10268-3
Pennebaker, J. W., Chung, C. K., Frazee, J., Lavergne, G. M., & Beaver, D. I. (2014). When Small
Words Foretell Academic Success: The Case of College Admissions Essays. PLOS ONE, 9(12),
e115844. https://doi.org/10.1371/journal.pone.0115844
Pennington, J., Socher, R., & Manning, C. (2014). Glove: Global Vectors for Word Representation.
Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing
(EMNLP), 1532–1543. https://doi.org/10.3115/v1/D14-1162
Perkins, D. F., Davenport, K. E., Morgan, N. R., Aronson, K. R., Bleser, J. A., McCarthy, K. J., Vogt,
D., Finley, E. P., Copeland, L. A., & Gilman, C. L. (2022). The influence of employment program
components upon job attainment during a time of identity and career transition. International Journal
for Educational and Vocational Guidance. https://doi.org/10.1007/s10775-022-09527-1
Peters, M. E., Neumann, M., Iyyer, M., Gardner, M., Clark, C., Lee, K., & Zettlemoyer, L. (2018).
Deep contextualized word representations (arXiv:1802.05365). arXiv. http://arxiv.org
/abs/1802.05365
Peterson, N. G., Mumford, M. D., Borman, W. C., Jeanneret, P. R., Fleishman, E. A., Levin, K. Y.,
Campion, M. A., Mayfield, M. S., Morgeson, F. P., Pearlman, K., Gowing, M. K., Lancaster, A. R.,
Silver, M. B., & Dye, D. M. (2001). Understanding Work Using the Occupational Information
Network (o*net): Implications for Practice and Research. Personnel Psychology, 54(2), 451–492.
https://doi.org/10.1111/j.1744-6570.2001.tb00100.x
186
Puad, M. H. M., & Desa, H. M. (2020). Dissecting Perceptions of New Graduates on Work
Orientation and Self-Confidence in Employability Skills Training Program. Universal Journal of
Educational Research, 8(1A), 70–75. https://doi.org/10.13189/ujer.2020.081310
Robertson, H. C., & Brott, P. E. (2013). Male Veterans’ Perceptions of Midlife Career Transition and
Life Satisfaction: A Study of Military Men Transitioning to the Teaching Profession. Adultspan
Journal, 12(2), 66–79. https://doi.org/10.1002/j.2161-0029.2013.00016.x
Rosenberg, S., Heimler, R., & Morote, E. (2012). Basic employability skills: A triangular design
approach. Education + Training, 54(1), 7–20. https://doi.org/10.1108/00400911211198869
Ross, C. M., & Young, S. J. (2005). Resume Preferences: Is It Really “Business as Usual”? Journal of
Career Development, 32(2), 153–164. https://doi.org/10.1177/0894845305279162
Routon, P. W. (2014). The Effect of 21st Century Military Service on Civilian Labor and Educational
Outcomes. Journal of Labor Research, 35(1), 15–38. https://doi.org/10.1007/s12122-013-9170-4
Royyan, A. R., & Setiawan, E. B. (2022). Feature Expansion Word2Vec for Sentiment Analysis of
Public Policy in Twitter | Jurnal RESTI (Rekayasa Sistem dan Teknologi Informasi).
http://www.jurnal.iaii.or.id/index.php/RESTI/article/view/3525
Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and
use interpretable models instead. Nature Machine Intelligence, 1(5), 206–215. https://doi.org/10.1038
/s42256-019-0048-x
Rutter, C. M. (n.d.). Military Veteran Transition to the Civilian Workforce: A Descriptive Study
[Ph.D., Grand Canyon University]. Retrieved March 23, 2023, from https://www.proquest.com
/docview/2465726378/abstract/C4BE6605BC164710PQ/1
Shah, N., Willick, D., & Mago, V. (2022). A framework for social media data analytics using
Elasticsearch and Kibana. Wireless Networks, 28(3), 1179–1187. https://doi.org/10.1007/s11276-
018-01896-2
Smith, P. A. C., & Sharma, M. (2002). Developing personal responsibility and leadership traits in all
your employees: Part 1 – shaping and harmonizing the high‐performance drivers. Management
Decision, 40(8), 764–774. https://doi.org/10.1108/00251740210441018
Squicciarini, M., & Nachtigall, H. (2021). Demand for AI skills in jobs: Evidence from online job
postings. https://doi.org/10.1787/3ed32d94-en
Stoklasa, J., Talášek, T., & Stoklasová, J. (2019). Semantic differential for the twenty-first century:
Scale relevance and uncertainty entering the semantic space. Quality & Quantity, 53(1), 435–448.
https://doi.org/10.1007/s11135-018-0762-1
Stone, C. B. (2016). The veteran myth: An experimental investigation of human resource managers’
perceptions of United States military veterans [Ph.D., The University of Texas at San Antonio]. In
187
ProQuest Dissertations and Theses.
https://www.proquest.com/docview/1793670665/abstract/78224DDD49F844C2PQ/1
Stone, P. J., & Hunt, E. B. (1963). A computer approach to content analysis: Studies using the
General Inquirer system. Proceedings of the May 21-23, 1963, Spring Joint Computer Conference,
241–256. https://doi.org/10.1145/1461551.1461583
Taherdoost, H. (2018). How to Design and Create an Effective Survey/Questionnaire; A Step by Step
Guide (SSRN Scholarly Paper No. 3224226). https://papers.ssrn.com/abstract=3224226
Valecha, R., Srinivasan, S. K., Volety, T., Kwon, K. H., Agrawal, M., & Rao, H. R. (2021). Fake
News Sharing: An Investigation of Threat and Coping Cues in the Context of the Zika Virus. Digital
Threats: Research and Practice, 2(2), 1–16. https://doi.org/10.1145/3410025
Valli, R. (2016). Creating a questionnaire for a scientific study. International Journal of Research
Studies in Education, 6(4). https://doi.org/10.5861/ijrse.2016.1584
Venkatesan, S., Valecha, R., Yaraghi, N., Oh, O., & Rao, H. R. (2021). Influence in Social Media: An
Investigation of Tweets Spanning the 2011 Egyptian Revolution. MIS Quarterly, 45(4), 1679–1714.
https://doi.org/10.25300/MISQ/2021/15297
Votto, A. M., Manuel, D., Valecha, R., Keeton, K., Rao, H. R. (2023). JC-Compass: A Framework
for Conducting Competency-Based Job Posting Research and Analysis. Manuscript in preparation.
Votto, A. M., Valecha, R., Najafirad, P., & Rao, H. R. (2021). Artificial Intelligence in Tactical
Human Resource Management: A Systematic Literature Review. International Journal of
Information Management Data Insights, 1(2), 100047. https://doi.org/10.1016/j.jjimei.2021.100047
188
VITA
Captain Alexis “Ali” Votto joined the UTSA Ph.D. program in Information Technology
as a student in August of 2020. She is one of the few Active-Duty members serving in the Air
Force to be selected for the highly competitive Secretary of the Air Force STEM Ph.D. program.
This program allows Active-Duty officers to attend a civilian institution in pursuit of a STEM-
related Ph.D. as their primary job for three years. She is a Force Support Officer (FSO) by trade,
which translates to Human Resource Management, Mortuary Affairs, and Manpower analytics as
her profession.
Having earned her Bachelor of Science in Mathematics and Master of Science in Human
Resource Management, she sought to marry these two academic backgrounds within her Ph.D. to
see how she can enhance the HRM profession and potential processes within the USAF. To her
knowledge, she is one of the very few FSOs in her career field to achieve a Ph.D. within a STEM
field. She hopes this achievement inspires others to compete for programs to further explore how
STEM exists in business management processes. Prior to joining the Ph.D. program, she served
in various leadership positions ranging from Section Command to Deputy Flight Command. She
has had administrative oversight for over 1,600 personnel as an executive officer and was
privileged with G-Series orders to issue administrative actions on behalf of the commander as a
section commander.
On a personal note, she calls San Antonio her home. This is the city where her career
started and where she hopes to return to once it’s time to retire from the USAF. Future endeavors
include serving at the Air Force Personnel Center at Randolph Air Force Base as a personnel
analyst. She is happily married to Joseph Votto and they have three fur-children, Koda, Hank,
and Butters.
ProQuest Number: 30575705
This work may be used in accordance with the terms of the Creative Commons license
or other rights statement, as indicated in the copyright statement or in the metadata
associated with this work. Unless otherwise specified in the copyright statement
or the metadata, all rights are reserved by the copyright holder.
ProQuest LLC
789 East Eisenhower Parkway
P.O. Box 1346
Ann Arbor, MI 48106 - 1346 USA