s40685-020-00134-w 2
s40685-020-00134-w 2
https://doi.org/10.1007/s40685-020-00134-w
ORIGINAL RESEARCH
Received: 15 October 2019 / Accepted: 1 November 2020 / Published online: 20 November 2020
The Author(s) 2020
1 Introduction
123
796 Business Research (2020) 13:795–848
123
Business Research (2020) 13:795–848 797
123
798 Business Research (2020) 13:795–848
123
Business Research (2020) 13:795–848 799
data (Canhoto and Clear 2020). Finally, reinforcement learning, as a separate group
of methods, is not based on fixed input/output data. Instead, the ML algorithm learns
behavior through trial-and-error interactions with a dynamic environment (Kael-
bling et al. 1996).
Furthermore, instead of grouping ML models as supervised, unsupervised, or
reinforcement type learning, the methodologies of algorithms may also be used to
categorize ML models. Examples are probabilistic models, which may be used in
supervised or unsupervised settings (Murphy 2012), or deep learning models (Lee
and Shin 2020), which rely on artificial neural networks and perform complex
learning tasks. In supervised settings, neural network models often determine the
relationship between input and output using network structures containing the so-
called hidden layers, meaning phases of transformation of the input data. Single
nodes of these layers (neurons) were first modeled after neurons in the human brain,
and they resemble human thinking (Bengio et al. 2017). In other settings, deep
learning may be used, for instance, to (1) process information through multiple
stages of nonlinear transformation; or (2) determine features, representations of the
data providing an advantage for, e.g., prediction tasks (Deng and Yu 2014).
For any estimation Yb of a random variable Y, bias refers to the difference between
the expected values of Yb and Y and is also referred to as systematic error
(Kauermann and Kuechenhoff 2010; Goodfellow et al. 2016). Cognitive biases,
specifically, are systematic errors in human judgment when dealing with uncertainty
(Kahneman et al. 1982). These cognitive biases are thought to be transferred to
algorithmic evaluations or predictions, where bias may refer to ‘‘computer systems
that systematically and unfairly discriminate against certain individuals or groups in
favor of others’’ (Friedman and Nissenbaum 1996, p. 332).
Algorithms are often characterized as ‘‘black box’’. In the context of HRM,
Cheng and Hackett (2019) characterize algorithms as ‘‘glass boxes’’, since some,
but not all, components of the theory are reflective. In this context, the consideration
and distinction of the three core elements are necessary, namely, transparency,
interpretability, and explainability (Roscher et al. 2020). Transparency is concerned
with the ML approach, while interpretability is concerned with the ML model in
combination with the data, which means the making sense of the obtained ML
model (Roscher et al. 2020). Finally, explainability comprises the model, the data,
and human involvement (Roscher et al. 2020). Concerning the former, transparency
can be distinguished at three different levels: ‘‘[…] at the level of the entire model
(simulatability), at the level of individual components, such as parameters
(decomposability), and at the level of the training (algorithmic transparency)’’
(Roscher et al. 2020, p. 4). Interpretability concerns the characteristics of an ML
model that need to be understood by a human (Roscher et al. 2020). Finally, the
element of explainability is paramount in HRM. Contextual information of human
and their knowledge from the domain of HRM are necessary to explain the different
sets of interpretations and derive conclusions about the results of the algorithms
123
800 Business Research (2020) 13:795–848
123
Business Research (2020) 13:795–848 801
Macready (1997)]. Here, the choice of the benchmark or rather the value indicating
the performance of the model is optimized through rotations of different
representations of the data and methods for prediction. For example, representative
bias might occur if females in comparison to males are underrepresented in the
training data of an algorithm. Hence, the outcome could be in favor of the
overrepresented group (i.e., males) and, hence, lead to discriminatory outcomes.
Technical bias may arise from technical constraints or technical consideration for
several reasons. For example, technical bias can originate from limited ‘‘[…]
computer technology, including hardware, software, and peripherals’’ (Friedman
and Nissenbaum 1996, p. 334). Another reason could be a decontextualized
algorithm that does not manage to treat all groups fairly under all important
conditions (Friedman and Nissenbaum 1996; Bozdag 2013). The formalization of
human constructs to computers can be another problem leading to technical bias.
Human constructs, such as judgments or intuitions, are often hard to quantify, which
makes it difficult or even impossible to translate them to the computer (Friedman
and Nissenbaum 1996). As an example, the human interpretation of law can be
ambiguous and highly dependent on the specific context, making it difficult for an
algorithmic system to correctly advise in litigation (c.f., Friedman and Nissenbaum
1996).
In the context of real users, emergent bias may arise. Typically, this bias occurs
after the construction as a result of changed societal knowledge, population, or
cultural values (Friedman and Nissenbaum 1996). Consequently, a shift in the
context of use might yield to problems and an emergent bias due to two reasons,
namely ‘‘new societal knowledge’’ and ‘‘mismatch between users and system
design’’ (see Table 1 in Friedman and Nissenbaum 1996, p. 335). If it is not possible
to incorporate new knowledge in society into the system design, emergent bias due
to new societal knowledge occurs. The mismatch between users and system design
can occur due to changes in state-of-the-art-research or due to different values. Also,
emergent bias can occur if a population uses the system with different values than
those assumed in the design process (Friedman and Nissenbaum 1996). Problems
occur, for example, when users originate from a cultural context that avoids
competition and promotes cooperative efforts, while the algorithm is trained to
reward individualistic and competitive behavior (Friedman and Nissenbaum 1996).
123
802 Business Research (2020) 13:795–848
Y 2 f0; 1g is a random variable describing, e.g., the recidivism of a subject, Yb its estimator and G 2
f0; 1g; describes whether a subject is a member of a certain protected group (G ¼ 1Þ or not ðG ¼ 0Þ
123
Business Research (2020) 13:795–848 803
likelihood of conscientious behavior and altruisms is higher for employees who feel
treated fairly (Cohen-Charash and Spector 2001). Conversely, unfairness can have
considerable adverse consequences. For example, in the recruitment context,
fairness perceptions of candidates during the selection process have important
consequences for decision to stay in the applicant pool or accept a job offer (Bauer
et al. 2001). Therefore, it is crucial to know how people feel about algorithmic
decision-making taking over managerial decisions formerly made by humans, since
the fairness perceptions during the recruitment process and/or training process have
essential and meaningful effects on attitudes, performance, morale, intentions, and
behavior (e.g., the acceptance or rejection of a job offer or job turnover, job
dissatisfaction, and reduction or elimination of conflicts) (Gilliland 1993; McCarthy
et al. 2017; Hausknecht et al. 2004; Cropanzano et al. 2007; Cohen-Charash and
Spector 2001). Moreover, negative experiences might damage the employers
image. Several online platforms offer the possibility of rating companies and their
recruitment and development process (Van Hoye 2013; Woods et al. 2020).
Considering justice and fairness in the organizational context (Gilliland 1993),
there are three core dimensions of justice: distributive, procedural, and interactional.
The three dimensions tend to be correlated. Distributive justice deals with the
outcome that some humans receive and some do not (Cropanzano et al. 2007). Rules
that can lead to distributive justice are ‘‘[…] equality (to each the same), equity (to
each in accordance with contributions, and need (to each in accordance with the
most urgency)’’ (Cropanzano et al. 2007, p. 37). To some extent, especially
concerning equity, this can be connected with individual fairness and group fairness
from Dwork et al. (2012) and equal opportunities from Hardt et al. (2016).
Procedural justice means that the process is consistent with all humans, not
including bias, accurate, and consistent with the ethical norms (Cropanzano et al.
2007; Leventhal 1980). Consistency plays an essential role in procedural justice,
meaning that all employees and all candidates need to receive the same treatment.
Additionally, the lack of bias, accuracy, representation of all parties, correction, and
ethics play an important role in achieving a high procedural justice (Cropanzano
et al. 2007). In contrast, interactional justice is about the treatment of humans,
meaning the appropriateness of the treatment from another member of the company,
the treatment with dignity, courtesy, and respect, and informational justice (share of
relevant information) (Cropanzano et al. 2007).
In general, algorithmic decision-making increases the standardization of
procedures, so that decisions should be more objective and less biased, and errors
should occur less frequently (Kaibel et al. 2019), since information processing by
human raters can be unsystematic, leading to contradictory and insufficient
evidence-based decisions (Woods et al. 2020). Consequently, procedural justice and
distributive justice are higher using algorithmic decision-making, because the
process is more standardized, which still not means that it is without bias.
However, especially in the context of an application or an employee evaluation, it
is not only about how fair the procedure itself is (according to fairness measures),
but it is also about how people involved in the decision process perceive the fairness
of the whole process. Often the personal contact, which characterizes the
123
804 Business Research (2020) 13:795–848
3 Methods
123
Business Research (2020) 13:795–848 805
insights (Siddaway et al. 2019; Lipsey and Wilson 2001; Ferguson and Brannick
2012). In April 2020, this search approach resulted in 3207 articles.
Following this initial identification, we manually screened each article (title and
abstract) to evaluate whether its content was fundamental relevant to impact bias,
discrimination, or fairness of algorithmic decision-making in HRM, especially in
recruitment, selection, development, and training in particular. The process of
123
806 Business Research (2020) 13:795–848
123
Business Research (2020) 13:795–848 807
Identification
Records identified through Additional records identified
database searching through other sources
(n = 3,136) (n = 71)
Studies included in
literature review
Included
(n = 36)
Fig. 1 PRISMA flow diagram illustrating the process. aTopic did not fit, mostly no HR and/or fairness,
no obvious discrimination context, bMostly no HR and/or fairness, no discrimination context after
reading the full text or not meeting the inclusion criteria
regarding the discrimination potential and fairness when using algorithmic decision-
making in HR recruitment and HR development.
This approach is not without limitations. First, the reliance on two databases might
be regarded as a limitation; however, the approach of selecting two broad and
common databases contributed to the validity and replicability of our findings due to
the extensive coverage of high-impact, peer-reviewed journals in these databases
(Podsakoff et al. 2005). Second, our review focused on two essential HR functions
that have severe consequences for individuals and society concerning ethics, namely
HR recruitment and HR development. We did not consider other areas of HRM,
since the focus of other HR functions is mainly the automation process (e.g., pay or
another administrative task). Thus, the situation is different in HR recruitment and
HR development, because societal decisions are made, which have crucial
consequences for the individual applicants and employees, such as job offer or
promotion opportunities. Especially when it comes to decisions about individuals
123
808 Business Research (2020) 13:795–848
and their potential, objective and perceived fairness is paramount (Ötting and Maier
2018; Lee 2018).
Moreover, only articles written in the English-language were part of the literature
review. Even though this procedure is accepted practice and there is some evidence
that including only English articles does not bias the results, it should be noted that
non-English articles were not included because English is the dominant language in
research (Morrison et al. 2012).
4 Descriptive results
The following section shows the current research landscape. We summarize the
main characteristics of the identified articles in Table 3 and present the main
findings in Table 4. This table reports the name of authors, year of publication, the
main focus of the study (i.e., focus on bias, discrimination, fairness, or perceived
fairness), applied method, the field of research, algorithmic decision-making
system, HR context (i.e., recruitment- distinguished between recruitment and
selection- or development), and the key findings. We analyze the main focus and the
key findings of the studies in the following sections. The table is sorted by the focus
of the article and whether it is on bias as a trigger for unfairness and discrimination
or specifically on fairness and discrimination.
Figure 2 illustrates the distribution of publications over time and the research
methods used. The first identified article in our sample of literature was published in
2014. From 2014 to 2016, only a few articles are published per year. From 2017,
interest in algorithmic decision-making and discrimination increased notably. As
shown in Fig. 2, there was enormous interest in the topic in 2019.
From a methodological perspective, another noteworthy result of this systematic
review is the predominance of non-empirical evidence, as Table 3 and Fig. 2 show
that the large majority of articles are non-empirical (i.e., conceptual paper, reviews,
and case studies). A reason for this is that scientific investigation of discrimination
by algorithmic decision-making represents a relatively new topic. However, the
number of quantitative papers increased from 2018. Most of the studies focused on
bias, discrimination, and objective fairness, while 12 studies examined perceived
fairness perceptions of applicants and employees (see Table 1). Furthermore, the
majority of studies are located in the area of recruitment and selection, whereby
these studies mostly focus on selection. Twelve studies are located in the area of HR
development. The majority of studies provided either no geographical specification
or were conducted in the USA (see Table 3).
Thirteen articles originate from management, and fourteen articles originate
from computer science, four articles originate from law, two from psychology, two
from information systems, and one from the behavioral sciences. This distribution
illustrates that the field does not have a core in business and management research
and is rather interdisciplinary. Nevertheless, the majority of articles originating from
management were published in high-ranked journals, such as Journal of Business
Ethics, Human Resource Management Review, Management Science, Academy of
Management Annals, and Journal of Management. The majority of these studies
123
Table 3 Overview of studies
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
development,
training
Naim et al. (2016) The automated B Empirical- Computer NLP, FEP Selection Recommends to speak USA
analysis of facial quantitative; science more fluently, use less
expressions, analysis of filler words, and smile
Business Research (2020) 13:795–848
123
Table 3 continued
810
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
123
development,
training
Mann and O’Neil Explains why B, D Non-empirical; Management General, hiring Selection Algorithms reflect USA
(2016) algorithms are conceptual algorithms human biases and
not neutral and paper prejudices that lead to
offers some machine learning
implications to mistakes and
reduce the risk of misinterpretations
biases of Bias and prejudice
algorithmic cannot be completely
decision-making eliminated from hiring
in the hiring
process HR professionals must
consider the
consequences of these
systems and ensure
they always reflect the
best human intentions
Kim (2017) Examines the use of B, D Non-empirical; Law General, Recruitment Because of the nature of USA
classification conceptual classification and data mining
schemes/data paper selection, techniques, employer
algorithms in training and reliance on these tools
terms of development poses novel
personnel challenges to
decisions workplace equality
Shows limitations in
existing law and
improvement
proposals
Business Research (2020) 13:795–848
Table 3 continued
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
development,
training
Rosenblat et al. Examines B, D Empirical- Computer Evaluation Development The need to exercise Not
(2016) discrimination qualitative; science systems quality control over a specified
thorough case study large disaggregated
evaluation in the workforce may permit
case of Uber the continued use of
drivers rating
Business Research (2020) 13:795–848
123
Table 3 continued
812
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
123
development,
training
Williams et al. Examines B, D Empirical- Information General Recruitment Algorithmic prediction USA
(2018) discrimination qualitative; and can include injustices
through the use of multiple case selection The predictions have to
algorithms in studies be checked for bias
decision-making and should be
processes corrected to avoid
Proposes strategies discrimination
for the prevention
of discrimination
Lambrecht and Examines gender B, D Empirical- Management Recommender Recruitment Fewer women saw 191
Tucker (2019) bias in delivery of quantitative; systems STEM ads than men countries
job ads field test An algorithm that
Conducts field test simply optimizes cost-
of how an effectiveness in ad
algorithm delivery will deliver
delivered ads ads that were intended
promoting job to be gender neutral in
opportunities in an apparently
the science, discriminatory way
technology,
engineering and
math fields
Business Research (2020) 13:795–848
Table 3 continued
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
development,
training
Sajjadiani et al. Prediction of future B, F Empirical- Psychology General Selection Algorithmic methods are USA
(2019) work outcomes quantitative, often calibrated
such as voluntary statistical exclusively to a
turnover, tests; specific application
involuntary machine pool
turnover, and learning Pairing established
Business Research (2020) 13:795–848
123
Table 3 continued
814
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
123
development,
training
Raghavan et al. Documents and B, F Empirical- Computer Employment Selection Target variables and Several
(2020) analyzes the qualitative; science assessment training data: most of countries
claims and case study the vendors offer
practices of customizable
companies assessment
offering Validation: vendor’s
algorithms for websites often do not
employment clarify whether they
assessment validate their models
Sánchez- Examines how three B, F Empirical- Computer Employment Selection Often lack of UK
Monedero et al. automated hiring qualitative; science assessment information on how
(2020) systems can be case study the system worked
understand and Claims and validation
attempt to are often vague
mitigate bias and
discrimination in
the UK
Business Research (2020) 13:795–848
Table 3 continued
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
development,
training
Stone et al. (2015) Reviews several of B, PF Non-empirical; Management Several electronic Recruitment There are still a number USA
the primary conceptual HRM tools and of questions about
forces that are paper selection, whether these new
presenting development systems enable
challenges for organizations to
HR research and achieve their primary
Business Research (2020) 13:795–848
practice HR goals
Tries to answer the
question whether
eHRM influences
organizational
effectiveness and
whether it enables
organizations to
achieve their HR
goals
Woodruff et al. Explores how B, PF Empirical- Computer General Recruitment Concept of algorithmic USA
(2018) members of qualitative; science and fairness is largely
potentially workshop; selection unfamiliar
affected interviews Learning about
communities in algorithmic
the USA feel (un)fairness elicited
about algorithmic negative feelings
fairness
Company handling of
algorithmic fairness
interacts significantly
with user trust
815
123
Table 3 continued
816
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
123
development,
training
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
development,
training
Rosenblat et al. Focuses on the new B, D, F Non-empirical; Computer General Recruitment Employers potentially Not
(2014) tool’s employers literature science and have access to more specified
use to sift through review selection comprehensive
job applications electronic profiles on
Addresses issues of job candidates than
privacy, fairness, has been traditionally
Business Research (2020) 13:795–848
123
Table 3 continued
818
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
123
development,
training
Persson (2016) Analyzes the B, D, F Empirical- Computer Data mining, Recruitment Employers might miss Not
problems of qualitative; science profiling and the best candidates, as specified
implicit bias in explorative selection the employed
algorithms analysis algorithms are tuned
regarding with limited and
recruitment outdated data
processes, The risk of directly or
discrimination, indirectly
and unfairness by discriminating
using algorithms candidates exists
Shows possible
solutions
Vasconcelos et al. Proposes a B, D, F Non-empirical; Computer General, hiring Selection Points out connections Not
(2017) structured conceptual science algorithms between bias of AI specified
approach to paper and the problem of
mitigate induction
discrimination Shows that there is a
and unfairness logical theory of
caused by bias in preferences
AI systems in
hiring decision
scenarios
Business Research (2020) 13:795–848
Table 3 continued
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
development,
training
Chen et al. (2018) Investigates gender- B, D, F Empirical- Computer Search engines Recruitment Individual fairness: even USA
based inequalities quantitative; science when controlling for
in the context of statistical all other visible
resume search tests candidate features,
engines, there is a slight
discrimination penalty against female
Business Research (2020) 13:795–848
candidates
Group fairness:
8.5–13.2% of job title/
city pairs show
statistically significant
group unfairness
Bogen (2019) Analysis of B, D, F Non-empirical; Management General, hiring Recruitment Most hiring algorithms USA
predictive tools conceptual algorithms and will drift toward bias
across the hiring paper selection by default
process to clarify Potential to help reduce
what hiring interpersonal bias
algorithms do, should not be
and where and discounted
how bias can
enter into the Only tools that
process proactively tackle
deeper disparities will
offer any hope that
predictive technology
can help promote
equity, rather than
erode it
819
123
Table 3 continued
820
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
123
development,
training
Simbeck (2019) Discusses the B, D, F Non-empirical; Management General Recruitment The rising data on Not
ethical conceptual and employees will specified
implications of paper selection, Lead to an
the application of development unprecedented
sophisticated transparency of
analytical employees
methods to
questions in HR Proposes to transfer key
management ethical concepts from
medical research,
artificial intelligence,
learning analytics, and
coaching to HR
analytics
Business Research (2020) 13:795–848
Table 3 continued
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
development,
training
Kellogg et al. Analyzes how the B, D, F Non-empirical; Management Algorithmic Development Algorithmic control in Not
(2020) implementation literature recommending, the workplace specified
of algorithmic review restricting, operates through six
technologies in recording, main mechanisms:
organizations rating, restricting,
may reshape replacing, and recommending,
Business Research (2020) 13:795–848
123
Table 3 continued
822
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
123
development,
training
Chamorro- Reviews three D Non-empirical; Behavioral Machine-learning Recruitment Talent analytics Not
Premuzic et al. innovations that conceptual science algorithms, and represent a more data- specified
(2017) have the potential paper social sensing selection, driven and evidence-
to revolutionize technology, development based approach than
the way gamified human intuition
organizations Assessment tools
identify, develop,
and engage talent
and are emerging
as tools used by
practitioners and
firms
Kim and Scott Examines D Empirical- Law Recommender Recruitment It is not certain whether Not
(2018) discrimination in qualitative systems current law is specified
the online adequate to reach all
application forms of targeted
process Recruitment with
Develops significant
mechanisms to discriminatory effects
check when
discrimination
occurs in online
recruitment
Business Research (2020) 13:795–848
Table 3 continued
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
development,
training
Cappelli (2019) Presents problems D Non-empirical; Management General Selection Possible solution might USA
which occur illustrative be tracking the
because of the case percentage of
rise of data openings filled from
science in hiring within, requiring that
processes by all openings be posted
Business Research (2020) 13:795–848
123
Table 3 continued
824
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
123
development,
training
Lee et al. (2015) Explores the impact F, PF Empirical- Computer Evaluation process Development Highlights how USA
of algorithmic, qualitative; science transparency of
data-driven semi- algorithmic
management in structured assignment influences
the context of interviews; employee’s
new ride sharing post analysis cooperation, work
services strategy, and
Highlights workaround creation
opportunities and These numeric systems
challenges in that made drivers
designing human- accountable for all
centered interactions were
algorithmic work sometimes seen as
assignment, unfair and ineffective
information, and and created negative
evaluation psychological feelings
in drivers
Business Research (2020) 13:795–848
Table 3 continued
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
development,
training
McCarthy et al. Comprehensive PF Non-empirical; Management Several selection Selection The field of applicant Not
(2017) review of narrative tools reactions has specified
research on review advanced
applicant considerably and
reactions to made substantial and
selection meaningful
Business Research (2020) 13:795–848
procedures contributions to
published since theory, methods, and
2000 (including practice of
algorithmic based recruitment and
selection tools) selection over the last
15 years
Langer et al. Analyzes the role of PF Empirical- Computer Several selection Selection Computer experience Not
(2018) computer quantitative; science tools did not affect the specified
experience and quasi- candidate reaction
information on experimental
applicant reaction design
towards novel
technologies for
personnel
selection
825
123
Table 3 continued
826
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
123
development,
training
Lee (2018) Explores the PF Empirical- Computer General Selection, People perceive USA
perceptions of quantitative; science development algorithmic decisions
algorithmic experimental as less trustworthy
management by design and more likely to
conducting an evoke negative
online emotion for tasks that
experiment using people think require
four managerial uniquely human skills
decisions
The decision-maker
(algorithmic or
human) was
manipulated, and
measured
perceived
fairness, trust, and
emotional
response
Ötting and Maier Examines the PF Empirical- Computer General Development Significant effects of Germany
(2018) effects of quantitative; science procedural justice on
procedural justice experimental employee behavior
and the type of design and attitudes,
decision agent on confirming the
employee importance of
behavior and procedural justice at
attitudes the workplace for both
human and system
decision agents
Business Research (2020) 13:795–848
Table 3 continued
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
development,
training
Kaibel et al. Analysis of the PF Empirical- Management Automated Selection Applicants perceived the Germany,
(2019) perception of quantitative; screening algorithmic selection USA
algorithm-based Experimental algorithm, as more consistent,
decisions versus design digital interview less personable
human-based The organizational
decisions in attractiveness is lower
Business Research (2020) 13:795–848
123
Table 3 continued
828
Author(s), year Main focus B/D/F/PF Method Field of System Recruitment, Key findings Geography
research selection,
123
development,
training
Leclercq- Investigates the Not Non-empirical; Management General Development Ethical issues are the France
Vandelannoitte moral and ethical specified case study ambivalence of the
(2017) implications of use of ubiquitous IT at
emerging forms work, subtlety of the
of control that control exerted by
have developed ubiquitous IT,
along with the invasiveness of
use of modern ubiquitous IT, and
ubiquitous self-reinforcement of
information ubiquitous IT-based
technology in the control
workplace
Recruitment Selection
CV curriculum vitae, FEP facial expression processing, NLP natural language processing, ML machine
learning
123
830 Business Research (2020) 13:795–848
14
12
10
0
2014 2015 2016 2017 2018 2019 2020
Fig. 2 Distribution of publications over time and research methods. Data on 2020 research articles are
based on our database search until April 2020
were published in 2019, which stresses the importance of fairness and discrimi-
nation as a recent topic in the management and HRM literature.
Our results suggest there is still room for academic researchers to complement
the literature and discussion on algorithmic decision-making and fairness. In the
following, we introduce some algorithmic decision tools used in HR recruitment and
HR development and their potential for discrimination.
5.1 HR recruitment
123
Business Research (2020) 13:795–848 831
multifaceted concept, not only the users (here: job seekers) need to be considered,
but also stakeholders (Burke et al. 2018). Hiring platforms, such as Xing and
LinkedIn, already implement predictive analytics. Their algorithms go through
thousands of job profiles to find the most eligible candidate for a specific job and
recommend this candidate to the recruiter (Carey and Smith 2016). Firms also
examine data about job seekers, analyze them based on past hiring decisions, and
then recommend only the applications that are a potential match (Kim 2016).
Consequently, firms can more precisely target potential candidates. These predic-
tions based on past decisions can unintentionally lead to companies using job
advertisements that strengthen gender and racial stereotypes, because if, for
example, in the past, more males were selected for high position jobs, the
advertisement is consequently shown to more males (historical bias). Thus, tension
exists between the goals of fairness and those of personalization (Burke et al. 2018).
In a non-empirical paper analyzing predictive tools in USA, Bogen (2019) gives
a prime example of algorithmic discrimination against other genders by demon-
strating that algorithms extrapolate based on patterns of the provided data. Thus, if
recruiters contacted males more frequently than females, the recommendation will
be to show job ads more often to males. An explanation could be that males are
more likely to click on high-paying job ads, and consequently, the algorithm learns
from this behavior (Burke et al. 2018).
Another example showed that targeted ads on Facebook were predominately
shown to females (85%), while jobs advertised by taxi companies were shown
mainly to males (Bogen 2019). In their field test of how an algorithm delivered ads
promoting job opportunities in the STEM fields, Lambrecht and Tucker (2019)
found in an empirical-quantitative field test among 191 countries that online job
advertisements in the science, technology, engineering, and math sector were more
likely shown to males than females. This gender bias in the delivery of job ads
occurs, because even if the job advertisement should be delivered explicitly gender
neutral, an algorithm that optimizes cost-effectiveness in ad delivery would deliver
ads discriminatorily due to crowding out (Lambrecht and Tucker 2019).
Platforms, such as Google, LinkedIn, and Facebook, offer advertisers the
possibility to target viewers based on sensitive attributes to exclude some job
seekers depending on their attributes (Kim and Scott 2018). For instance, Facebook
let firms choose among over 100 well-defined attributes (Ali et al. 2019). In this
case, humans interact and determine the output strategically (intentional discrim-
ination). For example, through their selection of personal traits, older potential
candidates are excluded from seeing the job advertisement. Companies make use of
targeted ads to attract job seekers who are most likely to have relevant skills, while
recommender systems can reject a large proportion of applicants (Kim and Scott
2018). Even if companies chose their viewers by relying on attributes that appear to
be neutral, these attributes can be closely related to protected traits, such as
ethnicity, and could allow biased targeting. Often, bias in recommender systems can
occur unintentionally and rely on attributes that are not obvious (Kim and Scott
2018). Kim and Scott (2018) analyzed in an empirical-qualitative paper that due to
spillover effects, it is more costly to serve ads to young females, because women on
Facebook are known to be more likely to click on ads (Kim and Scott 2018). Hence,
123
832 Business Research (2020) 13:795–848
algorithms that optimize cost efficiency may deliver ads more often to males,
because they are less expensive than females (Kim and Scott 2018). In summary,
these three studies based on non-empirical, empirical-qualitative, and empirical-
quantitative evidence show that historical biases and biases caused by cost-
effectiveness reasons occur in HR recruitment and selection.
With the help of search engines, recruiters proactively search for candidates who
use employment services on keywords and filters (Chen et al. 2018). The algorithm
rates applicants; consequently, the recruiter sees and more likely clicks on those at
the top. These rankings often take demographic features (e.g., name, age, country,
and education level) into account, and this can yield a disadvantage for some
candidates (Bozdag 2013; Chen et al. 2018). Other features are, for example, the
locations, previous search keywords, and the recent contacts in a user’s social
network. These service sites do now allow recruiters to filter search results by
demographics (e.g., gender, age, and ethnicity). Nonetheless, these variables exist
indirectly in other variables, such as years of experience as an indicator of age
(Chen et al. 2018). With the help of statistical tests and data on 855,000 USA job
candidates (search results for 35 job titles across 20 USA cities), Chen et al. (2018)
revealed in an empirical-qualitative single case study and review that the search
engines provided by Indeed, Monster, and CareerBuilder discriminate against
female candidates to a lesser extent.
5.2 HR selection
Striving for more efficiency due to time and cost pressures and limited resources by
simultaneously managing a large number of applications are among the main
reasons for the increasing use of algorithmic decision-making in the selection
context (Leicht-Deobald et al. 2019). Organizations are increasingly using
algorithmic decision tools, such as CV and résumé screening, telephone, or video
interviews, providing an algorithmic evaluation (Lee and Baykal 2017; Mann and
O’Neil 2016) before conducting face-to-face interviews (Chamorro-Premuzic et al.
2016; van Esch et al. 2019).
One possibility for using algorithmic decision-making in selection is the analysis
of the CV and résumé, with candidates entering their CVs or job preferences online,
and this information is subject to algorithmic analysis (Savage and Bales 2017).
Yarger et al. (2019) conceptually analyzed the fairness of talent acquisition software
in the USA and its potential to promote fairness in the selection process for
underrepresented IT professionals. The authors argue that it is necessary to audit
algorithms, because they are not neutral. One prominent example is the CV
screening tool of Amazon, which was trained on biased historical data that led to a
preference for male candidates based on the fact that, in the past, Amazon hired
more often males as software engineers as females and the algorithm has been
trained based on these data (historical bias) (Dastin 2018). Yarger et al. (2019)
suggest removing sources of human bias such as gender, race, ethnicity, religion,
sexual orientation, age, and information that can indicate membership in a protected
class. Text mining is often the foundation for the screening of CVs and résumés, an
approach to characterize and transform text using the words themselves as the unit
123
Business Research (2020) 13:795–848 833
123
834 Business Research (2020) 13:795–848
Turning the perspective from the employer to the candidates, especially the
perceived fairness of the candidates, plays an essential role in recruitment outcomes
(Gilliland 1993). Using a between-subject online experiment, Lee (2018) discovered
that people perceive human decisions to be fairer than algorithmic decisions in
hiring tasks. People think that the algorithm lacks the ability to discern
suitable applicants, because the algorithm makes judgments based on keywords
and does not take qualities that are hard to quantify into account. Participants do not
trust the algorithm, because it lacks human judgment and human intuition.
Contrasting findings are found in Suen et al.’s (2019) empirical-quantitative study
comparing synchronous videos to asynchronous videos analyzed by means of an AI;
they conclude that the videos analyzed by means of an AI did not negatively
influence perceived fairness in their Chinese sample.
Unlike the other studies, in an online experiment, Kaibel et al. (2019) recently
analyzed the perceived fairness of two different algorithmic decision tools, namely
initial screening and digital interviews. Results show that algorithmic decision-
making negatively affects personableness and the opportunity to perform during the
selection process, but it does not affect the perceived consistency. These
relationships are moderated by personal uniqueness and experienced discrimination.
5.3 HR development
123
Business Research (2020) 13:795–848 835
One of the methods that is used is data profiling, which is a special use of data
management. It aims to discover the meaningful features of data sets. The company
is provided with a broad picture of the data structure, content, and relationships
(Persson 2016). One company, for example, observed that the distance between the
workplace and home is a strong predictor of job tenure. If a hiring algorithm relied
on this aspect, discrimination based on residence occurs (Kim 2016). Additionally,
NLP is also used in the HR development. To identify skills and to support career
paths, some companies conduct interviews with their employees to create a
psychological profile (e.g., personality or cognitive ability) (Chamorro-Premuzic
et al. 2016).
Another approach is evaluation. For example, Rosenblat and Stark (2016)
examined in a case study the evaluation platform of the American passenger
transport mediation service company Uber and found that discrimination exists in
the evaluation of drivers. Uber tracks employees’ GPS positions and has
acceleration sensors integrated into the driver’s version of the Uber app to detect
heavy braking and speeding (Prassl 2018). Females are paid less than males,
because they drive slower. Consequently, the algorithm calculates a lower salary
due to slower driving for the same route.
To evaluate and promote employees, organizations increasingly rely on
recommender systems. For example, IBM offers IBM Watson Career Coach,
which is a career management solution that advises employees about online and
offline training based on their current job and previous jobs within the company and
based on the experiences of similar employees (IBM 2020). The pitfalls with respect
to recommender systems, as mentioned earlier, also apply in the development.
Regarding the perceived fairness, in an empirical-quantitative online experiment
Lee (2018) analyzed the fairness perception of managerial decisions (using a
customer service call center that uses NLP to evaluate the performance), whereby
the decision-maker was manipulated. Performance evaluations carried out by an
algorithm are less likely to be perceived as fair and trustworthy, and at the same
time, they evoke more negative feelings than human decisions.
6 Discussion
123
836 Business Research (2020) 13:795–848
review is the need for more quantitative research on the potential pitfalls of
algorithmic decision-making in the field of HRM.
Companies implement algorithmic decision-making to avoid or even overcome
human biases. However, our systematic literature review shows that algorithmic
decision-making is not a panacea for eliminating biases. Algorithms are vulnerable
to biases in terms of gender, ethnicity, sexual orientation, or other characteristics if
the algorithm builds upon inaccurate, biased, or unrepresentative input and training
data (Kim 2016). Algorithms replicate biases if the input data are already biased.
Consequently, there is a need for transparency; employees and candidates should
have the possibility to understand what happens within the process (Lepri et al.
2018).
Moreover, organizations need to consider the perceived fairness of employees
and applicants when using algorithmic decision-making in HR recruitment and HR
development. For companies, it is difficult to satisfy both computational fairness
from the computer science, which is defined by rules and formulas, and perceived
fairness from the management literature that is subjectively felt by potential and
current employees. To fulfill procedural justice and distributive justice, it is
important for organizations to reduce or avoid all types of biases and to achieve
subjective fairness, such as individual fairness, group fairness (Dwork et al. 2012),
and equal opportunity (Hardt et al. 2016). Companies need to continuously enhance
the perceived fairness of their HR recruitment and selection and HR training and
development process to avoid adverse impacts on the organization, such as
diminishing employer attractiveness, employer image, task performance, motiva-
tion, and satisfaction with the processes (Cropanzano et al. 2007; Cohen-Charash
and Spector 2001; Gilliland 1993).
With regard to fairness perceptions, it appears to be beneficial that humans make
the final decision if the decision is about the potential of employees or career
development (Lee 2018). At first glance, this partially contradicts previous findings
that the automated evaluation seems to be more valid, since human raters may
evaluate candidates inconsistently or without proper evidence (Kuncel et al. 2013;
Woods et al. 2020). However, while people accept that an algorithmic system
performs mechanical tasks (e.g., work scheduling), human tasks (e.g., hiring, work
evaluation) should be performed by humans (Lee 2018). Reasons for the lower
acceptance of algorithms in judging people and their potential are multifaceted. The
usage of this new technology in HRM, combined with a lack of knowledge and
transparency about how the algorithms work, increases emotional creepiness (e.g.,
Langer et al. 2019; Langer and König 2018) and decreases interpersonal treatment
and social interactions (e.g., Lee 2018) as well as fairness perceptions and the
opportunity to perform (e.g., Kaibel et al. 2019). To overcome these adverse
impacts of algorithmic decision-making in HRM, companies need to promote their
usage of algorithms (van Esch et al. 2019) and make the processes more transparent
of how algorithms are supporting the decisions of humans (Tambe et al. 2019). This
might help to create HR systems in recruitment and career development that are
both valid and perceived as fair. Nevertheless, a fruitful research avenue could be to
examine how companies should communicate or promote their usage of algorithms
123
Business Research (2020) 13:795–848 837
and whether employees and applicants accept a certain degree of algorithmic aid in
human decision-making.
In summary, companies should not solely rely on the information provided by
algorithms or even implement automatic decision-making without any control or
auditing by humans. While some biases might be more apparent, implicit
discrimination of less apparent personal characteristics might be more problematic,
because such implicit biases are more difficult to detect. In the following, we outline
theoretical and practical implications as well as future research directions.
This review reveals that current knowledge on the possible pitfalls of algorithmic
decision-making in HRM is still in an early stage, although we recently identified
increased attention to fairness and discrimination. Thus, the question arises about
what the most important future research priorities are (see Table 4 for exemplary
research questions). The majority of studies which we found concerning fairness
and discrimination were non-empirical. One reason for the paucity of empirical
research could be that algorithmic decision-making is a recent phenomenon in the
field of HR recruitment and HR development, which has not yet received much
attention from management scholars. Consequently, there is a need for more
sophisticated, theoretically, quantitative studies, especially in HR recruitment and
HR development, but also in HR selection. In this regard, a closer look reveals that
the majority of current research focuses on HR selection. However, also for HR
selection, only one or two studies per tool addressed fairness or perceived fairness.
In contrast, fairness perceptions and biases in HR recruitment and HR development
receive little attention (see Table 3).
The discussion on what leads to discrimination and its avoidance seems to be a
fruitful research avenue. Notably, the different types of algorithmic bias (see
Sect. 2.2) that can lead to (implicit) discrimination and unfairness need to be
considered separately. The existing studies mainly discuss bias, unfairness, and
discrimination in general, but rarely delve into detail by studying what kind of bias
occurred (e.g., historical bias or technical bias). Similarly, several studies
distinguished between mathematical fairness and perceived fairness, but did not
take a closer look at individual fairness, group fairness, or equal opportunity (see
Sect. 2.3).
Another prospective research area focuses on the difference in reliability and
validity between AI decision-makers and human raters (Suen et al. 2019). Many
studies found that an algorithm could be discriminatory, but the question remains
whether algorithms are fairer than humans are. However, this is important to address
to achieve the fairest possible decision-making process.
Another research avenue for new tools in HR recruitment and HR development
focuses on the individuals’ perspective and acceptance of algorithmic decision-
making. Only a few studies have examined the subjective fairness perceptions of
algorithmic decision-making in the HRM context. Thus, the way employees and
applicants perceive decisions made by an algorithm instead of humans is not fully
exploited (Lee 2018). In HR selection, a few studies have analyzed the perceived
123
838 Business Research (2020) 13:795–848
fairness. However, our systematic review underlines the recent calls by Hiemstra
et al. (2019) and Langer et al. (2018) for additional research to fully understand the
emotions and reactions of candidates and talented employees when using
algorithmic decision-making in HR recruitment or HR development processes.
Emotions and reactions can have important negative consequences for organiza-
tions, such as withdrawal from the application process or job turnover (Anderson
2003; Ryan and Ployhart 2000). In general, knowledge about applicants’ reactions
when using algorithmic decision-making is still limited (van Esch et al. 2019).
Previous studies analyzed a single algorithmic decision tool [see Kaibel et al. (2019)
for a recent exception]. Consequently, there is a need to examine applicants’
acceptance of algorithmic decision-making within the steps of the recruitment and
selection process (e.g., media content and recruitment tools on the employer’s
webpage, recommender systems in social media, screening and preselection,
telephone interview, and video interview).
Although there is some evidence that candidates react negatively to a decision
made by an algorithm (i.e., Kaibel et al. 2019; Ötting and Maier 2018; Lee 2018),
more research is needed on individuals’ acceptance of algorithms if algorithms
support the decisions by humans. Moreover, additional insights are needed into
whether transparency and more information about the algorithmic decision-making
process positively influences the fairness perception (Hiemstra et al. 2019). Finally,
while we found many studies examining the fairness perception of applicants (i.e.,
potential employees), the perspective of current employees on algorithmic decision-
making is still neglected in HRM research. Besides the threat of job loss due to
digitalization and automation, the question of how algorithms might help to assess,
promote, and retain qualified and talented employees remains important and will
become more important in the next decade. Thus, fairness and biases perceived by
current employees offer yet another fruitful research avenue in HR development.
Given that in many companies, the HR function has the main responsibility for
current and potential employees, our literature review shows that HR managers need
to be careful about implementing algorithmic decision-making, respecting privacy
and fairness concerns, and monitoring and auditing the algorithms that are used
(Simbeck 2019). This is accompanied by an obligation to inform employees and
applicants about the usage of the data and the potential consequences, for example,
forecasting career opportunities. Since the implementation of algorithmic decision-
making in HRM is a social process, employees should actively participate in this
process (Leicht-Deobald et al. 2019; Friedman et al. 2013; Tambe et al. 2019).
Moreover, applicants and employees must have the opportunity to not agree with
the proceedings (Simbeck 2019). A first step would be to implement company
guidelines for the execution and auditing of algorithmic decision-making and
transparent communication about data usage (Simbeck 2019; Cheng and Hackett
2019).
If companies implement an algorithm, the responsibility, accountability, and
transparency need to be clarified in advance. Members of the company need to have
123
Business Research (2020) 13:795–848 839
123
840 Business Research (2020) 13:795–848
less fair and trustworthy (Lee 2018). Moreover, pure algorithmic decisions evoke
negative feelings (Lee 2018). An implication to prevent anger among the applicants
or employees is a disclosure of the nature of the decision made by an algorithm
(Cheng and Hackett 2019). A short-term solution to avoid a decrease in the
acceptance could be a balanced approach between algorithmic and human decision-
making, which means that the algorithm makes a suggestion, but a human checks or
even makes the final decision. Hence, algorithmic decision-making seems to be an
indispensable tool for assistance in the decision, but human expertise is still
necessary (Yarger et al. 2019).
Of course, these practical implications are not limited to HR recruitment and HR
development; other HR functions might benefit from these insights, as well. In other
HR functions, employees should be informed and, if possible, involved in the
algorithms or AI’s implementation process. Responsibilities and accountability
should be clarified in advance, privacy should be respected, and the possibility for
employee voice should be acknowledged. Moreover, they should seek adequate
input data and implement data quality checks, which goes along with updating the
data regularly. If an external provider is in charge of programming and providing
the algorithm, the data and the algorithm should be adapted to the company and
should not be adopted without knowing the input data, the conditions for the
algorithmic outcomes, and the potential pitfalls of the algorithms.
7 Conclusion
Acknowledgements We thank Maike Giefers, Hannah Kaiser, and Anna Nieter, and Shirin Riazy for
their support.
Funding Not applicable for that section.
Data availability All material is available upon request.
123
Business Research (2020) 13:795–848 841
Conflict of interest The authors declare that they have no conflict of interest.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License,
which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as
you give appropriate credit to the original author(s) and the source, provide a link to the Creative
Commons licence, and indicate if changes were made. The images or other third party material in this
article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line
to the material. If material is not included in the article’s Creative Commons licence and your intended
use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain
permission directly from the copyright holder. To view a copy of this licence, visit http://
creativecommons.org/licenses/by/4.0/.
Appendix
Algorithm
– Algorithm*
– ‘‘Algorithmic model*’’
– ‘‘Data-algorithm*’’
– ‘‘Algorithmic decision-making’’, ‘‘algorithmic decision*’’
– ‘‘Artificial intelligence’’
– ‘‘Facial expression tool*’’, ‘‘facial expression processing*’’
– ‘‘Language processing*’’, ‘‘natural language processing*’’
– ‘‘Recommender system*’’
– ‘‘Search engine*’’
Discrimination
– Discrimination*
– Discriminat*
– Classification*, ‘‘classification problem*’’, ‘‘classification scheme*’’
– ‘‘Algorithmic discrimination*’’, ‘‘algorithmic bias discrimination*’’
– ‘‘Preventing discrimination*’’
– Anti-discrimination*, non-discrimination*
– Gender, age, sex, sexism, origin
– ‘‘Gender-based inequalities’’
– ‘‘Difference* among demographic group*’’
– Ethic*, ‘‘ethical implication*’’
– ‘‘Data mining discrimination*’’
– Favoritism, favouritism
– ‘‘Unfair treatment*’’
123
842 Business Research (2020) 13:795–848
Fairness
– Fair*, unfair*
– ‘‘Perceived fairness’’, ‘‘algorithmic fairness’’
– ‘‘Fairness word*’’, ‘‘fairness speech*’’, ‘‘fairness recommendation*’’
– Equal*, equit*, inequal*, ‘‘equal opportunit*’’
– Transparen*
– Legal*, right*
– Truth
– Impartial*
– Correct*
– Justicea
– Adverse impacta
Evaluation
– Evaluat*
– Judgement*, ‘‘algorithmic judgement*’’, ‘‘human judgement*’’, ‘‘mechanical
judgement*’’
– Rank*
– Rate*
– Measure*
– Valuation*
Bias
– Bias*
– ‘‘Algorithmic bias*’’, ‘‘national bias*’’, gender-bias*, ‘‘decision-making bias*’’,
‘‘human bias*’’, ‘‘technical bias*’’
– ‘‘Implicit bias* in algorithm*’’
– ‘‘Dealing with bias*’’
– ‘‘Pattern distortion*’’
– Pre-justice*
– Preconception*
– Tendenc*
– Prone*
Data mining
– Data*
– ‘‘Data set*’’
123
Business Research (2020) 13:795–848 843
HRM
References
123
844 Business Research (2020) 13:795–848
Carey, Dennis, and Matt Smith. 2016. How companies are using simulations, competitions, and analytics
to hire. Harvard Business Review. https://hbr.org/2016/04/how-companies-are-using-simulations-
competitions-and-analytics-to-hire.
Cascio, Wayne F., and Herman Aguinis. 2013. Applied psychology in human resource management.
London: Pearson Education.
Chalfin, Aaron, Oren Danieli, Andrew Hillis, Zubin Jelveh, Michael Luca, Jens Ludwig, and Sendhil
Mullainathan. 2016. Productivity and selection of human capital with machine learning. American
Economic Review 106 (5): 124–127.
Chamorro-Premuzic, Tomas, Dave Winsborough, Ryne A. Sherman, and Robert Hogan. 2016. New talent
signals: shiny new objects or a brave new world? Industrial and Organizational Psychology 9 (3):
621–640.
Chamorro-Premuzic, Tomas, Reece Akhtar, Dave Winsborough, Ryne A Sherman. 2017. The datafication
of talent: how technology is advancing the science of human potential at work. Current Opinion in
Behavioral Sciences 18:13–16.
Chander, Anupam. 2016. The racist algorithm. Michigan Law Review 115: 1023.
Chen, Le, Ruijun Ma, Anikó Hannák, and Christo Wilson. 2018. Investigating the impact of gender on
rank in resume search engines. In Proceedings of the 2018 chi conference on human factors in
computing systems: 1–14. https://doi.org/10.1016/j.hrmr.2019.100698.
Cheng, Maggie M., and Rick D. Hackett. 2019. A critical review of algorithms in HRM: definition,
theory, and practice. Human Resource Management Review 100698.
Citron, Danielle Keats, and Frank Pasquale. 2014. The scored society: due process for automated
predictions. Washington Law Review 89: 1.
Cohen-Charash, Yochi, and Paul E. Spector. 2001. The role of justice in organizations: a meta-analysis.
Organizational Behavior and Human Decision Processes 86 (2): 278–321.
Cropanzano, Russell, David E. Bowen, and Stephen W. Gilliland. 2007. The management of
organizational justice. Academy of Management Perspectives 21 (4): 34–48.
Crossan, Mary M., and Marina Apaydin. 2010. A multi-dimensional framework of organizational
innovation: a systematic review of the literature. Journal of Management Studies 47 (6): 1154–1191.
Danks, David, and Alex John London. 2017. Algorithmic bias in autonomous systems. In IJCAI: 4691-
4697.
Dastin, Jeffrey. 2018. Amazon scraps secret AI recruiting tool that showed bias against women. San
Fransico: Reuters.
Datta, Amit, Michael Carl Tschantz, and Anupam Datta. 2015. Automated experiments on ad privacy
settings. Proceedings on Privacy Enhancing Technologies 2015 (1): 92–112.
Daugherty, Paul R., and H.J. Wilson. 2018. Human? machine: reimagining work in the age of AI.
Boston: Harvard Business Press.
Deloitte. 2018. Mensch bleibt Mensch - auch mit algorithmen im recruiting. Wo der Einsatz von
Algorithmen hilfreich ist und wo nicht. https://www2.deloitte.com/de/de/pages/careers/articles/
algorithmen-im-recruiting-prozess.html. Accessed 12 Sept 2019.
Deloitte. 2020. State of AI in the enterprise – 3rd edition results of the survey of 200 AI experts on
artificial intelligence in German companies. https://www2.deloitte.com/content/dam/Deloitte/de/
Documents/technology-media-telecommunications/DELO-6418_State%20of%20AI%202020_KS4.
pdf. Accessed 10 Jun 2020.
Deng, Li., and Yu. Dong. 2014. Deep learning: methods and applications. Foundations and Trends in
Signal Processing 7 (3–4): 197–387.
Diakopoulos, Nicholas. 2015. Algorithmic accountability: journalistic investigation of computational
power structures. Digital Journalism 3 (3): 398–415.
Dreisbach, Caitlin, Theresa A. Koleck, Philip E. Bourne, Suzanne Bakken. 2019. A systematic review of
natural language processing and text mining of symptoms from electronic patient-authored text data.
International Journal of Medical Informatics 125:37–46.
Dwork, Cynthia, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness
through awareness. In Proceedings of the 3rd innovations in theoretical computer science
conference: ACM: 214–226.
Ferguson, Christopher J., and Michael T. Brannick. 2012. Publication bias in psychological science:
prevalence, methods for identifying and controlling, and implications for the use of meta-analyses.
Psychological Methods 17 (1): 120.
123
Business Research (2020) 13:795–848 845
Florentine, S. 2016. How artificial intelligence can eliminate bias in hiring. CIO Magazine. https://www.
cio.com/article/3152798/artificial-intelligence/how-artificial-intelligence-can-eliminate-bias-in-
hiring.html. Accessed 03 Mar 2020.
Friedman, Batya, and Helen Nissenbaum. 1996. Bias in computer systems. ACM Transactions on
Information Systems 14 (3): 330–347.
Friedman, Batya, Peter H. Kahn, Alan Borning, and Alina Huldtgren. 2013. Value sensitive design and
information systems. In Early engagement and new technologies: opening up the laboratory,
Dodrecht: Springer: 27–55.
Frijters, Paul. 1998. Discrimination and job-uncertainty. Journal of Economic Behavior & Organization
36 (4): 433–446.
Gil-Lafuente, Anna Marı́a, and Young Kyun Oh. 2012. Decision making to manage the optimal selection
of personnel in the hotel company applying the hungarian algorithm. The International Journal of
Management Science and Information Technology 6-(Oct-Dec): 27–42.
Gilliland, Stephen W. 1993. The perceived fairness of selection systems: an organizational justice
perspective. Academy of Management Review 18 (4): 694–734.
Goodfellow, Ian, Y. Bengio, and A. Courville. 2016. Machine learning basics. Deep Learning 1: 98–164.
Gough, David, Sandy Oliver, and James Thomas. 2017. An introduction to systematic reviews. London:
Sage.
Guchait, Priyanko, Tanya Ruetzler, Jim Taylor, and Nicole Toldi. 2014. Video interviewing: a potential
selection tool for hospitality managers–a study to understand applicant perspective. International
Journal of Hospitality Management 36: 90–100.
Hardt, Moritz, Eric Price, and Nati Srebro. 2016. Equality of opportunity in supervised learning. In
Advances in neural information processing systems: 3315–3323.
Hausknecht, John P., David V. Day, and Scott C. Thomas. 2004. Applicant reactions to selection
procedures: an updated model and meta-analysis. Personnel Psychology 57 (3): 639–683.
HireVue. 2019. https://www.hirevue.com. Accessed 01.Jan 2020.
Hiemstra, Annemarie MF., Janneke K. Oostrom, Eva Derous, Alec W. Serlie, and Marise Ph Born. 2019.
Applicant perceptions of initial job candidate screening with asynchronous job interviews: does
personality matter? Journal of Personnel Psychology 18 (3): 138.
Hoffmann, Anna Lauren. 2019. Where fairness fails: data, algorithms, and the limits of antidiscrimination
discourse. Information, Communication & Society 22 (7): 900–915.
Horton, John J. 2017. The effects of algorithmic labor market recommendations: Evidence from a field
experiment. Journal of Labor Economics 35 (2): 345–385.
Huselid, Mark A. 1995. The impact of human resource management practices on turnover, productivity,
and corporate financial performance. Academy of Management Journal 38 (3): 635–672.
IBM. 2020. IBM Watson Career Coach for career management. https://www.ibm.com/talent-
management/career-coach. Accessed 20 Apr 2020.
Kaelbling, Leslie Pack, Michael L. Littman, and Andrew W. Moore. 1996. Reinforcement learning: a
survey. Journal of Artificial Intelligence Research 4: 237–285.
Kahneman, Daniel, Stewart Paul Slovic, Paul Slovic, and Amos Tversky. 1982. Judgment under
uncertainty: heuristics and biases. Cambridge: Cambridge University Press.
Kaibel, Chris, Irmela Koch-Bayram, Torsten Biemann, and Max Mühlenbock. 2019. Applicant
perceptions of hiring algorithms-uniqueness and discrimination experiences as moderators. In
Academy of Management Proceedings: Academy of Management Briarcliff Manor, NY 10510.
Kaplan, Andreas, and Michael Haenlein. 2019. Siri, Siri, in my hand: who’s the fairest in the land? On the
interpretations, illustrations, and implications of artificial intelligence. Business Horizons 62 (1):
15–25.
Kauermann, Goeran, and Helmut Kuechenhoff. 2010. Stichproben: Methoden und praktische Umsetzung
mit R. Berlin: Springer.
Kellogg, Katherine C., Melissa A. Valentine, Angéle Christin. 2020. Algorithms at Work: The New
Contested Terrain of Control. Academy of Management Annals 14(1):366–410.
Kim, Pauline T. 2016. Data-driven discrimination at work. William & Mary Law Review 58: 857.
Kim, P. T. 2017. Data-Driven Discrimination at Work. William & Mary Law Review, 58(3):857.
Kim, Pauline T., and Sharion Scott. 2018. Discrimination in online employment recruiting. Louis ULJ 63:
93.
Kuncel, Nathan R., David M. Klieger, Brian S. Connelly, and Deniz S. Ones. 2013. Mechanical versus
clinical data combination in selection and admissions decisions: a meta-analysis. Journal of Applied
Psychology 98 (6): 1060.
123
846 Business Research (2020) 13:795–848
Lambrecht, Anja, and Catherine Tucker. 2019. Algorithmic bias? An empirical study of apparent gender-
based discrimination in the display of stem career ads. Management Science 65 (7): 2966–2981.
Langer, Markus, Cornelius J. König, and Andromachi Fitili. 2018. Information as a double-edged sword:
the role of computer experience and information on applicant reactions towards novel technologies
for personnel selection. Computers in Human Behavior 81: 19–30. https://doi.org/10.1016/j.chb.
2017.11.036.
Langer, Markus, Cornelius J. König, and Maria Papathanasiou. 2019. Highly automated job interviews:
acceptance under the influence of stakes. International Journal of Selection and Assessment. https://
doi.org/10.1111/ijsa.12246.
Leclercq-Vandelannoitte, Aurélie. 2017. An Ethical Perspective on Emerging Forms of Ubiquitous IT-
Based Control. Journal of Business Ethics 142 (1):139–154.
Lee, Min Kyung. 2018. Understanding perception of algorithmic decisions: fairness, trust, and emotion in
response to algorithmic management. Big Data & Society 5 (1): 2053951718756684.
Lee, Min Kyung, and Su Baykal. 2017. Algorithmic mediation in group decisions: fairness perceptions of
algorithmically mediated vs. discussion-based social division. In Proceedings of the 2017 ACM
conference on computer supported cooperative work and social computing: ACM: 1035-1048.
Lee, In., and Yong Jae Shin. 2020. Machine learning for enterprises: applications, algorithm selection,
and challenges. Business Horizons 63 (2): 157–170.
Leicht-Deobald, Ulrich, Thorsten Busch, Christoph Schank, Antoinette Weibel, Simon Schafheitle,
Isabelle Wildhaber, and Gabriel Kasper. 2019. The challenges of algorithm-based HR decision-
making for personal integrity. Journal of Business Ethics 160 (2): 377–392.
Lepri, Bruno, Nuria Oliver, Emmanuel Letouzé, Alex Pentland, and Patrick Vinck. 2018. Fair,
transparent, and accountable algorithmic decision-making processes. Philosophy & Technology 31
(4): 611–627.
Leventhal, Gerald S. 1980. What should be done with equity theory? In Social exchange, New York:
Springer: 27–55.
Lindebaum, Dirk, Mikko Vesa, and Frank den Hond. 2019. Insights from the machine stops to better
understand rational assumptions in algorithmic decision-making and its implications for organiza-
tions. Academy of Management Review. https://doi.org/10.5465/amr.2018.0181.
Lipsey, Mark W., and David B. Wilson. 2001. Practical meta-analysis. Thousand Oaks: SAGE
publications Inc.
Mann, Gideon, and Cathy O’Neil. 2016. Hiring algorithms are not neutral. Harvard Business Review 9.
https://hbr.org/2016/12/hiring-algorithms-are-not-neutral.
McCarthy, Julie M., Talya N. Bauer, Donald M. Truxillo, Neil R. Anderson, Ana Cristina Costa, and Sara
M. Ahmed. 2017. Applicant perspectives during selection: a review addressing ‘‘So what?’’,
‘‘What’s new?’’, and ‘‘Where to next?’’ Journal of Management 43 (6): 1693–1725.
McColl, Rod, and Marco Michelotti. 2019. Sorry, could you repeat the question? Exploring video-
interview recruitment practice in HRM. Human Resource Management Journal 29 (4): 637–656.
McDonald, Kathleen, Sandra Fisher, and Catherine E. Connelly. 2017. e-HRM systems in support of
‘‘smart’’ workforce management: an exploratory case study of system success. Electronic HRM in
the Smart Era 87–108. https://doi.org/10.1108/978-1-78714-315-920161004
Meade, Adam W., and Michael Fetzer. 2009. Test bias, differential prediction, and a revised approach for
determining the suitability of a predictor in a selection context. Organizational Research Methods
12 (4): 738–761.
Miller 2015. Can an algorithm hire better than a human. The New York Times. https://www.nytimes.com/
2015/06/26/upshot/can-an-algorithm-hire-better-than-a-human.html. Accessed 13 sep 2019.
Moher, David, Alessandro Liberati, Jennifer Tetzlaff, and Douglas G. Altman. 2009. Preferred reporting
items for systematic reviews and meta-analyses: the PRISMA statement. Annals of Internal
Medicine 151 (4): 264–269.
Möhlmann, M., and L. Zalmanson. 2017. Hands on the wheel: navigating algorithmic management and
Uber drivers’. In Autonomy’, in proceedings of the international conference on information systems
(ICIS), Seoul South Korea: 1–17.
Morrison, Andra, Julie Polisena, Don Husereau, Kristen Moulton, Michelle Clark, Michelle Fiander,
Monika Mierzwinski-Urban, Tammy Clifford, Brian Hutton, and Danielle Rabb. 2012. The effect of
English-language restriction on systematic review-based meta-analyses: a systematic review of
empirical studies. International Journal of Technology Assessment in Health Care 28 (2): 138–144.
Murphy, Kevin P. 2012. Machine learning: a probabilistic perspective. Cambridge: MIT press.
123
Business Research (2020) 13:795–848 847
Naim, Iftekhar, Md Iftekhar Tanveer, Daniel Gildea, and Mohammed Ehsan Hoque. 2016. Automated
analysis and prediction of job interview performance. IEEE Transactions on Affective Computing 9
(2): 191–204.
Ötting, Sonja K., and Günter. W. Maier. 2018. The importance of procedural justice in human–machine
interactions: intelligent systems as new decision agents in organizations. Computers in Human
Behavior 89: 27–39.
Paschen, Ulrich, Christine Pitt, and Jan Kietzmann. 2020. Artificial intelligence: Building blocks and an
innovation typology. Business Horizons 63 (2): 147–155.
Pasquale, Frank. 2015. The black box society. Cambridge: Harvard University Press.
Persson, Anders. 2016. Implicit bias in predictive data profiling within recruitments. In IFIP International
Summer School on Privacy and Identity Management. Springer.
Petticrew, Mark, and Helen Roberts. 2008. Systematic reviews in the social sciences: a practical guide.
Hoboken: John Wiley & Son.
Podsakoff, Philip M., Scott B. MacKenzie, Daniel G. Bachrach, and Nathan P. Podsakoff. 2005. The
influence of management journals in the 1980s and 1990s. Strategic Management Journal 26 (5):
473–488.
Prassl, Jeremias. 2018. Humans as a service: the promise and perils of work in the gig economy. Oxford:
Oxford University Press.
Precire. 2020. Precire technologies. https://precire.com/. Accessed 03 Jan 2020.
Raghavan, Manish, Solon Barocas, Jon Kleinberg, and Karen Levy. 2020. Mitigating bias in algorithmic
hiring: evaluating claims and practices. In Proceedings of the 2020 conference on fairness,
accountability, and transparency.
Roscher, Ribana, Bastian Bohn, Marco F. Duarte, and Jochen Garcke. 2020. Explainable machine
learning for scientific insights and discoveries. IEEE Access 8: 42200–42216.
Rosenblat, Alex, Tamara Kneese, and Danah Boyd. 2014. Networked employment discrimination. Open
Society Foundations’ Future of Work Commissioned Research Papers.
Rosenblat, Alex, and Luke Stark. 2016. Algorithmic labor and information asymmetries: a case study of
Uber’s drivers. International Journal of Communication 10: 27.
Roth, Philip L., Huy Le, Oh. In-Sue, Chad H. Van Iddekinge, and Steven B. Robbins. 2017. Who ru?: On
the (in) accuracy of incumbent-based estimates of range restriction in criterion-related and
differential validity research. Journal of Applied Psychology 102 (5): 802.
Russell, Stuart J., and Peter Norvig. 2016. Artificial intelligence: a modern approach. London: Pearson
Education Limited.
Ryan, Ann Marie, and Robert E. Ployhart. 2000. Applicants’ perceptions of selection procedures and
decisions: a critical review and agenda for the future. Journal of Management 26 (3): 565–606.
Sajjadiani, Sima, Aaron J. Sojourner, John D. Kammeyer-Mueller, and Elton Mykerezi. 2019. Using
machine learning to translate applicant work history into predictors of performance and turnover.
Journal of Applied Psychology. https://doi.org/10.1037/apl0000405.
Sánchez-Monedero, Javier, Lina Dencik, and Lilian Edwards. 2020. What does it mean to ’solve’ the
problem of discrimination in hiring? Social, technical and legal perspectives from the UK on
automated hiring systems. In Proceedings of the 2020 conference on fairness, accountability, and
transparency: 458–468.
Savage, David, and Richard A. Bales. 2017. Video games in job interviews: using algorithms to minimize
discrimination and unconscious bias. ABA Journal of Labor & Employment Law 32.
Siddaway, Andy P., Alex M. Wood, and Larry V. Hedges. 2019. How to do a systematic review: a best
practice guide for conducting and reporting narrative reviews, meta-analyses, and meta-syntheses.
Annual Review of Psychology 70: 747–770.
Silverman, Rachel Emma, and Nikki Waller. 2015. The algorithm that tells the boss who might quit. Wall
Street Journal. http://www.wsj.com/articles/the-algorithm-that-tells-the-boss-who-might-quit-
1426287935.
Simbeck, K. 2019. HR analytics and ethics. IBM Journal of Research and Development 63 (4/5): 1–9.
Stone, Diana L. Deadrick, Kimberly M. Lukaszewski, Richard Johnson. 2015. The influence of
technology on the future of human resource management. Human Resource Management Review 25
(2):216–231.
Suen, Hung-Yue., Mavis Yi-Ching. Chen, and Lu. Shih-Hao. 2019. Does the use of synchrony and
artificial intelligence in video interviews affect interview ratings and applicant attitudes? Computers
in Human Behavior 98: 93–101.
123
848 Business Research (2020) 13:795–848
Sumser, John. 2017. Artificial intelligence: ethics, liability, ownership and HR. Workforce Solutions
Review 8 (3): 24–26.
Suresh, Harini, and John V. Guttag. 2019. A framework for understanding unintended consequences of
machine learning. arXiv preprint arXiv:1901.10002.
Tambe, Prasanna, Peter Cappelli, and Valery Yakubovich. 2019. Artificial intelligence in human
resources management: challenges and a path forward. California Management Review 61 (4):
15–42.
van Esch, Patrick, J. Stewart Black, and Joseph Ferolie. 2019. Marketing AI recruitment: the next phase
in job application and selection. Computers in Human Behavior 90: 215–222.
Van Hoye, G. 2014. Word of mouth as a recruitment source: an integrative model. In Yu, K.Y.T. and
Cable, D.M. (eds), The Oxford Handbook of Recruitment. Oxford: Oxford University Press:
251–268.
Varghese, Jacob S., James C. Moore, and Andrew B. Whinston. 1988. Artificial intelligence and the
management science practitioner: rational choice and artificial intelligence. Interfaces 18 (4): 24–35.
Vasconcelos, Marisa, Carlos Cardonha, and Bernardo Gonçalves. 2018. Modeling epistemological
principles for bias mitigation in AI systems: an illustration in hiring decisions. In Proceedings of the
2018 AAAI/ACM Conference on AI, Ethics, and Society.
Veale, Michael, and Reuben Binns. 2017. Fairer machine learning in the real world: mitigating
discrimination without collecting sensitive data. Big Data & Society 4 (2): 2053951717743530.
Walker, Joseph. 2012. Meet the new boss: big data. Wall Street Journal. https://online.wsj.com/article/
SB10000872396390443890304578006252019616768.html. Accessed 13 Mar 2020
Williams, Betsy Anne, Catherine F Brooks, Yotam Shmargad. 2018. How Algorithms Discriminate
Based on Data They Lack: Challenges, Solutions, and Policy Implications. Journal of Information
Policy 8:78–115.
Wolpert, David H., and William G. Macready. 1997. No free lunch theorems for optimization. IEEE
Transactions on Evolutionary Computation 1 (1): 67–82.
Woodruff, Allison, Sarah E Fox, Steven Rousso-Schindler, and Jeffrey Warshaw. 2018. A qualitative
exploration of perceptions of algorithmic fairness. In Proceedings of the 2018 CHI Conference on
Human Factors in Computing Systems.
Woods, Stephen A., Sara Ahmed, Ioannis Nikolaou, Ana Cristina Costa, and Neil R. Anderson. 2020.
Personnel selection in the digital age: a review of validity and applicant reactions, and future
research challenges. European Journal of Work and Organizational Psychology 29 (1): 64–77.
Yarger, Lynette, Fay Cobb Payton, and Bikalpa Neupane. 2019. Algorithmic equity in the hiring of
underrepresented IT job candidates. Online Information Review. https://doi.org/10.1108/OIR-10-
2018-033. Accessed 3 Mar 2020.
Publisher’s Note Springer Nature remains neutral with regard to jurisdictional claims in published maps
and institutional affiliations.
123