Hiring, Algorithms, and Choice: Why Interviews Still Matter

Why do organizations conduct job interviews? The traditional view of interviewing holds that interviews are conducted, despite their steep costs, to predict a candidate’s future performance and fit. This view faces a twofold threat: the behavioral and algorithmic threats. Specifically, an overwhelming body of behavioral research suggests that we are bad at predicting performance and fit; furthermore, algorithms are already better than us at making these predictions in various domains. If the traditional view captures the whole story, then interviews seem to be a costly, archaic human resources procedure sustained by managerial overconfidence. However, building on T. M. Scanlon’s work, we offer the value of choice theory of interviewing and argue that interviews can be vindicated once we recognize that they generate commonly overlooked kinds of noninstrumental value. On our view, interviews should thus not be entirely replaced by algorithms, however sophisticated algorithms ultimately become at predicting performance and fit.

W hy do organizations conduct job interviews, despite the enormous costs associated with the interview process? At first blush, this does not seem like an especially challenging question. This is because a natural and seemingly obvious answer immediately comes to mind: interviews are for predicting a candidate's future performance and fit with respect to the hiring organization's requirements, values, and culture-that's why organizations conduct interviews, despite their costs (Cappelli, 2019b;Elfenbein & Sterling, 2018;Muehlemann & Strupler Leiser, 2018; Society for Human Resource Management [SHRM], 2017). This is also the traditional view of interviewing espoused by managers and is how the nature and function of interviews are characterized in human resource management (HRM) textbooks (Dessler, 2020;Mathis, Jackson, Valentine, & Meglich, 2016;Mondy & Martocchio, 2016). 1 Thus, although the costs may be undesirable, they are the price to pay, as it were, to be able to judge whether a candidate will match the needs of the role and the organization. 2 In this article, we suggest that the question of why to conduct interviews is a more difficult one than it first seems. The force of this question can be appreciated when juxtaposed against a twofold threat we argue the traditional view of interviewing faces. The first threat, the behavioral threat, holds that a large body of behavioral evidence suggests that we are poor predicters of future performance and bad judges of fit. This is for multiple reasons: the judgments of interviewers are riddled with biases, interviewers overestimate their assessment capacities, and organizations rarely assess the performance of candidates they might have passed on (in relation to the candidates they ultimately selected). As one HRM textbook notes, "traditionally, interviews have not been valid predictors of success on the job" (Mondy & Martocchio, 2016: 165). In short, those involved in making hiring decisions are demonstrably bad at predicting future performance and assessing fit.
The behavioral threat has brought some management theorists to suggest abandoning interviews as traditionally conceived (i.e., unstructured interviews) and moving toward structured interviews. Yet structured interviews, too, face problems: they can collapse into unstructured interviews, or alternatively, they start out unstructured either before or after the official start of the interview and, in doing so, increase exposure to the behavioral threat. More fundamentally, the behavioral threat is simply pushed back one step, to the point at which one decides the structure of the interview. Thus, although structured interviews may be an improvement upon unstructured interviews, they, too, do not fare especially well with respect to the behavioral threat.
A defender of the traditional view might acknowledge the force of the behavioral threat yet still respond, "We have no better alternative!" But this argumentative maneuver is cut off by the second threat the traditional view faces: the algorithmic threat. Algorithms already have a superior track record to humans, even expert humans, of predicting the performance and fit of candidates in a number of domains. Indeed, 67 percent of eighty-eight hundred recruiters and hiring managers globally surveyed by LinkedIn in 2018 noted that they use artificial intelligence (AI) tools to save time in sourcing and screening candidates (Ignatova & Reilly, 2018). So, where does this leave the practice of interviewing?
The behavioral and algorithmic threats, taken together, pose what we call the "interview puzzle" for the traditional view of interviewing. If the traditional view is correct about the nature and function of interviews-that interviews are for predicting the future performance and fit of a candidate with respect to the role's and organization's needs-then it seems as though the justification for the practice is predicting future performance and fit. The key point is that, even though a number of HRM scholars are skeptical of the efficacy of interviews at predicting future performance and fit, they nevertheless agree that the nature and function of interviews are for predicting future performance and for assessing candidate fit. 2 We note that with respect to a range of candidates, especially ones with more experience, the evaluation process is often mutual (i.e., a candidate may be evaluating whether a position at a given firm would satisfy the candidate's needs).
undermined. Not only is interviewing costly (Cappelli, 2020;Muehlemann & Strupler Leiser, 2018;SHRM, 2017) but we also are bad at it, and we may have better alternatives for predicting performance and fit (i.e., algorithms). Continuing to interview, then, if it is only about predicting performance and fit, seems to be at best an anachronistic human resources (HR) practice or at worst blatant wastefulness sustained by irrational managerial overconfidence. For these reasons, we argue that the traditional view of interviewing must be reexamined.
If interviews were singularly a means to predicting performance and fit, as the traditional view posits, we maintain that the justification for interviews is undermined. However, we argue that the antecedent in this conditional is false: interviews are not singularly a means to predicting performance and fit; rather, they are a much richer normative practice. In particular, we argue that interviews offer different kinds of value that have thus far been overlooked and thus the practice can be worth preserving, despite the behavioral and algorithmic threats. Something of normative significance would be lost were we to abandon the practice of interviewing, and this must be accounted for in our understanding of the nature of interviews.
In other words, we dissolve the interview puzzle by arguing that although the behavioral and algorithmic threats are indeed concerning, they only threaten to undermine our interview practices if the traditional view of interviewing is the whole story. But we argue that the traditional view of interviewing accounts for only part of its function-the parts it overlooks are the other kinds of value that interviews create, and these other kinds of value do not succumb to the behavioral and algorithmic threats. By reframing how we understand the nature of interviews, we advance a broader, normative conception of interviewing that suggests that our ability to choose whom we relate to in the workplace is an important source of value and that our work lives may be worse off without the practice.
We proceed as follows. In section 1, we characterize the traditional view of interviewing and discuss the costs of interviewing that are exhaustively documented in the HRM literature. In section 2, we discuss the behavioral and algorithmic threats and argue that they together undermine the traditional view of interviewing and thus generate the interview puzzle. In section 3, we introduce our value of choice theory of interviewing, grounded in the work of the philosopher T. M. Scanlon (1988Scanlon ( , 1998Scanlon ( , 2013Scanlon ( , 2019. We show how the interview puzzle can be dissolved once we grasp the inadequacy of the traditional view of interviewing: it fails to account for a broader range of contenders for the kinds of value that can be realized through interviewing. If the view we advance is correct, then the current understanding in HRM and management scholarship about the nature and function of interviews must be significantly expanded. In section 4, we offer several clarifications of our account and discuss some potential objections. In section 5, we discuss some new avenues of research that follow from our work. Finally, in section 6, we conclude.

THE TRADITIONAL VIEW OF INTERVIEWING
The traditional view of interviewing holds that interviews are one class of selection tools (among other tools, such as tests and background checks) that are useful for predicting a candidate's performance and fit. 3 In particular, a selection interview is defined as "a selection procedure designed to predict future job performance based on applicants' oral responses to oral inquiries" (Dessler, 2020: 207) and is considered a tool for assessing a candidate's knowledge, skills, abilities, and competencies in relation to what is required for the job (Dessler, 2020;Graves & Karren, 1996;McDaniel, Whetzel, Schmidt, & Maurer, 1994).
Interviews are widespread, in part, because of the belief that they are effective in simultaneously assessing candidates' ability, motivation, personality, aptitude, person-job fit, and person-organization fit (Highhouse, 2008). Several common assumptions sustain this belief: that making accurate predictions about candidates' future job performance is possible (Highhouse, 2008); that experience and intuition are necessary in effective hiring (Gigerenzer, 2007); that human beings (i.e., candidates) can be effectively evaluated only by equally sensitive complex beings (e.g., hiring managers), rather than by tests or algorithms (Highhouse, 2008); and that oral discussions with candidates can be revealing, as they allow for "reading between the lines" (Highhouse, 2008: 337).
Despite the widespread use of interviews, they are recognized to be a costly and time-consuming practice. The United States "fills a staggering 66 million jobs a year. Most of the $20 billion that companies spend on human resources vendors goes to hiring" (Cappelli, 2019b: 50). On average, employers in the United States spend approximately $4,000 per hire to fill non-executive-level positions and about $15,000 per hire to fill executive-level positions (SHRM, 2016(SHRM, , 2017, and a substantial portion of these costs is attributed to interviews. Outside the United States, employers report similar experiences. For example, in Switzerland, on average, employers spend as much as 16 weeks of wage payments to fill a skilled worker vacancy, of which 21 percent involves search costs, and roughly 50 percent of the search costs are direct interview costs (Muehlemann & Strupler Leiser, 2018). In addition, significant opportunity costs are associated with interviews for all parties involved (Muehlemann & Strupler Leiser, 2018).
With respect to the time spent on interviews, according to a recent talent acquisition benchmarking report, on average per job, US employers spend approximately eight days conducting interviews (SHRM, 2017). Employers report similar experiences outside the United States. For example, in Switzerland, on average, employers spend approximately 8.5 hours on job interviews per candidate (Muehlemann & Strupler Leiser, 2018).
Of course, the costs of hiring and interviewing are not uniform. The costs vary depending on the skill requirements of the job (Muehlemann & Strupler Leiser, 2018) and the degree of labor market tightness (Davis, Faberman, & Haltiwanger, 2012;Pissarides, 2009;Rogerson & Shimer, 2011), among other factors. That said, these costs on average remain substantial and are increasing-employers today spend twice as much time on interviews as they did in 2009 (Cappelli, 2019b). 4 As costly and time consuming as interviews are, there are also difficulties associated with verifying whether they are worth these costs. Indeed, "only about a third of US companies report that they monitor whether their hiring practices lead to good employees; few of them do so carefully, and only a minority even track cost per hire and time to hire" (Cappelli, 2019b: 50). Even if it were not so difficult to assess whether interviews are worth the costs with respect to the end posited by the traditional view (i.e., predicting performance and fit), two additional threats remain.

The Behavioral Threat
The traditional conception of interviews-as a means to predict a candidate's performance and fit in relation to a vacancy-hinges on an important assumption, namely, that performance and fit can be effectively predicted through interviewing. However, a considerable body of knowledge from the social sciences challenges this basic assumption and chronicles the poor track record of predicting performance and fit through interviews (Bishop & Trout, 2005;Bohnet, 2016;Chamorro-Premuzic & Akhtar, 2019;McCarthy, Van Iddekinge, & Campion, 2010;Rivera, 2012). Specifically, although there is empirical evidence that highlights the outsized role interviews have in the hiring process (Billsberry, 2007), interview-based hiring decisions have been found only to account for up to 10 percent of the variation in job performance (Conway, Jako, & Goodman, 1995). Additionally, biases pervade the process of predicting performance and fit through interviews, both in their unstructured and structured formats (Huffcutt, Roth, & McDaniel, 1996;McDaniel et al., 1994).

Unstructured Interviews
Unstructured interviews do not have a fixed format or a fixed set of questions, nor do they involve a fixed process for assessing the given responses (Schmidt & Hunter, 1998). During unstructured interviews, both the interviewer and the candidate investigate what seems most relevant at the time (Bohnet, 2016). This process often produces an overall rating for each applicant "based on summary impressions and judgments" (Schmidt & Hunter, 1998: 267). Unstructured interviews are often 4 Although the focus of our article is on employers, candidates bear significant costs too. For example, candidates must expend resources to sort through job opportunities, schedule commitments, and purchase new professional attire, among other costs. Relatedly, expending effort and time on interviewing could involve intangible short-and long-term opportunity costs that take candidates away from other productive activities. Furthermore, the psychological effects of the interview process can be onerous for the candidates. Although, in our article, we primarily highlight the costs employers bear, we acknowledge that the costs candidates bear ought to be taken seriously in their own right. assumed to be effective in concurrently assessing a range of dimensions associated with predicting performance and person-organization fit (Highhouse, 2008).
However, recent research shows that unstructured interviews may not in fact aid in hiring decisions. This research maintains that unstructured interviews are riddled with biases and are often swayed by the whims of the interviewers (Chamorro-Premuzic & Akhtar, 2019). Specifically, this research suggests that unstructured interviews are ineffective because interviewers tend to overlook the limits of their knowledge (Kausel, Culbertson, & Madrid, 2016), "decide on the fly" what questions to ask of which candidates and how to interpret responses (Cappelli, 2019b: 50), place disproportionate emphasis on a few pieces of information (Dawes, 2001), and confirm their own existing preferences (Chamorro-Premuzic & Akhtar, 2019). Subsequently, they become increasingly confident in the accuracy of their decisions, even when irrelevant information is introduced (Bohnet, 2016;Dawes, 2001). 5 One reason for interviewers' overconfidence regarding their predictive abilities is that they cannot often ascertain whether, absent interviews, their predictions would turn out to be better or worse, and they would generally lack a large enough sample to deduce any statistically valid inferences (Bishop & Trout, 2005).
While managers more heavily value a given trait or ability if evaluated by unstructured interviews rather than by alternative methods (e.g., paper-and-pencil tests) (Lievens, Highhouse, & DeCorte, 2005), a long-standing body of empirical evidence shows that unstructured interviews are unhelpful with selection decisions. For example, in the context of medical school applications, DeVaul, Jervey, Chappell, Caver, Short, and O' Keefe (1987) compare the students who were initially accepted versus those who were rejected for medical school and find that only 28 percent of the difference between these groups is related to academic and demographic factors and that 72 percent is related to the admissions committee's preferences developed through interviews. They report that when it comes to attrition and clinical performance during medical school and a subsequent year of postgraduate training, there are no significant differences between the accepted and the rejected groups, suggesting that interviews in this context are unhelpful to the decision-making process. In a similar fashion, Milstein, Wilkinson, Burrow, and Kessen (1981: 77) compare the performance of "a group of 24 applicants who were interviewed and accepted at the Yale University School of Medicine but went to other medical schools … with a group of 27 applicants who attended the same schools but had been rejected at Yale following an interview and committee deliberation." In this context, too, the researchers find no statistically significant relationship between admission decisions and performance, again pointing to the inefficacy of interviews in aiding the achievement of the decision-making ends. 6 5 Recent research suggests that part of why overconfidence persists, despite its considerable costs, is the status benefits it confers; moreover, these status benefits largely persist, even when the person's overconfidence is exposed (Anderson, Brion, Moore, & Kennedy, 2012;Kennedy, Anderson, & Moore, 2013). 6 See also Oskamp's (1965) study of the clinical decisions of psychologists, which shows that the accuracy of their decisions does not increase significantly with additional information from interviews (but confidence in their decision-making steadily increases).
Medical school admissions decisions are, of course, not hiring decisions, but similar results are seen in hiring contexts. In a study of the hiring practices at elite professional services firms, Rivera (2012) finds that employers often seek candidates who enjoy similar leisure pursuits and have shared experiences and selfpresentation styles. In doing so, Rivera shows that unstructured interviews may be less about assessing knowledge, skills, and abilities and more about exercising biases through replicating ourselves, including, but not limited to, our culture, gender, and ethnicity, in hiring decisions. Finally, through a meta-analysis, Schmidt and Hunter (1998) conclude that unstructured interviews are ineffective at predicting the performance of future employees.
Not only do we know that unstructured interviews are unhelpful in hiring decisions but there is also some empirical evidence that unstructured interviews reliably undermine those decisions (Bishop & Trout, 2005;DeVaul et al., 1987;Eysenck, 1954;Kausel et al., 2016;Milstein et al., 1981;Oskamp, 1965;Wiesner & Cronshaw, 1988). For example, as far back as the middle of the past century, in a large-scale empirical study, Bloom and Brundage (1947) found that the predictive gain in adding an interviewer's assessment of a candidate's experience, interest, and personality may well be negative. They specifically report that predictions based on test scores and interviewing were 30 percent worse than predictions based on test scores alone. More recently, Behroozi, Shirolkar, Barik, and Parnin (2020) have shown that even when tests are conducted in interview formats, such as "whiteboard technical interviews" common in software engineering, the mechanics and pressure of the interview context reduce the efficacy of the technical tests. This effect is heightened especially among minorities and other underrepresented groups (Munk, 2021). Other recent research reports similar findings: for example, research on human judgment documents that when decision makers (e.g., hiring managers, admissions officers, parole boards) judge candidates based on a dossier and an unstructured interview, their decisions tend to be worse than decisions based on the dossier alone (Bishop & Trout, 2005). In a similar fashion, Dana, Dawes, and Peterson (2013) show that adding an unstructured interview to diagnostic information when making screening decisions yields less accurate outcomes than not using an unstructured interview at all. In this case, even though the decision makers may sense that they are extracting useful information from unstructured interviews, in reality, that information is not useful (Dana et al., 2013).

Structured Interviews
Unlike the unstructured version, a structured interview involves a formal process that more systematically considers "rapport building, question sophistication, question consistency, probing, note taking, use of a panel of interviewers, and standardized evaluation" (Roulin, Bourdage, & Wingate, 2019: 37) in hiring decisions. In this interview format, to predict good hires, an expert interviewer systematically and consistently poses the same set of validated questions about past performance to all candidates and immediately scores each answer based on a set of predetermined criteria relevant to the tasks of the job (Cappelli, 2019b).
Although structured interviews are available and designed to standardize the hiring process and minimize subjectivity and bias (Bohnet, 2016;Reskin & McBrier, 2000), they are in effect not much more successful than unstructured interviews in aiding hiring decisions for at least three reasons. First, even though structured interviews, in theory, may be less biased 7 and a better predictor of future job performance 8 than their unstructured counterparts, they are not widely adopted in practice (König, Klehe, Berchtold, & Kleinmann, 2010;Roulin et al., 2019). The resistance to structuring interviews (Lievens et al., 2005;van der Zee, Bakker, & Bakker, 2002) is driven by interviewers' belief that a candidate's character is "far too complex to be assessed by scores, ratings, and formulas" (Highhouse, 2008: 339) that are predetermined in a structured format.
Second, even in cases when structured interviews are accepted, they are not well implemented for various reasons. For example, structured interviews tend to be more costly to construct (Schmidt & Hunter, 1998) in part because of the difficulties in designing and validating standardized questions and evaluation criteria (Bohnet, 2016;Roulin et al., 2019). Also, in reality, we rarely see structured interviews conducted by trained and experienced interviewers who manage to avoid having their idiosyncratic personalities distort the process (Roulin et al., 2019). Even when structured interviews are conducted by trained and experienced interviewers, the process sometimes deviates to a semistructured or unstructured format. For instance, in conforming to a predetermined set of questions, the flow of conversation in a structured interview might feel stilted, awkward, or uncomfortable for both the interviewer and the candidate, thereby inadvertently shifting the interview process to a less structured format (Bohnet, 2016).
Third, even when structured interviews are conducted by trained and experienced interviewers and the process does not deviate to an unstructured format, empirical evidence shows that structured interviews may not be systematic and free of bias because interviewers may used them to confirm their preexisting judgments rather than to evaluate the candidates-that is, a potential self-fulfilling prophecy (Dougherty, Turban, & Callender, 1994). On the candidates' side, there is also much room for introducing bias. For example, Stevens and Kristof (1995) show that applicants engage in significant impression management, even in structured interviews, thereby undermining the decision-making process. Furthermore, even when structured interviews are implemented properly, these issues and biases may not be eliminated: they may simply be shifted to the previous step of designing the interview and deciding its structure. Therefore not only are structured interviews rare but, even when they are used and properly implemented, they are afflicted with issues that complicate the evaluation of performance and fit. It is not surprising, then, that Cappelli (2019b: 56) argues that a structured interview is the "most difficult technique to get right." 7 The average validity of the structured interviews (at about 0.51) is greater than the average validity of the unstructured interviews (at about 0.38) and far greater than the average validity of poorly conducted unstructured interviews (Schmidt & Hunter, 1998: 267). 8 With respect to the predictive power of structured interviews, they "predict performance in job training programs with a validity of about .35" (Schmidt & Hunter, 1998: 267).
Although research shows that interviews can undermine the aims of the hiring process, interviews have remained a popular norm for employee selection for more than a hundred years (Buckley, Norris, & Wiese, 2000;van der Zee et al., 2002). They have remained popular not necessarily because the inefficacy of interviews is unknown. In fact, Rynes, Colbert, and Brown (2002) report that HR professionals appreciate the limitations of interviews. Still, hiring managers remain reluctant to outsource their judgment (Bohnet, 2016).

The Algorithmic Threat
Interviews, both in their unstructured and structured formats, if not by design, in practice are ineffective at assessing fit or predicting future performance and create a significant opportunity for bias in hiring decisions (Chamorro-Premuzic & Akhtar, 2019; Rivera, 2012). However, proponents of the traditional view of interviewing might respond that there are no alternatives. But this assertion falls short in the face of the second threat the traditional view faces, that is, the algorithmic threat. That is, algorithms, even simple ones, in a number of domains, already are no worse (and are at times superior) at predicting the performance and fit of candidates than humans, even expert humans (Bishop & Trout, 2005;Cappelli, 2020).
Algorithms can be an effective method for predicting future performance and fit primarily because the hiring challenge at its core is a prediction problem, and statistical algorithms are designed to take on and address prediction problems . For example, a simple statistical prediction rule (SPR) in a linear model is designed to predict a desired property P (e.g., future performance) based on a series of cues (e.g., education, experience, and past performance) such that P = w 1 (c 1 ) þ w 2 (c 2 ) þ w 3 (c 3 ) þ … þ w n (c n ), where c n and w n reflect the value and weight 9 of the nth cue (Bishop & Trout, 2005). Research shows that even this simple statistical algorithm is, at least in overall effect, better than humans in hiring predictions, in part because such a hiring algorithm is more consistent than humans (and cheaper, to boot). And, in practice, this algorithm can be better scaled and automated in a consistent way (Chamorro-Premuzic & Akhtar, 2019). Also, the increasing availability of good data, advances in statistical algorithms, and new capacities to analyze large-scale data have made this algorithmic route even more promising (Cappelli, 2020).
Indeed, more advanced statistical hiring algorithms based on machine learning can be better than humans at predicting performance and fit because they are specifically designed to "adaptively use the data to decide how to trade off bias and variance to maximize out-of-sample prediction accuracy" (Chalfin et al., 2016: 124). In this respect, for example, Cowgill (2019) finds that more advanced statistical hiring algorithms based on machine learning better predict job performance than humans because they lack some of the biases from which humans suffer. Also, Chalfin et al. (2016) find that, compared to the existing rank-ordering police hiring systems, machine learning algorithms that use sociodemographic attributes; prior behavior, including prior arrest records; and polygraph results would yield a 4.8 percent reduction in police shootings and physical and verbal abuse complaints.
In addition to the hiring domain, advanced statistical algorithms based on machine learning have been shown to be more effective than humans in a broader set of screening decisions where "a decision-maker must select one or more people from a larger pool on the basis of a prediction of an unknown outcome of interest" (Rambachan, Kleinberg, Ludwig, & Mullainathan, 2020: 91). For example, Kleinberg, Lakkaraju, Leskovec, Ludwig, and Mullainathan (2018) show that machine learning algorithms exhibit better performance than judges in bail decisions because they incorporate fewer irrelevant perceptions of the defendant (e.g., demeanor) into their decisions. Also, Dobbie, Liberman, Paravisini, and Pathania (2018) illustrate that machine learning algorithms minimize bias against certain types of applicants (e.g., immigrants). Other related studies in lending find that machine learning algorithms are better at predicting default (Fuster, Plosser, Schnabl, & Vickery, 2019) and are less discriminatory compared to face-to-face lenders (Bartlett, Morse, Stanton, & Wallace, 2019).
Critics of algorithmic decision-making in hiring (and elsewhere) raise at least two objections. The first objection pertains to the seeming ability of humans to pick up on soft, qualitative, or noncodifiable cues during interviews that are difficult to capture in algorithms (Gigerenzer, 2007;Highhouse, 2008). However, this is precisely where the research shows that there is a high likelihood and magnitude of bias clouding human decision-making. Indeed, the "speculation that humans armed with 'extra' qualitative evidence can outperform SPRs has been tested and has failed repeatedly" (Bishop & Trout, 2005: 33). Even if we grant that humans are skilled at inferring relevant information from subtle personality and intellect cues, as some research suggests (Gigerenzer, 2007), statistical algorithms often simply pull on the same cues. While many algorithms tend to draw on codifiable cues (rather than biasprone, noncodifiable cues), in contrast to humans, algorithms are more efficient and consistent, and they need not be managed with respect to their sense of self-esteem or self-importance (Chamorro-Premuzic & Akhtar, 2019).
The second objection regarding the algorithmic method of predicting future performance and assessing fit concerns fairness (Cappelli, Tambe, & Yakubovich, 2020;Newman, Fast, & Harmon, 2020;Raisch & Krakowski, 2021;Tambe, Cappelli, & Yakubovich, 2019). In this respect, although legitimate fairness concerns are associated with algorithmic predictions of human performance, research has shown that algorithms are often no worse than the alternative means of hiring, including using human judgment through interviews. For example, using data on teacher and police characteristics, Chalfin et al. (2016) show that statistical algorithms predict future performance better than humans. Though there are indeed fairness concerns with algorithms, these concerns are prevalent in human decision-making too . Specifically, Danieli et al. grant the prevalence of fairness issues in algorithms but also highlight several comparably concerning psychological biases in human judgment. For example, in hiring contexts, humans engage in bracketing (i.e., overemphasizing subsets of choices over the universe of all options), that is, choosing the top candidate who was interviewed on a given day instead of the top candidate interviewed throughout the search process . 10 In addition, Li (2020) summarizes research that shows how human judgment in hiring may discriminate based on race, religion, national origin, sex, sexual orientation, and age. Given this research, Cappelli (2020) warns us not to romanticize human judgment and to recognize "how disorganized most of our people management practices are now." He notes, "At least algorithms treat everyone with the same attributes equally, albeit not necessarily fairly." Indeed, a significant portion of the algorithmic fairness issues arguably stems from human actions, as well as the lack of diversity in the humans who designed them (Li, 2020) and the types of data with which humans trained them (Cappelli, 2020;De Cremer & De Schutter, 2021). For example, Dastin (2018) reports that Amazon's recruiting algorithm was biased against women because it was trained to assess candidates by discovering patterns in submitted résumés over a ten-year time frame-most of those résumés were submitted by men (see also Cappelli, 2019a). 11 As it turns out, recent research challenges the common assumption that biased data in the training stage of machine learning will lead to undesirable social outcomes. Specifically, Rambachan and Roth (2020) empirically examine the "bias in, bias out" assumption and highlight the conditions under which machine learning may reverse bias and ultimately prioritize groups that humans may have marginalized. More specifically, through mathematical modeling and simulation, they show 10 What about the possibility of complementing algorithmic predictions with human oversight? In other words, one might be tempted by the thought that a firm should use both algorithms and its own judgment; that is, one should consider the predictions of the algorithms, but vet these predictions against one's own assessment of the candidate. After all, algorithms will, at least on occasion, offer what seem to be obviously mistaken prescriptions. And if one's intuition contradicts what the algorithm is prescribing in such a case, one might defect from the algorithmic strategy.
Although tempting, this strategy faces serious problems. A crucial lesson from the literature on how to benefit from SPRs (including decision assistance algorithms) is that partial or selective compliance with the strategy results in significantly worse overall outcomes (Bishop & Trout, 2005;Dawes, Faust, & Meehl, 1989;Meehl, 1957). This has been confirmed on multiple occasions in the laboratory context and is a problem in contexts as wide-ranging as medical decision systems and criminal recidivism, as well as in interviews (Bishop & Trout, 2005: 46-47, 91;Goldberg, 1968;Leli & Filskov, 1984;Sawyer, 1966). Specifically, when one opts for a selection strategy based on a SPR (such as an algorithm), but then defects from this strategy on a case-by-case basis-because this particular case seems unique-this yields worse overall outcomes (Bishop & Trout, 2005). This is so even if there is a strong sense that the particular circumstance at hand is somehow exceptional (see the literature on the "broken leg problem" [Bishop & Trout, 2005: 45-46;Dawes et al., 1989;Meehl, 1957] when the decision maker "comes to believe she has strong evidence for defecting from the strategy" [Bishop & Trout, 2005: 46]). In other words, to secure the most overall instrumental benefits of an algorithm, its advice generally cannot be taken a la carte. 11 We recognize that, in some instances, algorithms risk amplifying our biases and can further entrench bad organizational cultures (because firms would use their own past HR decisions as data sets, which can in turn deepen morally untoward hiring practices). In such cases, this is indeed a significant added concern with using algorithms in lieu of humans. This, of course, would undermine the strength of our characterization of the algorithmic threat and, in turn, lessen the force of the puzzle we raise for the traditional view of interviewing, but it does not undermine our ultimate thesis that there are strong grounds for preserving the practice of interviewing-indeed, this would amount to a further independent consideration that supports our thesis. that, unlike the bias generated by measurement errors caused by mislabeled data, the bias generated by sample selection may be flipped by machine learning such that the machine learning outcomes would favor groups that encountered discrimination in the training data. 12 Rambachan and Roth argue that the bias reversal occurs because members of groups that are underrepresented in the original training data, for example, women, that make the cut are typically ones that are statistically outstanding performers. As such, in subsequent rounds of learning, the algorithm is fed data in which women are overly positively correlated with being outstanding performers. Rambachan and Roth show that this can ultimately reverse the underrepresentation in the data that is due to human decision makers.
We have thus far considered two objections to using algorithms instead of interviews, and we've suggested that these objections fall short. Yet one might correctly point out that many more objections to algorithms have recently appeared in the algorithmic ethics literature (Birhane, 2021;Hunkenschroer & Luetge, 2022;Martin, 2019;Müller, 2021;Tasioulas, 2019;Tsamados et al., 2022). For example, there are concerns related to algorithms systemically excluding certain individuals (Creel & Hellman, 2022), eliciting organizational monocultures (Kleinberg & Raghavan, 2021), or disproportionately harming marginalized groups (Birhane, 2021); worries related to the legitimacy and trustworthiness of algorithms (Benn & Lazar, 2022;Martin & Waldman, 2022;Tong, Jia, Luo, & Fang, 2021) and the lack of explainability in the case of opaque algorithms (Anthony, 2021;Kim & Routledge, 2022;Lu, Lee, Kim, & Danks, 2020;Rahman, 2021;Rudin, 2019;Selbst & Powles, 2017;Véliz, Prunkl, Phillips-Brown, & Lechterman, 2021;Wachter, Mittelstadt, & Floridi, 2017); 13 issues related to whether algorithms preclude us from taking people seriously as individuals (Lippert-Rasmussen, 2011;Susser, 2021); and concerns related to whether automated systems create responsibility or accountability gaps (Bhargava & Velasquez, 2019;Danaher, 2016;Himmelreich, 2019;Nyholm, 2018;Roff, 2013;Simpson & Müller, 2016;Sparrow, 2007;Tigard, 2021), among other concerns (Bedi, 2021;Tasioulas, 2019;Tsamados et al., 2022;Yam & Skorburg, 2021). In short, there's now a rich literature involving a wide range of concerns related to adopting algorithms in lieu of human decision makers (Hunkenschroer & Luetge, 2022;Martin, 2022;Müller, 2021;Tsamados et al., 2022). And the thought might be put more forcefully: insofar as these two aforementioned concerns could be objections to using algorithms (and in turn objections to the force of the interview puzzle), many more objections-like the ones articulated in the algorithmic ethics literature-may succeed. 14 We grant the force of this concern. Taken together, the arguments developed in the algorithmic ethics literature constitute a powerful concern regarding using algorithms in lieu of human decision makers. Furthermore, to the extent that these objections to algorithms succeed, it would weaken the strength of the algorithmic 12 The algorithm will continue to replicate and exacerbate any bias generated by measurement errors caused by mislabeled data. 13 See also the related debate concerning trade-offs between interpretability and accuracy (London, 2019). 14 We are grateful to an anonymous reviewer for raising this concern. threat (and, correspondingly, the force of the interview puzzle). However, for our ultimate aims, this does not concern us. This is because our broader project is not to defend algorithms-we do so in the context of the interview puzzle strictly for the sake of argument. Our ultimate aim is instead to argue that even if these wide-ranging objections to the use of algorithms fall short, there nevertheless remain independent moral considerations that tell against abdicating hiring choices to an algorithm. Crucially, the kinds of moral considerations on which we draw do not depend on certain bad outcomes that may arise due to algorithms. This is to say, even if algorithms were not systemically excluding individuals in arbitrary ways (Creel & Hellman, 2022), did not result in an organizational monoculture (Kleinberg & Raghavan, 2021), did not create responsibility gaps (Himmelreich, 2019;Johnson, 2015;Martin, 2019;Matthias, 2004;Roff, 2013;Sparrow, 2007), or did not elicit other morally untoward outcomes, there nevertheless remains an independent moral concern about firms abdicating their choices in the hiring domain to an algorithm. So, the argument we will now provide might be understood as providing further, independent grounds to resist using algorithms (at least in the context of hiring). Moreover, the arguments we offer do not hinge on certain bad outcomes arising due to using algorithms; as such, the force of our arguments remains, even if the bad outcomes associated with algorithms are ultimately engineered away.

Taking Stock of the Interview Puzzle
The behavioral and algorithmic threats present a significant twofold challenge and raise the interview puzzle for proponents of the traditional view of interviewing. To be sure, this does not mean that the traditional view is not, in part, correct. Finding high-performing candidates who fit the job requirements, as the traditional view posits, is plausibly an important end for firms to pursue. However, the behavioral and algorithmic threats, taken in conjunction, challenge whether interviews are a suitable means toward that end. Crucially, if interviews are only about this end, then the interview puzzle remains and threatens to undermine our justification for conducting interviews. We will now argue, however, that there is more to be said on behalf of interviews than the traditional view accounts for. Before proceeding, we offer a brief clarification about an assumption we make in the next section: we treat the interview process as equivalent to a hiring process with human decision makers. But, strictly speaking, this assumption is not always correct. Hiring processes with human decision makers can occur without interviews, because interviews are not the only available basis for selection. For example, tests or work samples might instead be used. However, tests and work samples are apt in a much narrower range of positions. Moreover, as HRM textbooks note, "interviews are one of the most common methods used for selection" (Mathis et al., 2016: 259), and "interviews continue to be the primary method companies use to evaluate applicants" (Mondy & Martocchio, 2016: 165). In fact, "while not all employers use tests, it would be very unusual for a manager not to interview a prospective employee" (Dessler, 2020: 192). For these reasons, we use "the interview process" interchangeably with "hiring process conducted by human decision makers." At the end of section 4, we briefly discuss the implications of relaxing this assumption.

THE VALUE OF CHOICE THEORY OF INTERVIEWS
The interview puzzle can be dissolved once we recognize that interviews play additional roles beyond predicting performance and fit. For this reason, even if the behavioral and algorithmic threats undermine the plausibility of interviews serving as a means toward the end of securing an employee who fits the role's and organization's needs, we need not conclude that the practice of interviewing is unjustified or something that ought to be abandoned: this is because interviews are a source of other kinds of value and are not exclusively a means for predicting performance and fit.
To be clear, on the view we develop, we do not challenge the importance of the end posited by the traditional view (i.e., the end of hiring an employee who fits the role's and organization's needs); rather, we argue that additional kinds of value are implicated in the practice of interviewing. Thus we offer a pluralistic theory of interviewing and argue that once we recognize the wider range of contenders for the kinds of value generated through interviewing, we can see that abandoning interviews would risk the loss of certain important kinds of value.
To understand the additional kinds of value implicated in the practice of interviews, we draw on philosopher T. M. Scanlon's (1988Scanlon's ( , 1998 account of the value of choice. Scanlon's (2013: 12) account "begins from the fact that people often have good reason to want what happens in their lives to depend on the choices they make, that is, on how they respond when presented with the alternatives." His work on the value of choice has been significant for debates and fields of inquiry as wide-ranging as paternalism (Cornell, 2015), bioethics (Walker, 2022), the freedom and moral responsibility debate (Duus-Otterström, 2011;Fischer, 2008), and contract theory (Dagan, 2019).
On the value of choice account, at least three different kinds of value can be generated when making a choice: instrumental, representative, and symbolic. The first is the instrumental value of a choice: if I am the one who makes the choice, I might make it more likely that I realize some end than were I not given the opportunity to choose. So, for example, if I'm a prospective car buyer and am given the choice over what color I want for my car, my making this choice realizes a certain instrumental value: of making it more likely that the car will satisfy my aesthetic preferences (in contrast to, for example, were the dealership to choose the color of the car on my behalf or were the color to be selected using a random color generator). So, the instrumental value in a choice is realized when it makes it more likely that a desired end of a prospective decision maker is achieved.
The second is the representative value of choice: this is the value that is generated when my making the choice alters the meaning of the outcome of the choicecrucially, this value is realized even if my making the choice is instrumentally worse at achieving certain ends than an alternative method of decision-making (e.g., an algorithm, a coin flip, deference to an expert). For example, it's important that I am the one who chooses a gift for my partner, not because I'm more likely to satisfy their preferences than they are (were they to choose the gift themselves), but rather because there is value in the fact that I was the one who chose it; in choosing the gift, I expressed myself (e.g., my desires, beliefs, and attitudes toward my significant other) through that act. More simply, representative value relates to how the outcome of the choice takes on a different meaning in virtue of who makes the choice.
The third is the symbolic value of choice: this is the value associated with certain choices reflecting that one is a competent member of the moral community who has standing that is "normally accorded an adult member of the society" (Scanlon, 1998: 253). For example, if I, as an adult, were not permitted to choose my bedtime, this would be demeaning and infantilizing. This is so even if a sleep specialist choosing my bedtime would result in outcomes better for my circadian rhythm and other physiological markers. My being able to choose reflects the judgment that I am a "competent, independent adult" (Scanlon, 1998: 253). This is the value that is risked when one is denied the opportunity to make certain choices, ones that, in a given social context, are choices that "people are normally expected to make … for themselves" (Scanlon, 1998: 253).
These are the three candidates for the value generated through making a choice. The first is instrumental, and the second two are noninstrumental sources of value. This may not exhaust the candidates for the kinds of value generated in making a choice, but it does taxonomize three important kinds of value that are generated in making a choice. Thus, if a choice is abdicated, (at least) these three kinds of value are at risk and are thus potential candidates for the value that would be lost.
Returning to the context of interviewing, when firms conduct interviews, they are making choices about whom to employ. So, let's now turn to how the value of choice account bears on interviewing. We will discuss each sort of value generated through choice-instrumental, representative, and symbolic-in turn.
The first is the instrumental value of choice. Securing instrumental value is the chief value with which the traditional view of interviewing is concerned. The thought goes as follows: interviewing realizes the instrumental value to the extent that it helps the firm predict a candidate's performance and fit. Those who are inclined to preserve interviews, on the basis of the traditional view of interviewing, might expect that the instrumental value of choice realized in interviewing-helping a firm better predict a candidate's performance and fit-is what both explains why we interview and also what justifies its costs.
Yet the instrumental value of interviewing is precisely what is called into question by the interview puzzle. Interviewing does not excel at generating the purported instrumental value that it is thought to elicit (namely, predicting future performance and fit). So, if the sole kind of value that could be generated through interviewing is instrumental value, then the grounds for the practice are undermined. But as the value of choice account tells us, there is a wider range of contenders for the kinds of value generated in making a choice. The critical oversight of the traditional view is its failure to recognize that the value generated through interviewing is not entirely conditional on the instrumental value of choice, given that there can be noninstrumental value generated through the choice.
This brings us to the second potential value-and one overlooked by the traditional view-that is realized through interviews: the representative value of choice. As Scanlon (1998, 253) points out, we value and want certain choices to "result from and hence to reflect [our] own taste, imagination, and powers of discrimination and analysis." In the interview context, we may value the fact that we are the ones choosing with whom we work, and there is value lost (i.e., representative value) when we abdicate that choice, even if our choosing does not as effectively realize the ends of predicting performance and fit as an algorithm. An algorithm might be better at predicting which romantic partner we should date, whom we should befriend, or which university we should attend-while this all might be correct, abdicating these choices and deferring to an algorithm would result in us losing something of value: representative value. Choosing to whom we relate in the workplace is a way "to see features of ourselves manifested in actions and their results" (Scanlon, 1998: 252). The representative value of a choice is the value that arises in virtue of the choice taking on a different meaning: because of both the fact of who makes the choice and the choice representing or expressing the person's judgments, desires, and attitudes.
The third value generated through interviewing, and another oversight of the traditional view of interviewing, is the symbolic value of choice. Scanlon (2019: 4) points out, "If it is generally held in one's society that it is appropriate for people in one's position to make certain decisions for themselves, then failing to make such a decision for oneself or being denied the opportunity to make it, can be embarrassing, or even humiliating." Thus the symbolic value of choice is what is lost when a person for whom it would be appropriate (in a given social context) to make a certain decision is precluded from making that decision. For example, to the extent that workplace norms in a given society involve members of an organization typically having a choice in their future colleagues-people with whom they would collaborate but also, in some cases, those whom they would befriend or with whom they would commiserate and form community (Casciaro, 2019;Estlund, 2003;Porter, Woo, Allen, & Keith, 2019)-through interviewing, depriving people of that choice may result in a loss of symbolic value. 15 Relatedly, a certain prestige and status are implicated in making certain choices (including selecting future colleagues through interviewing) that figure into the symbolic value of choice; this is especially vivid, for example, when alumni of a university are involved in on-campus recruiting at their alma mater (Binder, Davis, & Bloom, 2015). This prestige and status that are implicated in the symbolic value of choice are also part of what would be lost were firms to forsake interviews. Crucially, substituting interviews with algorithms can result in a loss of symbolic value even if, as a matter of fact, an algorithm may arrive at a better assessment of a candidate's expected performance and fit. 16 15 For a discussion of the downsides of workplace friendships for organizations, see Pillemer and Rothbard (2018). 16 It is worth noting that the term algorithm is often used to refer to multiple different kinds of processes, systems, and technologies (Leavitt, Schabram, Hariharan, & Barnes, 2021). For instance, some algorithms are rule based (or symbolic) systems, whereas others are association-based systems. Within these broad and rough categories are many varieties of algorithms and ways in which they might be combined and used. For the purposes of our argument, we put to one side the details regarding the technical specifications of algorithms while merely noting that the extent to which a value of choice is undermined by abdicating the choice to an algorithm may also depend on the type and nature of the algorithm.
Although the representative value of choice and the symbolic value of choice may seem similar, especially because, as Scanlon (1998: 253) puts it, "representative and symbolic value may be difficult to distinguish in some cases," they are not the same. Symbolic value concerns how making certain choices reflects one's standing, whereas representative value concerns how the meaning of a certain outcome depends on who is making the choice that elicited the outcome. Despite these differences, both are kinds of noninstrumental value, and neither depends on the instrumental effectiveness of the choice with respect to some end (Aristotle, 1962;Donaldson, 2021;Donaldson & Walsh, 2015;Gehman, Treviño, & Garud, 2013;Kant, 2012;O'Neill, 1992;Zimmerman & Bradley, 2019).
Our interviewing practices can be vindicated once we recognize that the choice involved in the interview process can realize both representative and symbolic value.
The key point is that "the reasons people have for wanting outcomes to be dependent on their choices often have to do with the significance that this dependence itself has for them, not merely with its efficacy in promoting outcomes that are desirable on other grounds" (Scanlon, 1998: 253). And the fact that representative and symbolic value are threatened when abdicating the choice involved in interviewing a candidate-the choice of whom to relate to in the workplace-generates pro tanto moral reason to preserve interviews as an organizational practice. Crucially, the representative and symbolic value undergirding our interview practices is not imperiled by the behavioral or algorithmic threats.
In other words, once we recognize the broader range of contenders for the kinds of value generated through interviewing, we can see that the behavioral and algorithmic threats only undermine part of the potential value in interviewing-its instrumental value. But we still have pro tanto moral reason to continue the practice of interviewing, given the noninstrumental value-representative and symbolic value -that may be lost were we to abandon the practice.

CLARIFICATIONS AND OBJECTIONS
We now turn our attention to a few clarifications and some potential objections. First, it's worth keeping in mind that even the noninstrumental values in a choice do not always tell in favor of preserving, rather than abdicating, a choice. For example, with respect to representative value, we might prefer, in some circumstances, for our choices not to reflect our judgments, desires, and attitudes. If one's organization is considering hiring one's close friend, one might prefer to have the "question of who will get a certain job (whether it will be my friend or some well-qualified stranger) not depend on how I respond when presented with the choice: I want it to be clear that the outcome does not reflect my judgment of their respective merits or my balancing of the competing claims of merit and loyalty" (Scanlon, 1998: 252). In other words, in circumstances that might present a conflict of interest, for example, there might be reasons related to representative value that tell against preserving the choice.
Second, the value of choice is not simply about having a greater number of options from which to select. This is to say, the value of choice generates reasons that "count in favor of 'having a choice,' but for reasons of all three kinds having more choice (over a wider range of alternatives) is not always better than less. Being faced with a wider range of alternatives may simply be distracting, and there are some alternatives it would be better not to have" (Scanlon, 2019: 4). So, in the context of interviewing, we remain agnostic about how the value of choice is affected by having more candidates from whom to select.
Third, one might doubt whether symbolic value would in fact be risked were we to forgo interviews. The point might be pressed as follows: because many (or even most) employees are not involved in hiring decisions, it is not clear that symbolic value would be lost (or that the failure to be involved in the interview process would be demeaning). 17 We grant that symbolic value may not be risked in many instances of abdicating a choice. But this clarification points the way to an advantage of our value of choice account: its contextual sensitivity. As Scanlon (1998: 253) notes, a key point with respect to whether symbolic value is risked in a given situation is whether the situation is one "in which people are normally expected to make choices of a certain sort for themselves." Ascertaining whether there is such an expectation in place in a given hiring context and, in turn, whether symbolic value would be lost will depend on certain sociological facts pertaining to the expectations in the given workplace and the norms governing that workplace culture, field, or industry. 18 This means that there is an important role for empiricists to play in ascertaining the workplace contexts, fields, or industries in which symbolic value is risked to a greater or lesser extent. And in contexts in which the strength of the norms associated with choosing members of one's organization are weaker, the reasons provided by the symbolic value of choice would be correspondingly weaker.
Fourth, one might raise the following question: what about organizations that outsource hiring to an external head-hunting firm? On our view, such an approach would, in effect, be morally akin to abdicating the choice to an algorithm, with respect to the value of choice. That said, there might be other sorts of considerations -for example, the various objections discussed in the algorithmic ethics literature 17 We are grateful to an anonymous reviewer for this point. We also acknowledge that many hiring decisions are made by internal HR divisions. But it is worth noting that even if these members of HR divisions may not ultimately work with the people they are hiring (unless, of course, the interview is for an HR position), the members of these HR divisions are themselves usually employees of the organization too. Moreover, in a number of fields, it is not uncommon in the final rounds of interviews for candidates to be interviewed by individuals who would be their immediate team members and managers if selected for the position.
18 Suppose a firm is deciding on candidates as a collective by using some sort of majoritarian procedure that nevertheless results in an outcome that is no individual's most preferred choice (List & Pettit, 2011;Pettit, 2007). First, does the individual's choice still matter? Our aim here in this article is not to enter the debate regarding the metaphysics and morality of group agents. That said, we note that the value of choice of the individual still matters, given that it is a key component of fixing the collective's choice. It is quite unlike cases in which an individual's choice (arguably) may not matter due to an outcome being causally overdetermined. That an individual's most preferred choice was not instantiated is a different matter from the value realized through making the choice. Second, such a collective decision procedure seems morally unobjectionable-could automating it render it objectionable? It may very well, albeit perhaps not for reasons related to the value of choice. This is because automating a procedure can change its very nature, morally speaking, for reasons of the sort discussed in the algorithmic ethics literature. We are grateful to an anonymous reviewer for these two questions. mentioned earlier-that make relying on algorithms morally worse than abdicating the choice to an external head-hunting firm. Still, it is quite right that the value of choice-related considerations would be morally akin. But this need not mean that there is no role for external head-hunting firms at all. This is because the concerns with respect to the value of choice primarily arise insofar as the firm defers to the judgment of the external head-hunting firm. This, however, does not preclude soliciting advice about hiring decisions from HR consultants or head-hunting firms. Notably, in the context of algorithms, deference to the algorithm is much more likely given that many algorithms are opaque. Moreover, failing to defer to the judgments of the algorithm-that is, picking and choosing on a case-by-case basis when to follow its prescriptions-drastically undercuts its overall instrumental benefits (Bishop & Trout, 2005).
Fifth, perhaps, all things considered, in some instances the costs of interviewing may be too burdensome and a firm might be forced to forgo the practice. Perhaps, in other instances, the importance of finding the right person is far too weighty-for example, selecting an airline pilot-for a human to make the decision if an algorithm would do so more effectively. But even in these cases, were we to abandon interviewing for a different selection method (e.g., an algorithm), it's worth keeping in mind that there may still be something of normative significance lost, that is, representative or symbolic value. 19 How might these trade-offs be managed? One potential approach might be as follows: suppose one regards instrumental value to be of much greater significance in the business realm than the sorts of noninstrumental value to which we've drawn attention. In such a case, a hybrid approach might be considered. Such an approach might involve conducting the initial screening with an algorithm and leaving the ultimate decision to a member of the organization. This may allow for reducing the potential trade-offs between the instrumental and the noninstrumental sources of value of choice. 20 In other words, our view is not that, in instances when an algorithm is vastly superior at achieving a given end, firms should pursue the drastically less instrumentally effective approach. As Scanlon (2019: 4) notes, the various reasons for the value of choice "can conflict with reasons of other kinds, particularly with instrumental reasons." So, we are not claiming that firms must always conduct interviews, instead of using algorithms. Nor are we claiming that the instrumental considerations are not of moral significance-in some instances, they may very well be of overriding moral importance. 21 Rather, our point is that multiple kinds of value can be 19 Quite apart from the representative or symbolic value that is risked when abdicating a choice to an algorithm are concerns about how doing so might undermine organizational learning (Balasubramanian, Ye, & Xu, 2022). 20 Of course, as earlier noted, picking and choosing when to comply with the predictions of the algorithm significantly undercuts the overall instrumental benefits of the algorithm (Bishop & Trout, 2005). Insofar as one pursues such a hybrid approach, it's worth keeping in mind that the various other moral objections to the use of algorithms discussed in the algorithmic ethics literature would still be relevant. 21 Suppose a physician faces two options: interpret medical images herself or rely on a predictive algorithm. Further suppose that the algorithm yields better instrumental results with respect to patient welfare. generated through the practice of interviewing-including sources of value that may generate conflicting reasons-and that an adequate theory of interviewing should not overlook this fact. If we are to abdicate interviews in a given context, we should do so in full view of the kinds of value that are risked. 22 Sixth, it's now worth revisiting the assumption we articulated at the end of section 2: treating the interview process as equivalent to a hiring process with human decision makers. As we acknowledged, this assumption is not always, strictly speaking, correct. A hiring process-including one in which humans are making the decisions-might not involve interviews at all; perhaps the hiring process involves choosing on the basis of work samples or tests.
So, when we relax this assumption, what follows? Our view would still imply that abdicating the hiring process entirely to algorithms would risk the various values of choice. However, our value of choice account does not entail a particular mode of choosing for a human decision maker-whether interviews, work samples, or tests. With respect to the narrow range of professions where work samples or tests can aptly be implemented, our value of choice arguments are neutral between choosing such an approach and interviewing (but of course, the value of choice account is not neutral between either of these routes and abdicating the choice to an algorithm). 23 Interviews are a way-the most prominent and common way, and the way most broadly applicable across a range of positions-for us to choose the members of our organizations, but they are indeed not the only way to choose in the hiring process.
To summarize, we have offered an account of some heretofore underappreciated normative dimensions of a widespread business practice, namely, interviewing. Our view helps address some of the challenges to which the traditional conception of interviewing succumbs. The traditional view has difficulty explaining why interviews persist and justifying why we should not abandon them, given their costs, our poor ability to predict performance and fit, and the presence of algorithmic alternatives. Our value of choice theory of interviewing both explains why interviews persist and justifies why there are grounds not to abandon the practice: interviews Must the physician insist on making the choice herself? Our view does not rule out that the physician should rely on the algorithm here-in other words, there may very well be cases that the good or bad at stake is so weighty that the instrumental value of relying on the algorithm swamps the various values of choice that may be realized in making the choice oneself. We are grateful to an anonymous reviewer for this example. 22 Our argument is neither about the badness of having fewer choices to make nor about the goodness of having more choices to make (nor is it about preserving the status quo number of choices one makes). With respect to the value of choice, that some other choice is made (e.g., to defer to an algorithm) has little bearing on whether, what kind, and the extent to which one of the values of choice would be undermined in abdicating this choice. Adding a choice elsewhere doesn't somehow replenish the value of choice that is undermined in no longer choosing one's colleagues. 23 Of course, the various ways in which we are bad at interviewing characterized in the behavioral threat section might tell in favor of choosing by way of these alternative modes of selection (e.g., tests or work samples) when possible. But we hesitate to make this judgment with confidence, given that different kinds of normative concerns may be associated with relying strictly on work samples or tests; for example, it potentially reduces people to a contrived and narrow set of criteria, rather than treating them with respect as individuals and as fellow members of the moral community. For an additional approach to hiring, see Sterling and Merluzzi's (2019) exploration of "tryouts" and their theoretical and practical potential. play an important normative function by securing noninstrumental sources of value in hiring.

FUTURE AVENUES OF RESEARCH
Our value of choice account of interviewing suggests several new avenues of research. First, a significant body of research in employment ethics primarily emphasizes the ethics of how employers ought to treat their employees (Arnold, 2010;Barry, 2007;Bhargava, 2020;Brennan, 2019;McCall, 2003;Werhane, Radin, & Bowie, 2004), but there is much less, apart from discrimination-related issues, surrounding the ethics of what is owed to prospective employees. Our work highlights the significance of a range of understudied issues to explore in this domain. Although some have explored the question of what is owed to former employees of a firm (Kim, 2014), what, if at all, is owed to potential employees, such as candidates who participate in interviews? Other such issues include, for example, the ethics of exploding offers, accepting applications from candidates that will never be considered, and alerting candidates of rejection. On the side of the candidate, issues include the ethics of feigning enthusiasm for an interview, pursuing an interview merely to solicit an external offer for negotiation leverage, and holding on to offers that one is confident one will not accept.
Second, our account of interviewing points the way to questions related to what may make employment relationships meaningful (Robertson, O'Reilly, & Hannah, 2020). Some contributors to the future of work scholarly conversation have argued that employers owe it to their employees to provide meaningful work (Bowie, 1998;Kim & Scheller-Wolf, 2019;Michaelson, 2021;Veltman, 2016). 24 By attending to the broader range of values associated with interviewing, managers may have the opportunity to make work and employment relationships more meaningful (Bartel, Wrzesniewski, & Wiesenfeld, 2012;Freeman, Harrison, Wicks, Parmar, & De Colle, 2010;Rosso, Dekas, & Wrzesniewski, 2010). So, an important question to address will be how the process of being selected for a position (i.e., through an interview or through selection by way of an algorithm) can contribute to preserving or promoting the meaningfulness of work (Carton, 2018;Grant, 2012;Jiang, 2021;Kim, Sezer, Schroeder, Risen, Gino, & Norton, 2021;Rauch & Ansari, 2022).
Third, there is a sense in which using algorithms in hiring decisions deepens the informational asymmetry between candidates and employers (Curchod, Patriotta, Cohen, & Neysen, 2020;Yam & Skorburg, 2021: 614). Switching to algorithms in hiring may prevent candidates from developing a better understanding of their prospective colleagues and the prospective employer's workplace culture and norms. On the other hand, if an interview was conducted, the candidate might have acquired this sort of valuable information, even if fallibly. Future scholars should explore the public policy implications of forgoing interviews, especially in jurisdictions with employment at will. The symmetrical right to exit is sometimes discussed as a potential justification for employment at will (Bhargava & Young, 2022;Hirschman, 1970;Maitland, 1989;Taylor, 2017). But when candidates and employers enter the employment relationship on starkly asymmetric informational grounds (Caulfield, 2021), it's worth exploring whether the fact of both parties having a right to exit the relationship loses some of its justificatory force with respect to employment at will and considering whether supplementary regulatory constraints would be in order.

CONCLUSION
The traditional view of interviewing espoused by both practitioners and management scholars alike holds that interviews are conducted-despite the steep costs associated with the process-to predict a candidate's performance and fit in relation to a vacancy. We argue that the traditional view faces a twofold threat: the behavioral and the algorithmic threats. The behavioral threat arises in virtue of a large body of behavioral evidence that points to us being poor predictors of future performance and bad judges of fit. The algorithmic threat arises in virtue of algorithms already being superior predictors of performance and fit than us in a number of domains, including the hiring domain.
If the traditional view of interviewing captures all there is to interviewing, then the justification for conducting interviews is undermined by the behavioral and algorithmic threats. However, we argue that the practice of interviewing can be vindicated once we recognize that there are a broader range of contenders for the kinds of value that can be realized through interviewing-crucially, some of these kinds of noninstrumental value that are realized through interviewing remain insulated from the behavioral and algorithmic threats. In short, we argue that even if algorithms are better predictors of performance and fit than us, it does not follow that we ought to abandon our interview practices: this is because important kinds of noninstrumental value are generated through interviewing that could be lost were we to forgo the practice.