Why Do We use Expert Interviews?
Experts are the observers of and mechanics behind what social scientists call “causal mechanisms” (Elster Reference Elster, Hedström and Swedberg1998). Where researchers often gather data of onset points and examine the outcomes of political decisions, rare are the possibilities of looking into the inner workings of political processes. Here interviewing experts becomes key to understanding how “X and Y interact” (Gerring Reference Gerring2017, 45). In this paper I provide suggestions on how the important data collection technique of qualitative expert interviews can be used in the most effective manner to discern decision-making and institutional behavior.
Current “quantitative and causal inference revolutions”Footnote 1 (Pepinsky Reference Pepinsky2019, 187; for a critique, see Elman and Burton Reference Elman and Burton2016) have dominated the methodological debate in the social sciences. Scholars conduct causal investigations with experimental or quasi-experimental designs (Dunning Reference Dunning2012) and with observational time-series cross-sectional data (Blackwell and Glynn Reference Blackwell and Glynn2018).Footnote 2 These developments have led to a shift in mainstream political science research toward investigating micro-level processes and from questions of external validity to ones of internal validity (Pepinsky Reference Pepinsky2019). This application of rigorous methods of causal inference has greatly expanded our knowledge about political behavior and its outcomes. However, despite the fact that quantitative approaches have driven the recent methodological debate in comparative politics and international relations,Footnote 3 the data collection technique of qualitative expert interviews still has an important and arguably even growing role to play “within a discipline addicted to causation” (Anderl Reference Anderl, Bennett and Checkel2015, 2).
As a tool to investigate causal mechanisms, expert interviews hold three advantages. First, they may add to experimental findings about micro processes and how decisions were made in practice (Fu and Simmons Reference Fu and Simmons2021). Researchers can blend general results with context-specific information that is often not in the public domain. In this way, qualitative information from expert interviews facilitates the interpretation of correlational analysis and can thereby improve causal inference by statistical means (Gerring Reference Gerring2017, 44–45; Glynn and Ichino Reference Glynn and Ichino2015; Kabeer Reference Kabeer2019). Second, political science recurrently deals with “big” questions that do not lend themselves easily to experimental or statistical analysis, particularly in cases where the number of observations is low. Experts can aggregate and weigh different pieces of information (refer to the section “Using expert interviews”). Third and relatedly, expert interviews can provide the data to link the macro and micro levels of analysis. For instance, in peace and conflict research interviewing experts may shed light on the impact of insurgents’ international sponsors on local peacebuilding dynamics (Balcells and Justino Reference Balcells and Justino2014) or on how the mobilization of individuals in civil war is linked to social structures (Shesterinina Reference Shesterinina2016).
However, despite its high practical relevance and the existence of classic works on the qualitative interview method (Aberbach and Rockman Reference Aberbach and Rockman2002; Dexter Reference Dexter2006 ; Tansey Reference Tansey2007), current reflection on the promises and limits of this technique is sparse. There is also little guidance offered on how to use the method most effectively (Fujii Reference Fujii2017; Lareau Reference Lareau2021). It is my key contention that pursuing an encompassing understanding of “experts,” one based on the knowledge of individuals, allows for the construction of a systematic “architecture” (Trachtenberg Reference Trachtenberg2006, 32) of data sources and increases expert interviews’ analytical value.
Oftentimes, researchers use qualitative expert interviews as a key method to gather information about political processes but only mention the technique in passing. The resulting problems are pertinent: 1) the selection of experts regularly does not follow clear guidelines; 2) experts’ personal biases are often not tackled; and 3) evidence is not captured systematically. These omissions severely inhibit the potential of qualitative expert interviews to trace causal mechanisms, which is their key potential strength.
In response, I seek to contribute to systematically applying qualitative expert interviews in a more effective manner. I situate these recommendations in respect to two particular bodies of work: a) an own research project on political interventions into the tax administration; and b) examples from the literature on political violence and civil wars. The paper’s contribution is threefold: My first argument is that for most theoretical and empirical research questions of interest, social scientists should focus on a broad notion of experts. When choosing experts, researchers should select both “inside” and “outside experts” and include them in an integrated framework. Second, I propose concrete measures for dealing with the technique’s main challenges of selection of experts, personal biases, and systematic capturing of evidence. Third, I argue that the combination of expert interviews with other methods such as list experiments and online interviewing are key means to mitigate practical challenges and to reduce social-desirability bias. These strategies serve to make qualitative expert interviews even more effective research tools.
Using Expert Interviews
Broadly understood, experts have specific knowledge about an issue, development, or event. Hence, following Dexter’s (Reference Dexter2006) classic understanding, an expert is any person who has specialized information on or who has been involved in the political or social process of interest. Consequently, I advocate an encompassing notion of experts. They “might be academics, practitioners, political elite, managers, or any other individuals with specialized experience or knowledge” (Maestas Reference Maestas, Atkeson and Alvarez2018, 585). Tapping into their insights responds to the fundamental challenge that a lot of issues of interest in political science and in the social sciences more generally are not directly observable, documented, or made transparent.
Most publications on qualitative interviews in political science focus on elites rather than experts. However, an encompassing notion of experts allows us to explicitly situate their status, knowledge, own interests, and potential biases in relation to each other, and to construct a coherent “architecture” of interview sources. With this “realist approach,” expertise is based on real knowledge (Collins and Evans Reference Collins and Evans2007, 3). Following Hafner-Burton et al. (Reference Hafner-Burton, Hughes and Victor2013, 369; also Pakulski Reference Pakulski and Darity2008), elites “occupy top positions in social and political structures” and “exercise significant influence over social and political change.” This status-oriented perspective is sensible for research that focuses on the perspective of a particular set of decision makers, for instance members of parliament (on Russia, see Rivera, Kozyreva, and Sarovskii Reference Rivera, Kozyreva and Sarovskii2002). However, members of the elite are not necessarily knowledgeable about the political process or event of interest. In this case, they do not qualify as experts and the researcher should desist from selecting them as interview partners. For most research objectives, the selection of expert interview partners should therefore be problem- and expertise-centered rather than status-oriented.Footnote 4 Yet selection based on expertise and on status may overlap. In other words, individuals can both hold expertise and be members of an elite group.
We can distinguish three broad functions of expert interviews: to 1) (inductively) explore a research topic/generate hypotheses; 2) collect data for qualitative and mixed-method designs (for instance to situate findings from experimental studies in real-world environments); and 3) generate quantitative data and allow for statistical inference. Based on the existing literature, I differentiate four main applications of qualitative expert interviews: assessment, aggregation, anticipation, and affirmation.Footnote 5 The first and most relevant is assessment, meaning that experts share their judgement on political and social processes. Most importantly, assessment entails that experts in an analytical way “reconstruct an event or set of events” (Tansey Reference Tansey2007, 766), often providing the empirical basis for the process tracing method (Beach and Pedersen Reference Beach and Pedersen2019; Bennett and Checkel Reference Bennett, Checkel, Bennett and Checkel2015).Footnote 6 Thus, the unit of analysis for expert interviews is a particular event, development, or decision making process.
The second, related function is aggregation, as experts are well-suited to reducing real-world complexity and bundling together multifaceted phenomena. Compared to the assessment function, aggregation can also entail the reconstruction of events and provision of information, but in a more descriptive manner. Third, experts can use their research or personal experience for anticipation and the prediction of events, of actors’ behavior, and/or of long-term developments—for instance the propensity for future violent conflict (Hanson et al. Reference Hanson2011; Meyer, De Franco, and Otto Reference Meyer, De Franco and Otto2019). However, the accuracy of expert prediction vis-à-vis other methods, in particular data-based approaches, is highly contested (Hanson et al. Reference Hanson2011; Hegre et al. Reference Hegre2021; Tetlock Reference Tetlock2005).
Finally, expert interviews may serve as a method of affirmation, meaning the confirmation or disproving of prior research results, information from other sources, or anecdotal evidence (Tansey Reference Tansey2007, 766–67). This function has a dark side to it: selectively invoking expert insights to support a partisan purpose or to undergird political action potentially distorts results and is unethical. As I will outline with examples from a research project on particularistic interventions into the tax administration as well as from the literature on political violence, purposefully selecting experts, dealing with their potential biases, and systematically relating different expert interviews to each other are key means of avoiding this pitfall.
The Selection of Experts
Qualitative expert interviews clearly lend themselves to purposeful, non-probability sampling (Goldstein Reference Goldstein2002; Tansey Reference Tansey2007), as expert judgements are inherently personal and not necessarily representative or replicable. Prior thorough reflection on the knowledge, but also potential information gaps and personal biases of experts must therefore guide the interviewee selection process. For this, I suggest a) to integrate inside and outside experts into one common analytical framework and b) to not only focus on high-level but also lower-level inside experts.
Inside versus outside experts: Most fundamentally, I suggest selecting both experts who make decisions (“inside experts”) and ones who analyze them (“outside experts”) as respondents (see table 1). Inside experts are decision makers who actually shaped the political or social process of interest. In this case, expert interviews are a proxy for participant observation (Pouliot Reference Pouliot, Bennett and Checkel2015, 247). Former tax administration officers, rebels, members of Congress, or international organization representatives can generally also be considered “insiders.” In fact, they oftentimes represent particularly reliable sources as they are more inclined to provide information after they have left the organization in question.Footnote 7 In contrast, external experts are outsiders to the process in question. They gain their expertise through research, experience, or interaction with policymakers and officials who took a decision. In contrast to insider experts, they themselves are not the object of analysis. Table 1 introduces guidelines for distinguishing inside from outside experts.
The categorization of inside and outside experts always needs to be made with respect to the particular research agenda and the function of expert interviews for the respective project. If political interventions into the tax administration are the focus of research, then respondents from the ministry of finance as well as tax officials on different hierarchical levels would clearly be regarded as inside experts, while taxpayers, advisors, and social scientists would be considered outside experts.
The differentiation between insiders and outsiders has key implications for expert interviews’ analytical value and the interpretation of the insights provided (Beach and Pedersen Reference Beach and Pedersen2019, 207–9). Interviewing inside and outside experts holds promises but also pitfalls (see the summary in table 1): First of all, while inside experts may provide detailed first-hand accounts and “hidden” knowledge that is not publicly available, they also have a stronger interest in withholding or molding information to “look good” and to inflate or downplay their influence. In contrast, outside experts may be in a better position to provide the “big picture,” meaning the assessment of processes and events as well as the aggregation of various sources. On the downside, they themselves rely on information from others and might lack knowledge of how exactly choices were made (e.g., which emotions were involved in a particular decision).
Given these advantages and disadvantages, I suggest that ideally researchers select both inside and outside experts and include them in an integrated analytical framework. The heuristic differentiation of both groups allows us to relate them to each other systematically. For their purposeful selection through the recommendations of other experts, snowball sampling and related techniques can be used (Goldstein Reference Goldstein2002; Heckathorn and Cameron Reference Heckathorn and Cameron2017; Shesterinina Reference Shesterinina2016, 415). Outside experts should be selected on the basis of their publication records, their (local) expertise, and, if applicable, their prior work for high-quality country-based indices such as the Bertelsmann Transformation Index (BTI 2021).
The case for mid- and low-level insiders: Oftentimes, scholars disregard the influence and informational advantage of inside experts who do not occupy top positions—be it election campaigners, civil society activists, or mid-level civil servants. For example, to investigate political intervention into the tax administration in two African countries (von Soest Reference von Soest2009, 47), I interviewed tax officers from the operational level and systematically related their perspectives to those from the tax agency leadership. It was thereby possible to contrast the information provided by the two groups and to get evidence on both the interference in the hiring and career trajectories as well as the day-to-day work of those officers who actually collect taxes and conduct tax audits on the ground. Respondents mentioned that several companies owned by government politicians had never been audited and at no time paid tax. “A common phrase employees heard from their managers in these cases was ‘don’t go to this man, this man is difficult’” (von Soest Reference von Soest2009, 107). On the other hand, the interviews pointed to instances of resistance and the insistence on preserving administrative standards.
As George and Bennett (Reference George and Bennett2005, 103) state: “Often, lower-level officials who worked on an issue every day have stronger recollections of how it was decided than the top officials who actually made the decision but who focused on the issues in question only intermittently.” In addition, lower-placed individuals are often less trained to provide a “polished” version of events. They are thus regularly more helpful in reconstructing political and social processes at the micro level. Thus, if possible, researchers should strive to interview inside experts—ones involved in the process in question—on different hierarchy levels (primary sources), in addition to outside experts (secondary sources).
In addition to the selection of both inside and outside experts, it is of crucial importance to limit the possibility that respondents’ personal biases compromise the results and thereby limit expert interviews’ analytical value.
Dealing with Personal Biases
Generally, a semi-structured format with defined topics and preformulated questions is most appropriate for expert interviews (Tansey Reference Tansey2007). They are guided by clear themes, keywords, and established questions while simultaneously allowing for follow-up enquiries and probes. In this way, they represent a useful combination of structure and flexibility, and facilitate both comparability and context sensitivity. To realize the full potential of this powerful data collection tool, I now put forward suggestions on how scholars can actively tackle two core personal biases when conducting expert interviews: reactivity and subjectivity.
Reactivity: The Relational Nature of Expert Interviews
Social scientists have long discussed the fact that individuals might adjust their behavior and utterances when participating in studies (Orne Reference Orne1962). Scholars—particularly those working with an interpretivist orientation—have stressed the “relational nature” of the interview process, meaning that both expert interviewees and researchers influence its outcomes (Berry Reference Berry2002; Dexter Reference Dexter2006; Fujii Reference Fujii2017; Peabody et al. Reference Peabody1990). An extremely insightful literature has reflected on the positionality of academics (often from the Global North) and of local interviewees, particularly in post-conflict contexts (e.g., Krause Reference Krause2021; Mwambari Reference Mwambari2019).Footnote 8 By design, the expert interview is a personal, face-to-face conversation between at least two individuals. It is principally asymmetric in that the researcher poses questions and the expert provides their assessment.
Yet interviewing needs to be understood as a two-way dialogue rather than a one-way interrogation (e.g., Cramer Reference Cramer2016; Fujii Reference Fujii2017). Doing research on the tax administration’s political environment in two African countries, even seemingly neutral outside experts (social scientists and employees of international organizations) were understandably well aware of their answers’ implications: a) despite prior information, they repeatedly asked me about the background to my research (“Are donors funding your project?”); b) some experts inquired about my and others’ judgements (“You have certainly done a lot of other interviews—what did they say?”) (von Soest Reference von Soest2009).
The interviewer influences the nature of the interview in at least two ways: by who they are and by how they pose questions. First, the researcher’s status, years of experience, gender, age, nationality, and further personal traits may affect the answers respondents give. Experienced researchers have an advantage in establishing their expertise—such as on the organization of the tax administration, United Nations peacekeeping missions, or the structuring of militant rebel groups—and consequently are taken seriously by the respondent, while inside experts—be they tax officers, soldiers, or (former) insurgents—might conversely see young academics as “harmless” outsiders and therewith more readily share information with them (Autesserre Reference Autesserre2014, 284–86). Researchers should therefore reflect on these matters ex ante and transparently report on how their positionality vis-à-vis experts—particularly insiders—might have influenced the interview situation and the information provided (George and Bennett Reference George and Bennett2005, 99).
Second, the questions posed and the interviewer’s own reactions may affect respondents’ answers. It is generally accepted that to elicit informed and unbiased insights, scholars should take a neutral stance and work with nonsuggestive questions. Yet the reactive nature of expert interviews works to the researcher’s advantage in a twofold manner. First, they can work with probes—following up on answers or making reference to other interviews or sources—to address contradictions. Furthermore, the researchers can use various strategies to obtain all the information the respondent is willing to share (Dunning Reference Dunning, Bennett and Checkel2015, 233). Using an approach from criminal investigation, researchers can for instance pose similar follow-up questions using different wordings and formulations. This strategy proves particularly useful when tracing sensitive issues such as political intervention into the tax administration or war crimes (von Soest Reference von Soest2007; Fujii Reference Fujii2010). Second, the researcher can use the respondent’s posture and (nonverbal) reactions to questions like nodding, hesitation, gestures, or expressively strong affirmation as a further source of information. This interview “meta-data” (Fujii Reference Fujii2010, 231) helps to trace the meaning respondents attach to events or processes (Pouliot Reference Pouliot, Bennett and Checkel2015).
Subjectivity: Lacking Memory and Active Misrepresentation
Even more so than other data sources such as archival material, expert interviews are acts of “purposeful communication” (George and Bennett Reference George and Bennett2005, 99). The information provided is always subjective and colored by the experts’ worldviews, interests, employment status, and cognitive abilities. Furthermore, as Berry (Reference Berry2002, 680) soberingly notes, “it is not the obligation of a subject to be objective and tell us the truth.” Experts can intentionally or unintentionally mispresent information (see table 2 for different potential biases).
This problem is particularly acute for internal experts, and even more so for those who are highly exposed to public scrutiny such as policymakers (Trachtenberg Reference Trachtenberg2006, 154). Yet external experts also have their ideological and personal predispositions that affect the answers they give. Analyzing highly contentious issues such as the support for violence or working in post-conflict and authoritarian environments makes social-desirability bias particularly salient (Lyall Reference Lyall, Bennett and Checkel2015, 204; Tripp Reference Tripp2018). Tax officers hardly instantly or directly spoke about corrupt practices in the tax administration (von Soest Reference von Soest2007). It also brings security concerns (for both respondents and researchers) and fundamental ethical issues to the fore (Clark Reference Clark2006; Fu and Simmons Reference Fu and Simmons2021; Parkinson Reference Parkinson2022; Wood Reference Wood2006). Furthermore, timing plays a crucial role; obviously, the further researchers go back in history, the higher the probability that respondents cannot remember certain events or processes.
Finally, respondents differ in their interpretations. The “Rashomon effect” denotes the simple fact that different participants in a process have alternate views as to what actually took place (George and Bennett Reference George and Bennett2005, 99f.). In consequence, the evidentiary value of just one expert interview is close to zero; it needs to be cross-checked with other interviews or data streams (“triangulation,” discussed later). I would furthermore advise scholars to conduct an assessment of the motives a respondent might have to distort or conceal information before interviews start (Beach and Pedersen Reference Beach and Pedersen2019, 210).
Combination with Other Data Collection Techniques/List Experiments
To overcome or at least limit the inherent dangers of personal bias, scholars can gain from combining qualitative expert interviews with other data collection techniques. My focus here lies on highly structured experimental methods; in my view, existing research has hardly explored the potential to integrate them with expert interviews. Experimental methods reduce incentives to distort or withhold information, particularly when the interviewer investigates controversial topics. For instance, while survey-based conflict research that directly asks about respondents’ exposure to violence is increasingly common, posing upfront questions on attitudes to violence is not (Balcells and Justino Reference Balcells and Justino2014). Social-desirability bias regularly renders responses invalid. To counter this, scholars have designed so-called list experiments to help examine “true biases and preferences that would be otherwise difficult to reveal” (Dietrich, Hardt, and Swedlund Reference Dietrich, Hardt and Swedlund2021, 603).
In a list experiment, the researcher asks the respondents how many preformulated statements they consider correct. These statements are taken from a list of assertions that includes a sensitive item, such as support for a violent organization (Swedlund Reference Swedlund2017). One group of interviewees receives the complete list of statements, while the other group gets the same one but without the sensitive statement. The comparison of the two groups’ mean answers provides “an estimate of how many respondents believe the sensitive item to be true” (Swedlund Reference Swedlund2017, 471).Footnote 9 The precondition for applying this indirect assessment method is a sufficiently high number of interviewees (Lyall Reference Lyall, Bennett and Checkel2015). Yet even if the number of respondents is too low to meaningfully calculate the difference between the two groups’ mean responses, simply comparing them gives an indication about the interviewed experts’ true preferences and the level of misrepresentation at play. Standardized methods such as these can be easily integrated into interviews with inside experts, particularly on contentious issues. For instance, in her research on the perceptions of officers at NATO headquarters, Hardt (Reference Hardt2018) started each interview with a five-minute paper-and-pen survey experiment. In this way, the cross-fertilization with standardized data collection techniques can strengthen the evidentiary value of expert interviews, in particular those conducted with policymakers (inside experts).
In addition to systematically selecting experts as well as dealing with the reactivity and subjectivity of the expert-interview method, the capturing of evidence should follow clear guidelines. Researchers should ensure that the material used is representative of the whole empirical corpus and triangulate expert interviews.
Representativeness: The Power of the Good Quote
Regularly, scholars use extracts or snippets from interviews to capture a certain aspect of the empirical reality in a concise and vivid manner. In doing so, researchers should relate these extracts to the whole empirical corpus and thereby counter the often-made accusation that quotes from expert interviews are “cherry-picked” (Dunning Reference Dunning, Bennett and Checkel2015, 232; Elman, Gerring, and Mahoney Reference Elman, Gerring and Mahoney2016, 383; Tripp Reference Tripp2018, 731, 735). This is all the more important as due to ethical and methodological constraints qualitative research is much harder to replicate than statistical analysis. Often, expert-interview transcripts cannot be made public.Footnote 10
To overcome the power of the good quote and avoid biases in using references, scholars should clearly catalogue the procedures guiding the aggregation and interpretation of information (Tripp Reference Tripp2018, 735–36). Two aspects are of crucial importance here. First, as outlined earlier, it starts with the balanced selection of inside and outside experts. As Schedler (Reference Schedler2012, 31) notes, the “quality of expert judgments … depends, first of all, on the quality of experts.” Second, when using extracts researchers should always state how strong expert consensus and how representative a particular quote is compared to all interview statements. Sentences such as “none of the outside experts interviewed for this study maintained that” or “this sentiment was widely shared among civil servants” (von Soest Reference von Soest2009) clearly situate the quote in the empirical corpus of expert interviews and other data sources.
With their comparatively high level of structuring, semi-structured expert interviews permit us to systematically assess the degree of “inter-expert agreement” (Dorussen, Lenz, and Blavoukos Reference Dorussen, Lenz and Blavoukos2005, 325), therewith putting the representativeness of selected interview extracts on firm grounds (Dunning Reference Dunning, Bennett and Checkel2015, 215, 232). The level of inter-expert agreement can then be reported as a measure for the reliability and validity of the obtained information. In doing so, the threshold for inter-expert agreement is considerably higher for outside experts than for inside ones.
As a general principle, the number of expert interviews should be as high as possible (Dorussen, Lenz, and Blavoukos Reference Dorussen, Lenz and Blavoukos2005; on expert surveys, see Maestas Reference Maestas, Atkeson and Alvarez2018). A large corpus of expert interviews facilitates situating individual perspectives, and thereby better gauging the representativeness of individual pieces of information. Of course, this requirement is dependent on: a) the relationship to other data streams (is information from other sources available?); b) the nature of the topic/research interest in question (for instance, hypothesis generation versus hypothesis testing); and c) whether the researcher deals with “hard-to-survey populations” (Tourangeau Reference Tourangeau2014). Hence, this guiding principle should not make scholars reject the method in difficult cases.
External and Internal Triangulation
The classic strategy to enhance confidence in the accuracy of the information from expert interviews is to blend it with other data streams (Denzin Reference Denzin1978). Expert interviews are rarely used as a stand-alone technique, but form part of an architecture of sources. My comparative study on political interventions into the tax administration in two African countries relied on over 150 semi-structured expert interviews that were complemented with the annual tax-administration reports, data on revenue performance, and secondary literature to strengthen the process tracing analysis (von Soest Reference von Soest2009; for another example, Lundgren Reference Lundgren2020). The precondition for this triangulation is that the sources do not depend on each other (Beach and Pedersen Reference Beach and Pedersen2019, 215).
In addition to the established triangulation of different data sources, I argue that more can be made of “internal triangulation”—meaning the organized cross-checking of information collected via expert interviews themselves. Analysis should systematically consider different expert groups to achieve as much control as possible. This includes both inside (actors) and outside (analysts) experts as well as, if applicable, insiders on different hierarchical levels. This strategy is key to adequately validating the information provided. To investigate political intervention into the tax administration (von Soest Reference von Soest2009, 47), internal control was sought by interviewing tax officers from the leadership versus from the operational level (middle- and even low-level tax officers). Indeed, the leadership overall painted a rosy picture, while tax officers in one country reported concrete interventions that politicians targeted their day-to-day work with. In addition, experts from outside the tax office were interviewed. These outsiders were businesspersons, tax advisors, civil society representatives, policymakers, and social scientists. The five respondent groups allowed me to control for the internal perspective and provided additional insights (von Soest Reference von Soest2009, 48).
This categorization of experts furthermore helps to refer to respondents while preserving their anonymity. As a convention, researchers could systematically designate sources as “insider” or “outsider” in using quotes and analyzing data. For instance, “According to inside expert A” or “Representing the majority of assessments, outside expert E stated that”. This categorization may complement or even supplement specific positional descriptions (which might at times be too revealing), such as “a mid-level tax officer” or “a leading activist.”
Online Expert Interviews
Expert interviews are designed to be in-person encounters. The face-to-face interview situation allows one to establish rapport with the respondent, notice cues, and record further non-verbal meta-data. Researchers are able to quickly adapt to the interview situation and flexibly pose follow-up questions. This ensures a high degree of validity. Also, being “in the field” eases or even is a precondition for the purposeful, encompassing, and balanced selection of respondents through referrals and snowball sampling (e.g., Driscoll Reference Driscoll2021; Goldstein Reference Goldstein2002; Heckathorn and Cameron Reference Heckathorn and Cameron2017). Yet in authoritarian and post-conflict environments access to the field is fundamentally restricted (de Vries and Glawion Reference de Vries and Glawion2021; Wackenhut Reference Wackenhut2018). More generally, conducting interviews on controversial topics might create serious risks for both respondents and researchers, even in otherwise peaceful contexts (Irgil et al. Reference Irgil, Kreft, Lee, Willis and Zvobgo2021). Most recently, the COVID-19 pandemic has drastically inhibited travel as well as access to individuals, whether within one’s country or abroad (Mwambari, Purdeková, and Bisoka Reference Mwambari, Purdeková and Bisoka2021; Schirmer Reference Schirmer2021).
Yet digital research methods have emerged as a viable alternative to in-person meetings. In particular, synchronous online interviewing through video-conferencing tools like Zoom or MS Teams are a useful supplement or even (if necessary due to security or other risks) replacement for in-person expert interviews. Scholars have identified particular methodological, security-related, and ethical challenges regarding remote meetings, such as the increased probability of sampling bias (some experts can hardly be reached through the Internet), security agencies’ online surveillance, and complicated trust-building in the absence of personal interaction. This could make expert interviews more superficial (Irgil et al. Reference Irgil, Kreft, Lee, Willis and Zvobgo2021, 1513; Mwambari, Purdeková, and Bisoka Reference Mwambari, Purdeková and Bisoka2021; van Baalen Reference van Baalen2018). Selecting interviewees and creating trust should be particular problems for decision makers; less so for external experts who routinely analyze the processes and events in question and are used to digital communication.
Recent experiences provide greater optimism and indicate that these challenges can be overcome. First, scrutiny of both researcher and participant perspectives suggests that scholars can also establish rapport via synchronous video-conferencing tools that transmit audio and pictures. Both interviewers and interviewees may feel comfortable using such technology (Archibald et al. Reference Archibald, Ambagtsheer, Casey and Lawless2019; Lo Iacono, Symonds, and Brown Reference Iacono, Valeria and David2016). Even referrals to further respondents seem possible; however, anecdotal evidence suggests that a prior on-site stay eases the selection of experts significantly (Schirmer Reference Schirmer2021). In addition, online meetings are less costly than in-person research (e.g., Irgil et al. Reference Irgil, Kreft, Lee, Willis and Zvobgo2021, 1513). The number of expert interviews can thereby be increased, and online conversations also be used as a follow-up to prior in-person meetings.
Thus, using digital technology for conducting expert interviews can supplement traditional in-person meetings. As for in-person interviews, while more difficult to achieve, it is imperative that researchers select a balanced sample of outside and, where available, inside experts. They should systematically reflect on and report potential biases in the information and assessments gained through online data collection techniques, and how these might have influenced findings.
Much of the current mainstream methodological discussion in political science has centered around the use of experimental methods and new ways of making causal inferences via statistical means. Further insightful research has discussed the methodological, psychological, and ethical issues of field-research methods such as immersion (most recently, Driscoll Reference Driscoll2021; Irgil et al. Reference Irgil, Kreft, Lee, Willis and Zvobgo2021; Krause Reference Krause2021; Parkinson Reference Parkinson2022). Yet considerations specifically focused on the data collection technique of qualitative expert interviews have attracted less attention.
In this paper I have made concrete suggestions on how to conduct expert interviews in a structured and transparent manner to strengthen their validity and reliability. This is all the more important as qualitative evidence—such as that from expert interviews—is much harder to replicate than statistical data. I argued that inside (actors) and outside (analysts) experts should both be included in an integrated analytical framework. To deal with expert interviews’ inherent challenges of selecting experts, countering their potential biases, and systematically capturing evidence, it is imperative to: a) reflect on the relational nature of the interview situation and to disclose one’s positionality; b) triangulate interviews and compare them with other data streams (internal and external triangulation); and c) systematically situate quotes in relation to the whole empirical corpus. Methods such as list experiments can reduce social-desirability bias, particularly for interviews with inside experts, while interviewing online is a useful supplement to traditional in-person meetings.
Conducted in such a structured manner, qualitative expert interviews have a crucial role to play not only in the current drive to assess causal mechanisms but also in generating important descriptive insights about the “what” and “how” of political processes and events (Fu and Simmons Reference Fu and Simmons2021). They help to examine how actors actually behaved in a decisional context and to discern what happened and why it happened. Findings from expert interviews thereby contribute to analyzing political science’s “big” questions and to linking the macro and micro levels of analysis. Future publications can present more in-depth advice on the structured analysis of expert-interview data and also formulate suggestions on how to bring journals’ preregistration and transparency requirements into line with this qualitative method’s specificities. Further discussing their systematic application will be crucial to making the most of expert interviews as a key data collection tool for social scientists.
The author thanks Julia Grauvogel and Oisín Tansey for their thorough and extremely insightful comments on earlier versions of the paper. He is also indebted to Tim Glawion, Bert Hoffmann, and Swantje Schirmer for their valuable advice on a revised version. Many thanks also to editor Michael Bernhard and the two anonymous reviewers for their constructive feedback and suggestions.