Forensic Science – Believe It or Not? Public Attitudes toward Forensic Evidence in Israel

Abstract Forensic science is undergoing an unprecedented period of reform. Wrongful convictions and errors of impunity have been attributed largely to forensic evidence, and concerns over the scientific foundations of many forensic disciplines have been raised in key official reports. In these turbulent times, it becomes particularly interesting to understand how forensic evidence is understood by the general public. Is it idealized? Are its inherent limitations recognized? The present study seeks to contribute to this growing body of work by addressing two main questions: (1) How does the general public perceive forensic science?; (2) How correct are individuals in their evaluations of specific types of forensic evidence? A survey of the Israeli public reveals considerable trust in the ability of forensics to reliably identify the perpetrator of a crime, although less trust is expressed when questions lead respondents to consider specific stages in the forensic process. Furthermore, respondents were often incorrect in their evaluations of the reliability of specific types of forensic evidence. The implications of these findings for police legitimacy, the practice of the criminal justice system, and the future study of attitudes toward forensic evidence, are discussed.


INTRODUCTION
Forensic science can be defined as the application of scientific methods and processes to solve crimes, often through the comparison of physical evidence gathered at a crime scene to that obtained from a suspect (Eckert 1996;Houck and Siegel 2009;Saferstein 2017).For most of the twentieth century, forensic methods were perceived to be trustworthy and reliable by courts, attorneys, jurors and the general public.As stated by Mnookin et al. (2011, 726): Long-used types of forensic sciencefingerprint examination, handwriting analysis, firearms and toolmark comparison, and other forms of pattern and impression evidenceare mainstays of criminal prosecution.For roughly a hundred years, these comparison and identification methods have regularly and routinely been employed as legal evidence.For most of that period, courts, attorneys, jurors, and the public, as well as forensic analysts themselves, have largely accepted this evidence as trustworthy and uncontroversial.
The idealization of forensic science changed radically in the early 2000s, when forensic methods came under growing public scrutiny and criticism.Professional carelessness, cognitive biases or downright incompetence have raised doubts about the trustworthiness of findings at several forensic laboratories (Mnookin et al. 2011).For example, in 2004, several senior Federal Bureau of Investigation (FBI) examiners mistakenly linked a fingerprint connected with the Madrid train bombing to Brandon Mayfield, an American lawyer and a convert to Islam (Mnookin et al. 2011).One of the greatest challenges to forensic practices has emerged from within forensic science itself, in the wake of advances in DNA analysis.Studies of wrongful convictions based on DNA findings revealed that errors in forensic tests constitute the second most important factor associated with such convictions (Saks and Koehler 2005;Innocence Project 2023).Concerns over the scientific foundations of many forensic disciplines were raised in two major official reports: the report of the Committee on Identifying the Needs of the Forensic Sciences Community, National Research Council (2009), and that of the President's Council of Advisors on Science and Technology (2016).
In light of these turbulent times and the shift in policy makers' approach to forensic science, it may well be asked how the general public perceives forensic evidence.Are ordinary citizens overcome by idealized views regarding the accuracy of forensic sciences (as might be expected given the surge of police investigation television shows; Cole and Dioso-Villa 2006;Schweitzer and Saks 2007), or have they become aware of the growing criticism and recognition of the field's limitations?Moreover, how accurate is the public in its evaluation of forensic sciences compared to the scientific assessment of these fields?Given the inherent connection between forensic science and the police (police experts gather forensic evidence and criminal investigators/detectives use it to build criminal cases; e.g.Williams 2008), such evaluations can be seen as an expression of "police legitimacy", a concept that has figured prominently in policing scholarship in recent decades (e.g.Nagin andTelep 2017, 2020;Weisburd and Majmundar 2018;Tyler and Nobo 2023).Furthermore, from the point of view of policy and practice, inaccurate perceptions (and particularly unfounded idealization of forensic evidence) could influence jury decisions (e.g.Winter and Greene 2007) and/or give rise to significant public pressure on the criminal justice system regarding particular cases (Roberts 2018).Thus, the present work seeks to contribute to the growing body of research on public attitudes toward forensic science, focusing specifically on the three types of attitudes examined in previous research: general assessments of forensic science, views regarding the particular stages of the forensic process (including perceived error and degree of human judgement) and evaluations of particular forensic methods (e.g.DNA, fingerprints, bite marks).
The present article begins with an explanation of the importance of public views of forensic science, followed by a summary of the available studies on the matter.We then review the research on the accuracy of specific forensic methods, and provide a summary of the studies that have examined the general public's understanding of the accuracy of these methods.This is followed by a description of the methodology of the present study (a community survey carried out in Israel) and an examination of the research findings.These reveal a strong public belief in the ability of forensic science as a whole, as well as specific forensic disciplines, to reliably identify the perpetrator of a crime, although less trust is expressed when questions lead respondents to consider the specific stages of the forensic process.In addition, while public views regarding the various forensic techniques appear to correspond with more objective and systematic assessments in the cases of DNA and fingerprint analysis, there is considerable divergence between public assessments and scientific findings concerning the other forensic disciplines.The implications of these findings are discussed, including their relationship to police legitimacy; the importance of measuring specific (rather than generalized) views of forensic evidence; the need to consider attitudes toward forensic evidence in the context of populism (the effort to advance policies that are likely to win public support; e.g.Roberts et al. 2002) and jury selection in order to reduce bias during trials; and the need to include detailed explanations and data indicating the reliability of different forensic techniques as part of the evidence presented to the court and the media.Taking a bird's-eye view, public attitudes toward forensic evidence are important because they are intrinsically linked to "police legitimacy", an idea that has attracted much attention in recent decades from police practitioners and researchers, policy makers and the general public (e.g.Nagin andTelep 2017, 2020;Weisburd and Majmundar 2018;Tyler and Nobo 2023).This considerable interest stems from the fact that a large body of work revealed a strong connection between citizens' perceptions of the police as trustworthy and legitimate, and various socially desirable outcomes.These include, for example, compliance with police requests in interpersonal interactions (Tyler and Huo 2002), in enforcement contexts (Dickson, Gordon, and Huber 2022) and in routine, daily behaviours (Sunshine and Tyler 2003;Tyler and Fagan 2008;Tyler and Jackson 2014), as well as a willingness to cooperate with law enforcement authorities (Mazerolle et al. 2012;Wolfe et al. 2016), and actual cooperation with them (Mastrofski 1996;Dai, Frank, and Sun 2011;Mazerolle et al. 2013;Desmond, Papachristos, and Kirk 2016; for a metaanalysis, see Walters and Bolger 2019).
Views of forensic science, in turn, can be seen as one (of several) expressions of police legitimacy.Trust in forensics indicates trust in a salient technique used by the police, and perhaps in their professionalism more generally.But no less important is the question of whether this trust is "deserved", that is whether public beliefs about these specific techniques are correct (Kaplan, Ling, and Cuellar 2020).It is important to emphasize that the present study makes no pretence to incorporate public views of forensic evidence into the well-established model of police legitimacy (as a component, predictor or outcome of legitimacy; e.g.Tyler and Nobo 2023).Alternatively, in framing the study and situating it in the literature on public attitudes toward the police, we take a broad perspective and consider the way a significant element of police work (forensics) is viewed by the public as inevitably and intrinsically linked to the popular legitimacy of law enforcement authorities.
Focusing more specifically on policy and practice considerations, public attitudes toward forensic science in countries that use a jury system are important because they may influence jury decisions of guilt or innocence (Winter and Greene 2007).Kaplan et al. (2020, 271-2) note that as the use of forensic science becomes more frequent in criminal cases that come before a jury, it is increasingly important to understand jurors' preconceptions about forensic investigation in order to minimize biases during proceedings; Kim, Barak, and Shelton (2009) have demonstrated that pre-trial juror expectations of forensic science may indeed affect jury decisions.Moreover, in her review of the literature, Eldridge (2019, 32) concluded that jurors: : : : often undervalue evidence, particularly if it is in a discipline that they may have previously considered to be less discriminating.They do not understand numerical testimony well, although they may prefer to hear it, and they vary widely in their interpretation of verbal expressions : : : In such constellations, general attitudes and biases concerning forensic evidence could make a heavy impact on jury decision-making.Indeed, prevailing cognitive models of juror decision-making (Pennington and Hastie 1981, 1986, 1988, 1993;Chen and Chaiken 1999) assert that the jurors' knowledge, expectations, attitudes and motivation influence their evaluation of the evidence and determination of a verdict (Winter and Greene 2007).
It is also important to note that public views may affect decision-making in the criminal justice system (Casillas, Enns, and Wohlfarth 2011;Roberts 2018).A useful example is the case of Joan Little, an African American women who was charged with first-degree murder in 1974 after stabbing a white prison guard who sexually assaulted her at North Carolina's Beaufort County jail.Black women and men protested on her behalf, and a committee in her name raised hundreds of thousands of dollars for her bond and legal fees (Black Women's Blueprint, Inc. 2019).Greene (2015, 428) argued that the "well-oiled defense fund" and "broad-based 'Free Joan Little' campaign" were instrumental in winning an acquittal.In her view, "without the funds and the activists' support, Little might well have received a death sentence".A more recent example can be found in Israel (the site of the present study), which relies on bench trials held before professional judges.In the context of the debate over reopening the murder trial of Zadorov in 2018, the Israel State Attorney stated: We have no intention of changing our position because of public pressure from one sector or another.Our responsibility is to follow the evidence.That is how we behaved in the past and that is how we will act in the future (Dolev 2018).This statement expressed the State Attorney's commitment to refrain from making decisions on the basis of public pressure, but the need to explicitly state this reflects recognition of the strong, potential effects of public pressure on the criminal justice system.
Such potential for public influence echoes the concept of penal populismthe promotion of a policy not because it was found to be effective in achieving its goals, but because of its expected political benefits.More specifically, penal populism refers to the effort to advance penal policies because they are more likely to garner votes or win public support, than policies that would effectively reduce crime or encourage truth-finding in the criminal justice system, but are less popular (e.g.Roberts et al. 2002).Similarly, recent studies have shown that public opinion favouring more severe punishment ("public punitiveness") has had considerable impact on court decision-making in the USA (Pickett 2019).As noted by Roberts (2018, 197,  To date, numerous studies have focused on the attitudes of the general public toward forensic science within the context of the so-called "CSI [crime scene investigation] effect" (Cole and Dioso-Villa 2006;Schweitzer and Saks 2007).This effect was named after the popular television programme CSI and its spin-offs, which bring forensic sciences into the spotlight and glorify their effectiveness, thus potentially creating unrealistic expectations regarding their utility and accuracy.These expectations, in turn, place a burden on the prosecution to bring forensic evidence indicating the suspect's guilt (i.e. if there is no damning forensic evidence, the accused person must be innocent).Alternatively, they may create an almost blind faith in the reliability of forensic practices, which places a burden on the defence (i.e. if there is damning forensic evidence, the accused person must be guilty; Cole and Dioso-Villa 2006;Schweitzer and Saks 2007).
Thus far, studies have failed to establish a consistent relationship between viewing CSI and jurors' verdict decisions (Podlas 2006;Shelton, Kim, and Barak 2006;Schweitzer and Saks 2007;Kim et al. 2009), and recent studies have found no evidence for such an effect at all (Klentz, Winters, and Chapman 2020;Lodge and Zloteanu 2020).Whatever the case may be, by focusing on direct exposure to forensic crime television dramas, this body of work does not account for the possible effects of other sources of forensic information, such as different crime-related programmes, crime fiction literature or media coverage of forensic issues.Moreover, this work focuses on jury verdicts.It does not examine whether watching CSI affects attitudes toward forensic science, which may, in turn, have an impact on juror decisions (Smith and Bull 2012).Therefore, given the possibility that pre-trial juror expectations of forensic science do in fact affect jury decisions (see Kim et al. 2009), Smith andBull (2012, 2014) developed a scale to measure pre-trial bias regarding forensic evidence, and confirmed the existence of "pro-defence" and "proprosecution" biases.Their study found that the British public placed only a moderate degree of trust in forensic science (slightly less than 2.5 on a scale of 5).
A second, related body of work focused more specifically on the different stages of the forensic process.Ribeiro, Tangen, and McKimmie (2019) examined perceptions of the likelihood of error and degree of human judgement involved at various stages of the forensic process (i.e.collection, storage, testing, analysis, reporting and presenting).They found that the forensics process was perceived to involve a considerable degree of human judgement and to be relatively error prone.Subsequently, Kaplan et al. (2020), using a modified version of the instrument developed in Ribeiro et al. (2019), found that people held a pessimistic view of forensic investigations, believing that an error could occur roughly half the time at each stage.
The Accuracy of Forensic Disciplines: Reality v. Public Perceptions How Reliable Are Forensic Techniques?
As already noted, for most of the past century, forensic methods were considered trustworthy and reliable.However, the Committee on Identifying the Needs of the Forensic Sciences Community, National Research Council (2009) report Strengthening Forensic Science in the United States: A Path Forward and the President's Council of Advisors on Science and Technology (PCAST) report (President's Council of Advisors on Science and Technology 2016) both found that the scientific foundations of many forensic disciplines were insufficient.PCAST evaluated specific forensic methods in order to determine whether they had been scientifically established as valid and reliable.The report reviewed research on seven forensic disciplines (DNA single-source and simple mixture, DNA complex mixture, bite marks, fingerprints, firearms, footwear and hair), and concluded that only two (DNA analysis of single-source and simple mixture samples and latent fingerprint analysis) were foundationally valid, meaning that they were repeatable, reproducible and accurate.With regard to the other five disciplines, some were not found to be foundationally valid, while others had not yet been subjected to objective research.
Regarding shoeprints, PCAST concluded that the foundational validity of footwear analysis associating shoeprints with particular shoes on the basis of specific identifying marks (known as "randomly acquired characteristics") lacked sufficient support in scientific research (President's Council of Advisors on Science and Technology 2016, 117).Bite mark analysis was dismissed as falling far short of the standards required for foundational validity (President's Council of Advisors on Science and Technology 2016, 87), with scientific evidence suggesting that examiners could not consistently or accurately agree on the identification or source of a possible bite mark.PCAST quoted the US Department of Justice (1987) guidelines regarding hair analysis, which noted that microscopic hair comparisons were insufficient for personal identification (President's Council of Advisors on Science and Technology 2016, 13).To date, no single review has provided a definitive answer about the foundational validity of techniques that were not included in the PCAST report (Kaplan et al. 2020).
As stated by PCAST, "the foundational validity of a subjective method can only be established through multiple, appropriately designed black-box studies" (President's Council of Advisors on Science and Technology 2016, 9).Black-box studies treat examiners as decision-making "black boxes" and measure the accuracy of their conclusions without considering how they arrived at their decisions.
In order to evaluate performance, examiner's decisions in specific cases are compared to the correct answers (of which they are of course unaware).Such studies are important for learning about the overall error (misclassification) rate of a forensic discipline.Recent black-box studies evaluating such errors in various forensic fields are summarized in Table 1, which also shows the empirical error rates in these practices.Table 1 reveals that the error rates of fingerprints and footwear disciplines are relatively close and are lower than handwriting and bloodstain pattern analyses.Hopefully, studies of error rates in additional disciplines will provide a more complete picture.
In sum, on the basis of PCAST and recent empirical research, certain forensic disciplines are regarded as highly reliable, such as DNA analysis and fingerprints, while others are highly questionable, such as bite mark and hair analysis.With regard to most other forensic disciplines (including footwear analysis), the degree of accuracy and reliability is still the subject of ongoing study, with promising results in some disciplines.But what does the public think about the reliability of different forensic methods?
How Does the Public Perceive the Accuracy of Forensic Science?
There appears to be considerable agreement across several studies that DNA analysis and fingerprint analysis are perceived by the general public as highly reliable techniques, with bite mark and hair analysis coming in second and shoeprints seen as considerably less reliable (Lieberman et al. 2008;Koehler 2017;Martire et al. 2019;Ribeiro et al. 2019;Kaplan et al. 2020).On the whole, nonforensic evidence is viewed as less trustworthy than most types of forensic analysis (Lieberman et al. 2008;Martire et al. 2019).A comparison of these perceptions with the scientific evaluation carried out as part of the PCAST report suggests that the public is well informed with regard to DNA and fingerprints.At the same time, there are broad misconceptions regarding bite mark analysis, which ranks third among the general public despite the fact that it has been scientifically discredited.Shoeprint analysis is perceived to be the least trustworthy discipline, in spite of PCAST's conclusion that more research is necessary before determining its status.The findings of five studies (Lieberman et al. 2008;Koehler 2017;Martire et al. 2019;Ribeiro et al. 2019;Kaplan et al. 2020) concerning the forensic disciplines evaluated by PCAST, as well as non-forensic disciplines considered in these studies (aside from alibi which was not examined in any of the studies), are shown in Table 2.For the sake of clarity, findings regarding forensic disciplines not included in the PCAST report have been omitted, as has information regarding firearms, a discipline included in the report but not examined in this study.
To summarize, we are not the first to grapple with questions concerning public attitudes toward forensic evidence.At the same time, this is still a very small body of work, originating primarily from English-speaking countries, clearly leaving much room for contribution.Moreover, we are unaware of studies that have examined all three relevant types of public attitudes toward forensic science within a single questionnaire (general evaluations, assessments of specific stages in the process of forensic investigation, attitudes toward specific forensic disciplines), making the comparison of responses across the different types extremely difficult.This, in turn, impedes our ability to conclude regarding similarities/differences between generalized and more focused views, and more broadlyto come to an integrative answer to the question "how does the public perceive forensic evidence?"Furthermore, until the publication of the PCAST report (President's Council of Advisors on Science and Technology 2016), it was not possible to compare public attitudes to objective scientific assessments of various forensic disciplines.In other words, it was not possible to answer the question "how accurate are citizens in their evaluations of specific forensic techniques?"Since 2016, this has only been done by Kaplan et al. (2020), and thus additional investigation is clearly warranted.Accordingly, the present study sets out to examine the three main types of public attitudes toward forensics identified in earlier work: (1) general assessments of the ability of forensic science to solve crime, (2) the degree to which each stage of the forensic process is prone to error, and the extent of human judgement involved; (3) the degree to which various forensic disciplines are error prone, while comparing these views to scientific evidence on the reliability of the different forensic disciplines.Finally, it should be noted that the present study relies on a relatively large population sample from a non-English-speaking country, thus expanding the context of this body of work.

Study Context
As noted above, the present study relies on data collected in Israel.It is thus important to provide some background on how forensic science is organized and handled in the Israeli criminal justice system.Most forensic investigations in Israel are carried out by the Israel Police Division of Identification and Forensic Science (DIFS), while certain areas (as suggested by their names) are handled by the National Center of  A second important characteristic of the Israeli context concerns the judicial system, which relies on bench trials before professional judges.There are no juries in Israel; professionally trained judges handle all aspects regarding the administration of justice.Unlike the judiciary in the United States, Israeli judges not only make decisions on legal issues, but operate as factfinders as well.Serious criminal cases are tried by a panel of three justices.In addition to the verdict, Israeli judges provide detailed opinions that can be passed on to superior courts (Straschnov 1999).Despite the lack of a jury system in the country, public attitudes toward forensic evidence in non-jury systems may bear important implications in terms of public pressure on the criminal justice system (Casillas et al. 2011;Roberts 2018).
To contextualize the findings, the highly publicized Zadorov case should be noted (e.g.Hovel 2013; Yanko and Raved 2022;Starr and Silkoff 2023) in which a young girl was murdered in the town of Katzrin in 2006.Roman Zadorov was convicted in 2010 and retried in 2021 after the Supreme Court found that there was sufficient reasonable doubt about his guilt.The 2010 conviction was based partly on prints found at the crime scene that seemed to match his shoes.However, in an unrelated case in December 2013, the Supreme Court ruled that such prints, although admissible as evidence, were problematic and thus of limited value (Hovel 2013).During the retrial (July 2021-March 2023), there was a considerable debate about inconclusive results of a mitochondrial DNA test conducted on hair found at the crime scene, which reignited speculation that Zadorov was innocent (Starr and Silkoff 2023).During the autumn of 2022, shortly before the present study was conducted, the debate concerning the mitochondrial DNA test attracted considerable media attention (e.g.Yanko and Raved 2022), raising questions about the significance of mitochondrial DNA taken from hair and reviving the earlier debate over the trustworthiness of footwear evidence.Thus, this case and its significant publicity may well have drawn public attention to forensic evidence, and specifically to hair analysis and footwear evidence, making the topic particularly salient among the general public.

Sampling and Participants
On 13 September 2022, an online survey measuring the attitudes of the Israeli public toward forensic evidence (see below) was conducted, using the services of "Midgam Project Web Panel", a survey platform frequently used by social scientists in Israel.Through its website, people sign up to participate in surveys for a fee.represent the adult population in Israel in terms of gender and age, based on data provided by the Israeli Central Bureau of Statistics (CBS).Unfortunately, national or religious minorities (e.g.Arabs, Druze) are not adequately represented in online survey platforms (e.g.Scherpenzeel and Bethlehem 2010), and thus the sample focuses on the adult Jewish population. 3 The initial sample included 1,507 participants.A reduction of potential carelessly invalid responses was conducted, using some of the detection methods articulated by Curran (2016).Huang et al. (2012) have suggested a conservative average cut-off time of 2 seconds per question, under which a response would be discarded as being careless.The current study has taken an even more conservative view and doubled this to 4 seconds per question, or 4.8 minutes for the entire survey.A second technique suggested by Curran (2016) examines strings of consistent responses.Currently there is no rule of thumb regarding the optimal length of such streams and various factors may be taken into consideration, such as the number of questions, the size of the scale and the nature of the question (Huang et al. 2012).Given the number of questions ( 72) and the five-point Likert scale, it was decided to set the cut-off string length at 20, so that any response with 20 or more consistent responses would be excluded.Additionally, it was decided to exclude participants who checked "don't know/irrelevant" in half or more of the questions.In total, 165 participants were excluded.This left a final sample of 1,342 participants, which are similar overall in their sociodemographic characteristics to the population from which they were drawn (see Table 3). 4A sensitivity analysis was conducted to evaluate the effect of excluding participants based on the chosen criteria by repeating the analysis on the entire sample without any exclusions.This analysis reveals that the exclusion did not affect the findings in a significant way.

The Survey Instrument
The questionnaire began with a consent form explaining the purpose of the study, along with its expected duration and identity of the researchers.Participants were notified that the survey is anonymous and voluntary, and that they may cease participation at any time and for any reason.Subsequently, they were presented with 72 survey items. 5Most items were phrased as statements, which respondents were asked to rank on a five-point Likert scale (1 = "completely disagree" to 5 = 3 For more details on Midgam's modus operandi, see Midgam (2024).4 Unfortunately, it was not possible to compare our data with that of the CBS regarding education due to a different definition of the minimum age included in the samples (i.e.18 and 15 years, respectively), and regarding income as the CBS does not provide raw data on this topic.

5
As mentioned, items were based on previous instruments developed by Smith andBull (2012, 2014) and Ribeiro et al. (2019).The survey items were translated into Hebrew, and the response scales were replaced with a five-point Likert scale.This was done for the following reasons: the 100-point scale in some of the original studies was seen to be unwieldly, and for the purposes of the present study, no more accurate than a five-point scale.Moreover, a standardized scale for all questions was chosen in order to simplify the questionnaire and make it more understandable for the respondents.Finally, not all types of forensic evidence included in Ribeiro et al. (2019) were included in the present studywe only focused on those examined in the PCAST report (President's Council of Advisors on Science and Technology 2016), in order to enable a comparison of perceptions with objective scientific assessments.Three common non-forensic types of evidence (alibi, confession and eyewitness testimony) were added to our study for purposes of comparison.
"completely agree" in some sections; 1 = "none at all" to 5 = "high degree" in others).The questions used in this study appear in Appendix 1.
The statements enquired about participants' general views of forensic science (based on the instrument developed by Smith andBull 2012, 2014).Additionally (based on the questions asked by Ribeiro et al. 2019), participants were asked to estimate the likelihood of error and degree of human judgement at each stage of the forensic investigation (e.g.To what extent is the collection process error prone?To what extent does the collection process rely on the individual judgement of the person collecting the evidence?),as well as the likelihood of error in the use of various types of forensic techniques (e.g.To what extent is shoeprint evidence error prone?), in addition to other types of evidence often presented at trials (e.g.eye witness testimonies, alibis and confessions).Data about participants' personal characteristics, such as gender, age and education, were obtained from "Midgam".

Analytical Strategy
In line with the research questions and past research in this area, the present study is a descriptive one.Four statements that capture broad views of forensic evidence are International Annals of Criminology presented at the outset.Then, respondents' views concerning specific stages of the forensic process are examined (the likelihood of error and degree of human judgement at each stage), followed by their views regarding the accuracy of different types of evidence, both forensic (five types) and non-forensic (three types).

General Views of Forensic Evidence
Respondents' general views of forensic science are presented in Figure 1.The percentage of participants per response (1 = "completely disagree" to 5 = "completely agree") is noted, along with the mean level of agreement and standard deviation for each statement.As can be seen from Figure 1, a large share of the respondents agree or completely agree (responses 4 and 5) with the first three statements.Specifically, 51% agree or completely agree that "Every crime can be solved with forensic science" compared to 24% who disagree or completely disagree (responses 1 and 2).Approximately 64% agree/completely agree that "Science is the most reliable way to identify perpetrators" compared to 13.5% who disagree/ completely disagree, and 46.7% agree/completely agree that "Forensics always identifies the guilty person" compared to 22.6% who disagree/completely disagree.
Respondents exhibited a more balanced reaction to the statement "Forensic evidence always provides a conclusive answer", with 33.6% agreeing or completely agreeing with the statement, and 35.2% disagreeing or completely disagreeing.Nevertheless, taken together, responses to these statements appear to indicate broad trust in the ability of forensic science to reliably identify the perpetrator of a crime.Moreover, Table 4, which shows a comparison between the current findings and those of Smith and Bull (2012), reveals that the participants of the current study exhibit a considerably higher degree of trust in forensic science (although a larger variability in responses) than those surveyed by Smith and Bull (2012).

Views of Specific Stages in the Forensic Process
Participants were asked to provide their opinions on the degree of error, as well as the extent of human judgement, at various stages in the forensic process of a criminal investigation.Table 5 presents the degree to which respondents believe that the forensic process is prone to error.It reveals that a relatively small share of the respondents (some 15% to 28%) believes that the different stages of the forensic process involve little or no error (responses 1 and 2).A much larger share (about 35% to 46%) believes that the process involves substantial error (responses 4 and 5).
The "examination, analysis and interpretation" stage is perceived as most prone to error, and the "presentation" stage as the least.Taken together, these responses reflect less trust in forensic science, or, put differently, more sophisticated, less idealized views than those expressed in response to the generalized questions about forensics reviewed above.
Regarding the role of human judgement, Table 6 reveals that respondents believe that a high level of human judgement is part of each stage of the forensic process.The percentage of participants who chose responses 4 and 5 (which indicate a high degree of human judgement) ranges from about 46% to 54%, with the exception of "storage", which is perceived by 37.7% to involve a high degree of human judgement.Thus, similar to the responses noted above regarding the perceived degree of error (but in contrast to general views of forensic evidence), these responses suggest recognition of the imperfections of the forensic process, or, in other words, less trust in the process.

Views of Specific Forensic Techniques
Table 7 presents the perceived degree of error (1 = "none at all"; 5 = "high degree") embedded in eight types of evidence (five forensic and three non-forensic), in the order of perceived accuracy, from most accurate (lowest mean error score) to the least (highest mean error score).As can be seen from Table 7, DNA is viewed as Percentage of participants per level of perceived error (1 = "none at all" to 5 = "high degree"), for each stage of the forensic process.
b Within-subjects analysis of variance was conducted resulting in a p value ≤ 0.001.The p value was validated using a permutation test which is robust to normality and sphericity assumptions.For more details see Welch (1990).Percentage of participants per perceived level of human judgement (1 = "none at all" to 5 = "high degree"), for each stage of the forensic process.
b Within-subjects analysis of variance was conducted resulting in a p value ≤ 0.001.The p value was validated using a permutation test which is robust to normality and sphericity assumptions.For more details, see Welch (1990). c All post hoc comparisons (using Bonferroni adjustment) are statistically significant except for Examination, analysis and interpretation v. Reporting, and Examination, analysis and interpretation v. Presentation.
the most accurate type of evidence, followed by fingerprints, hair comparison, bite marks and shoeprints.The non-forensic types of evidencealibi, confession and eyewitness testimonyare at the bottom of the table, meaning that all types of forensic evidence were perceived by respondents to be more accurate than other forms of evidence.This ranking reflects considerable trust in forensic techniques.
As stated above, it is of interest to compare public views to more objective, systematic assessments of the various forensic disciplines.Table 8 compares the Percentage of participants per level of perceived error (1 = "none at all" to 5 = "high degree"), for each type of evidence.The different types of evidence are ordered by perceived accuracy (low to high mean error score).
b Within-subjects analysis of variance was conducted resulting in a p value ≤ 0.001.The p value was validated using a permutation test which is robust to normality and sphericity assumptions.For more details, see Welch (1990).2016,120), was perceived by respondents as the third most reliable discipline.The size of the gaps between the disciplines is also noteworthy: hair analysis is perceived to be only slightly more error prone than fingerprints, a fact that might suggest that the Israeli public does not clearly discern between the most and least reliable forensic disciplines.Taken together, these rankings portray a somewhat naive view of forensic evidence, or, in other words, the public is often incorrect in the legitimacy it ascribes the police through the prism of forensic science.

DISCUSSION
Given the intrinsic relationship between forensic sciences and the police, and the significant (and growing) role of science and technology in policing more generally (Williams 2008), we found it important to examine how the general public perceives forensic evidence, under the assumption that trust in forensic evidence sheds light on police legitimacy more generally.It would be difficult to place little trust in police methods and technologies, but at the same time view the police as a legitimate authority (and vice versa).Nevertheless, we also found it important to examine whether the legitimacy that citizens ascribe to the police through the prism of forensic evidence is justified.
The analysis of the survey data reveals rather complex findings: when asked generalized questions about forensic evidence, respondents display considerable trust in the ability of forensics to reliably identify the perpetrator of a crime.At the same time, in line with the findings of Ribeiro et al. (2019), when asked about specific stages of the forensic process, a large share of the respondents expressed the view that substantial human judgement and considerable error are part of all stages of the forensic process.This might come as a surprise given the fact that forensic science is often perceived to be rooted in technology and hence largely independent of human judgement, a likely outcome of the CSI effect (Ribeiro et al. 2019).These more critical views are, to a certain degree, more "correct", as they are more in line with the National Research Council and PCAST reports (Committee on Identifying the Needs of the Forensic Sciences Community, National Research Council 2009; President's Council of Advisors on Science and Technology 2016).
Nevertheless, a relatively high percentage of respondents perceive a low degree of error across all of the forensic techniques.Generally, the public considers forensic evidence (including techniques known to be untrustworthy) to be more reliable than non-forensic evidence, such as eyewitness testimony.It is particularly noteworthy that even bite mark analysis, which was discredited by the President's Council of Advisors on Science and Technology (2016, 87), is considered to be more reliable than shoeprints and all types of non-forensic evidence.This overestimation of the reliability of bite marks is in line with the findings of Ribeiro et al. (2019) and Kaplan et al. (2020).Overall, it appears that the public is not well informed about the relative reliability of different forensic disciplines.Aside from DNA and fingerprint analysisindividuals assign similar scores to disciplines that are regarded very differently by PCAST, and even rank some of the less reliable ones above others, which are considered by PCAST to be more reliable.
Whether or not respondents are correct in their evaluations of specific forensic techniques, the findings of the present study overall appear to indicate considerable public trust in forensic science as a field, as well as in the specific forensic disciplines.This reflects positively on police legitimacy, at least through the prism of forensics.At the same time, as shown above, the public is often incorrect in its assessments of forensic evidence, meaning that the popular legitimacy that the police gain from their involvement in forensics may not be justified.Additionally, through the lens of populism and potential public pressure (Roberts et al. 2002;Roberts 2018), the findings suggest that the public may be influenced by misinformation and is not fully aware of the field's limitations.Populist attitudes based on insufficient or inaccurate information could have a negative impact on the weight accorded to various types of forensic evidence in the criminal justice system.
It is interesting to note that a comparison between the current findings and those of Smith and Bull (2012) from the UK reveals that the participants of the current study exhibit a considerably higher degree of trust in forensic science (although a larger variability in responses).This finding is particularly interesting given that trust in the Israel Police is generally lower than that in the UK, USA and many other countries (Hadar 2009).In contrast to what might be expected, it may be that Israeli citizens do not associate forensic science with the police, which raises the broader question of how attitudes toward forensic evidence develop and whether such attitudes are actually connected to police legitimacy.It is further advised to advance research in this area by measuring attitudes toward police legitimacy and how they correlate with attitudes toward forensic science.
Beyond the main finding, our analysis draws attention to the difference between responses to the survey items that enquired about forensic science in general, and responses to more specific questions about the different stages of the forensic process.As noted above, respondents expressed general, positive orientation toward forensic science, but when asked specific questions that prompted them to delve more deeply into the forensic process and envision the possibility of error at its various stages, they express more critical views.This finding is in line with the conclusions of Pickett (2019), who surveyed public views of punitive policies, and found that specific questions encouraged respondents to think in a more explicit and critical manner.His explanation echoes the psychological literature on attitudebehaviour relations (for a review, see Kruglanski et al. 2015), and specifically the notion of behaviour focus, which maintains that general attitudes are only weakly related to behaviour, contrary to more specific attitudes about the behaviour itself (Fishbein and Ajzen 1975;Ajzen 1980Ajzen , 1985Ajzen , 2012Ajzen , 2015)).Thus, these findings underscore the importance of using specific, nuanced questions in research on public attitudes, especially when their behavioural ramifications are of interest.Interestingly, in contrast to the questions about the stages of the forensic process, the questions about the specific forensic disciplines did not generate more critical views.We suspect that the overall trust in forensic science was not challenged by these questions, because they enquired about each discipline as a whole and thus did not "force" the respondent into more explicit and critical thinking.In this sense, they behaved as "general" rather than "specific" questions.
In terms of implications, as forensic techniques are used in court despite the fact that the reliability and validity of many are unknown, it is important to continue the effort to study the accuracy and error rates of forensic sciences, as called for in the PCAST report (President's Council of Advisors on Science and Technology 2016).In addition, it is recommended that the findings of these studies be given wide circulation, in order to provide the public with accurate information and contain the potential negative effects of populism.When forensic evidence is presented in court, it should be accompanied by detailed explanations and data regarding the reliability of the different techniques, including error rates.In cases where these data are not available, such evidence should be presented along with a note of caution.In countries that use a jury system, this information should be presented to the jurors, and in countries that use a bench trial, it is advised to make this information publicly available.
A few limitations of the present study should be noted.First, our sample consists of the adult Jewish population in Israel.We thus encourage future research to replicate this study in other communities, as well as in countries that use different judicial systems.Second, local events may have had an impact on the respondents' evaluations of various forensic disciplines.As noted earlier, controversy over the Zadorov case centred around shoeprint and hair evidence.The weight of shoeprint evidence was debated repeatedly during the court proceedings (Hovel 2013), potentially undermining respondents' trust in this specific forensic discipline.This could explain why shoeprints were regarded as the least accurate type of forensic evidence in our study.With regard to hair analysis, which was ranked third for accuracy after DNA and fingerprints, it is possible that the media preoccupation with mitochondrial DNA from hair led to an erroneous public association of hair analysis with DNA.This may have created a "halo effect" (Thorndike 1920;Nisbett and Wilson 1977), which enhanced the perceived accuracy of this practice.Thus, we again encourage replications of our study and analysis in various social contexts.

CONCLUSIONS
The popular legitimacy of the police is anchored, among other things, in the (perceived) reliability of police techniques.Accordingly, this study sets out to illuminate the attitudes of the general public toward the trustworthiness of forensic evidence, and the extent to which these views correspond with objective assessments of the reliability of specific forensic techniques.Our findings reveal that the public places much trust in forensic science in general, as well as in specific forensic disciplines.At the same time, less trust is expressed when questions lead respondents to consider the details of the forensic process at different stages.How "correct" is the public in its assessments?The high regard for DNA and fingerprint analysis corresponds to the findings of objective research, but the public appears to be less aware of the limitations of other forensic techniques.Thus, through the prism of forensics, the public appears to attribute considerable legitimacy to the police, albeit not always justifiably.It is therefore essential to continue research on the reliability of different forensic techniques as called for in the PCAST report (President's Council of Advisors on Science and Technology 2016), andno less importantto make these findings publicly available, both in reports to juries and in public statements.

Figure 1 .
Figure 1.General attitudes toward forensic science (n = 1,342).Histograms show the level of agreement with four statements reflecting general attitudes toward forensics.The statement is noted on the top of each histogram, along with the percentage of participants at each level of agreement, ranging from 1 = "none at all" to 5 = "high degree".The means and standard deviations (SD) are noted at the top left of each histogram.

c
Statistically significant post hoc comparisons (using Bonferroni adjustment): Reporting v. all other stages and Presentation v. all other stages.

c
All post hoc comparisons (using Bonferroni adjustment) are statistically significant except for Hair comparison v. Bite marks and Alibi v. Confession.
The Study of Public Attitudes toward Forensic Evidence Why are Public Attitudes toward Forensic Evidence Important?
citing US Department of Justice 1987):Many sentencing commissions and judges are to some degree affected by public pressure to make sentences harsher.The Chairman of the U.S. Federal Sentencing Guidelines Commission has stated that: "Public input has played a pivotal role in the formulation of the sentencing guidelines."

Table 1 .
Error Rates in Selected Forensic Disciplines a a False positive = incorrect decision that two samples from different sources originate from the same source; false negative = incorrect decision that two samples from the same source originate from different sources.

Table 2 .
Actual and Perceived Reliability of Different Forensic TechniquesForensic Medicine 1 and the Clinical Toxicology and Pharmacology Laboratory 2 (Public Committee for the Prevention of False Convictions and their Correction 2021).DIFS is an internationally recognized body and an active member of the European Network of Forensic Science Institutes.

Table 3 .
Sample Characteristics a aThe population data were obtained from the Israeli Central Bureau of Statistics in 2020.

Table 5 .
Perceived Level of Error at Each Stage of the Forensic Process (n = 1,342) a

Table 6 .
Perceived Level of Human Judgement at Each Stage of the Forensic Process (n = 1,342) a

Table 7 .
Perceived Level of Error by Type of Evidence (n = 1,342) a

Table 8 .
A Comparison of Scientific Assessment and Perceived Reliability of Forensic Disciplines Identified in this Study President's Council of Advisors on Science and Technology (2016).Mean perceived error scores are also noted in Table7; lower score = higher perceived reliability.Foundationally valid = the method is repeatable, reproducible and accurate; for a detailed explanation, see President's Council of Advisors onScience and Technology (2016, 47).mean scores presented in Table7with the PCAST assessment (President's Council of Advisors on Science and Technology 2016).As can be seen from Table8, the Israeli public's views correspond with the PCAST assessment with regard to DNA and fingerprint analysis (President's Council of Advisors on Science and Technology 2016, 75, 101), but there is considerable divergence between them concerning the other disciplines.Bite marks, which were discredited in the PCAST report (President's Council of Advisors on Science and Technology 2016, 87), were ranked as less error prone than footwear analysis, which "lacks sufficient support" (but is not discredited) according to PCAST (President's Council of Advisors on Science and Technology 2016, 117).Hair analysis, which was also dismissed by PCAST for the purpose of personal identification (President's Council of Advisors on Science and Technology a PCAST = b c