Literature DB >> 35634572

Advancing a paradigm shift in evaluation of forensic evidence: The rise of forensic data science.

Geoffrey Stewart Morrison1,2.   

Abstract

Widespread practice across the majority of branches of forensic science uses analytical methods based on human perception, and interpretive methods based on subjective judgement. These methods are non-transparent and are susceptible to cognitive bias, interpretation is often logically flawed, and forensic-evaluation systems are often not empirically validated. I describe a paradigm shift in which existing methods are replaced by methods based on relevant data, quantitative measurements, and statistical models; methods that are transparent and reproducible, are intrinsically resistant to cognitive bias, use the logically correct framework for interpretation of evidence (the likelihood-ratio framework), and are empirically validated under casework conditions.
© 2022 The Author.

Entities:  

Keywords:  EWG, Expert Working Group on Human Factors in Latent Print Analysis; FSR, Forensic Science Regulator for England & Wales; Forensic data science; Forensic science; HoL, House of Lords Science and Technology Select Committee; Likelihood ratio; PCAST, President’s Council of Advisors on Science and Technology; Paradigm shift; TRL, technology readiness level; UKRI, United Kingdom Research and Innovation; Validation

Year:  2022        PMID: 35634572      PMCID: PMC9133770          DOI: 10.1016/j.fsisyn.2022.100270

Source DB:  PubMed          Journal:  Forensic Sci Int Synerg        ISSN: 2589-871X


Introduction

The present paper is a written version of a keynote presentation given at the European Academy of Forensic Science 2022 conference. It discusses an ongoing paradigm shift in evaluation of forensic evidence. It describes: the current state of affairs (staus quo) the new paradigm (quo vadis?) obstacles to the advancement of the paradigm shift (impedimenta) a strategy to advance the paradigm shift (via progredi)

A paradigm shift in evaluation of forensic evidence

Status quo

Curran [1]: Is forensic science the last bastion of resistance against statistics? UK House of Lords Science and Technology Select Committee (HoL) [2]: In regard to pattern comparison methods, … “the comparison of fingerprints, toolmarks, footwear, tire marks and ballistics” [are] “spot-the-difference” techniques in which “there is little, if any, robust science involved in the analytical or comparative processes used and as a consequence there have been questions raised around the reproducibility, repeatability, accuracy and error rates of such analysis.” (§155) In forensic science, the process of evaluation of strength of evidence consists of: analysis, i.e., extraction of information from items of interest (the evidence)1; and interpretation, i.e., drawing inferences with respect to the meaning of the information extracted by the analysis. Items of interest may be, for example: a fingermark of questioned source recovered from a crime scene and a fingerprint collected from a known individual a recording of a speaker of questioned identity on an intercepted telephone call and a recording of a police interview with a speaker of known identity a fired cartridge case recovered from a crime scene and cartridge cases fired in a forensic laboratory from a gun found in the possession of a suspected shooter Forensic practitioners conduct evaluations in order to assist legal-decision makers to make decisions with respect to questions of legal concern such as:2 Do the fingermark and fingerprint originate from the same finger? Is the speaker of questioned identity on the intercepted recording the same as the speaker of known identity? Was the cartridge case recovered from the crime scene fired from the suspect's gun? Currently, across the majority of branches of forensic science, widespread practice is that analysis is conducted using human perception, and interpretation is conducted using subjective judgement. Even in branches of forensic science in which analysis is conducted using instrumental measurement, interpretation is commonly based on subjective judgement, e.g., by eyeballing graphical representations of the measured values. Human-perception-based analysis methods and subjective-judgement-based interpretation methods are non-transparent and are susceptible to cognitive bias. Currently, across the majority of branches of forensic science, even branches of forensic science in which interpretation is conducted using statistical models, interpretation of evidence is often logically flawed, and forensic-evaluation systems (the end-to-end combination of analysis and interpretation methods) are often not empirically validated or not adequately empirically validated.3

Quo vadis?

Introduction

Saks & Koehler [4]: we envision a paradigm shift in the traditional forensic identification sciences in which untested assumptions and semi-informed guesswork are replaced by a sound scientific foundation and justifiable protocols. Although obstacles exist both inside and outside forensic science, the time is ripe for the traditional forensic sciences to replace antiquated assumptions of uniqueness and perfection with a more defensible empirical and probabilistic foundation. (p. 895) US President's Council of Advisors on Science and Technology (PCAST [5]): neither experience, nor judgment, nor good professional practice … can substitute for actual evidence of foundational validity and reliability. The frequency with which a particular pattern or set of features will be observed in different samples, which is an essential element in drawing conclusions, is not a matter of “judgment.” It is an empirical matter for which only empirical evidence is relevant. (p. 6) Objective methods are, in general, preferable to subjective methods. Analyses that depend on human judgment (rather than a quantitative measure …) are obviously more susceptible to human error, bias, and performance variability across examiners. In contrast, objective, quantified methods tend to yield greater accuracy, repeatability and reliability, including reducing variation in results among examiners. Subjective methods can evolve into or be replaced by objective methods. (p. 47) A paradigm shift in evaluation of forensic evidence is ongoing in which methods based on human perception and subjective judgement are replaced by methods based on relevant data, quantitative measurements, and statistical models/machine-learning algorithms; methods that: are transparent and reproducible (§2.2.2); are intrinsically resistant to cognitive bias (§2.2.3); use the logically correct framework for interpretation of evidence (the likelihood-ratio framework) (§2.2.4); and are empirically validated under casework conditions (§2.2.5). I address each of these elements of the new paradigm in the following four subsections. They are, in turn, followed by a subsection which discusses the applicability to this paradigm shift of Kuhn's [6] description of a paradigm shift, and the implications thereof (§2.2.6).

Transparency and reproducibility

Methods dependent on human perception and subjective judgement are intrinsically non-transparent and therefore not reproducible by others. Human introspection is often mistaken, hence a forensic practitioner's explanation of how they reached their conclusion may not reflect how they actually reached that conclusion (Edmond et al. [7]). In contrast, methods based on data, quantitative measurement, and statistical models are transparent and reproducible: measurement (feature-extraction) and statistical-modelling/machine-learning methods can be described in detail, and data and software tools can potentially be shared with others.

Cognitive bias

There has been a great deal of concern about cognitive bias in forensic science (National Research Council [8]; Expert Working Group on Human Factors in Latent Print Analysis, EWG [9]; Found [10]; Stoel et al. [11]; PCAST [5]; Edmond et al. [7]; Cooper & Meterko [12]; Expert Working Group on Human Factors in Handwriting Examination [13]; Spellman et al. [14]). Cognitive bias is subconscious bias, it cannot be controlled by strength of will. Forensic practitioners are susceptible to cognitive bias when making perceptual observations: their degree of belief in the probability that a hypothesis is true can affect their analysis of the evidence. Since the output of the analysis is the input to the interpretation, bias in the former affects the latter. Forensic practitioners are susceptible to cognitive bias when they are making subjective judgements and are exposed to information that could influence their degree of belief in the probability that a hypothesis is true but that would not logically affect the probability of obtaining the evidence conditional on whether the hypothesis were true. Some potentially biasing information is task-irrelevant and should be withheld from practitioners, but some potentially biasing information is task-relevant and practitioners employing human-perception and subjective-judgement methods will need to be exposed to it at some point in the evaluation process, e.g., practitioners who visually compare known-source fingerprints and questioned-source fingermarks must be exposed to both, but exposure to a higher-quality print may bias their analysis of ambiguous details in a lower-quality mark. Systems in which the strength-of-evidence conclusion is directly the result of subjective judgement are particularly susceptible to cognitive bias. Systems based on quantitative measurements and statistical models require subjective judgements on decisions such as whether the data used for training the system and the data used for validating the system are sufficiently representative of the relevant population for the case and sufficiently reflective of the conditions of the items of interest in the case so that the output of the system will be a meaningful answer to the question posed in the case and so that the results of the validation will provide a meaningful indication of the performance of the systems under the conditions of the case. These decisions, however, are made at the beginning of the process before the practitioner has analyzed the items of interest, hence the practitioner cannot know what effect these decisions will have on the strength-of-evidence conclusion. The remainder of the evaluation process is automated, hence not susceptible to cognitive bias.

Likelihood-ratio framework

In current practice, interpretation of evidence is often logically flawed, e.g., it is based on the uniqueness or the individualization fallacy (Saks & Koehler [15]; Cole [16], [17]), and conclusions are often expressed categorically, e.g., “identification”, “inconclusive”, “exclusion” (i.e., posterior probability of 1 or 0 with respect to the same-source hypothesis, with “inconclusive” meaning no conclusion rather than an intermediate probability), or using some form of uncalibrated verbal posterior-probability scale, e.g., “identification”, “probable identification”, “inconclusive”, “probable exclusion”, “exclusion”. Jackson [18] and Kaye [19] review these and other commonly used but logically flawed conclusions. In contrast, the likelihood-ratio framework is advocated as the logically correct framework for evaluation of evidence by the vast majority of experts in forensic inference and statistics (including: Aitken et al. [20]; Morrison et al. [21]; Morrison et al. [22]; with 31, 19 and 20 authors and supporters respectively), and by key organizations including: Association of Forensic Science Providers of the United Kingdom and of the Republic of Ireland [23]. Royal Statistical Society (Aitken et al. [24]) European Network of Forensic Science Institutes (Willis et al. [25]) National Institute of Forensic Science of the Australia New Zealand Policing Advisory Agency (Ballantyne et al. [26]) American Statistical Association (Kafadar et al. [27]) Forensic Science Regulator for England & Wales (FSR) [28]. The likelihood-ratio framework requires assessment of: the probability of obtaining the evidence if one hypothesis were true versus the probability of obtaining the evidence if an alternative hypothesis were true The two hypotheses must be mutually exclusive. One hypothesis should represent the position of the prosecution in the case, and the other the position of the defence, e.g., the fingermark of questioned origin was deposited by a finger of a particular known individual, versus the fingermark of questioned origin was deposited by a finger of some other individual selected at random from the relevant population. In this example: the numerator of the likelihood ratio quantifies the similarity between the mark and the print the denominator quantifies the typicality of the mark with respect to the relevant population For continuously-valued data, likelihood ratios can be calculated as the ratio of two probability-density functions evaluated at the value of the evidence.4

Empirical validation

Empirical validation under conditions reflecting those of the case to which a forensic-evaluation system is to be applied is the only way to know how well that system performs under the conditions of the case. The need for validation under casework conditions has been emphasized by FSR [31], and by PCAST [5]: Without appropriate estimates of accuracy, an examiner’s statement that two samples are similar—or even indistinguishable—is scientifically meaningless: it has no probative value, and considerable potential for prejudicial impact. Nothing—not training, personal experience nor professional practices—can substitute for adequate empirical demonstration of accuracy. (p. 46) Protocols for validating systems that output likelihood ratios have been developed, including metrics and graphics appropriate for representing the results of such validations (Meuwly [32]; Brümmer & du Preez [33]; Morrison [34]; Meuwly et al. [35]; Ramos et al. [36]; Morrison et al. [22]). Much of the latter work has been conducted in the context of forensic voice comparison, but the results are applicable across forensic science in general.

A Kuhnian paradigm shift

The idea that evaluation of forensic evidence is undergoing a paradigm shift is not new. The most famous article heralding a paradigm shift is Saks & Koehler [4]. Allowing for differences in wording and level of detail and the passage of time, I believe that Saks & Koehler [4] and the present paper describe the same paradigm shift. In contrast to Saks & Koehler's [4] statement that they intended “paradigm shift” as a metaphor, however, I view the paradigm shift in evaluation of forensic evidence as a true Kuhnian paradigm shift (Kuhn [6]) in the sense that5 it requires rejection of existing methods and the ways of thinking that underpin them, and rejection of the idea that progress can be made by incremental improvements to existing methods. Instead, it requires the wholesale adoption of an entire constellation of new methods and new ways of thinking. That a paradigm shift requires the wholesale adoption of an entire constellation of new methods and new ways of thinking remains the case irrespective of whether one considers the shift to be from one paradigm to another or to be from a pre-paradigm to a paradigm period of science. As suggested in Saks & Koehler [4], a pre-paradigm period would seem to be a more accurate description of the status quo, with multiple traditions of evaluation of evidence used both within individual branches of forensic science and across different branches of forensic science, and hence there being no established widely-accepted overarching paradigm in use. Some authors have used the term “paradigm shift” in relation to a single element or a subset of the elements of the paradigm shift as I have outlined it above, but I believe that all of these elements are required as part of the constellation. This may be viewed as a radical stance, and it faces resistance, but, over the last decade and a half, my colleagues and I have had substantial success in contributing to advancing this paradigm shift in forensic voice comparison.

Impedimenta

The paradigm shift in evaluation of forensic evidence is ongoing, but progress is slow or stalling for multiple reasons including the following: The new paradigm has only been adopted in a few branches of forensic science, and only by a minority of researchers and practitioners (§2.3.2). Only some elements of the new paradigm have been adopted as part of incremental change (§2.3.3). There is misunderstanding of the new paradigm and resistance to its adoption (§2.3.4). Research is often not informed by practice and has no impact on practice (§2.3.5). It is difficult to obtain funding for evidential-forensic-science research (§2.3.6). There are genuine practical impediments to implementing the new paradigm (§2.3.7). I discuss each of these impediments in the following six subsections.

The new paradigm has only been adopted in a few branches of forensic sciences, and only by a minority of researchers and practitioners

In the 1990s, the new paradigm was widely adopted for forensic evaluation of DNA (Foreman et al. [37]). Although the volume and importance of casework in this branch of forensic science makes it influential, single-source DNA profiles are invariant and discrete. They therefore have a very different structure from the continuously-valued data with within-source variability that results from analyses in most other branches of forensic science. The situation is more complex for low-template DNA and for DNA mixtures, but there is still a difference in data structure. Interpretation of DNA profiles is also dependent on well-developed theory of genetic inheritance, whereas interpretation in most branches of forensic science will have to be data driven (as is the case in machine learning in general, including in biometric applications). The potential for transfer of new-paradigm knowledge and methods from DNA to other branches of forensic science is therefore limited. Since around 2000, a growing number of researchers and practitioners in forensic voice comparison have developed and adopted methods for calculation of likelihood ratios based on acoustic measurements and statistical models / machine-learning algorithms, and have developed and adopted methods for calibration and validation of likelihood-ratio systems under casework conditions. This has included adoption of state-of-the-art machine-learning approaches to automatic speaker recognition (Lee et al. [38]; Matějka et al. [39]; Villalba et al. [40]; Morrison et al. [41]; Morrison et al. [42]; Weber et al. [43]). At present, however, only a minority of practitioners have adopted the new paradigm. Survey results published in 2011, 2016, and 2019 (Gold & French [44]; Morrison et al. [45]; Gold & French [46]) suggest that, although the proportion of practitioners who have adopted human-supervised-automatic approaches and numeric likelihood ratios is growing, they are still a minority.6 In addition, inconsistent with the new paradigm, most respondents in the most recent survey who reported using a human-supervised-automatic approach also reported that they combined it with human-perception-based analysis and subjective-judgement-based interpretation. Data in human-supervised-automatic approaches are continuously valued and have intrinsic within-source variability, a data structure shared with many other branches of forensic science. Compared to DNA, new-paradigm knowledge and methods from forensic voice comparison, including statistical models and calibration and validation methods, should therefore be easier to transfer to and adapt for use other branches of forensic science. For an example of such transfer and adaptation, see Basu et al. [47]. Forensic voice comparison is, however, a relatively niche branch of forensic science, which limits the extent to which developments in forensic voice comparison are noticed and adopted by researchers and practitioners in other branches of forensic science. Curran [1] lamented that only 13% of laboratories surveyed used the likelihood-ratio framework for glass evidence, but this may be one of the highest rates of adoption after DNA. In many other branches of forensic science the rate of adoption of the likelihood-ratio framework by practitioners is near zero (Bali et al. [48]; Cole & Barno [49]).

Only some elements of the new paradigm have been adopted as part of incremental change

Although in the short term adopting some elements of the new paradigm as part of incremental change may be a viewed as a step in the right direction, in the long term it may actually impede a paradigm shift. Just because it is a transition between incommensurables, the transition between competing paradigms cannot be made a step at a time, … Like the gestalt switch, it must occur all at once (though not necessarily in an instant) or not at all. (Kuhn [6], p. 149) Some practitioners assign likelihood-ratio values based on subjective judgement, and the values they assign are not subject to empirical calibration or empirical validation (see Risinger [50]; Morrison & Thompson [51]; Morrison et al. [52]). Some authors emphasize the logic of the likelihood-ratio framework and consider subjective assignment of likelihood ratio an acceptable end goal or consider it a step in the right direction, but such incremental steps potentially impede a paradigm shift which requires the abandonment of interpretation methods that are entirely dependent on subjective judgement.7 In addition, placing an emphasis on subjectivist concepts of probability is detrimental to attempts to encourage practitioners to adopt methods based on relevant data, quantitative measurements, and statistical models, and to adopt empirical validation under casework conditions (Morrison [54]). The majority of proposals to address cognitive bias in forensic science (e.g., EWG [9]; Stoel et al. [11]; Thompson et al. [55]; FSR [56]) assume the continued use of human-perception- and subjective-judgement-based methods. Although this may be necessary in the short term, it potentially impedes a paradigm shift to quantitative-measurement- and statistical-model-based methods. Some practitioners make use of systems based on quantitative measurements and statistical models, but do not empirically calibrate or validate their system using data that reflect the relevant population and the conditions for the case, and/or rather than directly report the output of the system, they use it as input to a subjective-judgement process that also considers other information including from human-perception-based analyses (see Morrison & Thompson [51]; Morrison [57], [58]). Such approaches are pernicious in that use of technology may give the false impression of scientific validity, and reaction against this may impede a paradigm shift that includes adoption of quantitative measurements and statistical models.

There is misunderstanding of the new paradigm and resistance to its adoption

As with all Kuhnian paradigm shifts, there is misunderstanding of the new paradigm and resistance to its adoption. Some resitance is cultural. The cultures of some branches of forensic science seem to be especially resistant to the adoption of statistical-model-based methods and of validation (see Mnookin et al. [59]; Curran [1]; Morrison [60]; Morrison & Stoel [61]; Swofford et al. [62]). Practitioners in multiple branches of forensic science often claim that training and experience provide sufficient warrant for their conclusions (see Mnookin et al. [59]; Risinger [50]; PCAST [5]; Morrison & Thompson [51]), and deny or obfuscate about the need for validation (see Cole [63]; Morrison [60]; PCAST [5]; Koehler [64]; Morrison et al. [52]), or propose lax validation protocols that do not require demonstration of performance under casework conditions (see Morrison et al. [65], [66]). People in general tend to prefer methods which involve greater human input even when validation results indicate that data-driven methods lead to better results. Over time, however, people can come to accept data-driven methods (Swofford & Champod [67]). There is a belief that likelihood ratios are difficult to understand (see Bali et al., 2020; Swofford et al. [62]; Swofford & Champod [68]). Commonly occurring misunderstandings have even been given names, e.g., the “prosecutor's fallacy” and the “defense attorney's fallacy” (Thompson & Schurmann [69]). There are many examples of legal rulings in which judges have misunderstood the meaning of a likelihood ratio (the England & Wales Court of Appeal 2010 ruling in R v T is an infamous example; see, e.g., Berger et al. [70]; Redmayne et al. [71]; Morrison [72]; Thompson [73]). Results of empirical research on lay understanding of expressions of strength of evidence are mixed (Eldridge [74]; Martire & Edmond [75]). Despite legal rulings and recommendations concerning the need for validation, courts often do not understand empirical validation and its necessity, and accept testimony based on forensic-evaluation methods that have not been validated under conditions reflecting those of the case under consideration, or even that have not been empirically validated at all (see Bernstein [76]; Morrison [60], [57], [58]; Cooper [77]; Edmond [78]).

Research is often not informed by practice and has no impact on practice

Research that appears to be about forensic science may not actually be about solving real forensic-science problems. For example, research may really be about a method in statistics or machine learning with an apparent forensic-science problem being used as an example of the application of that method. Research in forensic science is sorely needed, but it should address primarily forensic science questions—not questions relating to the application of chemistry, biology, statistics, or psychology. (Margot [79], p. 801) it is critical that researchers and funding bodies understand the importance of conducting research that is informed by practice and can be translated into practical applications. (Roux & Weyermann [80], p. 2) Research that is divorced from forensic practice may lead to academic papers but nothing else. Even research that does address real forensic-science problems will fail to have impact unless it involves genuine collaboration in which researchers understand the demands of practice, and in which practitioners are willing to embrace research-informed change (Curran [1]).

It is difficult to obtain funding for evidential-forensic-science research

Few funding agencies have sustained funding targeted at forensic science, and funding agencies seldom have panels of reviewers knowledgeable about evidential forensic science. Applications for funding for evidential-forensic-science research made to non-forensic-science-targeted calls are often rejected because reviewers do not understand the epistemology or value of forensic-science research. Applications are often rejected because their goals are to improve forensic science, which is an applied science, and funding-agency criteria or reviewers often do not value applied science. The larger scientific community must now come to the aid of our forensic colleagues in advocating both for: (i) the research and financial support that is so clearly needed to advance the field and (ii) the requirement for empirical testing that is so clearly needed to advance the cause of justice …. Forensic scientists have long complained that their work is not always valued by their scientific colleagues because of its applied nature; it is time for the scientific community to move beyond that conceit. (Bell et al. [81]) At the other extreme, when there are calls specific to forensic science, they usually focus exclusively or primarily on short-term goals related to law-enforcement investigative applications rather than on courtroom-evidential applications (investigative and evidential applications have very different requirements), and they usually focus on technology rather than on forensic inference. technology-oriented development … often overrul[es] the importance of appropriate scientific reasoning to solve actual problems (Roux et al. [82], p. 679) Research calls requiring deliverables with a high technology readiness level (TRL) are not sources of funding for paradigm-shifting research. In 2018, United Kingdom Research and Innovation (UKRI) informed the UK House of Lords Science and Technology Select Committee that UKRI had invested GBP 56 M over 10 years in forensic-science research (less than 0.1% of UKRI's total budget), but on closer inspection most of the funding counted to obtain that figure was not for research projects that actually focused on (or even made any contribution to) forensic science: only about GBP 17 M went to forensic-science focussed research, and GBP 15 M of that went to TRL research, only GBP 2 M to foundational research (HoL [2]; Morgan & Levin [83]). HoL [2] recommended that UKRI “urgently and substantially increase the amount of dedicated funding allocated to forensic science” (§187), but, more than 3 years after the publication of the HoL report, this has not (yet) happened.

There are genuine practical impediments to implementing the new paradigm

Even if practitioners want to adopt the new paradigm, they will be unable to do so unless they are provided with the quantitative-measurement and statistical-modelling/machine-learning tools and the case-relevant data necessary to calculate likelihood ratios and validate systems under the conditions of the cases on which they work. Practitioners will also not be able to adopt the new paradigm unless they are provided with training on understanding the new paradigm in general and on how to implement it for the types of cases they work on.

Via progredi

Kuhn [6]: The transfer of allegiance from paradigm to paradigm is a conversion experience that cannot be forced …. a generation is sometimes required to effect the change … Conversions will occur a few at a time until, after the last holdouts have died, the whole profession will again be practicing under a single, but now a different, paradigm. (pp. 150–151) Kuhnian paradigm shifts are not rapid and individuals cannot be forced to embrace the new paradigm, but my aim is to facilitate and thereby advance the adoption of the new paradigm. My strategy is to work with researchers and practitioners who want to adopt the new paradigm, to work with them on addressing practical impediments to applying the new paradigm in casework, i.e.: To provide researchers, practitioners, and lawyers with training leading to understanding of the new paradigm; To collaborate with researchers and practitioners on building relevant databases and on developing and validating statistical models applicable in their particular branches of forensic science; and To conduct research on how to present likelihood ratios and validation results so as to maximize understanding by laypeople, and thereby provide guidance to forensic practitioners on how to communicate forensic-evaluation results to legal-decision makers. I invite others to consider whether this is a strategy that they would also be interested in adopting, either in part or in whole. Element 2 of the strategy requires collaboration between researchers with expertise in forensic data science and researchers and practitioners with expertise in particular branches of forensic science. Academic publications are unlikely to convince practitioners to adopt the new paradigm, but other practitioners successfully applying the new paradigm are potentially convincing. In any branch of forensic science, the number of practitioners who initially want to adopt the new paradigm and who want to collaborate on this endeavour will almost certainly be a very small minority, but it will be more productive to work with a small minority on developing practical solutions than to try to convince the majority of practitioners without providing practical solutions. Once the practical solutions are being used by the small minority, use of the new paradigm has the potential to spread. Even then, I do not expect adoption of the new paradigm to be rapid, but I do expect higher rates of adoption among newer practitioners and trainees, leading to a generational shift.

Conclusion

A paradigm shift in evaluation of forensic evidence is ongoing. The shift is away from methods based on human perception and subjective judgement, to methods based on: relevant data, quantitative measurements, and statistical models. New paradigm methods: are transparent and reproducible are intrinsically resistant to cognitive bias use the logically correct framework for interpretation of evidence (the likelihood-ratio framework) are empirically validated under casework conditions This is a Kuhnian paradigm shift, which requires: rejection of existing methods and the ways of thinking that underpin them rejection of the idea that progress can be made by incremental improvements to existing methods the wholesale adoption of an entire constellation of new methods and new ways of thinking Some branches of forensic science, such as forensic voice comparison, are more advanced in the paradigm shift than others. Knowledge gained in advancing the paradigm shift in forensic voice comparison can assist in advancing the paradigm shift in other branches of forensic science. Validation protocols, metrics, and graphics developed in the context of forensic voice comparison are immediately applicable in other branches of forensic science. Statistical models / machine-learning algorithms used in forensic voice comparison can even be transferred and adapted for use in other branches of forensic science. My strategy for advancing the paradigm shift requires collaboration between researchers with expertise in forensic data science and researchers and practitioners with expertise in particular branches of forensic science. My strategy is to work with researchers and practitioners who want to adopt the new paradigm. My strategy is to work with them on addressing practical impediments to applying the new paradigm in casework, i.e.: To provide researchers, practitioners, and lawyers with training leading to understanding of the new paradigm; To collaborate with researchers and practitioners on building relevant databases and on developing and validating statistical models applicable in their particular branches of forensic science; and To conduct research on how to present likelihood ratios and validation results so as to maximize understanding by laypeople, and thereby provide guidance to forensic practitioners on how to communicate forensic-evaluation results to legal-decision makers. In any branch of forensic science, the number of practitioners who initially want to adopt the new paradigm and who want to collaborate on this endeavour will almost certainly be very small, but it will be more productive to work with a small minority on developing practical solutions than to try to convince the majority of practitioners without providing practical solutions. Once the practical solutions are being used by the small minority, use of the new paradigm has the potential to spread. Even then, I do not expect adoption of the new paradigm to be rapid, but I do expect higher rates of adoption among newer practitioners and trainees, leading to a generational shift. I have been asked several times over the years whether I could suggest a name for the new paradigm other than “new”. Here, I propose that the new paradigm could be called “forensic data science”. My hope is that, after the paradigm shift is complete, it will simply be called “forensic science”.

Disclaimer

All opinions expressed in the present paper are those of the author, and, unless explicitly stated otherwise, should not be construed as representing the policies or positions of any organizations with which the author is associated.

Declaration of competing interest

The author declares that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  32 in total

1.  Measuring the validity and reliability of forensic likelihood-ratio systems.

Authors:  Geoffrey Stewart Morrison
Journal:  Sci Justice       Date:  2011-04-14       Impact factor: 2.124

2.  Cognitive bias research in forensic science: A systematic review.

Authors:  Glinda S Cooper; Vanessa Meterko
Journal:  Forensic Sci Int       Date:  2019-01-22       Impact factor: 2.395

3.  A comment on the PCAST report: Skip the "match"/"non-match" stage.

Authors:  Geoffrey Stewart Morrison; David H Kaye; David J Balding; Duncan Taylor; Philip Dawid; Colin G G Aitken; Simone Gittelson; Grzegorz Zadora; Bernard Robertson; Sheila Willis; Susan Pope; Martin Neil; Kristy A Martire; Amanda Hepler; Richard D Gill; Allan Jamieson; Jacob de Zoete; R Brent Ostrum; Amke Caliebe
Journal:  Forensic Sci Int       Date:  2016-10-26       Impact factor: 2.395

Review 4.  Distinguishing between forensic science and forensic pseudoscience: testing of validity and reliability, and approaches to forensic voice comparison.

Authors:  Geoffrey Stewart Morrison
Journal:  Sci Justice       Date:  2013-08-13       Impact factor: 2.124

5.  INTERPOL survey of the use of speaker identification by law enforcement agencies.

Authors:  Geoffrey Stewart Morrison; Farhan Hyder Sahito; Gaëlle Jardine; Djordje Djokic; Sophie Clavet; Sabine Berghs; Caroline Goemans Dorny
Journal:  Forensic Sci Int       Date:  2016-04-01       Impact factor: 2.395

6.  A guideline for the validation of likelihood ratio methods used for forensic evidence evaluation.

Authors:  Didier Meuwly; Daniel Ramos; Rudolf Haraksim
Journal:  Forensic Sci Int       Date:  2016-04-26       Impact factor: 2.395

7.  Probabilistic reporting and algorithms in forensic science: Stakeholder perspectives within the American criminal justice system.

Authors:  H Swofford; C Champod
Journal:  Forensic Sci Int Synerg       Date:  2022-02-12

8.  Challenges to reasoning in forensic science decisions.

Authors:  Barbara A Spellman; Heidi Eldridge; Paul Bieber
Journal:  Forensic Sci Int Synerg       Date:  2021-12-20

9.  Communicating forensic science opinion: An examination of expert reporting practices.

Authors:  Agnes S Bali; Gary Edmond; Kaye N Ballantyne; Richard I Kemp; Kristy A Martire
Journal:  Sci Justice       Date:  2019-12-23       Impact factor: 2.124

Review 10.  Implementation of algorithms in pattern & impression evidence: A responsible and practical roadmap.

Authors:  H Swofford; C Champod
Journal:  Forensic Sci Int       Date:  2021-02-18       Impact factor: 2.395

View more
  2 in total

1.  The opacity myth: A response to Swofford & Champod (2022).

Authors:  Geoffrey Stewart Morrison; Nabanita Basu; Ewald Enzinger; Philip Weber
Journal:  Forensic Sci Int Synerg       Date:  2022-06-19

2.  Forensic comparison of fired cartridge cases: Feature-extraction methods for feature-based calculation of likelihood ratios.

Authors:  Nabanita Basu; Rachel S Bolton-King; Geoffrey Stewart Morrison
Journal:  Forensic Sci Int Synerg       Date:  2022-05-27
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.