Why do latent fingerprint examiners differ in their conclusions?

•Differing conclusions explained in terms of image effects and examiner effects.•Variability among examiners due to implicit individual decision thresholds.•Innovative method of comparing performance of human forensic examiners.•3-level conclusion scale does not precisely represent examiners’ conclu...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Forensic science international 2020-11, Vol.316, p.110542-110542, Article 110542
Hauptverfasser: Hicklin, R. Austin, Ulery, Bradford T., Ausdemore, Madeline, Buscaglia, JoAnn
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:•Differing conclusions explained in terms of image effects and examiner effects.•Variability among examiners due to implicit individual decision thresholds.•Innovative method of comparing performance of human forensic examiners.•3-level conclusion scale does not precisely represent examiners’ conclusions. Forensic latent print examiners usually but do not always reproduce each other’s conclusions. Using data from tests of experts conducting fingerprint comparisons, we show the extent to which differing conclusions can be explained in terms of the images, and in terms of the examiners. Some images are particularly prone to disagreements or erroneous conclusions; the highest and lowest quality images generally result in unanimous conclusions. The variability among examiners can be seen as the effect of implicit individual decision thresholds, which we demonstrate are measurable and differ substantially among examiners; this variation may reflect differences in skill, risk tolerance, or bias. Much of the remaining variability relates to inconsistency of the examiners themselves: borderline conclusions (i.e., close to individual decision thresholds) often were not repeated by the examiners themselves, and tended to be completed more slowly and rated difficult. A few examiners have significantly higher error rates than most: aggregate error rates of many examiners are not necessarily representative of individual examiners. The use of a three-level conclusion scale does not precisely represent the underlying agreements and disagreements among examiners. We propose a new method of quantifying examiner skill that would be appropriate for use in proficiency tests. These findings are operationally relevant to staffing, quality assurance, and disagreements among experts in court.
ISSN:0379-0738
1872-6283
DOI:10.1016/j.forsciint.2020.110542