- Original article
- Open Access
Gaze behavior and cognitive states during fingerprint target group localization
© This is a U.S. government work and not under copyright protection in the U.S.; foreign copyright protection may apply 2019
- Received: 8 August 2018
- Accepted: 20 February 2019
- Published: 5 April 2019
The comparison of fingerprints by expert latent print examiners generally involves repeating a process in which the examiner selects a small area of distinctive features in one print (a target group), and searches for it in the other print. In order to isolate this key element of fingerprint comparison, we use eye-tracking data to describe the behavior of latent fingerprint examiners on a narrowly defined “find the target” task. Participants were shown a fingerprint image with a target group indicated and asked to find the corresponding area of ridge detail in a second impression of the same finger and state when they found the target location. Target groups were presented on latent and plain exemplar fingerprint images, and as small areas cropped from the plain exemplars, to assess how image quality and the lack of surrounding visual context affected task performance and eye behavior. One hundred and seventeen participants completed a total of 675 trials.
The presence or absence of context notably affected the areas viewed and time spent in comparison; differences between latent and plain exemplar tasks were much less significant. In virtually all trials, examiners repeatedly looked back and forth between the images, suggesting constraints on the capacity of visual working memory. On most trials where context was provided, examiners looked immediately at the corresponding location: with context, median time to find the corresponding location was less than 0.3 s (second fixation); however, without context, median time was 1.9 s (five fixations). A few trials resulted in errors in which the examiner did not find the correct target location. Basic gaze measures of overt behaviors, such as speed, areas visited, and back-and-forth behavior, were used in conjunction with the known target area to infer the underlying cognitive state of the examiner.
Visual context has a significant effect on the eye behavior of latent print examiners. Localization errors suggest how errors may occur in real comparisons: examiners sometimes compare an incorrect but similar target group and do not continue to search for a better candidate target group. The analytic methods and predictive models developed here can be used to describe the more complex behavior involved in actual fingerprint comparisons.
- Eye tracking
- Latent fingerprints
Visual localization is a critical task within the fingerprint comparison process, in which the fingerprint examiner analyzes and memorizes a specific area of detail in one fingerprint and searches for the corresponding area in another fingerprint. This study isolates individual localization tasks, and details how eye-gaze behavior can be used to describe and analyze how fingerprint examiners perform localization.
This study is the first detailed evaluation of localization in fingerprint comparison. Latent fingerprint comparison is critically important within the criminal justice system. A variety of studies have shown that the accuracy and reliability of conclusions by fingerprint examiners are imperfect (e.g., Neumann, Champod, Yoo, Genessay, & Langenburg, 2013; Pacheco, Cerchiai, & Stoiloff, 2014; Ulery, Hicklin, Buscaglia, & Roberts, 2011). Such studies show that some examiners (rarely) make erroneous identifications, erroneous exclusions are much more prevalent than erroneous identifications, and examiners often are inconclusive on comparisons that resulted in identifications from other examiners. Understanding why such errors and disagreements occur requires detailed evaluation of fingerprint examiners. Here we evaluate a critical subtask of fingerprint comparison in order to gain a greater understanding of how experts compare fingerprints, and why they differ.
Trained latent print examiners visually compare fingerprints to determine whether the prints share enough detail in agreement to conclude that they came from the same source. Currently, human experts (not computers) make the final conclusions on latent print comparisons, using their expertise rather than a quantitative standard to determine if the information content is sufficient to make a decision. The entire fingerprint examination process involves a number of perceptual and memory-based mechanisms, but one of the fundamental subtasks is target group localization: selecting a small area of distinctive features (a target group), analyzing and memorizing that specific area of detail in one image, and searching for the corresponding area in another image (Ashbaugh, 1999; Expert Working Group on Human Factors in Latent Print Analysis, 2012; National Institute of Justice, 2011). During comparison, the examiner iteratively selects and compares additional target groups, accumulating evidence to make a decision as to whether the two impressions came from the same or different sources. One challenge when working with fingerprint experts is that there are substantial differences among examiners: previous work has shown that latent print examiners vary significantly in their determinations (Pacheco et al., 2014; Ulery et al., 2011; Ulery, Hicklin, Buscaglia, & Roberts, 2012), as well as in the features used as the bases for their decisions (Langenburg, Champod, & Genessay, 2012; Neumann et al., 2013; Ulery, Hicklin, Roberts, & Buscaglia, 2014, 2016). Because target group localization is at the core of the fingerprint comparison process, understanding localization should provide insight into that process, and provide a basis for understanding how and why examiners vary in their conclusions. The fingerprint discipline can use such information in improving processes and minimizing sources of errors — and the broader forensic science and legal communities may use such information to more completely understand the reliability of conclusions by fingerprint examiners.
Deconstructing tasks into constituent elements
Isolating specific subtasks can make the study of complex visual behavior tractable (Hayhoe, 2000; Johnson, Sullivan, Hayhoe, & Ballard, 2014). A full fingerprint comparison involves the localization of multiple target groups, which in aggregate provide support for (or against) a source determination. Each time a target group is selected, those features are placed into visual working memory (Baddeley, 1986), which is used when comparing against details in the other print. Establishing correspondence relies on the quality of the impressions and the distinctiveness of the configuration of the features in the target group. Eye behavior is constrained by the small features (minutiae), the narrow visual field of the fovea, and the capacity and persistence of visual working memory. The process of finding an area that corresponds to a selected target group typically requires multiple fixations on the latent impression, and then multiple fixations on the exemplar impression (Busey et al., 2011; Busey, Yu, Wyatte, & Vanderkolk, 2013; Busey, Swofford, Vanderkolk, & Emerick, 2015).
It may be instructive to compare this target group localization task to a “Where’s Waldo” task (e.g., Credido, Teixeira, Reis, Moreira, & Andrade, 2012) using elements of Kundel’s deconstruction of detection (Kundel, Nodine, & Carmody, 1978). These elements include orientation (“what do I have to work with?” — assessing the overall image); scanning (“where is it?” — looking for potential locations); recognition (“is this it?” — recognizing as a potential target when seen); and decision (“am I sure?” — deciding whether it is the correct target). A number of authors have extended Kundel’s idea of inferring underlying cognitive states from eye-gaze behavior and it has proven to be useful when deconstructing a task into constituent elements (e.g., Kardan, Berman, Yourganov, Schmidt, & Henderson, 2015; Land, Mennie, & Rusted, 1999; Marshall, 2007). Ballard, Hayhoe, and Pelz (1995) used a block copying task to identify when participants relied on working memory and when they chose to externalize this memory to the physical world. Later work by Hayhoe (2000) extended this approach and identified specific visual routines as well as the information used by each when performing a block copying or driving task. Similar cognitive states may underlie fingerprint target localization as examiners progress from a search phase to a decision phase, and there may be characteristic eye behaviors associated with each phase.
The role of context
In the scanning state of “Where’s Waldo,” knowing what Waldo looks like does not provide any information regarding where to look in the image because context is deliberately minimized or eliminated by the artist. With a large fingerprint impression, however, the overall ridge flow and pattern type provides extensive contextual information useful in localizing a target group. This contextual information need not be explicitly provided: skilled examiners often can look at a portion of a fingerprint and infer the approximate location on the finger from its local ridge details. In addition, context may improve the interpretation of the details (Barenholtz, 2014). In recognition, the Where’s Waldo pictures contain similar-looking decoys that slow the process of finding the target; as we will see, fingerprints may also contain such decoys. After Waldo has been recognized, it is generally trivial to decide that it is Waldo and not a decoy; however, two impressions of the same finger may vary considerably in appearance and quality, so the process of deciding whether two target groups correspond is often not trivial. Relational information from macro features, such as the core or delta, may help the examiner eliminate some decoys, whereas relational information is made deliberately uninformative in the Where’s Waldo task. Thus, whereas Where’s Waldo can be viewed as a fairly traditional visual search task, fingerprint comparisons may involve a tight interplay between the appearance of individual features and the broader context in which they appear.
Natural scenes may serve as a good model for fingerprints because they also contain structural information that provides context that guides the interpretation of individual features. Of the few studies that have addressed the eye-gaze behavior of fingerprint examiners (Busey et al., 2011, 2013, 2015), none has addressed the effect of context. However, working with structured displays, Chun and Jiang (1998, 1999) found that spatial configuration information can indirectly guide search to the target because the visual covariation of the structure of a display is learned; similar experience may help fingerprint examiners infer the general location of a region from a partial print (e.g., deltas are generally found below the core). Neider and Zelinsky (2006) addressed the role of context in natural scene perception using target images (blimps and cars) that were either presented in their correct context or in an incorrect context (sky or ground). They argued that contextual information creates expectations that guide search behavior by biasing searches toward target-consistent regions. Similar effects were modeled by Oliva and Torralba (2007) to demonstrate how the scene structure and prior knowledge of the world can be used to construct a saliency map that adds contextual influences to bottom-up processes. The general consensus of the literature is that objects presented in context are recognized faster and more accurately than objects that are not, and that context might influence the search and localization of a target feature through top-down mechanisms (but see Heeger, 2017 for a bi-directional approach).
Fingerprint impressions contain local pattern information that could provide the basis for expectations that guide search to regions that are more likely to contain a target region. Similar biasing was identified by Pomplun (2006), who asked participants to first encode a small target and then search for that target in a naturalistic image such as a clock face or landscape. His results suggested that subjects were sensitive to the similarity of the target and various regions of the image, and quantified this using a measure termed saccadic selectivity. Observers tended to saccade to regions of the searched image that bear similarity to the target, although this can depend on the relevant dimension in a particular task. For example, if color is irrelevant, observers will shift to other dimensions, such as shape, as the basis for their saccades. Based on his results, we would expect clusters of distributions on regions that are similar in appearance to our target region, although surrounding spatial context may help eliminate some decoy regions.
Assess the effects of context: our first goal is to assess the effects of visual context on the eye behavior of examiners, and what this reveals about the perceptual and cognitive mechanisms that support localization
Deconstruct the localization task: our second goal is to assess whether we can deconstruct the localization task, mapping observable eye-gaze behavior to underlying cognitive states such as searching or deciding
Develop analytic tools: our third goal is to develop tools to describe and analyze eye-gaze behavior in fingerprint comparison tasks
To accomplish these goals, we proceed in several stages. First, in the “Describing eye-gaze behavior in localization” section, we explore a variety of descriptive methods: these methods and the resulting findings have both theoretical and applied interest to the scientific and forensic communities. In the “Characterizing eye-gaze behavior in localization” section, we define methods to categorize and characterize eye-gaze behavior during localization. In the “Inferring cognitive states” section, we then use these methods to infer underlying cognitive states from the overt behaviors.
Looking beyond the current study, the longterm purpose of this work will be to use the results, lessons learned, methods, and metrics from this experiment as the basis for ongoing and future analyses of fingerprint comparisons, specifically in order to understand the extent to which errors and differences in conclusions can be explained by differences in behavior.
“Find the target group — You will be shown an image on the left with a target group already identified/indicated (a square about eight ridges on a side). The target group may be shown as an outlined area within a full exemplar or latent, or as a small cropped area shown by itself. You will then be shown a rolled exemplar from the same source (same finger, same subject, with the same orientation). Find the indicated target group, and when you are confident that you have found it, say DONE. The target group will always be present. (The purpose is so that we can understand in isolation the behavior of your eyes when memorizing a group of features, and searching for that group of features — which we expect to see often in standard comparisons. We are not trying to trick you (the source is always there): we are trying to see how you memorize and search for a target group.)”
Eight image sets were selected from the “Ground Truth” dataset distributed with the Federal Bureau of Investigation’s (FBI) Universal Latent Workstation (ULW, n.d.). Each image set was constructed from three impressions of the same finger (latent, plain (flat) exemplar, rolled exemplar) scanned at 39.4 pixels per millimeter (ppmm) — note that the images were scanned at 1000 pixels per inch in accordance with the prevailing standard (ANSI/NIST-ITL, 2013), and therefore metric equivalents are rounded. When presented to an examiner, the image on the left was either (a) the latent image, (b) the plain exemplar image, or (c) a small area cropped from the plain exemplar image; the mated, high-clarity rolled exemplar was always shown on the right side of the screen. The images from two image sets are shown in Fig. 1.
A small area of each finger was selected as the “target group.” This target group was shown outlined in yellow on the latent and plain image; the cropped image was identical to the target group outlined on the plain image but without the surrounding context. Each target group area was a 150 × 150 pixel square (3.8 mm × 3.8 mm; approximately eight ridges across assuming an average ridge-ridge distance of 0.56 mm). Because of the plasticity of the skin, the areas on the latent and plain images were not strictly identical. Thus, eight image sets (FT1–FT8) were defined; within each set three image pairs were defined, with the left image varying based on the task type (latent, plain, cropped), but with the same corresponding rolled exemplar used as the right image. We refer to image pairs as FT1Latent, FT1Plain, FT1Crop, etc. The image in the plain and cropped tasks was identical except for the visual context included in the plain tasks, so that comparing plain to cropped would isolate the effect of context.
The target areas were selected to provide a moderate amount of pattern-level information, so that the task was tractable but not obvious. We considered selecting low-information target groups (e.g., areas where ridges are relatively parallel with few minutiae, or targets cropped from the latent prints), but rejected that because it would change the scenario to one where the participants might not have enough information to complete the task. Conversely, we also avoided highly distinctive areas: the reason that we stopped using target groups FT7 and FT8 was because they were too obvious (clearly core and delta formations, respectively).
In order to isolate the specific task of finding the target and determine the role of context, we deliberately only used fingerprints from the same source (mated image pairs), and told the examiners that they were mated. Using only mated image pairs was necessary: if examiners were also deciding whether the images were mated, they presumably would have compared in detail regions outside the target group. This would have turned our task into a full examination, instead of the intended subset of the comparison process.
When assessing whether a fixation is considered in the target area, we add a margin of 30 pixels (just over one ridge width) to the 150 × 150 pixel target to allow for factors such as eye-tracker measurement imprecision (e.g., calibration), foveal field of view, distortion, and skin elasticity.
Data collection setup
Presentation code was written in MATLAB (MathWorks, 2012) using functions from the Psychtoolbox for image presentation (Kleiner, Brainard, & Pelli, 2007; Pelli, 1997), and the Eyelink Toolbox for coordination with the eye tracker (Cornelissen, Peters, & Palmer, 2002). The software interface allowed examiners to zoom in and out at the mouse location using keyboard presses, which were also used to pan the image. The two images were presented separately on each half of the monitor and could be zoomed independently. A software tool to mark and link features such as minutiae was available, and was used on 10% of the target-finding trials. Contrast inversion was also available but seldom used for the present task. The default zoom level was 1:1 (one screen pixel to one image pixel); because the screen resolution was 3.7 ppmm and the image resolution was 39.4 ppmm, there was an effective magnification of 11x. The current zoom levels and markup mode were displayed in the upper-left portion of the monitor as shown in Fig. 2; fixations near that textbox were excluded from our analyses. There were no other user-interface elements (e.g., scrollbars, toolbars, buttons) displayed on the screen.
At the start of the trial, the left image was displayed. When the examiner verbally indicated that they were ready to proceed with the Comparison phase of the trial, the experimenter displayed the right image. When the examiner said that they had found the target, the experimenter terminated the trial and triggered the final drift correction dots.
Gaze location (from both eyes) was recorded at 1 kHz using an EyeLink 1000 eye tracker. The EyeLink camera and illuminator were positioned immediately in front of the monitor on the table, and could be adjusted left or right if necessary to eliminate glare on glasses. We generally used 75% illumination within the EyeLink system, but bifocal contacts or thick glasses that tended to block IR illumination sometimes required 100% illumination.
At the start of, and periodically during, the experiment, the experimenter calibrated the eye tracker. Calibration was done within the EyeLink software using 13 calibration dots; gaze locations were tracked using the centroid measurement technique. Calibration was re-attempted as necessary, with a goal of obtaining an average error calculation of 0.5° for both eyes. A value of 0.8° was considered acceptable, although one examiner was only able to achieve 0.94° and 0.97° after seven calibrations, but was still included in the analyses. The calibration error goal of 0.5° corresponds to about ± one ridge (25 screen pixels) at a 1:1 zoom level (68% of fixations in this task were at 1:1). The foveal field of view (area of high acuity) typically corresponds to about 2°, or a diameter of about 100 screen pixels; at a 1:1 zoom level, this is 2.5 mm in image coordinates, or about 4.5 ridges.
In addition to the overall calibration that was managed within the eye tracker, we collected additional data to allow us to correct for gaze drift in post-processing. Each trial began and ended with seven drift correction dots presented in sequence; the examiner was instructed to fixate on each dot of known location for 1.25 s. Post-processing corrected for systematic drift on each trial by comparing gaze points within a threshold distance (2°) from the dot during the 1.25-s interval against the ground truth location. A clustering algorithm based on the mean shift algorithm (Cheng, 1995) was used to determine the largest cluster of at least 20 successive 1-kHz gaze points within the 2° radius; this cluster center was taken as the intended location of gaze for that drift correction dot. By working with raw gaze points, we take advantage of the density of the gaze points (potentially over multiple fixations and micro-fixations) to determine the intent of gaze during the drift correction procedure. Final drift correction was done using QR decomposition (Francis, 1961) that found a second-order polynomial transformation in both dimensions (six total coefficients) that allowed for translation, rotation, and scaling of the gaze points to the ground truth locations. This transformation was then applied to all fixations extracted from the raw gaze stream, as described next.
Each trial consists of a series of eye locations over time: an (x,y,t) path of raw 1-kHz data, which was processed to differentiate saccades and fixations. Because visual information is only coarsely represented during saccades (Ross, Morrone, Goldberg, & Burr, 2001), and because much of the relevant information for fingerprint comparisons is found in smaller details (minutiae), fixations were the fundamental unit of analysis for most of the study results. The raw gaze stream (1-kHz sample data) was partitioned into fixations and saccades using the following approach. We used the Engbert-Mergenthaler saccade detector (Engbert & Mergenthaler, 2006) with a velocity threshold of 8 pixels/s (0.16°/s) and a saccade duration minimum threshold of 9 ms to identify long saccades. As in Port, Trimberger, Hitzeman, Redick, and Beckerman (2016), a saccade required that the eye remain at rest for at least 20 ms within a ± 0.25° X-Y positional window. The Engbert-Mergenthaler saccade detector (Engbert & Mergenthaler, 2006) does a very good job of detecting saccades in which the eye moves more than 1.5° (Port et al., 2016). However, saccades of less than 1.5° tend to include a mixture of what subjectively appear to be even shorter saccades and fixations that are close together. Differentiating saccades from fixations is further complicated by the fact that fingerprint examiners may make very regular, closely spaced, fixations when counting or following ridges. A single saccade detector with a high-velocity threshold may risk grouping together several close fixations into a single fixation (of long duration) that may not accurately represent the detailed behavior. However, a low threshold risks producing many spurious fixations and saccades. Thus, we modified the Engbert-Mergenthaler saccade detector using a variation of the double-threshold algorithm, which is popular in many image-processing applications (e.g., Chen, Sun, Heng, & Xia, 2008) where distance can complement a threshold applied to some other value such as image intensity. Similar approaches have been used for eye-tracking analyses on fingerprint examiners (Busey et al., 2013, 2015; Parada et al., 2015).
Fixations were required to be longer than 80 ms, although Manor and Gordon (2003) argued that a minimum of 100 ms can also be justified. The centroid of the fixation was computed as the median x and y location of all of the raw gaze points that are identified as part of the fixation. The resulting mean fixation duration was 320 ms (median 270 ms; quartiles 198–373 ms). The median saccade duration was 21 ms within an image; median duration for crossing saccades (between the left and right image) was 76 ms. Details in Additional file 1: Appendix SI-4.1.
The above procedures determine fixation centroids projected onto the monitor coordinates. These coordinates were then adjusted by the aforementioned drift correction procedures. Because our software allows for scaling and panning of the images, an additional step was required to project the drift-corrected fixations into the image coordinates using the scaling and panning parameters active during that fixation. Additional information regarding fixation extraction methods used and their accuracy are shown in Additional file 1: Appendix SI-4.2.
Each participant was assigned a sequence of fingerprint comparisons, interspersed with three types of directed tasks. In addition to the find-the-target task that is the focus of this paper, the directed tasks included ridge following and ridge counting; results of the fingerprint comparisons and other directed tasks will be reported in subsequent papers. Testing occurred in June–August 2016 in six locations in Ohio, Indiana, Virginia, Kentucky, and Georgia. Participants were provided with written instructions prior to the test. An experimenter then verbally summarized the instructions and answered any questions. Participants were requested to continue testing for 2 h or until all of the assigned trials were completed; however, participants were permitted to stop early or continue after the 2-h time period.
Participation was open to practicing latent print examiners who are currently doing casework or have done casework within the last year. Participants gave informed consent after reviewing a human subject consent form approved by the Federal Bureau of Investigation Institutional Review Board prior to the start of the study. A total of 122 examiners participated: 39% were from federal agencies, 31% state, 22% local, 5% international, and 2% private. Seventy-nine percent were from accredited laboratories. Seventy-six percent had 5 or more years of experience as a latent print examiner; none had less than 1 year. Nineteen percent wore glasses, 29% had contact lenses, and 7% had LASIK. No participants were required by their employers to participate. Participants were assured that their results would remain anonymous; a coding system was used to ensure anonymity during our analyses and in reporting. Usable eye-tracking data for the present task was collected from 117 participants: of the total 122, four participants were tested during an initial phase of data collection in which find-the-target tasks were not assigned, and data from one participant was unusable due to a corrupt file. (See Additional file 1: Appendix SI-2 for further details on participants.)
The 117 participants completed a total of 675 valid trials (two invalid trials were omitted). Each examiner was assigned only one type of task (latent, plain, cropped) from each image set (FT1–FT8); therefore, no examiner was assigned two trials involving the same source finger. Because participants were allowed to stop early or continue, some completed as few as two or as many as eight trials; most participants (87) completed at least two trials of each type (latent, plain, cropped). Image sets FT7 and FT8 were retired early during the course of testing to better use the limited time with each examiner; due to the smaller resulting sample sizes, these two image sets are omitted from analyses aggregated by image pair. From 32 to 39 examiners completed each of the tasks from FT1Latent through FT6Cropped (18 image pairs); only seven to nine examiners completed each of FT7Latent through FT8Cropped (six image pairs). The 675 trials included a total of 53,093 valid fixations; for analyses omitting FT7 and FT8, there were 630 trials with 49,242 valid fixations. (See Additional file 1: Appendix SI-5 for further details regarding omitted data and test yield.)
In this section, we describe the behavior of examiners during the localization process, and show the effects of context. We also showcase techniques that we developed to visualize and describe eye-gaze behavior.
Context had a significant effect on the number and spatial distribution of comparison fixations in the right image. With context (latent and plain tasks), most fixations were in or near the target area (for both left and right images): less than 4% of fixations were farther than one target width (approximately 3° of visual angle) from the target area; examiners needed few fixations far from the target in the right image to get an overall sense (gestalt) of the print. Without context (cropped tasks), examiners looked at much more of the right image, with far more fixations, and with more areas considered by many examiners: 36% of fixations were more than one target width away from the target area. The median area visited by examiners in the right image on cropped tasks was more than triple that of plain tasks, and nearly triple that of latent tasks (as measured by the count of grid cells visited; details in Additional file 1: Appendix SI-6.3).
Most image sets included areas in the right image not adjacent to the target that drew the attention of many or most examiners on the cropped trials. Such “decoy” areas tended to have pattern flow similar to the target area. In Fig. 4, the ridges in the target area converge from the top right to bottom left; such ridge flow is found in two areas in the fingerprint, the left side of the delta (where the target is) and the lower left of the core (“decoy” areas). Pomplun (2006) characterized such decoy areas by calculating the saccadic selectivity, and found that observers were more likely to fixate on regions in natural images that are similar to the target along dimensions such as intensity, contrast, orientation, or spatial frequency. In our stimuli, other elements, such as curvature and ridge flow, are also likely guiding saccadic behavior. Additional file 1: Appendix SI-6.1 shows other clear examples of saccadic selectivity, in which the decoy areas attract a large number of fixations in the cropped task. However, the addition of context almost completely eliminates the fixations to decoy areas, as well as much of the scanning and searching behavior seen in cropped trials. This illustrates the powerful constraints provided by context in our task.
Time to complete
We expect that the amount of time spent by each examiner was affected by individual motivational factors such as competitiveness, and their interpretation of the appropriate level of diligence for the task.
Some examiners started fixating on the right image 150 ms after it was presented (regardless of task type). In approximately half of trials, examiners first fixated on the right image within 250 ms after it was presented. At approximately 0.5 s, some examiners had already turned their attention back to the left image; after 2 s, examiners in about half of trials had done this, including 15% who had looked right-left-right (details in Additional file 1: Appendix SI-8).
How is localization accomplished so rapidly when context is provided? In plain or latent tasks, the paucity of fixations outside the target area on both the left and right images indicates that examiners either know where to expect the target area to be shown on the screen (even prior to the right image being shown), or acquire coarse information regarding ridge flow from the right image from peripheral vision prior to the first fixation in the right image. Drew et al. (2013) observed such rapid understanding of an image with expert radiologists, noting that experts use nonselective processing that “extracts information from global or statistical information without selecting specific objects.” Greene and Oliva (2009) describe similar rapid image processing using a set of global primitives such as “openness” (for outdoor scenes) or “transience” (for images such as waves or waterfalls) that allow for rapid image classification. Such similar primitives may exist for fingerprints, such that examiners may quickly determine the Henry classification type (e.g., whorls, loops, and arches (Henry, 1900)), orientation, or whether an impression overlaps another print. Within the medical imaging literature, several authors have argued for an initial holistic processing of medical images followed by a more deliberate, slower search and discovery process (Kundel, Nodine, Conant, & Weinstein, 2007; Kundel, Nodine, Krupinski, & Mello-Thoms, 2008; Nodine, Kundel, Lauver, & Toto, 1996). In Kundel et al. (2008), the authors found that participants made a saccade to more than half of all tumors within the first second, which is consistent with the rapid localization seen in Fig. 7.
Errors finding the target
Our objective was to assess how examiners find target groups — not whether they would succeed. Given that the participants are practicing latent print examiners, the occurrence of any errors on such a simple task is notable. Examiners apparently failed to find the correct target in six of the 675 trials, all in cropped tasks (twice by one participant). In each of these apparent errors, the areas that appear to have misled examiners were the “decoy” areas of the print (areas viewed by many or most other examiners); each of these areas had ridge flow similar to the target. Haste does not appear to have been the cause for four of the six apparent errors, as those trials took longer than the median Comparison time for those image pairs; the two trials that were faster than the median Comparison time were made by the same examiner, who had faster than median Comparison times overall. Miscalibration does not appear to have been a concern for these trials because drift correction was verified prior to and after each trial. Four of the five examiners who made errors (including the examiner with two errors) had less than 5 years of experience; 24% of all participants had less than 5 years of experience. Additional file 1: Appendix SI-9 shows the fixations and images for these six trials.
We bring attention to these localization errors because they suggest how erroneous conclusions in actual comparisons may occur. Although we detected few localization errors, they are still notable: we cannot expect to have rare events in quantity (for comparison, erroneous identification rates are estimated at approximately 0.1% (Ulery et al., 2011)). If, in a real comparison, an examiner selects and compares an incorrect but similar target group, and either does not notice or discounts the correct target group, that could start a path of reasoning that could result in an erroneous identification (if the examiner discounts or does not notice discrepancies) or an erroneous exclusion (by expanding from the wrong location then noting invalid “discrepancies”). Ulery et al. have previously observed (Ulery et al., 2014, 2016) examples of erroneous conclusions where examiners’ markups revealed misassociated features. Given the seriousness (albeit rarity) of erroneous identifications, and a higher than expected erroneous exclusion rate in latent print examination (Pacheco et al., 2014; Ulery et al., 2011), it is important to highlight faulty localization as a behavior that could help explain such erroneous conclusions.
There was substantial variation in Analysis times. We found no strong correlation between Analysis and Comparison times (there was a very weak positive correlation; see Additional file 1: Appendix SI-7.1). We might have expected a negative correlation if lengthy analysis were to result in faster target finding or decision-making, or a positive correlation if individual examiners were consistently slower or faster in both phases. The weak observed correlation could reflect both opposing influences
Examiners generally spent more Comparison time in the right image than the left image: median 82% of Comparison time was in the right image for cropped tasks, 76% for plain tasks and 67% for latent tasks — presumably because the left image in latent tasks is poorer quality and, therefore, requires a greater proportion of Comparison time. However, the repeated returns to the left image suggests that subjects are reluctant to trust short-term memory, or the contents of short-term memory are fading. Thus, they use the physical presence of the left image to reduce the demands on short-term memory (Ballard et al., 1995). During Comparison, examiners frequently switched back and forth between the left and right images. Examiners generally spent 1 to 3 s on an image before looking at the other image, except for the right image in cropped trials, in which they generally spent 2 to 6 s (inter-quartile ranges; details in Additional file 1: Appendix SI-13.3)
In cropped tasks (but not latent or plain tasks) many examiners spent an extended period outside the target prior to a final period in the target area. Much of the variability in the Comparison time of cropped trials can be accounted for as differences in the amount of time spent outside the target. The total amount of time spent in the target area prior to announcing that the target was found was similar across trials and tasks: in plain and latent trials examiners generally fixated in the target area in the right image for 3 to 6 s (inter-quartile range) or 4 to 11 s for cropped trials (details in Additional file 1: Appendix SI-11).
The spatial plots and timelines are powerful visualization techniques for exploring this complex dataset; they can be used as qualitative descriptions and to develop hypotheses and general explanations. However, in order to measure the behavioral patterns we observed, we need additional tools to characterize eye-gaze behavior. In this section, we partition trials into subphases and define behavioral metrics that are associated with these subphases.
Summarizing trials in terms of subphases
Definitions of subphases of trials
A period at the start of Comparison prior to the examiner first looking at the target area, in which all fixations in the right image are outside the target area. Because Subphase A was often very brief, we use “AL” and “AS” to indicate whether the subphase was longer or shorter than 1 s
In Fig. 9, for trials in which the first right fixation is red (out of the target area), Subphase A is the period prior to the first continuous 0.5 s of blue (in)
Any period of time not in Subphase A or C
In Fig. 9, Subphase B is a period alternating between blue and red (after initial red and final blue periods)
A period at the end of Comparison when the examiner is only looking at the target area, in which all fixations in the right image are in the target area in the right image
In Fig. 9, for trials in which the last right fixation is blue (in the target area), Subphase C is the period after the last 0.5 s of red (out), until the end of the trial; for trials that have no red, the entire Comparison phase is considered Subphase C
Categorization of trials in terms of presence or absence of subphases, with proportions of trials and median comparison times. Rows with similar subphase combinations are grouped. Median times based on three or fewer trials are shown in italics. (n = 675 trials)
% of trials
Median comparison time (secs)
Only looked in target
Quickly found target, then only looked in target
Took time to find target, then only looked in target
Started in target, continued looking elsewhere, then only looked in target
Quickly found target, continued looking elsewhere, then only looked in target
Took time to find target, continued looking elsewhere, then only looked in target
Started in target, but did not end up there
Quickly found target, but did not end up there
Took time to find target, but did not end up there
Quickly completed but never found target
Took time, but never found target
Although here Subphases A and C are defined in terms of the start and end of an isolated find-the-target task, full fingerprint comparisons will generally consist of multiple localization events in series (e.g., ABCACABCAC), and, therefore, Subphase C would often be followed by another Subphase A.
Characterizing transitory overt behaviors
Speed3 measures the speed of eye movement within an image in image pixels per second. For each fixation, Speed3 is measured over a series of up to ± three fixations, as the sum of inter-fixation distances, divided by the time from the start of the first fixation to the end of the last fixation in the series. Fixations that are near left-right transitions will have fewer than ± three fixations in the series
PercentOfFixesInCell measures the percentage of fixations in a trial that are located in each cell (as defined in the “Spatial analyses” section). For each fixation, PercentOfFixesInCell is calculated for the cell in which the fixation is located. This local spatial density measure may reflect an examiner’s level of interest in an area
We use “image visit” to refer to a consecutive series of fixations in an image (left or right), bounded by the transitions into and out of that image.
○ TimeInImage is the duration of time spent in each image visit, from the start of the first to the end of the last fixation in that image
○ DetailedBackAndForth is a count of image visits in which the examiner returns to the same small area in consecutive image visits in that image. The examiner is considered to have returned to the same small area if the distance between the centers of mass of consecutive image visits in the same image is within 88 pixels (approximately four ridges), and the maximum distance between any two fixations in each image visit is no more than twice that distance
There are strong relations between these metrics for overt behaviors and subphases, which we will use as a basis for inferring cognitive states in the next section. The speed of eye movement is generally slower in Subphase C, faster in AL, and much faster in AS (medians: AS = 11.9 ridges/s; AL = 10.3, B = 7.6, C = 4.5). During Comparison, the median TimeInImage was about 3 s for the right image in cropped tasks; 1–2 s otherwise. Subphases C and AS have shorter TimeInImage than AL or B (medians: Subphase AS = 1.7 s, AL = 3.2, B = 3.8, C = 1.9). DetailedBackAndForth is nonzero in a quarter of the fixations in Subphases B and C, but is almost always zero in Subphase A. PercentOfFixesInCell was notably higher for cells in the target area (and, therefore, for Subphase C). Details are in Additional file 1: Appendix SI-13.
Having developed a set of measures that characterize performance in our target localization task, we now turn to the goal of inferring the underlying cognitive states that generated the observable gaze behavior. During an actual comparison, localization decisions may occur multiple times (unlike in this directed task); in the more naturalistic setting of actual comparisons we would like to separate the continuous stream of fixations into discrete cognitive states. However, one challenge is that the subphases described in the “Summarizing trials in terms of subphases” section are defined relative to the known target location, which will not be available in casework-like comparisons. To address this challenge, we explore the relation between the metrics developed in the “Characterizing transitory overt behaviors” section and the subphases to determine how accurately we can predict the underlying subphase given a set of observable metrics that do not depend on knowing the target location. We develop this relation so that we can address this question in casework-like comparisons: if an examiner has one or more target groups in memory and is searching for a correspondence between two impressions, can we determine when the examiner believes that they have found a correspondence?
A first step in this modeling process is to determine a measure of cognitive state. We viewed talk-aloud protocols as too intrusive and not representative of casework-like settings. Instead, subphases serve as proxies for an examiners’ cognitive state or intent. Subphase A (especially AS) usually has fast eye movement with fixations far apart and relatively little back-and-forth movement, indicating a “where is it?” period of looking for potential locations, roughly corresponding to Kundel’s “Scanning” (Kundel et al., 1978). Subphase C generally has slow eye movement and fixations close together, often with detailed back-and-forth to the same location, indicating detailed work, consistent with an “am I sure?” period of deciding whether it is the correct target, roughly corresponding to Kundel’s “Decision.” Subphase B appears to be an “Is this it?” period, alternating between AL-like behavior and C-like behavior.
To provide a link between overt behavior and underlying cognitive states, we developed a set of logistic regression models to assess the association of behaviors with subphases. We are particularly interested in predicting Subphase C (indicating that an observer has entered a decision or correspondence portion of the target group comparison process) and Subphase A (indicating scanning for a target group). We do not explicitly model Subphase B because its relation to any underlying cognitive state(s) is ambiguous, and, therefore, Subphase B is less theoretically interesting.
Each right-image fixation is labeled with a subphase according to the definitions in Table 1. We evaluate several logistic regression models to provide an indication of the strength of association between the measures introduced in the “Characterizing transitory overt behaviors” section and underlying cognitive states, as estimated by subphases. We predict Subphases A and C in separate binary models: we model A vs. all other fixations (B and C), and separately model C vs. all other fixations (A and B). By modeling in this way, we can identify the strength of the relation between our gaze-behavior measures and what we interpret as the Scanning and Deciding cognitive states.
Performance of logistic regression models predicting the subphase or location of fixations (n = 22,747 Comparison-phase fixations on the right image from 675 trials). Model performance is compared using area under the (receiver operating characteristic) curve (AUC). No cross-terms were used in these models: each model had one degree of freedom for each predictor term
Latent and Plain
Speed3 + PercentOfFixesInCell
Speed3 + TimeInImage
Speed3 + DetailedBackAndForth
TimeInImage + DetailedBackAndForth
Speed3 + TimeInImage + DetailedBackAndForth
Speed3 + PercentOfFixesInCell + TimeInImage + DetailedBackAndForth
The logistic regression results show that the subphases are associated with basic differences in eye behavior and our metrics are effective in detecting these differences. More broadly, we believe that the association between real-world gaze behaviors (as measured by these metrics) and the underlying cognitive states (as approximated by the subphases) is robust enough to be useful when analyzing gaze behavior from casework-like comparisons. In such analyses, models developed from the associations shown here can be used as potential indicators of Scanning and Deciding cognitive states, providing a heretofore unavailable means of assessing how fingerprint examiners conduct comparisons and make determinations.
The purpose of this study is to describe and characterize the eye behavior of experts in a localization task, which we intend to serve as a basis for the deconstruction of the overall fingerprint comparison process. In particular, we develop methods for evaluating eye-gaze behavior during localization, describe the behaviors observed, and develop models that link the observable eye-gaze behavior to underlying cognitive states, providing an indication of when an observer believes that they have found a corresponding location for a target region.
There were significant differences in the eye behavior of examiners for tasks with or without visual context. The absence of visual context (cropped tasks) was generally associated with longer Comparison times, and larger areas of the image viewed. For the same image pair and task (particularly in cropped tasks), there was substantial variability among different trials in terms of which areas of the image were viewed, and in Analysis and Comparison times. Trials generally had relatively little variation in the amount of time spent looking in the target area: most of the variability in Comparison time can be attributed to differences in the time spent looking in nontarget areas. When visual context was provided, examiners usually located the correct target area within 1 s. When context was not explicitly provided, examiners appeared to infer information about the context from characteristics in the cropped target area, such as ridge curvature and patterns of convergence. Several of the fingerprints included areas with pattern flow similar to the target (“decoy” areas) that attracted the attention/gaze of a majority of examiners on cropped trials. The large number of fixations to these decoy areas in the cropped task is consistent with visual feature guidance based on similarity along basic visual dimensions (Pomplun, 2006), although the relevant dimensions for fingerprints may be different than those for natural images.
Beyond the information provided by explicit spatial context (i.e., the position of the target area in the left impression), there may be more experience-based contextual information that guides eye movements, even in the cropped images that do not provide explicit content. Examiners may acquire global shape information from expertise-based expectations of fingerprint structure and peripheral mechanisms that provide overall shape of ridge flow, both of which are likely to be much more effective when context is provided. This global shape information may be used to localize rapidly and to avoid decoy areas. These same peripheral mechanisms likely support the re-acquisition of a potentially corresponding location when the examiner must make a saccade back to the left image.
In several trials, examiners failed to locate the correct target area: these examiners located incorrect but similar “decoy” target groups, and declared that they found the target. Although we observed few errors, it is notable that practicing latent print examiners made any such errors. We bring attention to these errors because they suggest how erroneous conclusions in actual comparisons may occur: erroneous identifications, erroneous exclusions, or inappropriate inconclusives could all result from incorrect localizations.
To infer the cognitive state of fingerprint examiners, we developed a method to partition the localization process into subphases; although the subphases are defined here based on ground-truth knowledge of the correct target locations, the subphases act as proxies for these cognitive states. We then demonstrated an ability to predict these cognitive states directly from eye-tracking data (not using the ground-truth knowledge), indicating potential for use in flagging when an examiner appears to have found a corresponding location in conventional fingerprint comparisons. These methods provide a tool for determining the extent to which incorrect correspondences of target groups can explain differences in examiners’ conclusions, particularly erroneous conclusions. The analytical methods developed here to assess this simple but critical process are intended as a stepping stone for use in future analyses of examiner gaze behavior in the more complex processes involved in fingerprint comparisons, which generally can be expected to consist of multiple localization events in series. In such analyses, models developed from the associations shown here can be used as potential indicators of Scanning and Deciding cognitive states, providing a heretofore unavailable means of assessing how fingerprint examiners conduct comparisons and make determinations.
Understanding how and why expert latent print examiners make errors and reach different conclusions requires detailed evaluation of fingerprint examiners. This study is a first step in that process: to describe and characterize the eye behavior of experts in a localization task, serving as a basis for the deconstruction of the overall fingerprint comparison process. We intend for the findings and methods discussed here to be used as tools in the evaluation of full fingerprint comparisons — using eye-gaze data collected in the same sitting as these localization trials. When evaluating eye-gaze behavior in full fingerprint comparisons the models developed here can be used to isolate localization behavior, indicating the areas that examiner considers to be corresponding between the prints being compared. Differences among examiners in what they consider correspondences may be used to explain differences in fingerprint comparison conclusions. Lack of localization behavior is also of interest: if an examiner concludes a comparison with little or no localization behavior, the resulting decision is holistic (as might be expected when excluding fingerprints or unrelated patterns). In ongoing and future work, we will use the localization behavior of individual examiners to understand the extent to which errors and differences in conclusions can be explained by differences in behavior.
We thank the latent print examiners who participated in these studies; Brandi Emerick, Mac Vogelsang, Haley Drabek, and Kelly Carter (for their assistance in data collection); and Madeline Ausdemore (for assistance in analyses). This is publication number 18-20 of the FBI Laboratory Division. Names of commercial manufacturers are provided for identification purposes only and inclusion does not imply endorsement of the manufacturer or its products or services by the FBI. This work was funded in part under a contract award to Noblis, Inc. from the FBI Laboratory, with funding from the FBI CJIS Division Biometric Center of Excellence. The views expressed are those of the authors and do not necessarily reflect the official policy or position of the FBI or the U.S. Government.
RAH, BTU, and TAB were funded under a contract award to Noblis, Inc. from the FBI Laboratory, with funds from the FBI CJIS Division Biometric Center of Excellence. MAR and JB are FBI employees.
Availability of data and materials
The table of fixations used for analyses in this study is included as Additional file 2. The raw eye-gaze data can be made available on request to the corresponding author. The attribution of results to individual examiners cannot be made available because the study design, and Institutional Review Board approval, assured the anonymity of the participants.
All authors were responsible for planning, designing, and conducting the study. TAB was responsible for configuring eye-tracking equipment, collecting eye-tracking data, and extracting fixation and saccade data from the raw eye-tracking data. RAH, BTU, and TAB conducted data analyses. RAH and BTU were the primary writers of the manuscript, with contributions from all authors. All authors reviewed and approved the final manuscript.
Ethics approval and consent to participate
Use of human subjects in this study was approved by the FBI Institutional Review Board. Informed consent was obtained from all fingerprint examiners who participated.
Consent for publication
All images were selected from the FBI’s Universal Latent Workstation Ground Truth Dataset (ULW, n.d.), which includes higher-resolution scans of the same images used in NIST’s Special Database 27 (NIST SD27, n.d.). Both datasets have been publicly distributed for many years.
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
- ANSI/NIST-ITL (2013) National Institute of Standards: American National Standard for Information Systems: Data format for the interchange of fingerprint, facial & other biometric information. ANSI/NIST-ITL 1-2013.Google Scholar
- Ashbaugh, D. R. (1999). Quantitative-qualitative friction ridge analysis : an introduction to basic and advanced ridgeology. Boca Raton: CRC Press.View ArticleGoogle Scholar
- Baddeley, A. (1986). Oxford psychology series, No. 11. Working memory. New York, NY, US: Clarendon Press/Oxford University Press.Google Scholar
- Ballard, D. H., Hayhoe, M. M., & Pelz, J. B. (1995). Memory Representations in Natural Tasks. Journal of Cognitive Neuroscience, 7(1), 66–80. https://doi.org/10.1162/jocn.1918.104.22.168.View ArticlePubMedGoogle Scholar
- Barenholtz, E. (2014). Quantifying the role of context in visual object recognition. Visual Cognition, 22(1), 30–56.View ArticleGoogle Scholar
- Blignaut, P. (2009). Fixation identification: The optimum threshold for a dispersion algorithm. Attention, Perception, & Psychophysics, 71(4), 881–895. https://doi.org/10.3758/App.71.4.881.View ArticleGoogle Scholar
- Busey, T., Yu, C., Wyatte, D., Vanderkolk, J. R., Parada, F. J., & Akavipat, R. (2011). Consistency and variability among latent print examiners as revealed by eye tracking methodologies. Journal of Forensic Identification, 61(1), 60–91.Google Scholar
- Busey, T. A., Swofford, H. J., Vanderkolk, J., & Emerick, B. (2015). The impact of fatigue on latent print examinations as revealed by behavioral and eye gaze testing. Forensic Science International, 251, 202–208. https://doi.org/10.1016/j.forsciint.2015.03.028.View ArticlePubMedGoogle Scholar
- Busey, T. A., Yu, C., Wyatte, D., & Vanderkolk, J. (2013). Temporal sequences quantify the contributions of individual fixations in complex perceptual matching tasks. Cognitive Science, 37(4), 731–756. https://doi.org/10.1111/cogs.12029.View ArticlePubMedGoogle Scholar
- Chen, Q., Sun, Q. S., Heng, P. A., & Xia, D. S. (2008). A double-threshold image binarization method based on edge detector. Pattern Recognition, 41(4), 1254–1267. https://doi.org/10.1016/j.patcog.2007.09.007.View ArticleGoogle Scholar
- Cheng, Y. Z. (1995). Mean Shift, Mode Seeking, and Clustering. IEEE Transactions on Pattern Analysis and Machine Intelligence, 17(8), 790–799. https://doi.org/10.1109/34.400568.View ArticleGoogle Scholar
- Chun, M. M., & Jiang, Y. H. (1998). Contextual cueing: Implicit learning and memory of visual context guides spatial attention. Cognitive Psychology, 36(1), 28–71. https://doi.org/10.1006/cogp.1998.0681.View ArticlePubMedGoogle Scholar
- Chun, M. M., & Jiang, Y. H. (1999). Top-down attentional guidance based on implicit learning of visual covariation. Psychological Science, 10(4), 360–365. https://doi.org/10.1111/1467-9280.00168.View ArticleGoogle Scholar
- Cornelissen, F. W., Peters, E. M., & Palmer, J. (2002). The Eyelink Toolbox: Eye tracking with MATLAB and the psychophysics toolbox. Behavior Research Methods, Instruments, & Computers, 34(4), 613–617. https://doi.org/10.3758/Bf03195489.View ArticleGoogle Scholar
- Credido, H. F., Teixeira, E. N., Reis, S. D. S., Moreira, A. A., & Andrade, J. S. (2012). Statistical patterns of visual search for hidden objects. Scientific Reports, 2, 920. https://doi.org/10.1038/srep00920.View ArticleGoogle Scholar
- Donovan, T., & Litchfield, D. (2013). Looking for Cancer: Expertise Related Differences in Searching and Decision Making. Applied Cognitive Psychology, 27(1), 43–49. https://doi.org/10.1002/acp.2869.View ArticleGoogle Scholar
- Drew, T., Evans, K., Võ, ML., Jacobson, FL., Wolfe, JM. (2013). Informatics in radiology: what Can You See in a Single Glance and How Might This Guide Visual Search in Medical Images? RadioGraphics, 33(1):263-74. https://doi.org/10.1148/rg.331125023
- Engbert, R., & Mergenthaler, K. (2006). Microsaccades are triggered by low retinal image slip. Proceedings of the National Academy of Sciences of the United States of America, 103(18), 7192–7197. https://doi.org/10.1073/pnas.0509557103.View ArticlePubMedPubMed CentralGoogle Scholar
- Expert Working Group on Human Factors in Latent Print Analysis (2012). Latent print examination and human factors : improving the practice through a systems approach : the report of the Expert Working Group on Human Factors in Latent Print Analysis. Washington, D.C.: NIST NIJ, National Institute of Justice.Google Scholar
- Francis, J. G. F. (1961). The QR Transformation, I. The Computer Journal, 4(3), 265–271.View ArticleGoogle Scholar
- Greene, M. R., & Oliva, A. (2009). Recognition of natural scenes from global properties: Seeing the forest without representing the trees. Cognitive Psychology, 58(2), 137–176. https://doi.org/10.1016/j.cogpsych.2008.06.001.View ArticlePubMedGoogle Scholar
- Hayhoe, M. (2000). Vision using routines: A functional account of vision. Visual Cognition, 7(1–3), 43–64. https://doi.org/10.1080/135062800394676.View ArticleGoogle Scholar
- Heeger, D. J. (2017). Theory of cortical function. Proceedings of the National Academy of Sciences of the United States of America, 114(8), 1773–1782.View ArticleGoogle Scholar
- Henry, E. R. (1900). Classification and Uses of Fingerprints, (1st ed., ). London: Routledge & Sons.Google Scholar
- Johnson, L., Sullivan, B., Hayhoe, M., & Ballard, D. (2014). Predicting human visuomotor behaviour in a driving task. Philosophical Transactions of the Royal Society, B: Biological Sciences, 369(1636). https://doi.org/10.1098/rstb.2013.0044 ARTN 20130044.
- Kardan, O., Berman, M. G., Yourganov, G., Schmidt, J., & Henderson, J. M. (2015). Classifying Mental States From Eye Movements During Scene Viewing. Journal of Experimental Psychology: Human Perception and Performance, 41(6), 1502–1514.PubMedGoogle Scholar
- Kleiner, M., Brainard, D., & Pelli, D. (2007). What's new in Psychtoolbox-3? Perception, 36, 14–14.Google Scholar
- Kundel, H. L., Nodine, C. F., & Carmody, D. (1978). Visual Scanning, Pattern-Recognition and Decision-Making in Pulmonary Nodule Detection. Investigative Radiology, 13(3), 175–181. https://doi.org/10.1097/00004424-197805000-00001.View ArticlePubMedGoogle Scholar
- Kundel, H. L., Nodine, C. F., Conant, E. F., & Weinstein, S. P. (2007). Holistic component of image perception in mammogram interpretation: Gaze-tracking study. Radiology, 242(2), 396–402. https://doi.org/10.1148/radiol.2422051997.View ArticlePubMedGoogle Scholar
- Kundel, H. L., Nodine, C. F., Krupinski, E. A., & Mello-Thoms, C. (2008). Using gaze-tracking data and mixture distribution analysis to support a holistic model for the detection of cancers on mammograms. Academic Radiology, 15(7), 881–886. https://doi.org/10.1016/j.acra.2008.01.023.View ArticlePubMedGoogle Scholar
- Land, M., Mennie, N., & Rusted, J. (1999). The roles of vision and eye movements in the control of activities of daily living. Perception, 28(11), 1311–1328.View ArticleGoogle Scholar
- Langenburg, G., Champod, C., & Genessay, T. (2012). Informing the judgments of fingerprint analysts using quality metric and statistical assessment tools. Forensic Science International, 219(1–3), 183–198. https://doi.org/10.1016/j.forsciint.2011.12.017 Epub 2012 Jan 24.View ArticlePubMedGoogle Scholar
- Manor, B. R., & Gordon, E. (2003). Defining the temporal threshold for ocular fixation in free-viewing visuocognitive tasks. Journal of Neuroscience Methods, 128(1–2), 85–93. https://doi.org/10.1016/S0165-0270(03)00151-1.View ArticlePubMedGoogle Scholar
- Marshall, S. P. (2007). Identifying cognitive state from eye metrics. Aviation, Space and Environmental Medicine, 78(5), B165–B175.Google Scholar
- MathWorks (2012). MATLAB and Statistics Toolbox Release 2012b. Natick: The MathWorks, Inc.Google Scholar
- National Institute of Justice (2011). Fingerprint Sourcebook. https://www.ncjrs.gov/pdffiles1/nij/225320.pdf. Accessed 11 Mar 2019
- Neider, M. B., & Zelinsky, G. J. (2006). Scene context guides eye movements during visual search. Vision Research, 46(5), 614–621. https://doi.org/10.1016/j.visres.2005.08.025.View ArticlePubMedGoogle Scholar
- Neumann C., Champod C., Yoo M., Genessay T., & Langenburg G. (2013). Improving the Understanding and the Reliability of the Concept of "Sufficiency" in Friction Ridge Examination. National Institute of Justice, 12 July 2013. https://www.ncjrs.gov/pdffiles1/nij/grants/244231.pdf. Accessed 11 Mar 2019
- NIST SD27 (n.d.). National Institute of Standards and Technology. Special Database 27. [Digital dataset]. https://www.nist.gov/itl/iad/image-group/nist-special-database-2727a. Accessed 11 Mar 2019
- Nodine, C. F., Kundel, H. L., Lauver, S. C., & Toto, L. C. (1996). Nature of expertise in searching mammograms for breast masses. Academic Radiology, 3(12), 1000–1006. https://doi.org/10.1016/S1076-6332(96)80032-8.View ArticlePubMedGoogle Scholar
- Nodine, C. F., Mello-Thoms, C., Kundel, H. L., & Weinstein, S. P. (2002). Time course of perception and decision making during mammographic interpretation. American Journal of Roentgenology, 179(4), 917–923. https://doi.org/10.2214/ajr.179.4.1790917.View ArticlePubMedGoogle Scholar
- Oliva, A., & Torralba, A. (2007). The role of context in object recognition. Trends in Cognitive Sciences, 11(12), 520–527. https://doi.org/10.1016/j.tics.2007.09.009.View ArticlePubMedGoogle Scholar
- Pacheco, I., Cerchiai, B., Stoiloff, S. (2014). Miami-Dade research study for the reliability of the ACE-V process: Accuracy & precision in latent fingerprint examinations. NIJ Technical Report 2010-DN-BX-K268. http://www.ncjrs.gov/pdffiles1/nij/grants/248534.pdf. Accessed 11 Mar 2019
- Parada, F. J., Wyatte, D., Yu, C., Akavipat, R., Emerick, B., & Busey, T. (2015). ExpertEyes: Open-source, high-definition eyetracking. Behavior Research Methods, 47(1), 73–84. https://doi.org/10.3758/s13428-014-0465-z.View ArticlePubMedGoogle Scholar
- Pelli, D. G. (1997). The VideoToolbox software for visual psychophysics: Transforming numbers into movies. Spatial Vision, 10(4), 437–442. https://doi.org/10.1163/156856897x00366.View ArticlePubMedGoogle Scholar
- Pomplun, M. (2006). Saccadic selectivity in complex visual search displays. Vision Research, 46(12), 1886–1900. https://doi.org/10.1016/j.visres.2005.12.003.View ArticlePubMedGoogle Scholar
- Port, N. L., Trimberger, J., Hitzeman, S., Redick, B., & Beckerman, S. (2016). Micro and regular saccades across the lifespan during a visual search of "Where's Waldo" puzzles. Vision Research, 118, 144–157. https://doi.org/10.1016/j.visres.2015.05.013.View ArticlePubMedGoogle Scholar
- Ross, J., Morrone, M. C., Goldberg, M., & Burr, D. (2001). Changes in visual perception at the time of saccades. Trends in Neurosciences, 24(2), 113–121.View ArticleGoogle Scholar
- Ulery, B. T., Hicklin, R. A., Buscaglia, J., & Roberts, M. A. (2011). Accuracy and reliability of forensic latent fingerprint decisions. Proceedings of the National Academy of Sciences of the United States of America, 108(19), 7733–7738. https://doi.org/10.1073/Pnas.1018707108.
- Ulery, B. T., Hicklin, R. A., Buscaglia, J., & Roberts, M. A. (2012). Repeatability and Reproducibility of Decisions by Latent Fingerprint Examiners. PLOS ONE, 7(3). https://doi.org/10.1371/journal.pone.0032800 ARTN e32800.
- Ulery, B. T., Hicklin, R. A., Roberts, M. A., & Buscaglia, J. (2014). Measuring what latent fingerprint examiners consider sufficient information for individualization determinations. PLOS ONE, 9(11), e110179. https://doi.org/10.1371/journal.pone.0110179.
- Ulery, B. T., Hicklin, R. A., Roberts, M. A., & Buscaglia, J. (2016). Interexaminer variation of minutia markup on latent fingerprints. Forensic Science International, 264, 89–99. https://doi.org/10.1016/j.forsciint.2016.03.014.View ArticlePubMedGoogle Scholar
- ULW (n.d.). Federal Bureau of Investigation; Universal Latent Workstation (ULW) [Computer Software]. https://www.fbibiospecs.cjis.gov/Latent/PrintServices. Accessed 11 Mar 2019