- Original article
- Open Access
The impact of speed and bias on the cognitive processes of experts and novices in medical image decision-making
Cognitive Research: Principles and Implications volume 3, Article number: 28 (2018)
Training individuals to make accurate decisions from medical images is a critical component of education in diagnostic pathology. We describe a joint experimental and computational modeling approach to examine the similarities and differences in the cognitive processes of novice participants and experienced participants (pathology residents and pathology faculty) in cancer cell image identification. For this study we collected a bank of hundreds of digital images that were identified by cell type and classified by difficulty by a panel of expert hematopathologists. The key manipulations in our study included examining the speed-accuracy tradeoff as well as the impact of prior expectations on decisions. In addition, our study examined individual differences in decision-making by comparing task performance to domain general visual ability (as measured using the Novel Object Memory Test (NOMT) (Richler et al. Cognition 166:42–55, 2017). Using signal detection theory and the diffusion decision model (DDM), we found many similarities between experts and novices in our task. While experts tended to have better discriminability, the two groups responded similarly to time pressure (i.e., reduced caution under speed instructions in the DDM) and to the introduction of a probabilistic cue (i.e., increased response bias in the DDM). These results have important implications for training in this area as well as using novice participants in research on medical image perception and decision-making.
The ability to classify and interpret medical images is critical in the diagnosis of many diseases. Despite significant improvements in imaging assays as well as meticulous education and training, diagnostic errors still occur. In order to improve diagnostic decision-making based on medical images, it is critical to understand the cognitive processes involved in these decisions. This research borrows well-validated experimental and computational methods from perceptual decision-making and applies them to investigate cancer cell image identification. Using both non-experts as well as pathologists (residents and faculty), we examine the impact of time pressure and externally imposed bias on the identification of single cell images related to cancer diagnosis. Using computational modeling techniques, we find that these manipulations have important impacts on diagnostic decisions. Specifically, we find similarities in how novices and pathologists trade off speed and accuracy instructions as well as how they respond to externally imposed bias. In addition, we find that participants with better domain general visual ability perform better at the task. In sum, these results shed light on the cognitive mechanisms that play a role in medical image perception and decision-making. In the future, this knowledge could be used to improve training and education, and this method of investigation could lead to new insights about the cognitive processes involved in image-based decisions.
Accurate interpretation and classification of medical images is an important component of the diagnosis and treatment of numerous diseases. A wide range of medical disciplines (Samei & Krupinski, 2010) ranging from pathology (our focus here), to radiology, to ophthalmology rely on expert analysis of images to detect abnormalities. While the exact rate of diagnostic errors is unknown, consistent evidence suggests that error rates are > 10% (Goldman et al., 1983; Hoff, 2013; Kirch & Schafii, 1996; Shojania, Burton, McDonald, & Goldman, 2003; Sonderegger-Iseli, Burger, Muntwyler, & Salomon, 2000). It is thus critical that we understand how people make perceptual decisions from medical images in order to improve training and minimize the occurrence of misdiagnoses. This requires investigation of the cognitive processes underlying decision-making in this domain and how those processes evolve with training and experience. The goal of this paper is to use experimental methods and computational tools developed in the area of perceptual decision-making to probe the cognitive processes involved in pathology image-based decisions in novices and experts.
Decisions based on medical images have a number of parallels with perceptual decision-making, where people make choices based on sensory information. The investigation of perceptual decision-making has a rich history in psychology, cognitive science, and neuroscience. In aggregate, this research has shown that perceptual decisions are typically based on the accumulation of information over time. Such accumulated perceptual information is thought to be related to neural activity in multiple cortical and subcortical brain areas (Gold & Shadlen, 2007; Heekeren, Marrett, & Ungerleider, 2008; Summerfield & de Lange, 2014; Summerfield & Egner, 2009). This accumulation process is known to be influenced by external factors such as time pressure and expectations (Egner, Monti, & Summerfield, 2010; Leite & Ratcliff, 2011; Maddox & Bohil, 1998; Mulder, Wagenmakers, Ratcliff, Boekel, & Forstmann, 2012). Computational modeling has shown that these different external factors influence different latent components of the decision process. In particular, time pressure affects response caution (quantified by the amount of information needed to make a decision), while prior expectations impact internal biases (e.g., bias toward reporting the presence of an abnormality even before viewing an image) (Leite & Ratcliff, 2011; Mulder et al., 2012).
However, perceptual decision-making of medical images in clinical settings has received less attention. Numerous studies have probed the perceptual processes involved in image-based decisions, particularly in the context of radiology (Bertram, Helle, Kaakinen, & Svedstrom, 2013; Krupinski, 2010; van der Gijp et al., 2017). However, these studies have largely focused on how medical image observers perform visual search (Bertram et al., 2013; Krupinski, 2010; Krupinski et al., 2006; Krupinski, Graham, & Weinstein, 2013; van der Gijp et al., 2017). Eventually, a decision must be made, and understanding the cognitive processes involved in these decisions is the main objective of this paper.
Here, we present a study investigating the cognitive processes underlying cancer image detection in diagnostic pathology. More specifically, we investigate how various external factors influence the ability of novice undergraduate students and pathologists (residents and faculty) to distinguish between normal cells (standard white blood cells such as lymphocytes, monocytes, or neutrophils) and abnormal cancer cells (“blast” cells, associated with acute leukemia) in clinical images. Toward this end, we take a joint experimental and modeling approach utilizing experimental paradigms and modeling methods previously developed in the course of basic research on perceptual decision-making (Ratcliff & Smith, 2004; Schouten & Bekker, 1967; Wickelgren, 1977).
To investigate this process experimentally, we passively collected a large bank of digital images of both blast and non-blast white blood cells drawn from patients at the Vanderbilt University Medical Center (all images were obtained as part of routine clinical care). A panel of expert pathologists classified each of these images, providing a fully annotated data set consisting of hundreds of images of varying type and level of difficulty. Using this image bank, we developed a perceptual decision-making experiment to investigate how time pressure and externally imposed bias influence individuals’ behavior.
We chose to examine the speed-accuracy tradeoff (SAT) (Reed, 1973; Wickelgren, 1977) as well as the impact of external bias because these factors have relevance in the clinical context. With the current and projected shortage of medical technologists and pathologists (Allen, 2002; Bennett, Thompson, Holladay, Bugbee, & Steward, 2015; Lewin, 2016; Sullivan, 2016) coupled with a desire to improve throughput and turnaround times and reduce costs, many laboratories hope to increase productivity by using automated basic recognition sorters. In essence, automated systems have the potential to offset some of the human workload in order to increase productivity, which is largely dependent on the speed with which slides are screened. For example, the Food and Drug Administration (FDA) increased the workload for cytotechnologists from 100 slides per day to 200 slides per day if they are using the ThinPrep imaging system, an automated system used for gynecologic cytology (Elsheikh et al., 2013). However, it is unclear how this increase in workload (even though it comes with the assistance of an automated system) influences diagnostic decisions. In particular, research has shown that decreasing screening times for cytotechnologists from 5 min per slide to 3.7 min per slide resulted in a lower detection of abnormal findings (10.4–8.3%) and an increase in false negatives (3.8–7.0%) (Elsheikh et al., 2010). In other words, the cytotechnologists were trading off speed and accuracy, even when they had access to an automated system. More generally, as machine learning and artificial intelligence (AI) become more integrated into the diagnostic process, the desire for increased productivity is likely to result in higher workloads for medical image observers. While machines will likely be able to process images faster than humans, human observers will still need to be part of the diagnostic process (at least for the foreseeable future). Thus, it is critical to understand how medical image observers trade off speed and accuracy in diagnostic decision-making.
In addition, prior expectations and biases are likely to play a significant role in medical image-based decision-making. In diagnostic pathology, images may be passed through automated basic recognition sorters and/or analyzed by medical technologists and residents before being analyzed by senior faculty experts. In this diagnostic chain, images that clearly lack abnormalities are rarely passed up the chain. Thus, an image that has made it to a senior faculty expert’s desk may in and of itself be a cue, setting expectations before an image is even seen.
In addition to testing participants’ ability to discriminate between and classify images of blast and non-blast cells, we also investigate how participants’ domain general visual ability affects their performance on this task. Toward this end, we employ a second task, the Novel Object Memory Test (NOMT), to assess each participant’s general ability to learn and recognize objects with which they have no prior experience (Richler, Wilmer, & Gauthier, 2017). We use this to probe to what extent general object recognition, which has been studied in much more detail in lab settings (Gauthier et al., 2014; Hildebrandt, Wilhelm, Herzmann, & Sommer, 2013; McGugin et al., 2012a, 2012b), correlates with or affects participants’ efficacy on the blast cell identification task.
To gain further insight into the cognitive processes underlying decisions on this task, we utilize computational modeling linked with results of this experiment. One of the benefits of quantitative modeling, and the reason we use it here, is that it provides a way to quantify latent cognitive processes and statistically separate the different components of the decision process (caution, bias, and rate of information uptake) that are not accessible through traditional statistical methods alone. For this, we utilize a version of the classic diffusion decision model (DDM) (Ratcliff, 1978; Ratcliff & McKoon, 2008; Ratcliff, Smith, Brown, & McKoon, 2016), which has been shown to account for detailed patterns of behavior across a wide range of decision-making paradigms (Ratcliff, Love, Thompson, & Opfer, 2012; Ratcliff, Thapar, & McKoon, 2001, 2004, 2010), to model the choice and response time behavior of participants on this task and extract these underlying cognitive parameters.
We recruited both novice and medical professionals to complete the experiment. Thirty-seven undergraduate students at Vanderbilt University participated in exchange for course credit. In addition, 19 pathologists from the Vanderbilt University Medical Center (VUMC) participated in exchange for a $15 gift card. We recruited pathologists with different levels of experience ranging from first year pathology residents to faculty pathologists. We targeted about equal numbers of “experienced” and “inexperienced” practitioners, defined by the number of hematopathology rotations completed. All pathology residents at VUMC must complete at least four rotations by the end of their residency. We classified individuals who completed all four mandatory rotations as “experienced” and those who had not as “inexperienced.” We had 9 “experienced” and 10 “inexperienced” participants. Note that our sample sizes were based on convenience (in the case of the pathologists) as well as modeling requirements. The typical sample size for experiments using similar modeling methods is 20–40 participants (Dutilh et al., 2012; Holmes, Trueblood, & Heathcote, 2016; White & Poldrack, 2014). The data are available on the Open Science Framework at https://osf.io/r3gzs/.
To create the stimuli, we collected a bank of 840 digital images of Wright-stained white blood cells taken from anonymized patient peripheral blood smears at VUMC. The images were taken by a CellaVision DM96 automated digital cell morphology instrument (CellaVision AB, Lund, Sweden). This instrument, with its accompanying software, identifies and images single white blood cells and classifies them into one of 17 cell types based on morphologic characteristics. The classification of each cell is confirmed by a trained medical technologist.
A ratings panel of three hematopathology faculty from the Department of Pathology at VUMC was used to identify and rate each image. The raters first identified each image as a blast or a non-blast cell. Following this identification, they were asked to provide a difficulty rating for each image on a 1–5 scale. If the raters identified the image as a blast cell, they were asked to rate how similar the image was to a classic blast cell (with a rating of 1 being “not similar” and a rating of 5 being “very similar”). Raters were told that a classic blast cell image is one that might be used in a textbook. If raters identified the image as a non-blast cell, they were asked to rate how morphologically similar the cell is to a blast cell (with a rating of 1 being “not similar” and a rating of 5 being “very similar”).
Out of the original set of 840 images, the three expert raters agreed on the cell type (i.e., blast or non-blast) for 633 images. From this set, we grouped the images into four types based on the difficulty ratings. The average rating for the blast images was 4.40 (standard deviation (SD) = 0.46), and the average rating for the non-blast images was 1.68 (SD = 0.55). Blast images that received an average rating of 4.66 or greater were categorized as easy blasts (151 images, Fig. 1a). Blasts images that received an average rating of 4 or less were categorized as hard blasts (98 images, Fig. 1b). We selected 4.66 and 4 as the cutoff points for easy and hard images, respectively, because these values represented the 0.75 and 0.25 quantiles of the ratings for the blast images, respectively. Images with an average rating between 4 and 4.66 were not included, as we wanted clear separation between easy and hard images. Non-blast images that received an average rating of 1.66 or less were categorized as easy non-blasts (129 images, Fig. 1c). Non-blast images that received an average rating of 2 or greater were categorized as hard non-blasts (108 images, Fig. 1d). We selected 1.66 and 2 as the cutoff points for easy and hard images because these values represented the 0.25 and 0.75 quantiles of the ratings for the non-blast images, respectively. Images with an average rating between 1.66 and 2 were not included, as we wanted clear separation between easy and hard images. For the experiment, we selected 75 images from each category for a total of 300 unique images. This bank of images was used for all aspects of the experiment including training, practice, and the main task. Note that we have more than 300 trials in the experiment, so some images are repeated (this includes repeating images from the training and practice in the main task). However, no images were repeated until all of the images from a category had been shown.
In the main task, participants first completed a training stage to familiarize themselves with blast cells (both novices and experts completed the training for consistency). The training focused on teaching participants to identify blast cells and was patterned on the training in the NOMT. There were four blocks of training trials. Each block started with participants studying five images of blast cells one at a time. After studying these five images, participants then completed 15 trials where they were presented three images (one blast image and two non-blast images) and asked to choose the image they thought was the blast cell. The four training blocks had the following structure of blast and non-blast images: block 1 was easy blast versus easy non-blast, block 2 was easy blast versus hard non-blast, block 3 was hard blast versus easy non-blast, and block 4 was hard blast versus hard non-blast. Note that the image training used a total of 180 unique images (60 blast images and 120 non-blast images) from the original set of 300.
After completing the four training blocks, participants completed a practice block of 60 trials to familiarize themselves with the main task. Each trial started with a fixation cross displayed for 250 ms. After fixation, participants were shown a single image and had to identify it as a blast or non-blast cell. Participants received trial-by-trial feedback about their choices in this block; thus, these trials acted as additional training for the two categories of images. In this practice block, half of the trials were blast cells and half were non-blast cells. Thus, participants had an equal amount of practice with each category. Across both the training and practice blocks, participants completed 120 trials (60 training and 60 practice) before starting the main task. These 120 trials used a total of 150 non-blast images (corresponding to all of the non-blast images in our original set of 300 images) and 90 blast images.
The main task consisted of six blocks with 100 trials in each block. The main task was the same as the practice block, where participants were asked to identify single images. However, participants did not receive trial-by-trial feedback about their choices. They received feedback about their performance at the end of each block. The 100 trials in each block were composed of equal numbers of easy blast images, hard blast images, easy non-blast images, and hard non-blast images, fully randomized.
There were three manipulations across blocks: accuracy, speed, and bias. In the accuracy blocks, participants were instructed to respond as accurately as possible and were given 5 s to respond. In the speed block, participants were instructed to respond quickly and were given 1 s to respond. If they responded after the deadline, they received the message “Too Slow!” The 5-s and 1-s response windows for the accuracy and speed conditions, respectively, were based on the response time data from the three expert raters. The 0.975 quantile of the expert raters’ response times was 4.96 s; thus, we set the accuracy response window to 5 s. The 0.5 quantile of the expert raters’ response times was 1.04 s; thus, we set the speed response window to 1 s.
In the bias blocks, participants were shown a probabilistic cue on half of the trials. The cue was a red dot that was shown after fixation for 500 ms. The cue identified the upcoming image as most likely being a blast cell. The cue was valid 65% of the time, and participants were instructed about the validity at the start of the block. The validity of the cue was based on previous literature using similar cueing manipulations (Dunovan, Tremel, & Wheeler, 2014; Forstmann, Brown, Dutilh, Neumann, & Wagenmakers, 2010; Glockner & Hochman, 2011). In particular, we selected a cue with low validity because we hypothesized that a low validity cue might have a larger impact on novice participants than pathologists. That is, novices might rely more on the cue as compared to experts, who might simply ignore the cue because of its low validity. The order of the first three blocks was randomized but with the constraint that there was one block for each type of manipulation (i.e., accuracy, speed, and bias). The order of the last three blocks was identical to the order of the first three blocks.
After completing the main task, participants completed a version of the NOMT (Richler et al., 2017). The NOMT is modeled after the Cambridge Face Memory Test (Duchaine & Nakayama, 2006) and provides a measure of domain general visual ability. In our experiment, we used two categories of novel objects (Ziggerins, shown in Fig. 1e, and Greebles, shown in Fig. 1f). For each category, participants started with a learning phase where a target object was shown in three views followed by three test trials where the target was shown alongside two distractor objects. Participants received trial-by-trial feedback during these trials. This learning procedure was repeated for six target objects (the six targets for each category are shown in Fig. 1e and f), where each target object was slightly different from the other targets in the same category. Following the learning phase, participants completed 54 test phase trials where they had to select which of three objects was any one of the six studied targets.
Signal detection theory (SDT)
We fit an equal-variance form of SDT to the data using hierarchical Bayesian methods (Lee & Wagenmakers, 2013). SDT has two main parameters of interest: discriminability and criterion. We performed separate hierarchical fits to the novice participants, inexperienced pathologists (less than four hematopathology rotations), and experienced pathologists (four or more hematopathology rotations).
Diffusion decision model
To gain insight into the participants’ decision process beyond what is possible with signal detection theory and statistical analysis of behavioral results alone, we use the canonical diffusion decision model (DDM) of decision-making. The DDM posits that, over the course of a decision, evidence stochastically accumulates over time until a confidence threshold is reached and a decision is initiated. This model has three core elements that make it valuable in assessing participants’ behavior on the blast identification task. (1) How fast people accumulate evidence over time is encoded in an accumulation rate parameter (d). A high/low rate indicates better/worse performance on the task. (2) The level of confidence a person requires to make a decision (i.e., level of caution) is encoded in a threshold parameter (a). (3) Finally, any initial preference for responding one way or the other prior to seeing the stimulus is described by a bias parameter (z). These are the three critical characteristics/parameters in the model that we will rely on to make inferences. See Fig. 2 for a schematic description of the DDM.
The full version of the DDM that we use here comprises 9 (or 10) free parameters: accumulation rates for easy and hard blast images (dBE, dBH), accumulation rate for easy and hard non-blast images (dNBE, dNBH), trial-to-trial variability in those accumulation rates (sd), start point (z, which determines the initial bias), trial-to-trial variability in the start point (sz), evidence threshold (a), encoding and response time (tND). There is also a parameter encoding within-trial stochasticity (s). However as is common, we fix this parameter to a value s = 0.1 to avoid parameter degeneracy in the model (one parameter must be fixed). For the cueing instruction data, we introduce an additional parameter to denote the bias on trials where the cue is actually shown (zcue). This will allow us to determine if the cue has any discernible effect on initial bias. Given that the speed, accuracy, and cueing instruction conditions all have the potential to influence people’s behavior in different ways, we fit each instruction condition separately and do not assume up front that any model parameters are the same across experimental conditions.
We use a hierarchical Bayesian algorithm to fit the DDM to the participants’ data, providing an account of the choices made and the full distribution of response times at both the individual and population levels. For purposes of hierarchical DDM model fitting, we grouped all 19 pathologists (experienced and inexperienced) into a single medical population and all 37 novices into a single novice population. These two populations were fit independently. The (in)experienced medical participants were grouped together due to the practical limitations of hierarchical modeling; 9 and 10 participants in each subgroup, respectively, are insufficient to define a hierarchical population with the DDM. Given the high level of correlation between model parameters in this model, we utilize the differential evolution Markov chain Monte Carlo (DEMCMC) method (Turner & Sederberg, 2012) to carry out this Bayesian estimation. Since the DDM does not have an analytically tractable closed-form likelihood function, we utilize a recently developed approximation, the probability density approximation (PDA) method (Holmes, 2015; Holmes & Trueblood, 2017; Turner & Sederberg, 2014), to approximate the likelihood of each parameter set sampled.
Results and discussion
We first examined average accuracy on the 60 practice trials preceding the main blast identification task to see how well participants learned to identify the images. For novice participants, the proportion of trials answered correctly in the practice block was 0.73 (SD = 0.09). We removed three participants with accuracy less than two standard deviations below the average, because these participants were likely not engaged in the task. For the pathologists, the proportion of trials answered correctly in the practice block was 0.90 (SD = 0.08). One of the experienced pathologists was removed due to a computer error that affected data recording.
For the behavioral analyses, we used Bayesian statistics implemented using the open source software package JASP (Team, 2016). For each test, we report the Bayes factor (BF), which is the ratio quantifying the evidence in the data favoring one hypothesis relative to another (when comparing the alternative hypothesis to the null, we calculate BF10, where the subscript “10” indicates evidence for the alternative “1” to the null “0”). While BFs are directly interpretable, labels for the strength of BFs have been proposed. In particular, BFs greater (less) than 1, 3 (1/3), 10 (1/10), 30 (1/30), and 100 (1/100) are considered Anecdotal, Moderate, Strong, Very Strong, and Extreme evidence, respectively (Kass & Raftery, 1995).
First, we examined whether or not novice participants learned to generalize information about blast cells from training and practice to the main test trials. Because many of the images used in training and practice were also used in the main trials, it is possible that novices simply remembered specific images and their corresponding labels rather than learning general characteristics of blast versus non-blast cells. To examine this issue, we compared accuracy between the “old” blast images (the 90 images used in training and practice) and the “new” blast images (the additional 60 images not seen in training or practice) during the main trials. Overall, the accuracy on “old” blast images during the main trials was 0.76 (SD = 0.14), and the accuracy on “new” blast images was 0.75 (SD = 0.13). This difference was not statistically significant (BF10 = 0.49; t(33) = 1.47, p = 0.152). Thus, we can conclude that participants did learn general characteristics of blast images during training and practice and were generalizing this information to new images during the main trials. Note that all 150 non-blast images were used in training and practice, and thus this comparison is not possible for these images. However, we believe that similar learning most likely occurred for the non-blast images rather than participants remembering individual images.
Next, we examined the hit and false alarm rates for the three groups of participants across all trials and conditions. We compared the hit rates for the three groups of participants using a Bayesian analysis of variance (ANOVA). This analysis showed that the alternative model was strongly preferred to the null (BF10 = 6032.67). In particular, the hit rate for both groups of pathologists was greater than the hit rate for novices (BF10 = 221.5 for novices as compared to experienced pathologists; BF10 = 262.3 for novices as compared to inexperienced pathologists), but there was no difference between the two groups of pathologists (BF10 = 0.44). Next we compared the false alarm rates for the three groups using a Bayesian ANOVA and found that the alternative model was preferred to the null (BF10 = 7.67). Specifically, experienced pathologists had a lower false alarm rate than novices (BF10 = 38.63). However, there was no difference between the false alarm rate for novices and inexperienced pathologists (BF10 = 0.43). There was also very little difference between the two groups of pathologists (BF10 = 1.5).
Signal detection theory results
We fit SDT to each of the speed, accuracy, and bias instruction conditions separately. We examined the best-fit values for the two key model parameters: discriminability and criterion. Figure 3 shows group-level posterior distributions (best-fit parameter distributions) for these parameters in each of the three instruction conditions and for the three groups of participants, and Table 1 lists the corresponding means. As shown in both Fig. 3 and Table 1, experience leads to increased discriminability, but no change in criterion.
Next we analyzed differences in parameter values across instruction conditions by conducting Bayesian t tests on the group-level posterior distributions and reported the corresponding Bayesian p values. For the two groups of pathologists, there was no significant difference in discriminability between speed and accuracy conditions (p = 0.18 for experienced pathologists and p = 0.15 for inexperienced pathologists). However, discriminability was significantly larger under accuracy instructions as compared to speed for the novice participants (p = 0.03). There was no difference in the criterion for accuracy and speed instructions (all p values were greater than 0.25).
For the bias condition, we fit trials where the cue was present and absent separately. Bayesian t tests on the posterior distributions showed no difference in discriminability when the cue was present as compared to absent (all p values were greater than 0.4). A Bayesian t test on the posterior distributions showed that the criterion was marginally lower when the cue was present as compared to absent for novices (p = .096). There was no difference in the criterion when the cue was present as compared to absent for the two pathologist groups (both p values were greater than 0.195).
In sum, the SDT analysis shows that expertise influences discriminability and not criterion. However, we did not find any differences in the two key parameters across the instruction conditions (except for lower discriminability under speed instructions for novices). The lack of differences among instruction conditions is not surprising. SDT provides only a limited analysis of underlying cognitive processes, in part because it does not take into account response times. Below, we analyze the data using the DDM, which takes into account both choice and response time data.
Comparison of SDT parameters with visual ability (NOMT)
For all participants, the average proportion of correct responses on the NOMT was 0.73 (SD = 0.10). We compared participants’ performance on the NOMT with the discriminability and criterion parameters from the SDT modeling. We calculated Bayesian Pearson correlations between NOMT accuracy and SDT parameters separately for accuracy, speed, and bias blocks since the model was fit separately to these conditions. The correlations are provided in Table 2. Overall, we found positive correlations between discriminability and NOMT accuracy. The correlations were the strongest for the speed and bias conditions (i.e., the BFs for these conditions indicated Moderate evidence for the correlations). There was no evidence for correlations between NOMT accuracy and criterion. In sum, specific ability on the task (measured by discriminability) is positively related to domain general visual ability (measured by the NOMT).
Diffusion decision modeling results
Similar to SDT, we fit the DDM to each of the speed, accuracy, and bias instruction conditions separately. It is in principle possible to fit the totality of the data at once, as is often done. Typically this is accomplished by fixing certain parameters (accumulation rates, for example) to be the same across instruction conditions while others (threshold, for example) are condition dependent. This however restricts up front the properties of the model that can vary between conditions. By fitting the three conditions separately, we allow maximal model flexibility, so that the data can determine what is the same or different across conditions.
In addition, we also fit the novice participants and pathologists separately. Note that for pathologists we only fit the hard trials in the accuracy condition. We did this because the pathologists made almost no errors on the easy trials in the accuracy condition, and the DDM has difficulty fitting data when choice proportions are near ceiling (i.e., perfect performance), since errors are required to inform some parameters. For the speed and bias conditions, the pathologists made a sufficient number of errors on the easy trials that we were able to include them in the fitting of these conditions.
To determine if the model was able to capture the data (which is necessary for it to be useful), we (1) extracted the mean parameters for fits to each of the conditions, (2) calculated the predicted choice proportion and mean response time for each condition, and (3) compared those predictions to choice proportions and mean response times from data. Results (Fig. 4) show that the model provides a good accounting of most aspects of the observed data. In each of these figure panels, the diagonal line represents the line of perfect agreement (prediction = observation), with results lying close to this line in most cases. Note that there is more spread in the fits for the bias condition because this condition has half as many observations as the speed and accuracy conditions (due to the presence of cue/no cue trials). Thus, increased noise in the data would be expected. Also, the model has some trouble accounting for the long response times, which is a common issue with DDM and other similar models since it is a model of speeded decision-making. Overall, the model fits the observed data well; thus, we will further analyze the model results.
We next look at the best-fit values for the three key model parameters linked to behavioral characteristics of interest here (accumulation rate, threshold, bias). Figure 5 shows group-level posterior distributions (best-fit parameter distributions) for these parameters under the speed and accuracy conditions. Analysis of the drift rate estimates shows a consistent pattern across conditions. For the novice participants, evidence accumulation rates are lower for more difficult images, regardless of instruction condition and difficulty. Note that we do not show the drift rates for the easy trials for the pathologists. These were not estimated for the accuracy condition, and the drift rate posteriors for the speed and bias conditions were too broad (due to the small number of errors) to draw a strong conclusion. Interestingly, there is a significant difference between participants’ ability to perceive the characteristics of blast and non-blast images, respectively, as evidenced by the fact that dNB ≠ dB. For novices, it appears that the characteristics of hard non-blast images are the most difficult to discern, while the characteristics of easy non-blast are the simplest. In both novice participants (Fig. 5 top panels) as well as pathologists (Fig. 5 bottom panels), hard non-blast images were more difficult to discern than hard blast images.
Results additionally show that there is no detectable bias in the speed or accuracy conditions. That is, participants had no implicit preference for identifying cells as either blast or non-blast (i.e., posterior of the start-point bias includes 0). Comparison of the threshold parameters between the speed and accuracy conditions suggests that the speed instruction predominantly influences the threshold parameter. Thus, under speed instructions, it appears that both novice participants and pathologists become less cautious.
Figure 6 shows posterior distributions (best-fit parameter distributions) for the bias condition. The introduction of a cue indicating a higher likelihood that the subsequent image is a blast does appear to introduce a small bias in both novices and pathologists. Specifically, the start-point parameter shifts toward the threshold in the presence of a cue, suggesting that participants have a prior bias to respond “blast” before seeing an image. In addition, the drift rates in the bias condition show a similar pattern to those in the speed and accuracy conditions.
Comparison of DDM parameters with visual ability (NOMT)
Next, we compared participants’ performance on the NOMT with measures of speed, accuracy, and bias derived from this modeling. To do so, we used Bayesian linear regression to predict NOMT performance using the best-fit parameters from the DDM for each individual (we included both novices and pathologists). We carried out the linear regression analyses separately for accuracy, speed, and bias blocks since the model was fit separately to these conditions. For the accuracy condition, there were 5 predictors (tND, dBH, dNBH, a, bias (a – z/2)) since dBE and dNBE were not estimated for the pathologists. For the speed condition, there were 7 predictors (tND, dBE, dBH, dNBE, dNBH, a, bias (a – z/2)). For the bias condition, there were 8 predictors since there were two different biases in the model (one for cued trials and one for uncued trials). We examined all possible combinations of predictors (25 = 32 models were fit for the accuracy condition, 27 = 128 models were fit for the speed condition, and 28 = 256 models were fit for the bias condition).
For the accuracy condition, no model was strongly preferred to the null model (for all models, BF10 < 1.5). For the speed condition, the preferred model was the one with only dNBE (BFModel = 10.41 and BF10 = 115.91, R2 = 0.242). In particular, participants with larger dNBE parameter values had better performance on the NOMT. For the bias condition, the preferred model was one with both non-blast drift rates (dNBE, dNBH) and the start-point bias parameter when the cue was absent (BFModel = 23.82, BF10 = 158.26, R2 = 0.336). Similar to the speed condition, participants with larger non-blast drift rates had better performance on the NOMT. In addition, better NOMT performance was associated with a smaller bias parameter value when the cue was absent. Overall, these results show that the primary cognitive DDM parameter that correlates with NOMT performance is the evidence accumulation rates on non-blast images. As compared to SDT, the DDM provides a more nuanced correlation between task-specific ability and the NOMT, showing that the relationship is predominately driven by ability on non-blast images. These results suggest that the NOMT might have limited ability to identify individuals who make minimal detection errors (as this relationship seems to be confined to only non-blast images).
As a final note, we acknowledge that the difference in compensation between the pathologists and novices is a possible confound in our study. The pathologists were compensated with gift cards, whereas the novices were compensated through course credit. While it is possible that our results were influenced by the difference in compensation, we feel that this effect was at most minor. In particular, the pathologists did not receive performance-based compensation. All pathologists received a gift card worth the same amount regardless of performance.
In this study, we applied a joint experimental and modeling approach to investigate the cognitive processes involved in cancer cell image detection in diagnostic pathology. To probe the differences between the underlying cognitive processes of novices and experts, we used SDT and DDM analyses to assess the influence of two common cognitive manipulations that are relevant in the clinical context: speed-accuracy tradeoff and prior expectations. Many medical image observers are facing increasing workloads due to the current and projected shortages of medical technologists and pathologists (Allen, 2002; Bennett et al., 2015; Lewin, 2016; Sullivan, 2016) along with desires to improve turnaround times and reduce costs. The aim to increase productivity can result in decreased screening times and ultimately a tradeoff between speed and accuracy. The increased reliance on automated and AI systems has the (counterintuitive) potential to compound the problem. Even though these systems can offset some of the human workload, humans still play an integral role in diagnosis (at least for the near future). In the human-machine diagnostic team, it is often assumed that humans are doing less work per case and thus can increase the overall number of cases reviewed within a given day (e.g., the FDA increased the workload for cytotechnologists from 100 slides per day to 200 slides per day if they are using the ThinPrep imaging system). However, such an increase in workload (even though it comes with the assistance of a machine) can potentially exacerbate the tradeoff between speed and accuracy.
Additionally, we assessed the influence of prior expectations on performance. In diagnostic pathology, images are often analyzed by medical technologists, residents, and/or automated basic recognition sorters before being seen by senior pathologists. Images that clearly lack abnormalities are rarely passed on to a senior expert. Thus, the mere presence of an image on an expert’s desk is a cue, potentially setting expectations before the image is viewed. To examine the influence of prior expectations, we assessed how participants responded to the presence of a probabilistic cue. We also examined individual differences in decision-making by measuring domain general visual ability using the Novel Object Memory Test (NOMT).
To assess the influence of these manipulations, we used two common modeling frameworks intended to extract cognitive parameters associated with task performance, signal detection theory (SDT) and the diffusion decision model (DDM). Each of these models was fit to participant data to assess how the parameters change in response to different instruction conditions (i.e., speed, accuracy, and bias conditions) as well as how parameter values relate to experience.
SDT shows a strong dependence of discriminability on expertise with increased expertise being associated with a higher degree of discriminability. There was no difference in the criterion parameter for different levels of experience. The SDT analysis also showed very little influence of instructions on parameters (though speed instructions appear to impact discriminability for novices). This finding is not surprising given the restricted nature of SDT, which only has two cognitive parameters to account for a wide array of potential effects and thus can lead to multiple effects being conflated. In particular, it has no mechanism for quantifying the effect of changes in cognitive strategies associated with response caution (which often occur under time pressure) or response biases. In addition, we found that NOMT performance was positively correlated with discriminability and not criterion.
DDM results paint a more detailed picture of the influence of the key manipulations (speed, accuracy, and bias) on cognitive processes. Results show that speed instructions lead to a significant reduction in caution in both novices and experts. We note that this finding is at odds with other literature suggesting that experts can become more accurate under speed instructions (Beilock, Bertenthal, Hoerger, & Carr, 2008; Beilock, Bertenthal, McCoy, & Carr, 2004). In addition, estimates of the start-point bias parameter indicate that the presence of a probabilistic cue biases participants to respond “blast” before viewing the image in both novices and pathologists. Finally, drift rate estimates show distinct differences between accumulation rates associated with blast and non-blast images. For difficult conditions, blast cells appear to be more discernable (higher associated drift rate) than non-blast cells for both novices and experts. In contrast, on easy conditions, non-blasts appear to be more discernable than blast cells for novices. We also examined the relationship between DDM parameters and NOMT accuracy. This analysis revealed that the primary DDM parameter that correlates with NOMT performance is the evidence accumulation rates on non-blast images. As compared to SDT, these results paint a more nuanced picture of the relationship between task-specific ability and the NOMT, suggesting that the NOMT might be limited in assessing individual differences in this task.
In aggregate, these results suggest the following conclusions. First, novices and experts have similar behavioral characteristics. While experts are clearly superior at the task (i.e., greater discriminability), both novices and experts respond to time pressure and external cues in similar ways, and they both exhibit asymmetric responses to blast and non-blast stimuli. This suggests that while experiments with trained expert participants will always be the gold standard for research in this field, there is merit in working with novice participants, who are easier to recruit and allow for a wider array of studies. In addition, these results have important implications for training in this area. Clearly, expertise alone is not sufficient in altering the cognitive strategies and biases that are used when participants face time pressure and external cues. Second, our results show that individual differences in diagnostic decision-making are due in part to differences in visual ability (as measured by the NOMT), but these results are limited since the relationship is mainly driven by ability on non-blast images (as assessed by the DDM). Understanding individual differences is the first step in developing and improving individualized training. Future research could further explore the manipulations introduced here as well as the impact of individual differences in medical image decision-making.
Allen, K. (2002). Cytologist shortage harms patient health. ASCT News, 22, 33–35.
Beilock, S. L., Bertenthal, B. I., Hoerger, M., & Carr, T. H. (2008). When does haste make waste? Speed-accuracy tradeoff, skill level, and the tools of the trade. Journal of Experimental Psychology: Applied, 14(4), 340–352. https://doi.org/10.1037/a0012859.
Beilock, S. L., Bertenthal, B. I., McCoy, A. M., & Carr, T. H. (2004). Haste does not always make waste: expertise, direction of attention, and speed versus accuracy in performing sensorimotor skills. Psychonomic Bulletin & Review, 11(2), 373–379. https://doi.org/10.3758/Bf03196585.
Bennett, A., Thompson, N. N., Holladay, B., Bugbee, A., & Steward, C. A. (2015). ASCP wage and vacancy survey of US medical laboratories. Laboratory Medicine, 40(3), 133–141.
Bertram, R., Helle, L., Kaakinen, J. K., & Svedstrom, E. (2013). The effect of expertise on eye movement behaviour in medical image perception. PLoS One, 8(6), e66169. https://doi.org/10.1371/journal.pone.0066169.
Duchaine, B., & Nakayama, K. (2006). The Cambridge Face Memory Test: Results for neurologically intact individuals and an investigation of its validity using inverted face stimuli and prosopagnosic participants. Neuropsychologia, 44(4), 576–585. https://doi.org/10.1016/j.neuropsychologia.2005.07.001.
Dunovan, K. E., Tremel, J. J., & Wheeler, M. E. (2014). Prior probability and feature predictability interactively bias perceptual decisions. Neuropsychologia, 61, 210–221. https://doi.org/10.1016/j.neuropsychologia.2014.06.024.
Dutilh, G., Vandekerckhove, J., Forstmann, B. U., Keuleers, E., Brysbaert, M., & Wagenmakers, E. J. (2012). Testing theories of post-error slowing. Attention, Perception & Psychophysics, 74(2), 454–465. https://doi.org/10.3758/s13414-011-0243-2.
Egner, T., Monti, J. M., & Summerfield, C. (2010). Expectation and surprise determine neural population responses in the ventral visual stream. The Journal of Neuroscience, 30(49), 16601–16608. https://doi.org/10.1523/JNEUROSCI.2770-10.2010.
Elsheikh, T. M., Austin, R. M., Chhieng, D. F., Miller, F. S., Moriarty, A. T., Renshaw, A. A., & American Society of Cytopathology (2013). American Society of Cytopathology workload recommendations for automated pap test screening: developed by the productivity and quality assurance in the era of automated screening task force. Diagnostic Cytopathology, 41(2), 174–178. https://doi.org/10.1002/dc.22817.
Elsheikh, T. M., Kirkpatrick, J. L., Cooper, M. K., Johnson, M. L., Hawkins, A. P., & Renshaw, A. A. (2010). Increasing cytotechnologist workload above 100 slides per day using the ThinPrep imaging system leads to significant reductions in screening accuracy. Cancer Cytopathology, 118(2), 75–82. https://doi.org/10.1002/cncy.20065.
Forstmann, B. U., Brown, S., Dutilh, G., Neumann, J., & Wagenmakers, E. J. (2010). The neural substrate of prior information in perceptual decision making: a model-based analysis. Frontiers in Human Neuroscience, 4. https://doi.org/10.3389/Fnhum.2010.00040.
Gauthier, I., McGugin, R. W., Richler, J. J., Herzmann, G., Speegle, M., & Van Gulick, A. E. (2014). Experience moderates overlap between object and face recognition, suggesting a common ability. Journal of Vision, 14(8). https://doi.org/10.1167/14.8.7.
Glockner, A., & Hochman, G. (2011). The interplay of experience-based affective and probabilistic cues in decision making arousal increases when experience and additional cues conflict. Experimental Psychology, 58(2), 132–141. https://doi.org/10.1027/1618-3169/a000078.
Gold, J. I., & Shadlen, M. N. (2007). The neural basis of decision making. Annual Review of Neuroscience, 30, 535–574. https://doi.org/10.1146/annurev.neuro.29.051605.113038.
Goldman, L., Sayson, R., Robbins, S., Cohn, L. H., Bettmann, M., & Weisberg, M. (1983). The value of the autopsy in three medical eras. The New England Journal of Medicine, 308(17), 1000–1005. https://doi.org/10.1056/NEJM198304283081704.
Heekeren, H. R., Marrett, S., & Ungerleider, L. G. (2008). The neural systems that mediate human perceptual decision making. Nature Reviews Neuroscience, 9(6), 467–479. https://doi.org/10.1038/nrn2374.
Hildebrandt, A., Wilhelm, O., Herzmann, G., & Sommer, W. (2013). Face and object cognition across adult age. Psychology and Aging, 28(1), 243–248. https://doi.org/10.1037/a0031490.
Hoff, S. R. (2013). Breast cancer: missed interval and screening-detected cancer at full-field digital mammography and screen-film mammography—results from a retrospective review. Radiology, 264(1), 378–386. https://doi.org/10.1148/radiol.12124051.
Holmes, W. R. (2015). A practical guide to the probability density approximation (PDA) with improved implementation and error characterization. Journal of Mathematical Psychology, 68-4, 13–24. https://doi.org/10.1016/j.jmp.2015.08.006.
Holmes, W. R., & Trueblood, J. S. (2017). Bayesian analysis of the piecewise diffusion decision model. Behavior Research Methods. https://doi.org/10.3758/s13428-017-0901-y.
Holmes, W. R., Trueblood, J. S., & Heathcote, A. (2016). A new framework for modeling decisions about changing information:the piecewise linear ballistic accumulator model. Cognitive Psychology, 85, 1–29. https://doi.org/10.1016/j.cogpsych.2015.11.002.
Kass, R. E., & Raftery, A. E. (1995). Bayes factors. Journal of the American Statistical Association, 90(430), 773–795. https://doi.org/10.1080/01621459.1995.10476572.
Kirch, W., & Schafii, C. (1996). Misdiagnosis at a university hospital in 4 medical eras - report on 400 cases. Medicine, 75(1), 29–40. https://doi.org/10.1097/00005792-199601000-00004.
Krupinski, E. A. (2010). Current perspectives in medical image perception. Attention, Perception, & Psychophysics, 72(5), 1205–1217. https://doi.org/10.3758/APP.72.5.1205.
Krupinski, E. A., Graham, A. R., & Weinstein, R. S. (2013). Characterizing the development of visual search expertise in pathology residents viewing whole slide images. Human Pathology, 44(3), 357–364. https://doi.org/10.1016/j.humpath.2012.05.024.
Krupinski, E. A., Tillack, A. A., Richter, L., Henderson, J. T., Bhattacharyya, A. K., Scott, K. M., … Weinstein, R. S. (2006). Eye-movement study and human performance using telepathology virtual slides: Implications for medical education and differences with experience. Human Pathology, 37(12), 1543–1556. https://doi.org/10.1016/j.humpath.2006.08.024.
Lee, M. D., & Wagenmakers, E.-J. (2013). Bayesian cognitive modeling: a practical course. Cambridge: Cambridge University Press.
Leite, F. P., & Ratcliff, R. (2011). What cognitive processes drive response biases? A diffusion model analysis. Judgment and Decision Making, 6(7), 651–687.
Lewin, D. N. (2016). The future of the pathology workforce. Critical Values, 9(3), 6–8.
Maddox, W. T., & Bohil, C. J. (1998). Base-rate and payoff effects in multidimensional perceptual categorization. Journal of Experimental Psychology. Learning, Memory, and Cognition, 24(6), 1459–1482.
McGugin, R. W., Gatenby, J. C., Gore, J. C., & Gauthier, I. (2012). High-resolution imaging of expertise reveals reliable object selectivity in the fusiform face area related to perceptual performance. Proceedings of the National Academy of Sciences of the United States of America, 109(42), 17063–17068. https://doi.org/10.1073/pnas.1116333109.
McGugin, R. W., Richler, J. J., Herzmann, G., Speegle, M., & Gauthier, I. (2012). The Vanderbilt Expertise Test reveals domain-general and domain-specific sex effects in object recognition. Vision Research, 69, 10–22. https://doi.org/10.1016/j.visres.2012.07.014.
Mulder, M. J., Wagenmakers, E. J., Ratcliff, R., Boekel, W., & Forstmann, B. U. (2012). Bias in the brain: A diffusion model analysis of prior probability and potential payoff. Journal of Neuroscience, 32(7), 2335–2343. https://doi.org/10.1523/JNEUROSCI.4156-11.2012.
Ratcliff, R. (1978). Theory of memory retrieval. Psychological Review, 85(2), 59–108. https://doi.org/10.1037//0033-295x.85.2.59.
Ratcliff, R., Love, J., Thompson, C. A., & Opfer, J. E. (2012). Children are not like older adults: a diffusion model analysis of developmental changes in speeded responses. Child Development, 83(1), 367–381. https://doi.org/10.1111/j.1467-8624.2011.01683.x.
Ratcliff, R., & McKoon, G. (2008). The diffusion decision model: theory and data for two-choice decision tasks. Neural Computation, 20(4), 873–922. https://doi.org/10.1162/Neco.2008.12-06-420.
Ratcliff, R., & Smith, P. L. (2004). A comparison of sequential sampling models for two-choice reaction time. Psychological Review, 111(2), 333–367. https://doi.org/10.1037/0033-295X.111.2.333.
Ratcliff, R., Smith, P. L., Brown, S. D., & McKoon, G. (2016). Diffusion decision model: current issues and history. Trends in Cognitive Sciences, 20(4), 260–281. https://doi.org/10.1016/j.tics.2016.01.007.
Ratcliff, R., Thapar, A., & McKoon, G. (2001). The effects of aging on reaction time in a signal detection task. Psychology and Aging, 16(2), 323–341.
Ratcliff, R., Thapar, A., & McKoon, G. (2004). A diffusion model analysis of the effects of aging on recognition memory. Journal of Memory and Language, 50(4), 408–424. https://doi.org/10.1016/j.jml.2003.11.002.
Ratcliff, R., Thapar, A., & McKoon, G. (2010). Individual differences, aging, and IQ in two-choice tasks. Cognitive Psychology, 60(3), 127–157. https://doi.org/10.1016/j.cogpsych.2009.09.001.
Reed, A. V. (1973). Speed-accuracy trade-off in recognition memory. Science, 181(4099), 574–576. https://doi.org/10.1126/Science181.4099.574.
Richler, J. J., Wilmer, J. B., & Gauthier, I. (2017). General object recognition is specific: evidence from novel and familiar objects. Cognition, 166, 42–55. https://doi.org/10.1016/j.cognition.2017.05.019.
Samei, S., & Krupinski, E. (2010). The handbook of medical image perception and techniques, (1st ed., ). Cambridge: Cambridge University Press.
Schouten, J. F., & Bekker, J. A. (1967). Reaction time and accuracy. Acta Psychologica, 27, 143–153.
Shojania, K. G., Burton, E. C., McDonald, K. M., & Goldman, L. (2003). Changes in rates of autopsy-detected diagnostic errors over time: a systematic review. Journal of the American Medical Association, 289(21), 2849–2856. https://doi.org/10.1001/jama.289.21.2849.
Sonderegger-Iseli, K., Burger, S., Muntwyler, J., & Salomon, F. (2000). Diagnostic errors in three medical eras: a necropsy study. Lancet, 355(9220), 2027–2031.
Sullivan, H. C. (2016). A changing workforce brings new challenges, opportunities. Critical Values, 9(3), 14–16.
Summerfield, C., & de Lange, F. P. (2014). Expectation in perceptual decision making: Neural and computational mechanisms. Nature Reviews Neuroscience, 15(11), 745–756. https://doi.org/10.1038/nrn3838.
Summerfield, C., & Egner, T. (2009). Expectation (and attention) in visual cognition. Trends in Cognitive Sciences, 13(9), 403–409. https://doi.org/10.1016/j.tics.2009.06.003.
JASP Team. (2018). JASP (Version 0.9)[Computer software]. https://jasp-stats.org/.
Turner, B. M., & Sederberg, P. B. (2012). Approximate Bayesian computation with differential evolution. Journal of Mathematical Psychology, 56(5), 375–385. https://doi.org/10.1016/j.jmp.2012.06.004.
Turner, B. M., & Sederberg, P. B. (2014). A generalized, likelihood-free method for posterior estimation. Psychonomic Bulletin & Review, 21(2), 227–250. https://doi.org/10.3758/s13423-013-0530-0.
van der Gijp, A., Ravesloot, C. J., Jarodzka, H., van der Schaaf, M. F., van der Schaaf, I. C., van Schaik, J. P. J., & Ten Cate, T. J. (2017). How visual search relates to visual diagnostic performance: a narrative systematic review of eye-tracking research in radiology. Advances in Health Sciences Education: Theory and Practice, 22(3), 765–787. https://doi.org/10.1007/s10459-016-9698-1.
White, C. N., & Poldrack, R. A. (2014). Decomposing bias in different types of simple decisions. Journal of Experimental Psychology-Learning Memory and Cognition, 40(2), 385–398. https://doi.org/10.1037/a0034851.
Wickelgren, W. A. (1977). Speed-accuracy tradeoff and information-processing dynamics. Acta Psychologica, 41(1), 67–85. https://doi.org/10.1016/0001-6918(77)90012-9.
The authors would like to thank Isabel Gauthier for advice on using the NOMT.
This work was supported by a Clinical and Translational Research Enhancement Award from the Department of Pathology, Microbiology, and Immunology, Vanderbilt University Medical Center. JST and WRH were supported by National Science Foundation grant SES-1556415.
Availability of data and materials
All de-identified data are available on the Open Science Framework at https://osf.io/r3gzs/.
Ethics approval and consent to participate
This study and all of its materials and consent documents were approved prior to the initiation of data collection by the Vanderbilt University Institutional Review Board (IRB # 161767).
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Trueblood, J.S., Holmes, W.R., Seegmiller, A.C. et al. The impact of speed and bias on the cognitive processes of experts and novices in medical image decision-making. Cogn. Research 3, 28 (2018) doi:10.1186/s41235-018-0119-2
- Medical decision-making
- Diagnostic pathology
- Signal detection theory
- Diffusion decision model
- General object recognition
- Cancer image detection