Main Navigation

Missing the radiological forest for the trees

Clear abnormalities can missed even by experienced radiologists if they're looking for something else

There’s a classic video demonstrating how our brains process information and allocate attention in which people bounce and pass basketballs and the viewer is asked to count the passes.

If you haven’t seen it, go watch it here and then come back. Go ahead. I’ll wait.

The experiment highlights a phenomenon called inattentional blindness. We can’t pay attention to everything at once, so our brains have to filter information. In the situation in the video, the stakes were low. But what if inattentional blindness causes a radiologist, for example, to miss something obvious and serious?

A study from University of Utah researchers Lauren Williams, Trafton Drew and colleagues finds that even experienced radiologists, when looking for one abnormality, can completely miss another. The results, published in Psychonomic Bulletin & Review, show that inattentional blindness can befall even experts.

“Inattentional blindness reveals the limits of human cognition,” says Williams, a recent U graduate and now a postdoctoral scholar at the University of California, San Diego, “and this research demonstrates that even highly trained experts are bound by the same machinery as everyone else.”

“If even these experts miss these seemingly obvious findings,” adds Drew, associate professor of psychology, “it suggests that this is something really critical we need to understand about how all of us perceive the world.”

Missing the gorilla

By some estimates medical errors, including missed radiological abnormalities, are the third leading cause of death in the United States. “We’ve known for a long time that many errors in radiology are retrospectively visible,” Drew says. “This means if something goes wrong with a patient, you can often go back to the imaging for that patient and see that there were visible signs—say, a lung nodule—on something like a chest CT.”

So, in 2013, Drew and colleagues conducted an experiment to understand how trained experts could miss those clear signs. In that study, the authors presented radiologists with chest computed tomography (CT) scans and asked them to look for lung cancer nodules. But the authors had also placed an image of a gorilla into the scan—something that obviously doesn’t belong in a lung. Drew and his colleagues found that 83% of radiologists did not notice the gorilla.

But that’s a gorilla. How would the results be different, they wondered, if instead of a gorilla it was an abnormality that could plausibly come up on a CT scan?

So Williams, Drew and their colleagues from UCLA and Macquarie University set up another experiment. They asked 50 radiologists to evaluate seven chest CT scans for lung cancer, but this time the final scan included two clear abnormalities: a significant breast mass and a lymphadenopathy (an abnormal lymph node). Two-thirds of the radiologists did not notice the potentially cancerous mass. A third did not notice the lymphadenopathy.

Like anyone that experiences inattentional blindness, I think many radiologists were simply surprised to learn they had missed something,” says Williams, who administered the experiment. “Our intuition tells us that if something is fully visible, we’ll detect it, but we’ve all experienced the feeling of missing important information that is retrospectively obvious when our attention is focused elsewhere.”

Experience wasn’t a factor in whether or not the radiologists noticed the abnormalities, the researchers found, suggesting that years of experience doesn’t outweigh universal cognitive truths, and that missing the abnormalities isn’t a reflection on the competence or skill of the radiologist.

“It suggests that understanding the situation that led to the missed abnormality may be far more important than focusing on the experience of the individuals that missed it,” Williams says.

The radiologist’s perspective

U radiologist Megan Mills, a co-author of the study, says she was not surprised by the results. Directed and accurate clinical information from the physician who ordered the scan is key to correctly diagnosing an abnormality. “In the absence of clinical information, radiologists are searching for the proverbial needle in the haystack,” she says. “With a specific indication or reason for the exam, the haystack is much more manageable.”

Finding an abnormality outside of what the ordering physician was looking for is called an “incidental finding,” and it can be a good thing if it’s a finding that can improve a patient’s care. But in some cases, following up on incidental findings can require additional testing that may or may not be needed and would incur additional cost. “This is one of the many balancing acts of radiology,” Mills says, “identifying those findings that will make a positive impact in care without over-diagnosing irrelevant or unimportant findings.”

Radiologist and co-author William Auffermann adds that radiologists are faced with a huge volume of information each day. “For example, a standard CT of the chest may contain about 500 images,” he says. “In a typical day, one may be expected to read 20-30 or more chest CTs, corresponding to 10,000-15,000 or more images per day; a daunting task.”

So missing an incidental finding does not necessarily equate to a medical error, Mills says. “Once you know there is a gorilla, it is easy to pick it out of the crowd, but just because something is retrospectively apparent, it doesn’t mean that someone should be expected to see the gorilla the first time around.”

Seeing the gorilla by knowing where to look

In a subsequent experiment, however, instead of asking the radiologists to look for lung cancer nodules, the researchers asked radiologists to look at the same scan and report on a broader range of abnormalities. This time, only 3% missed the mass and 10% missed the lymphadenopathy.

“There a huge amount of information in the ever-growing amount of data we gather on each patient, and what we actually notice depends very strongly on what you are looking for,” Drew says. “Cataloging how often radiologists miss something in plain sight misses a really important piece of the puzzle: What were they looking for when they missed the thing in plain sight?”

“Our research demonstrates that focusing narrowly on one task may cause radiologists to miss unexpected abnormalities, even if those abnormalities are critical for patient outcomes,” Williams adds. “However, focused attention is probably beneficial when the abnormalities match the radiologist’s expectations.” Any changes to clinical process would need to find the balance between the two, she says. Some possibilities might be a general assessment of a scan before looking for specific abnormalities, or using checklists to scan for commonly missed findings.

Would the use of artificial intelligence, which doesn’t have the same cognitive limitations that humans do, resolve the problem of inattentional blindness? Not necessarily, Drew says. AI is only as good as its training and programming. Algorithms are good at finding narrowly defined abnormalities, he says, but not as good at detecting all possible findings on a scan.

“Radiologists might benefit from being thoughtful about what they are looking for rather assuming that if they see it they will perceive it,” Drew says. “AI has in some ways, the same limitation: it’s only going to be good at detecting what it has been taught to detect.”

Auffermann says that the current use of AI in radiology is often as an adjunct to the radiologist. “An AI algorithm may act as a virtual second reader of a study and flag potential abnormalities for the radiologist to consider further,” he says. “That said, the capabilities and role of AI in radiology and medicine has been evolving rapidly.”

Williams says that advancements in radiological technology have produced increasingly clear medical imaging. “However, if radiologists frequently miss a large, clearly visible abnormality when their attention is focused on another task, it suggests that having a clear image is not enough.”

Drew says the study can help us understand how we often find only what we’re looking for.

“Everyone, even experts, can miss things that seem really obvious if we are not looking for them,” he says. “If you’ve searched through your whole apartment for your phone, you might assume you would have noticed your keys during that search. Our research suggests a reason why you will probably have to search again specifically for the keys.”

Find the full study here.

This release has been modified from its original version.