Case Western Reserve University’s diagnostic imaging computers outperform human counterparts, but offer tools for better, less costly care
The “deep learning” computers in Anant Madabhushi’s diagnostic imaging lab at Case Western Reserve University routinely defeat their human counterparts in diagnosing heart failure, detecting various cancers and predicting their strength.
But Madabhushi—even as he gladly touts three recent examples of apparent cyber superiority that played out in his lab—also dismisses any implication of a coming future when such machines replace pathologists and radiologists.
“There’s initially always going to be some wincing and anxiety among pathologists and radiologists over this idea—that our computational imaging technology can outperform us or even take our jobs,” said Madabhushi, whose center has made significant diagnostic advances in cardiovascular disease and also brain, lung, breast, prostate and head and neck cancers since opening in 2012.
Madabhushi, the F. Alex Nason Professor II of biomedical engineering at the Case School of Engineering, contends that his research is not only introducing invaluable diagnostic tools, but also helping to identify those patients with less aggressive disease who may not need more aggressive therapy.
Since 2016, Madabhushi and his team have received over $9.5 million from the National Cancer Institute to develop computational tools for analysis of digital pathology images of breast, lung, and head and neck cancers to identify which patients with these diseases could be spared aggressive radiotherapy or chemotherapy.
“It’s not so much that we were able to ‘beat’ the pathologist or the radiologist, but rather that the machine was able to add value to what they can offer,” he said. “There is desperate need for better decision-support tools that allows them to serve patients, especially in places where there are very few pathologists or radiologists.
“By providing them with decision support, we can help them become more efficient. For instance, the tools could help reduce the amount of time spent on cases with no obvious disease or obviously benign conditions and instead help them focus on the more confounding cases.”
Those tools have been producing exceptionally accurate results at Madabhushi’s Center for Computational Imaging and Personalized Diagnostics (CCIPD) at Case Western Reserve.
Consider three recent examples:
- The computational-imaging system in Madabhushi’s lab correctly predicted with a 97-percent accuracy which among 105 patients were already showing evidence of pending heart failure. By comparison, two pathologists were correct 74 percent and 73 percent, respectively. The results were recently published in the journal PLOS ONE.
- Madabhushi and co-investigators—supported by a $608,000 U.S. Department of Defense, Congressionally Directed Medical Research Program grant—show that while human radiologists could flag up to half of all nodules that show up in a CAT scan as “suspicious” or “indeterminate,” about 98 percent of those nodules actually turn out to be benign. In a recent study published in the Journal of Medical Imaging, Madabhushi and his group showed that their computational imaging technique was between 5 and 8 percent superior compared to two human experts in distinguishing benign from malignant lung nodules on CAT scans.
- In an international study of prostate cancer scans in the U.S., Finland and Australia, the computational imaging algorithms outperformed their human counterparts in two ways, detailed in a study recently published in the Journal of Magnetic Resonance Imaging.
- In more than 70 percent of cases where radiologists missed the presence of clinically significant prostate cancer on a magnetic resonance imaging (MRI) scan, the machine algorithm caught it.
- In half of cases where radiologists mistakenly identified the presence of clinically significant prostate cancer on a MRI scan, the machine was able to correctly identify that no clinically significant disease was present.
“This is all very exciting data for us, but now we need more validation and to demonstrate these results on larger cohorts,” Madabhushi said. “But we really believe this is more evidence of what computational imaging of pathology and radiology images can do for cardiovascular and cancer research and practical use among pathologists and radiologists.”
Radiomics and Pathomics: The tech behind the triumph
So, what exactly are these supercomputers doing that humans can’t that creates such a wide margin in diagnostic success?
The short answer could be said for virtually all computer advantages in the last half century: The machines do work at far greater speed and volume.
The precise difference here is that the diagnostic imaging computers at the CCIPD can read, log, compare and contrast literally hundreds of slides of tissue samples in the amount of time a pathologist might spend on a single slide.
Then, they rapidly and completely catalogue characteristics like texture, shape and structure of glands, nuclei and surrounding tissue to determine the aggressiveness and risk associated with certain diseases.
This is where the “deep learning” comes in: From all of that, they create algorithms which can look beyond what the human eye can see in comparing and contrasting those multitudes of images. Finally, they are working toward predicting everything from how aggressive a disease is going to be to whether a scanned nodule is likely to even turn cancerous.
In the end, all of this new information should help pathologists and radiologists with the interpretations of slides and scans, but more critically can help clinicians make more-informed treatment recommendations.
Madabhushi said that can help a single pathologist do her work more efficiently by more accurately triaging patients by true need for care—or provide hope to an entire nation.
“I always use the example of Botswana, where they have a population of 2 million people—and only one pathologist that we are aware of,” he said. “From that one example alone, you can see that this technology can help that one pathologist be more efficient and help many more people.”
For more information, contact Mike Scott at firstname.lastname@example.org or 216.368.1004.
This article was originally published April 27, 2018.