How man and machine can work together to diagnose diseases in medical scans
- Written by Anant Madabhushi, Professor of Biomedical Engineering, Case Western Reserve University
With artificial intelligence, machines can now examine thousands of medical images – and billions of pixels within these images – to identify patterns too subtle for a radiologist or pathologist to identify.
The machine then uses this information to identify the presence of a disease or estimate its aggressiveness, likelihood of survival or potential response to treatment.
We are engineers at the Center for Computational Imaging and Personalized Diagnostics[1]. Our team works with physicians and statisticians to develop and validate these kinds of tools.
Many worry that this technology aims to replace doctors. But we believe the technology will work in tandem with humans, making them more efficient and helping with decisions on complicated cases.
Machine learning and medical images
In one study[2], researchers at Stanford showed that machines were just as accurate as trained dermatologists in distinguishing skin cancers from benign lesions in 100 test images.
In another[3], computer scientists at Google used an AI approach called deep learning to accurately identify which patients had diabetic retinopathy – a constellation of changes in the retina due to diabetes – from high-resolution photographs of the retina. Another deep learning project[4] at Google successfully predicted cardiovascular disease risk from retina images.
Our group has been developing new ways to identify disease in scans like MRI and CT, as well as digitized tissue slide images.
In biopsied images of heart tissue from 105 patients with heart disease, our algorithms predicted[5] with high accuracy which patients would go on to have heart failure.
In another study[6] involving MRI scans from prostate cancer patients, our computer algorithms identified clinically significant disease in more than 70 percent of cases where radiologists missed it. In half of the cases where radiologists mistakenly thought that the patient had aggressive prostate cancer, the machine was able to correctly identify that no clinically significant disease was present.
Wellcome Collection, CC BY[7][8]Predicting outcomes and treatment response
Our team has also been developing approaches to predict a patient’s response to specific therapies and monitor early treatments.
Take the case of immunotherapy. Immunotherapy drugs boost the body’s own immune defenses to fight against the cancer. They have shown tremendous promise in comparison to traditional chemotherapy[9], but certain caveats limit their widespread use. Only about 1 in 5 patients with lung cancer[10] might actually respond to immunotherapy regimens. Additionally, these therapies cost upwards of US$200,000 per patient per year and can have toxic effects on the immune system. Physicians need a way to determine accurately which patients might benefit.
To identify the likelihood of a successful response before therapy begins, our lab is building software to examine routine diagnostic CT scans of lung tumors. The software looks at tumor texture, intensity and shape, as well as the shape of vessels feeding the nodules[11]. This information might help oncologists optimize the therapy dosage or alter a patient’s treatment plan.
There are many cancers and other diseases where computational tools to predict disease aggressiveness or treatment response could aid physicians. For example, a study published this June[12] compared women with breast cancer who had been treated with adjuvant chemotherapy or standard endocrine therapy. For about 70 percent of the patients, the chemotherapy had no demonstrated benefit compared to the standard approach. Preventing unnecessary and often deleterious chemotherapy thus becomes a key issue for doctors. However, currently, the only way to predict outcome depends upon expensive genomic tests that destroy tissue.
We are currently working on a new way to interrogate digitized tissue images from biopsies of the breast. Our new project, starting this year, will test the technology on women with breast cancers at Tata Memorial Hospital in Mumbai, India.
Making it possible
There is a massive opportunity for clinicians, radiologists and pathologists to enrich their decisions with artificial intelligence. That’s particularly true when it comes to building treatment plans tailored to the individual patient.
Before such technology can be used in hospitals, researchers like ourselves need to do further tests to ensure it’s reliable and valid. This can be done by carrying out tests at multiple medical institutions.
It’s also important for physicians to be able to interpret the technology. They’re unlikely to adopt technology that cannot be explained by existing biology research. For example, our lung tumor software looks at vessel shape because studies show[13] that the degree of convolutedness of the vessels feeding the tumor can negatively affect drug delivery.
That’s why it’s crucial artificial intelligence researchers like ourselves engage clinicians early in the development process as equal collaborators.
References
- ^ Center for Computational Imaging and Personalized Diagnostics (ccipd.case.edu)
- ^ one study (doi.org)
- ^ In another (doi.org)
- ^ deep learning project (doi.org)
- ^ our algorithms predicted (doi.org)
- ^ In another study (doi.org)
- ^ Wellcome Collection (wellcomecollection.org)
- ^ CC BY (creativecommons.org)
- ^ tremendous promise in comparison to traditional chemotherapy (doi.org)
- ^ Only about 1 in 5 patients with lung cancer (dx.doi.org)
- ^ the shape of vessels feeding the nodules (ascopubs.org)
- ^ a study published this June (doi.org)
- ^ studies show (ascopubs.org)
Authors: Anant Madabhushi, Professor of Biomedical Engineering, Case Western Reserve University