AI Approach Outperformed Human Experts in Identifying Cervical Precancer

A research team led by investigators from the National Institutes of Health and Global Good has developed a computer algorithm that can analyze digital images of a woman's cervix and accurately identify precancerous changes that require medical attention. This artificial intelligence (AI) approach, called automated visual evaluation, has the potential to revolutionize cervical cancer screening, particularly in low-resource settings.

To develop the method, researchers used comprehensive datasets to "train" a deep, or machine, learning algorithm to recognize patterns in complex visual inputs, such as medical images. The approach was created collaboratively by investigators at the National Cancer Institute (NCI) and Global Good, a project of Intellectual Ventures, and the findings were confirmed independently by experts at the National Library of Medicine (NLM). The results appeared in the Journal of the National Cancer Institute on January 10, 2019. NCI and NLM are parts of NIH.

"Our findings show that a deep learning algorithm can use images collected during routine cervical cancer screening to identify precancerous changes that, if left untreated, may develop into cancer," said Mark Schiffman, M.D., M.P.H., of NCI's Division of Cancer Epidemiology and Genetics, and senior author of the study. "In fact, the computer analysis of the images was better at identifying precancer than a human expert reviewer of Pap tests under the microscope (cytology)."

The new method has the potential to be of particular value in low-resource settings. Health care workers in such settings currently use a screening method called visual inspection with acetic acid (VIA). In this approach, a health worker applies dilute acetic acid to the cervix and inspects the cervix with the naked eye, looking for "aceto whitening," which indicates possible disease. Because of its convenience and low cost, VIA is widely used where more advanced screening methods are not available. However, it is known to be inaccurate and needs improvement.

Automated visual evaluation is similarly easy to perform. Health workers can use a cell phone or similar camera device for cervical screening and treatment during a single visit. In addition, this approach can be performed with minimal training, making it ideal for countries with limited health care resources, where cervical cancer is a leading cause of illness and death among women.

To create the algorithm, the research team used more than 60,000 cervical images from an NCI archive of photos collected during a cervical cancer screening study that was carried out in Costa Rica in the 1990s. More than 9,400 women participated in that population study, with follow up that lasted up to 18 years. Because of the prospective nature of the study, the researchers gained nearly complete information on which cervical changes became precancers and which did not. The photos were digitized and then used to train a deep learning algorithm so that it could distinguish cervical conditions requiring treatment from those not requiring treatment.

Overall, the algorithm performed better than all standard screening tests at predicting all cases diagnosed during the Costa Rica study. Automated visual evaluation identified precancer with greater accuracy (AUC=0.91) than a human expert review (AUC=0.69) or conventional cytology (AUC=0.71). An AUC of 0.5 indicates a test that is no better than chance, whereas an AUC of 1.0 represents a test with perfect accuracy in identifying disease.

"When this algorithm is combined with advances in HPV vaccination, emerging HPV detection technologies, and improvements in treatment, it is conceivable that cervical cancer could be brought under control, even in low-resource settings," said Maurizio Vecchione, executive vice president of Global Good.

The researchers plan to further train the algorithm on a sample of representative images of cervical precancers and normal cervical tissue from women in communities around the world, using a variety of cameras and other imaging options. This step is necessary because of subtle variations in the appearance of the cervix among women in different geographic regions. The ultimate goal of the project is to create the best possible algorithm for common, open use.

Liming Hu, David Bell, Sameer Antani, Zhiyun Xue, Kai Yu, Matthew P Horning, Noni Gachuhi, Benjamin Wilson, Mayoore S Jaiswal, Brian Befano, L Rodney Long, Rolando Herrero, Mark H Einstein, Robert D Burk, Maria Demarco, Julia C Gage, Ana Cecilia Rodriguez, Nicolas Wentzensen, Mark Schiffman.
An Observational Study of Deep Learning and Automated Evaluation of Cervical Images for Cancer Screening, JNCI: Journal of the National.
Cancer Institute, doi: 10.1093/jnci/djy225.

Most Popular Now

MEDICA 2024 + COMPAMED 2024: Adapted Hal…

11 - 14 November 2024, Düsseldorf, Germany. The final preparations for MEDICA 2024 and COMPAMED 2024 in Düsseldorf have begun. A total of more than 5,500 exhibitors from approximately 70 countries...

AI does Not Necessarily Lead to more Eff…

The use of artificial intelligence (AI) in hospitals and patient care is steadily increasing. Especially in specialist areas with a high proportion of imaging, such as radiology, AI has long...

Commission Joins Forces with Venture Cap…

The Commission has launched a Trusted Investors Network bringing together a group of investors ready to co-invest in innovative deep-tech companies in Europe together with the EU. The Union's investment...

An AI-Powered Pipeline for Personalized …

Ludwig Cancer Research scientists have developed a full, start-to-finish computational pipeline that integrates multiple molecular and genetic analyses of tumors and the specific molecular targets of T cells and harnesses...

Philips and Medtronic Advocacy Partnersh…

Royal Philips (NYSE: PHG, AEX: PHIA), a global leader in health technology, and Medtronic Neurovascular, a leading innovator in neurovascular therapies, today announced a strategic advocacy partnership. Delivering timely stroke...

Wearable Cameras Allow AI to Detect Medi…

A team of researchers says it has developed the first wearable camera system that, with the help of artificial intelligence (AI), detects potential errors in medication delivery. In a test whose...

AI could Transform How Hospitals Produce…

A pilot study led by researchers at University of California San Diego School of Medicine found that advanced artificial intelligence (AI) could potentially lead to easier, faster and more efficient...

New AI Tool Predicts Protein-Protein Int…

Scientists from Cleveland Clinic and Cornell University have designed a publicly-available software and web database to break down barriers to identifying key protein-protein interactions to treat with medication. The computational tool...

AI for Real-Rime, Patient-Focused Insigh…

A picture may be worth a thousand words, but still... they both have a lot of work to do to catch up to BiomedGPT. Covered recently in the prestigious journal Nature...

Start-Ups will Once Again Have a Starrin…

11 - 14 November 2024, Düsseldorf, Germany. The finalists in the 16th Healthcare Innovation World Cup and the 13th MEDICA START-UP COMPETITION have advanced from around 550 candidates based in 62...

New Research Shows Promise and Limitatio…

Published in JAMA Network Open, a collaborative team of researchers from the University of Minnesota Medical School, Stanford University, Beth Israel Deaconess Medical Center and the University of Virginia studied...

G-Cloud 14 Makes it Easier for NHS to Bu…

NHS organisations will be able to save valuable time and resource in the procurement of technologies that can make a significant difference to patient experience, in the latest iteration of...