machine learning
Using Artificial Intelligence to Detect Cervical Cancer
Posted on by Dr. Francis Collins

My last post highlighted the use of artificial intelligence (AI) to create an algorithm capable of detecting 10 different kinds of irregular heart rhythms. But that’s just one of the many potential medical uses of AI. In this post, I’ll tell you how NIH researchers are pairing AI analysis with smartphone cameras to help more women avoid cervical cancer.
In work described in the Journal of the National Cancer Institute [1], researchers used a high-performance computer to analyze thousands of cervical photographs, obtained more than 20 years ago from volunteers in a cancer screening study. The computer learned to recognize specific patterns associated with pre-cancerous and cancerous changes of the cervix, and that information was used to develop an algorithm for reliably detecting such changes in the collection of images. In fact, the AI-generated algorithm outperformed human expert reviewers and all standard screening tests in detecting pre-cancerous changes.
Nearly all cervical cancers are caused by the human papillomavirus (HPV). Cervical cancer screening—first with Pap smears and now also using HPV testing—have greatly reduced deaths from cervical cancer. But this cancer still claims the lives of more than 4,000 U.S. women each year, with higher frequency among women who are black or older [2]. Around the world, more than a quarter-million women die of this preventable disease, mostly in poor and remote areas [3].
These troubling numbers have kept researchers on the lookout for low cost, but easy-to-use, tools that could be highly effective at detecting HPV infections most likely to advance to cervical cancer. Such tools would also need to work well in areas with limited resources for sample preparation and lab analysis. That’s what led to this collaboration involving researchers from NIH’s National Cancer Institute (NCI) and Global Good, Bellevue, WA, which is an Intellectual Ventures collaboration with Bill Gates to invent life-changing technologies for the developing world.
Global Good researchers contacted NCI experts hoping to apply AI to a large dataset of cervical images. The NCI experts suggested an 18-year cervical cancer screening study in Costa Rica. The NCI-supported project, completed in the 1990s, generated nearly 60,000 cervical images, later digitized by NIH’s National Library of Medicine and stored away safely.
The researchers agreed that all these images, obtained in a highly standardized way, would serve as perfect training material for a computer to develop a detection algorithm for cervical cancer. This type of AI, called machine learning, involves feeding tens of thousands of images into a computer equipped with one or more high-powered graphics processing units (GPUs), similar to something you’d find in an Xbox or PlayStation. The GPUs allow the computer to crunch large sets of visual data in the images and devise a set of rules, or algorithms, that allow it to learn to “see” physical features.
Here’s how they did it. First, the researchers got the computer to create a convolutional neural network. That’s a fancy way of saying that they trained it to read images, filter out the millions of non-essential bytes, and retain the few hundred bytes in the photo that make it uniquely identifiable. They fed 1.28 million color images covering hundreds of common objects into the computer to create layers of processing ability that, like the human visual system, can distinguish objects and their qualities.
Once the convolutional neural network was formed, the researchers took the next big step: training the system to see the physical properties of a healthy cervix, a cervix with worrisome cellular changes, or a cervix with pre-cancer. That’s where the thousands of cervical images from the Costa Rican screening trial literally entered the picture.
When all these layers of processing ability were formed, the researchers had created the “automated visual evaluation” algorithm. It went on to identify with remarkable accuracy the images associated with the Costa Rican study’s 241 known precancers and 38 known cancers. The algorithm’s few minor hiccups came mainly from suboptimal images with faded colors or slightly blurred focus.
These minor glitches have the researchers now working hard to optimize the process, including determining how health workers can capture good quality photos of the cervix with a smartphone during a routine pelvic exam and how to outfit smartphones with the necessary software to analyze cervical photos quickly in real-world settings. The goal is to enable health workers to use a smartphone or similar device to provide women with cervical screening and treatment during a single visit.
In fact, the researchers are already field testing their AI-inspired approach on smartphones in the United States and abroad. If all goes well, this low-cost, mobile approach could provide a valuable new tool to help reduce the burden of cervical cancer among underserved populations.
The day that cervical cancer no longer steals the lives of hundreds of thousands of women a year worldwide will be a joyful moment for cancer researchers, as well as a major victory for women’s health.
References:
[1] An observational study of Deep Learning and automated evaluation of cervical images for cancer screening. Hu L, Bell D, Antani S, Xue Z, Yu K, Horning MP, Gachuhi N, Wilson B, Jaiswal MS, Befano B, Long LR, Herrero R, Einstein MH, Burk RD, Demarco M, Gage JC, Rodriguez AC, Wentzensen N, Schiffman M. J Natl Cancer Inst. 2019 Jan 10. [Epub ahead of print]
[2] “Study: Death Rate from Cervical Cancer Higher Than Thought,” American Cancer Society, Jan. 25, 2017.
[3] “World Cancer Day,” World Health Organization, Feb. 2, 2017.
Links:
Cervical Cancer (National Cancer Institute/NIH)
Global Good (Intellectual Ventures, Bellevue, WA)
NIH Support: National Cancer Institute; National Library of Medicine
Using Artificial Intelligence to Catch Irregular Heartbeats
Posted on by Dr. Francis Collins

Thanks to advances in wearable health technologies, it’s now possible for people to monitor their heart rhythms at home for days, weeks, or even months via wireless electrocardiogram (EKG) patches. In fact, my Apple Watch makes it possible to record a real-time EKG whenever I want. (I’m glad to say I am in normal sinus rhythm.)
For true medical benefit, however, the challenge lies in analyzing the vast amounts of data—often hundreds of hours worth per person—to distinguish reliably between harmless rhythm irregularities and potentially life-threatening problems. Now, NIH-funded researchers have found that artificial intelligence (AI) can help.
A powerful computer “studied” more than 90,000 EKG recordings, from which it “learned” to recognize patterns, form rules, and apply them accurately to future EKG readings. The computer became so “smart” that it could classify 10 different types of irregular heart rhythms, including atrial fibrillation (AFib). In fact, after just seven months of training, the computer-devised algorithm was as good—and in some cases even better than—cardiology experts at making the correct diagnostic call.
EKG tests measure electrical impulses in the heart, which signal the heart muscle to contract and pump blood to the rest of the body. The precise, wave-like features of the electrical impulses allow doctors to determine whether a person’s heart is beating normally.
For example, in people with AFib, the heart’s upper chambers (the atria) contract rapidly and unpredictably, causing the ventricles (the main heart muscle) to contract irregularly rather than in a steady rhythm. This is an important arrhythmia to detect, even if it may only be present occasionally over many days of monitoring. That’s not always easy to do with current methods.
Here’s where the team, led by computer scientists Awni Hannun and Andrew Ng, Stanford University, Palo Alto, CA, saw an AI opportunity. As published in Nature Medicine, the Stanford team started by assembling a large EKG dataset from more than 53,000 people [1]. The data included various forms of arrhythmia and normal heart rhythms from people who had worn the FDA-approved Zio patch for about two weeks.
The Zio patch is a 2-by-5-inch adhesive patch, worn much like a bandage, on the upper left side of the chest. It’s water resistant and can be kept on around the clock while a person sleeps, exercises, or takes a shower. The wireless patch continuously monitors heart rhythms, storing EKG data for later analysis.
The Stanford researchers looked to machine learning to process all the EKG data. In machine learning, computers rely on large datasets of examples in order to learn how to perform a given task. The accuracy improves as the machine “sees” more data.
But the team’s real interest was in utilizing a special class of machine learning called deep neural networks, or deep learning. Deep learning is inspired by how our own brain’s neural networks process information, learning to focus on some details but not others.
In deep learning, computers look for patterns in data. As they begin to “see” complex relationships, some connections in the network are strengthened while others are weakened. The network is typically composed of multiple information-processing layers, which operate on the data and compute increasingly complex and abstract representations.
Those data reach the final output layer, which acts as a classifier, assigning each bit of data to a particular category or, in the case of the EKG readings, a diagnosis. In this way, computers can learn to analyze and sort highly complex data using both more obvious and hidden features.
Ultimately, the computer in the new study could differentiate between EKG readings representing 10 different arrhythmias as well as a normal heart rhythm. It could also tell the difference between irregular heart rhythms and background “noise” caused by interference of one kind or another, such as a jostled or disconnected Zio patch.
For validation, the computer attempted to assign a diagnosis to the EKG readings of 328 additional patients. Independently, several expert cardiologists also read those EKGs and reached a consensus diagnosis for each patient. In almost all cases, the computer’s diagnosis agreed with the consensus of the cardiologists. The computer also made its calls much faster.
Next, the researchers compared the computer’s diagnoses to those of six individual cardiologists who weren’t part of the original consensus committee. And, the results show that the computer actually outperformed these experienced cardiologists!
The findings suggest that artificial intelligence can be used to improve the accuracy and efficiency of EKG readings. In fact, Hannun reports that iRhythm Technologies, maker of the Zio patch, has already incorporated the algorithm into the interpretation now being used to analyze data from real patients.
As impressive as this is, we are surely just at the beginning of AI applications to health and health care. In recognition of the opportunities ahead, NIH has recently launched a working group on AI to explore ways to make the best use of existing data, and harness the potential of artificial intelligence and machine learning to advance biomedical research and the practice of medicine.
Meanwhile, more and more impressive NIH-supported research featuring AI is being published. In my next blog, I’ll highlight a recent paper that uses AI to make a real difference for cervical cancer, particularly in low resource settings.
Reference:
[1] Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network. Hannun AY, Rajpurkar P, Haghpanahi M, Tison GH, Bourn C, Turakhia MP, Ng AY.
Nat Med. 2019 Jan;25(1):65-69.
Links:
Arrhythmia (National Heart, Lung, and Blood Institute/NIH)
Video: Artificial Intelligence: Collecting Data to Maximize Potential (NIH)
Andrew Ng (Palo Alto, CA)
NIH Support: National Heart, Lung, and Blood Institute
Attending NIH Artificial Intelligence Workshop
Posted on by Dr. Francis Collins

A lot of great discussion took place at the NIH Workshop on Artificial Intelligence: Collecting Data to Maximize Potential.” Craig Mundie (left), president of Mundie and Associates, San Francisco, gave a tremendous keynote address. The one-day workshop was held on July 23, 2018. Credit: NIH
Teaching Computers to “See” the Invisible in Living Cells
Posted on by Dr. Francis Collins

Credit: Steven Finkbeiner, University of California, San Francisco and the Gladstone Institutes
For centuries, scientists have trained themselves to look through microscopes and carefully study their structural and molecular features. But those long hours bent over a microscope poring over microscopic images could be less necessary in the years ahead. The job of analyzing cellular features could one day belong to specially trained computers.
In a new study published in the journal Cell, researchers trained computers by feeding them paired sets of fluorescently labeled and unlabeled images of brain tissue millions of times in a row [1]. This allowed the computers to discern patterns in the images, form rules, and apply them to viewing future images. Using this so-called deep learning approach, the researchers demonstrated that the computers not only learned to recognize individual cells, they also developed an almost superhuman ability to identify the cell type and whether a cell was alive or dead. Even more remarkable, the trained computers made all those calls without any need for harsh chemical labels, including fluorescent dyes or stains, which researchers normally require to study cells. In other words, the computers learned to “see” the invisible!
New ‘Liquid Biopsy’ Shows Early Promise in Detecting Cancer
Posted on by Dr. Francis Collins

Caption: Liquid biopsy. Tumor cells shed protein and DNA into bloodstream for laboratory analysis and early cancer detection.
Early detection usually offers the best chance to beat cancer. Unfortunately, many tumors aren’t caught until they’ve grown relatively large and spread to other parts of the body. That’s why researchers have worked so tirelessly to develop new and more effective ways of screening for cancer as early as possible. One innovative approach, called “liquid biopsy,” screens for specific molecules that tumors release into the bloodstream.
Recently, an NIH-funded research team reported some encouraging results using a “universal” liquid biopsy called CancerSEEK [1]. By analyzing samples of a person’s blood for eight proteins and segments of 16 genes, CancerSEEK was able to detect most cases of eight different kinds of cancer, including some highly lethal forms—such as pancreatic, ovarian, and liver—that currently lack screening tests.
In a study of 1,005 people known to have one of eight early-stage tumor types, CancerSEEK detected the cancer in blood about 70 percent of the time, which is among the best performances to date for a blood test. Importantly, when CancerSEEK was performed on 812 healthy people without cancer, the test rarely delivered a false-positive result. The test can also be run relatively cheaply, at an estimated cost of less than $500.
Creative Minds: Using Machine Learning to Understand Genome Function
Posted on by Dr. Francis Collins
Science has always fascinated Anshul Kundaje, whether it was biology, physics, or chemistry. When he left his home country of India to pursue graduate studies in electrical engineering at Columbia University, New York, his plan was to focus on telecommunications and computer networks. But a course in computational genomics during his first semester showed him he could follow his interest in computing without giving up his love for biology.
Now an assistant professor of genetics and computer science at Stanford University, Palo Alto, CA, Kundaje has received a 2016 NIH Director’s New Innovator Award to explore not just how the human genome sequence encodes function, but also why it functions in the way that it does. Kundaje even envisions a time when it might be possible to use sophisticated computational approaches to predict the genomic basis of many human diseases.
Autism Spectrum Disorder: Progress Toward Earlier Diagnosis
Posted on by Dr. Francis Collins

Stockbyte
Research shows that the roots of autism spectrum disorder (ASD) generally start early—most likely in the womb. That’s one more reason, on top of a large number of epidemiological studies, why current claims about the role of vaccines in causing autism can’t be right. But how early is ASD detectable? It’s a critical question, since early intervention has been shown to help limit the effects of autism. The problem is there’s currently no reliable way to detect ASD until around 18–24 months, when the social deficits and repetitive behaviors associated with the condition begin to appear.
Several months ago, an NIH-funded team offered promising evidence that it may be possible to detect ASD in high-risk 1-year-olds by shifting attention from how kids act to how their brains have grown [1]. Now, new evidence from that same team suggests that neurological signs of ASD might be detectable even earlier.
Brain Scans Show Early Signs of Autism Spectrum Disorder
Posted on by Dr. Francis Collins

Source: Getty Images
For children with autism spectrum disorder (ASD), early diagnosis is critical to allow for possible interventions at a time when the brain is most amenable to change. But that’s been tough to implement for a simple reason: the symptoms of ASD, such as communication difficulties, social deficits, and repetitive behaviors, often do not show up until a child turns 2 or even 3 years old.
Now, an NIH-funded research team has news that may pave the way for earlier detection of ASD. The key is to shift the diagnostic focus from how kids act to how their brains grow. In their brain imaging study, the researchers found that, compared to other children, youngsters with ASD showed unusually rapid brain growth from infancy to age 2. In fact, the growth differences were already evident by their first birthdays, well before autistic behaviors typically emerge.
If I Only Had a Brain? Tissue Chips Predict Neurotoxicity
Posted on by Dr. Francis Collins

Caption: 3D neural tissue chips contain neurons (green), glial cells (red), and nuclei (blue). To take this confocal micrograph, developing neural tissue was removed from a chip and placed on a glass-bottom Petri dish.
Credit: Michael Schwartz, Dept. of Bioengineering, University of Wisconsin-Madison
A lot of time, money, and effort are devoted to developing new drugs. Yet only one of every 10 drug candidates entering human clinical trials successfully goes on to receive approval from the Food and Drug Administration (FDA) [1]. Many would-be drugs fall by the wayside because they prove toxic to the brain, liver, kidneys, or other organs—toxicity that, unfortunately, isn’t always detected in preclinical studies using mice, rats, or other animal models. That explains why scientists are working so hard to devise technologies that can do a better job of predicting early on which chemical compounds will be safe in humans.
As an important step in this direction, NIH-funded researchers at the Morgridge Institute for Research and University of Wisconsin-Madison have produced neural tissue chips with many features of a developing human brain. Each cultured 3D “organoid”—which sits comfortably in the bottom of a pea-sized well on a standard laboratory plate—comes complete with its very own neurons, support cells, blood vessels, and immune cells! As described in Proceedings of the National Academy of Sciences [2], this new tool is poised to predict earlier, faster, and less expensively which new or untested compounds—be they drug candidates or even ingredients in cosmetics and pesticides—might harm the brain, particularly at the earliest stages of development.
Previous Page