AI model identifies certain breast tumor stages more likely to progress to invasive cancer

Ductal carcinoma in situ (DCIS) is a form of preinvasive tumor that sometimes progresses to a highly deadly type of breast cancer. It accounts for about 25 percent of all breast cancer diagnoses.

Since it is difficult for clinicians to find out the sort and stage of DCIS, patients with DCIS are sometimes overtreated. To deal with this, an interdisciplinary team of researchers from MIT and ETH Zurich developed an AI model that may discover the several stages of DCIS from an inexpensive and easy-to-obtain breast tissue image. Their model shows that each the state and arrangement of cells in a tissue sample are necessary for determining the stage of DCIS.

Because such tissue images are really easy to acquire, the researchers were capable of construct certainly one of the most important datasets of its kind, which they used to coach and test their model. After they compared its predictions to conclusions of a pathologist, they found clear agreement in lots of instances.

In the longer term, the model may very well be used as a tool to assist clinicians streamline the diagnosis of simpler cases without the necessity for labor-intensive tests, giving them more time to guage cases where it’s less clear if DCIS will turn into invasive.

“We took step one in understanding that we must be the spatial organization of cells when diagnosing DCIS, and now we have now developed a method that’s scalable. From here, we actually need a prospective study. Working with a hospital and getting this all of the option to the clinic shall be a vital step forward,” says Caroline Uhler, a professor within the Department of Electrical Engineering and Computer Science (EECS) and the Institute for Data, Systems, and Society (IDSS), who can be director of the Eric and Wendy Schmidt Center on the Broad Institute of MIT and Harvard and a researcher at MIT’s Laboratory for Information and Decision Systems (LIDS).

Uhler, co-corresponding creator of a paper on this research, is joined by lead creator Xinyi Zhang, a graduate student in EECS and the Eric and Wendy Schmidt Center; co-corresponding creator GV Shivashankar, professor of mechogenomics at ETH Zurich jointly with the Paul Scherrer Institute; and others at MIT, ETH Zurich, and the University of Palermo in Italy. The open-access research was published July 20 in Nature Communications.

Combining imaging with AI   

Between 30 and 50 percent of patients with DCIS develop a highly invasive stage of cancer, but researchers don’t know the biomarkers that would tell a clinician which tumors will progress.

Researchers can use techniques like multiplexed staining or single-cell RNA sequencing to find out the stage of DCIS in tissue samples. Nevertheless, these tests are too expensive to be performed widely, Shivashankar explains.

In previous work, these researchers showed that an inexpensive imagining technique often known as chromatin staining may very well be as informative because the much costlier single-cell RNA sequencing.

For this research, they hypothesized that combining this single stain with a fastidiously designed machine-learning model could provide the identical details about cancer stage as costlier techniques.

First, they created a dataset containing 560 tissue sample images from 122 patients at three different stages of disease. They used this dataset to coach an AI model that learns a representation of the state of every cell in a tissue sample image, which it uses to infer the stage of a patient’s cancer.

Nevertheless, not every cell is indicative of cancer, so the researchers needed to aggregate them in a meaningful way.

They designed the model to create clusters of cells in similar states, identifying eight states which are necessary markers of DCIS. Some cell states are more indicative of invasive cancer than others. The model determines the proportion of cells in each state in a tissue sample.

Organization matters

“But in cancer, the organization of cells also changes. We found that just having the proportions of cells in every state just isn’t enough. You furthermore may need to know how the cells are organized,” says Shivashankar.

With this insight, they designed the model to think about proportion and arrangement of cell states, which significantly boosted its accuracy.

“The interesting thing for us was seeing how much spatial organization matters. Previous studies had shown that cells that are near the breast duct are necessary. But it is usually necessary to think about which cells are near which other cells,” says Zhang.

After they compared the outcomes of their model with samples evaluated by a pathologist, it had clear agreement in lots of instances. In cases that weren’t as clear-cut, the model could provide details about features in a tissue sample, just like the organization of cells, that a pathologist could use in decision-making.

This versatile model may be adapted to be used in other kinds of cancer, and even neurodegenerative conditions, which is one area the researchers are also currently exploring.

“We now have shown that, with the appropriate AI techniques, this straightforward stain will be very powerful. There remains to be far more research to do, but we’d like to take the organization of cells into consideration in additional of our studies,” Uhler says.

This research was funded, partly, by the Eric and Wendy Schmidt Center on the Broad Institute, ETH Zurich, the Paul Scherrer Institute, the Swiss National Science Foundation, the U.S. National Institutes of Health, the U.S. Office of Naval Research, the MIT Jameel Clinic for Machine Learning and Health, the MIT-IBM Watson AI Lab, and a Simons Investigator Award.