Discovering correspondences between molecular profiles and morphological features via deep learning (Conference Presentation)

Author(s):  
Richard Chen ◽  
Faisal Mahmood
2020 ◽  
Vol 6 (10) ◽  
pp. 101
Author(s):  
Mauricio Alberto Ortega-Ruiz ◽  
Cefa Karabağ ◽  
Victor García Garduño ◽  
Constantino Carlos Reyes-Aldasoro

This paper describes a methodology that extracts key morphological features from histological breast cancer images in order to automatically assess Tumour Cellularity (TC) in Neo-Adjuvant treatment (NAT) patients. The response to NAT gives information on therapy efficacy and it is measured by the residual cancer burden index, which is composed of two metrics: TC and the assessment of lymph nodes. The data consist of whole slide images (WSIs) of breast tissue stained with Hematoxylin and Eosin (H&E) released in the 2019 SPIE Breast Challenge. The methodology proposed is based on traditional computer vision methods (K-means, watershed segmentation, Otsu’s binarisation, and morphological operations), implementing colour separation, segmentation, and feature extraction. Correlation between morphological features and the residual TC after a NAT treatment was examined. Linear regression and statistical methods were used and twenty-two key morphological parameters from the nuclei, epithelial region, and the full image were extracted. Subsequently, an automated TC assessment that was based on Machine Learning (ML) algorithms was implemented and trained with only selected key parameters. The methodology was validated with the score assigned by two pathologists through the intra-class correlation coefficient (ICC). The selection of key morphological parameters improved the results reported over other ML methodologies and it was very close to deep learning methodologies. These results are encouraging, as a traditionally-trained ML algorithm can be useful when limited training data are available preventing the use of deep learning approaches.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Dmitrii Bychkov ◽  
Nina Linder ◽  
Aleksei Tiulpin ◽  
Hakan Kücükel ◽  
Mikael Lundin ◽  
...  

AbstractThe treatment of patients with ERBB2 (HER2)-positive breast cancer with anti-ERBB2 therapy is based on the detection of ERBB2 gene amplification or protein overexpression. Machine learning (ML) algorithms can predict the amplification of ERBB2 based on tumor morphological features, but it is not known whether ML-derived features can predict survival and efficacy of anti-ERBB2 treatment. In this study, we trained a deep learning model with digital images of hematoxylin–eosin (H&E)-stained formalin-fixed primary breast tumor tissue sections, weakly supervised by ERBB2 gene amplification status. The gene amplification was determined by chromogenic in situ hybridization (CISH). The training data comprised digitized tissue microarray (TMA) samples from 1,047 patients. The correlation between the deep learning–predicted ERBB2 status, which we call H&E-ERBB2 score, and distant disease-free survival (DDFS) was investigated on a fully independent test set, which included whole-slide tumor images from 712 patients with trastuzumab treatment status available. The area under the receiver operating characteristic curve (AUC) in predicting gene amplification in the test sets was 0.70 (95% CI, 0.63–0.77) on 354 TMA samples and 0.67 (95% CI, 0.62–0.71) on 712 whole-slide images. Among patients with ERBB2-positive cancer treated with trastuzumab, those with a higher than the median morphology–based H&E-ERBB2 score derived from machine learning had more favorable DDFS than those with a lower score (hazard ratio [HR] 0.37; 95% CI, 0.15–0.93; P = 0.034). A high H&E-ERBB2 score was associated with unfavorable survival in patients with ERBB2-negative cancer as determined by CISH. ERBB2-associated morphology correlated with the efficacy of adjuvant anti-ERBB2 treatment and can contribute to treatment-predictive information in breast cancer.


2021 ◽  
Vol 36 (Supplement_1) ◽  
Author(s):  
A Le ◽  
I Miyatsuka ◽  
J Otsuki ◽  
M Shiotani ◽  
N Enatsu ◽  
...  

Abstract Study question Can deep learning (DL) algorithms trained on time-lapse videos be used to detect and track the size and gender of pronuclei in developing human zygotes? Summary answer Our DL algorithm not only outperforms state-of-the-art models in detecting the pronuclei but can also accurately identify and track its gender and size over time. What is known already Recent researches have explored the use of DL to extract key morphological features of human embryos. Existing studies, however, focus either on blastocysts’ morphological measurements (Au et al. 2020) or on embryos’ general developmental stages classification (Gingold et al. 2018, Liu et al. 2019, Lau et al. 2019). So far, only one paper attempted to evaluate zygotes’ morphological components but stopped short of identifying the existence and location of their pronuclei (Leahy et al. 2020). We address this research gap by training a DL model that can detect, classify the gender, and quantify the size of zygotes’ pronuclei over time. Study design, size, duration A retrospective analysis using 91 fertilized oocytes from infertile patients undergoing IVF or ICSI treatment at Hanabusa Women’s Clinic between January 2011 and August 2019 was conducted. Each embryo was time-lapse monitored using Vitrolife which records an image every 15 minutes at 7 focal planes. For our study, we used videos of the first 1–2 days of the embryo from its 3 central focal planes, corresponding to 70–150 images per focal plane. Participants/materials, setting, methods All 273 timelapse videos were split into 30,387 grayscale still images at a 15-minute interval. Each image was checked and annotated by experienced embryologists where every pixel of the image was classified into 3 categories: male pronuclei, female pronuclei, and others. Images were converted into grayscale, resized into 500x500 pixels, and then fed into a neural network with the Mask R-CNN architecture and a ResNet101 backbone to produce a pronuclei instance segmentation model. Main results and the role of chance The 91 embryos were split into training (∼70% or 63 embryos) and validation (∼30% or 28 embryos). Our pronuclei model takes as input a single image and outputs a bounding box, mask, category, confidence score, and size measured in terms of pixel for each detected candidate. For prediction, we run the model on the 3 middle focal planes and merge candidates by using the one with the highest confidence score. We used the mean-average precision (mAP) score to evaluate our model’s ability to detect pronuclei and used the mean absolute percentage error (MAPE) between the actual size (as annotated by the embryologist) and the predicted one to check the model’s performance in tracking the pronuclei’s size. The mAP for detecting pronuclei, regardless of its gender, achieved by our model was 0.698, higher than the 0.680 value reported in the Leahy et al. paper (2020). Breakdown by gender, our model’s mAP for male and female pronuclei are 0.734 and 0.661 respectively. The overall MAPE for tracking pronuclei’s size is 21.8%. Breakdown by gender, our model’s MAPE for male and female pronuclei are 19.4% and 24.3% respectively. Limitations, reasons for caution Samples were collected from one clinic with videos recorded from one time-lapse system which can limit our results’ reproducibility. The accuracy of our DL model is also limited by the small number of embryos that we used. Wider implications of the findings: Even with a limited training dataset, our results indicate that we can accurately detect and track the gender and the size of zygotes’ pronuclei using time-lapse videos. In future models, we will increase our training dataset as well as include other time-lapse systems to improve our models’ accuracy and reproducibility. Trial registration number Not applicable


Blood ◽  
2020 ◽  
Vol 136 (Supplement 1) ◽  
pp. 10-12
Author(s):  
John-William Sidhom ◽  
Ingharan J Siddarthan ◽  
Bo-Shiun Lai ◽  
Adam Luo ◽  
Bryan Hambley ◽  
...  

Acute Promyelocytic Leukemia (APL) is a subtype of Acute Myeloid Leukemia (AML), classified by a translocation between chromosomes 15 and 17 [t(15;17)], that is notably distinguished clinically by a rapidly progressive and fatal course. Due to the acute nature of its presentation, prompt and accurate diagnosis is required to initiate appropriate therapy that can be curative. However, the gold standard genetic tests can take days to confirm a diagnosis and thus therapy is often initiated on high clinical suspicion based on both clinical presentation as well as direct visualization of the peripheral smear. While there are described cellular morphological features that distinguish APL, there is still considerable difficulty in diagnosing APL from direct visualization of a peripheral smear by a hematopathologist. We hypothesized that deep learning pattern recognition would have greater discriminatory power and consistency compared to humans to distinguish t(15;17) translocation positive APL from t(15;17) translocation negative AML. To best tackle the problem of diagnosing APL rapidly from a peripheral smear, study patients with APL and AML were identified via retrospective chart review from a list of confirmed FISH t(15;17)-positive (n = 34) and -negative (n = 72) patients presenting at The Johns Hopkins Hospital (JHH). Additional inclusion criteria included new disease diagnosis, no prior treatment, and availability of peripheral blood smear image uploaded to CellaVision. Patients were separated into a discovery cohort presenting prior to 1/2019 (APL, n = 22; AML, n=60) and a validation cohort presenting on or after 1/2019 (APL, n = 12; AML, n = 12). A multiple-instance deep learning model employing convolutional layers at the per-cell level (Figure 1A) was trained on the discovery cohort and then tested on the independent prospective validation cohort to assess generalizability of the model. When compared to 10 academic clinicians (denoted with red +) who consisted of leukemia-treating hematologists, oncologists, and hematopathologists, the deep learning model was equivalent or outperformed 9/10 readers (Figure 1B) with an AUC of 0.861. We further looked at the performance of using proportion of promyelocytes (per CellaVision classification) as a biomarker of APL which had an AUC of 0.611. Finally, we applied integrated gradients, a method by which to extract per-pixel importance to the classification probability to identify and understand the morphological features the model was learning and using to distinguish APL (Figure 1C). We noted that the appearance of the chromatin in the non-APL leukemias was more dispersed and focused at the edge of the cell whereas in APL, the chromatin was more condensed and focused at the center of the cell. These morphological features, taught to us by the model, have not been previously reported in the literature as being useful for distinguishing APL from non-APL. Our work presents a deep learning model capable of rapid and accurate diagnosis of APL from universally available peripheral smears. In addition, explainable artificial intelligence is provided for biological insights to facilitate clinical management and reveal morphological concepts previously unappreciated in APL. The deep learning framework we have delineated is applicable to any diagnostic pipeline that can leverage a peripheral blood smear, potentially allowing for efficient diagnosis and early treatment of disease. Figure 1. Disclosures Streiff: Bayer: Consultancy, Speakers Bureau; Dispersol: Consultancy; BristolMyersSquibb: Consultancy; Janssen: Consultancy, Research Funding; Pfizer: Consultancy, Speakers Bureau; Portola: Consultancy; Boehringer-Ingelheim: Research Funding; NHLBI: Research Funding; PCORI: Research Funding; NovoNordisk: Research Funding; Sanofi: Research Funding. Moliterno:Pharmessentia: Consultancy; MPNRF: Research Funding. DeZern:MEI: Consultancy; Abbvie: Consultancy; Astex: Research Funding; Celgene: Consultancy, Honoraria. Levis:Astellas: Honoraria, Research Funding; Menarini: Honoraria; Amgen: Honoraria; FujiFilm: Honoraria, Research Funding; Daiichi-Sankyo: Honoraria.


2020 ◽  
Vol 19 (1) ◽  
Author(s):  
Lal Hussain ◽  
Tony Nguyen ◽  
Haifang Li ◽  
Adeel A. Abbasi ◽  
Kashif J. Lone ◽  
...  

Abstract Background The large volume and suboptimal image quality of portable chest X-rays (CXRs) as a result of the COVID-19 pandemic could post significant challenges for radiologists and frontline physicians. Deep-learning artificial intelligent (AI) methods have the potential to help improve diagnostic efficiency and accuracy for reading portable CXRs. Purpose The study aimed at developing an AI imaging analysis tool to classify COVID-19 lung infection based on portable CXRs. Materials and methods Public datasets of COVID-19 (N = 130), bacterial pneumonia (N = 145), non-COVID-19 viral pneumonia (N = 145), and normal (N = 138) CXRs were analyzed. Texture and morphological features were extracted. Five supervised machine-learning AI algorithms were used to classify COVID-19 from other conditions. Two-class and multi-class classification were performed. Statistical analysis was done using unpaired two-tailed t tests with unequal variance between groups. Performance of classification models used the receiver-operating characteristic (ROC) curve analysis. Results For the two-class classification, the accuracy, sensitivity and specificity were, respectively, 100%, 100%, and 100% for COVID-19 vs normal; 96.34%, 95.35% and 97.44% for COVID-19 vs bacterial pneumonia; and 97.56%, 97.44% and 97.67% for COVID-19 vs non-COVID-19 viral pneumonia. For the multi-class classification, the combined accuracy and AUC were 79.52% and 0.87, respectively. Conclusion AI classification of texture and morphological features of portable CXRs accurately distinguishes COVID-19 lung infection in patients in multi-class datasets. Deep-learning methods have the potential to improve diagnostic efficiency and accuracy for portable CXRs.


2019 ◽  
Vol 517 ◽  
pp. 128-137 ◽  
Author(s):  
Cuong Ly ◽  
Adam M. Olsen ◽  
Ian J. Schwerdt ◽  
Reid Porter ◽  
Kari Sentz ◽  
...  

2020 ◽  
Author(s):  
Sahil S. Nalawade ◽  
Fang F. Yu ◽  
Chandan Ganesh Bangalore Yogananda ◽  
Gowtham K. Murugesan ◽  
Bhavya R. Shah ◽  
...  

AbstractDeep learning has shown promise for predicting glioma molecular profiles using MR images. Before clinical implementation, ensuring robustness to real-world problems, such as patient motion, is crucial. We sought to evaluate the effects of motion artifact on glioma marker classifier performance and develop a deep learning motion correction network to restore classification accuracies. T2w images and molecular information were retrieved from the TCIA and TCGA databases. Three-fold cross-validation was used to train and test the motion correction network on artifact-corrupted images. We then compared the performance of three glioma marker classifiers (IDH mutation, 1p/19q codeletion, and MGMT methylation) using motion-corrupted and motion-corrected images. Glioma marker classifier performance decreased markedly with increasing motion corruption. Applying motion correction effectively restored classification accuracy for even the most motion-corrupted images. For IDH classification, an accuracy of 99% was achieved, representing a new benchmark in non-invasive image-based IDH classification and exceeding the original performance of the network. Robust motion correction can enable high accuracy in deep learning MRI-based molecular marker classification rivaling tissue-based characterization.STATEMENT OF SIGNIFICANCEDeep learning networks have shown promise for predicting molecular profiles of gliomas using MR images. We demonstrate that patient motion artifact, which is frequently encountered in the clinic, can significantly impair the performance of these algorithms. The application of robust motion correction algorithms can restore the performance of these networks, rivaling tissue-based characterization.


2021 ◽  
Vol 13 (12) ◽  
pp. 6859
Author(s):  
Chenyi Cai ◽  
Zifeng Guo ◽  
Baizhou Zhang ◽  
Xiao Wang ◽  
Biao Li ◽  
...  

The study of urban morphology contributes to the evolution of cities and sustainable development. Urban morphological feature extraction and similarity analysis represents a practical framework in many studies to interpret and introduce the current built environment to aid in proposing novel designs. In conventional methods, morphological features are represented based on qualitative descriptions, symbolical interpretation, or manually selected indicators. However, these methods could cause subjective bias and limit the generalizability. This study proposes a hybrid data-driven approach to support quantitative morphological descriptions and multi-dimensional similarity analysis for urban design decision-making and to further morphology-related studies using information abundance via a deep-learning approach. We constructed a dataset of 3817 residential plots with geometrical and related infrastructure information. A deep convolutional neural network, GoogLeNet, was implemented with the plots’ figure–ground images, by quantifying the morphological features into 2048-dimensional feature vectors. We conducted a similarity analysis of the plots by calculating the Euclidean distance between the high-dimensional feature vectors. Then, a comparison study was performed by retrieving cases based on the plot shape and plots with buildings separately. The proposed method considers the overall characteristics of the urban morphology and social infrastructure situations for similarity analysis. This method is flexible and effective. The proposed framework indicates the feasibility and potential of integrating task-oriented information to introduce custom and adequate references via deep learning methods, which could support decision making and association studies on morphology with urban consequences. This work could serve as a basis for further typo-morphology studies and other morphology-related ecological, social, and economic studies for sustainable built environments.


Sign in / Sign up

Export Citation Format

Share Document