Construction of high-performance systems with neural networks for detection of abnormal areas from chest X-ray images

Author(s):  
Takahiro Sasaki ◽  
Kentaro Kinoshita ◽  
Satoru Kishida ◽  
Yoshiharu Hirata ◽  
Seigo Yamada
Author(s):  
Saleh Albahli ◽  
Waleed Albattah

Objective: Automatic prediction of COVID-19 using deep convolution neural networks based pre-trained transfer models and Chest X-ray images. Method: This research employs the advantages of computer vision and medical image analysis to develop an automated model that has the clinical potential for early detection of the disease. Using Deep Learning models, the research aims at evaluating the effectiveness and accuracy of different convolutional neural networks models in the automatic diagnosis of COVID-19 from X-ray images as compared to diagnosis performed by experts in the medical community. Result: Due to the fact that the dataset available for COVID-19 is still limited, the best model to use is the InceptionNetV3. Performance results show that the InceptionNetV3 model yielded the highest accuracy of 98.63% (with data augmentation) and 98.90% (without data augmentation) among the three models designed. However, as the dataset gets bigger, the Inception ResNetV2 and NASNetlarge will do a better job of classification. All the performed networks tend to over-fit when data augmentation is not used, this is due to the small amount of data used for training and validation. Conclusion: A deep transfer learning is proposed to detecting the COVID-19 automatically from chest X-ray by training it with X-ray images gotten from both COVID-19 patients and people with normal chest Xrays. The study is aimed at helping doctors in making decisions in their clinical practice due its high performance and effectiveness, the study also gives an insight to how transfer learning was used to automatically detect the COVID-19.


2020 ◽  
Vol 112 (5) ◽  
pp. S50
Author(s):  
Zachary Eller ◽  
Michelle Chen ◽  
Jermaine Heath ◽  
Uzma Hussain ◽  
Thomas Obisean ◽  
...  

Author(s):  
Aleksei Aleksandrovich Rumyantsev ◽  
Farkhad Mansurovich Bikmuratov ◽  
Nikolai Pavlovich Pashin

The subject of this research is medical chest X-ray images. After fundamental pre-processing, the accumulated database of such images can be used for training deep convolutional neural networks that have become one of the most significant innovations in recent years. The trained network carries out preliminary binary classification of the incoming images and serve as an assistant to the radiotherapist. For this purpose, it is necessary to train the neural network to carefully minimize type I and type II errors. Possible approach towards improving the effectiveness of application of neural networks, by the criteria of reducing computational complexity and quality of image classification, is the auxiliary approaches: image pre-processing and preliminary calculation of entropy of the fragments. The article provides the algorithm for X-ray image pre-processing, its fragmentation, and calculation of the entropy of separate fragments. In the course of pre-processing, the region of lungs and spine is selected, which comprises approximately 30-40% of the entire image. Then the image is divided into the matrix of fragments, calculating the entropy of separate fragments in accordance with Shannon’s formula based pm the analysis of individual pixels. Determination of the rate of occurrence of each of the 255 colors allows calculating the total entropy. The use of entropy for detecting pathologies is based on the assumption that its values differ for separate fragments and overall picture of its distribution between the images with the norm and pathologies. The article analyzes the statistical values: standard deviation of error, dispersion. A fully connected neural network is used for determining the patterns in distribution of entropy and its statistical characteristics on various fragments of the chest X-ray image.


2019 ◽  
Vol 38 (5) ◽  
pp. 1197-1206 ◽  
Author(s):  
Hojjat Salehinejad ◽  
Errol Colak ◽  
Tim Dowdell ◽  
Joseph Barfett ◽  
Shahrokh Valaee

Proceedings ◽  
2020 ◽  
Vol 54 (1) ◽  
pp. 31
Author(s):  
Joaquim de Moura ◽  
Lucía Ramos ◽  
Plácido L. Vidal ◽  
Jorge Novo ◽  
Marcos Ortega

The new coronavirus (COVID-19) is a disease that is caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2). On 11 March 2020, the coronavirus outbreak has been labelled a global pandemic by the World Health Organization. In this context, chest X-ray imaging has become a remarkably powerful tool for the identification of patients with COVID-19 infections at an early stage when clinical symptoms may be unspecific or sparse. In this work, we propose a complete analysis of separability of COVID-19 and pneumonia in chest X-ray images by means of Convolutional Neural Networks. Satisfactory results were obtained that demonstrated the suitability of the proposed system, improving the efficiency of the medical screening process in the healthcare systems.


2021 ◽  
Author(s):  
Sandi Baressi Šegota ◽  
◽  
Simon Lysdahlgaard ◽  
Søren Hess ◽  
Ronald Antulov

The fact that Artificial Intelligence (AI) based algorithms exhibit a high performance on image classification tasks has been shown many times. Still, certain issues exist with the application of machine learning (ML) artificial neural network (ANN) algorithms. The best known is the need for a large amount of statistically varied data, which can be addressed with expanded collection or data augmentation. Other issues are also present. Convolutional neural networks (CNNs) show extremely high performance on image-shaped data. Despite their performance, CNNs exhibit a large issue which is the sensitivity to image orientation. Previous research shows that varying the orientation of images may greatly lower the performance of the trained CNN. This is especially problematic in certain applications, such as X-ray radiography, an example of which is presented here. Previous research shows that the performance of CNNs is higher when used on images in a single orientation (left or right), as opposed to the combination of both. This means that the data needs to be differentiated before it enters the classification model. In this paper, the CNN-based model for differentiation between left and right-oriented images is presented. Multiple CNNs are trained and tested, with the highest performing being the VGG16 architecture which achieved an Accuracy of 0.99 (+/- 0.01), and an AUC of 0.98 (+/- 0.01). These results show that CNNs can be used to address the issue of orientation sensitivity by splitting the data in advance of being used in classification models.


Author(s):  
René Hosch ◽  
Lennard Kroll ◽  
Felix Nensa ◽  
Sven Koitka

Purpose Detection and validation of the chest X-ray view position with use of convolutional neural networks to improve meta-information for data cleaning within a hospital data infrastructure. Material and Methods Within this paper we developed a convolutional neural network which automatically detects the anteroposterior and posteroanterior view position of a chest radiograph. We trained two different network architectures (VGG variant and ResNet-34) with data published by the RSNA (26 684 radiographs, class distribution 46 % AP, 54 % PA) and validated these on a self-compiled dataset with data from the University Hospital Essen (4507, radiographs, class distribution 55 % PA, 45 % AP) labeled by a human reader. For visualization and better understanding of the network predictions, a Grad-CAM was generated for each network decision. The network results were evaluated based on the accuracy, the area under the curve (AUC), and the F1-score against the human reader labels. Also a final performance comparison between model predictions and DICOM labels was performed. Results The ensemble models reached accuracy and F1-scores greater than 95 %. The AUC reaches more than 0.99 for the ensemble models. The Grad-CAMs provide insight as to which anatomical structures contributed to a decision by the networks which are comparable with the ones a radiologist would use. Furthermore, the trained models were able to generalize over mislabeled examples, which was found by comparing the human reader labels to the predicted labels as well as the DICOM labels. Conclusion The results show that certain incorrectly entered meta-information of radiological images can be effectively corrected by deep learning in order to increase data quality in clinical application as well as in research. Key Points:  Citation Format


Sign in / Sign up

Export Citation Format

Share Document