scholarly journals FFU-Net: Feature Fusion U-Net for Lesion Segmentation of Diabetic Retinopathy

2021 ◽  
Vol 2021 ◽  
pp. 1-12
Author(s):  
Yifei Xu ◽  
Zhuming Zhou ◽  
Xiao Li ◽  
Nuo Zhang ◽  
Meizi Zhang ◽  
...  

Diabetic retinopathy is one of the main causes of blindness in human eyes, and lesion segmentation is an important basic work for the diagnosis of diabetic retinopathy. Due to the small lesion areas scattered in fundus images, it is laborious to segment the lesion of diabetic retinopathy effectively with the existing U-Net model. In this paper, we proposed a new lesion segmentation model named FFU-Net (Feature Fusion U-Net) that enhances U-Net from the following points. Firstly, the pooling layer in the network is replaced with a convolutional layer to reduce spatial loss of the fundus image. Then, we integrate multiscale feature fusion (MSFF) block into the encoders which helps the network to learn multiscale features efficiently and enrich the information carried with skip connection and lower-resolution decoder by fusing contextual channel attention (CCA) models. Finally, in order to solve the problems of data imbalance and misclassification, we present a Balanced Focal Loss function. In the experiments on benchmark dataset IDRID, we make an ablation study to verify the effectiveness of each component and compare FFU-Net against several state-of-the-art models. In comparison with baseline U-Net, FFU-Net improves the segmentation performance by 11.97%, 10.68%, and 5.79% on metrics SEN, IOU, and DICE, respectively. The quantitative and qualitative results demonstrate the superiority of our FFU-Net in the task of lesion segmentation of diabetic retinopathy.

2020 ◽  
Vol 14 ◽  
Author(s):  
Charu Bhardwaj ◽  
Shruti Jain ◽  
Meenakshi Sood

: Diabetic Retinopathy is the leading cause of vision impairment and its early stage diagnosis relies on regular monitoring and timely treatment for anomalies exhibiting subtle distinction among different severity grades. The existing Diabetic Retinopathy (DR) detection approaches are subjective, laborious and time consuming which can only be carried out by skilled professionals. All the patents related to DR detection and diagnoses applicable for our research problem were revised by the authors. The major limitation in classification of severities lies in poor discrimination between actual lesions, background noise and other anatomical structures. A robust and computationally efficient Two-Tier DR (2TDR) grading system is proposed in this paper to categorize various DR severities (mild, moderate and severe) present in retinal fundus images. In the proposed 2TDR grading system, input fundus image is subjected to background segmentation and the foreground fundus image is used for anomaly identification followed by GLCM feature extraction forming an image feature set. The novelty of our model lies in the exhaustive statistical analysis of extracted feature set to obtain optimal reduced image feature set employed further for classification. Classification outcomes are obtained for both extracted as well as reduced feature set to validate the significance of statistical analysis in severity classification and grading. For single tier classification stage, the proposed system achieves an overall accuracy of 100% by k- Nearest Neighbour (kNN) and Artificial Neural Network (ANN) classifier. In second tier classification stage an overall accuracy of 95.3% with kNN and 98.0% with ANN is achieved for all stages utilizing optimal reduced feature set. 2TDR system demonstrates overall improvement in classification performance by 2% and 6% for kNN and ANN respectively after feature set reduction, and also outperforms the accuracy obtained by other state of the art methods when applied to the MESSIDOR dataset. This application oriented work aids in accurate DR classification for effective diagnosis and timely treatment of severe retinal ailment.


Sensors ◽  
2021 ◽  
Vol 21 (11) ◽  
pp. 3922
Author(s):  
Sheeba Lal ◽  
Saeed Ur Rehman ◽  
Jamal Hussain Shah ◽  
Talha Meraj ◽  
Hafiz Tayyab Rauf ◽  
...  

Due to the rapid growth in artificial intelligence (AI) and deep learning (DL) approaches, the security and robustness of the deployed algorithms need to be guaranteed. The security susceptibility of the DL algorithms to adversarial examples has been widely acknowledged. The artificially created examples will lead to different instances negatively identified by the DL models that are humanly considered benign. Practical application in actual physical scenarios with adversarial threats shows their features. Thus, adversarial attacks and defense, including machine learning and its reliability, have drawn growing interest and, in recent years, has been a hot topic of research. We introduce a framework that provides a defensive model against the adversarial speckle-noise attack, the adversarial training, and a feature fusion strategy, which preserves the classification with correct labelling. We evaluate and analyze the adversarial attacks and defenses on the retinal fundus images for the Diabetic Retinopathy recognition problem, which is considered a state-of-the-art endeavor. Results obtained on the retinal fundus images, which are prone to adversarial attacks, are 99% accurate and prove that the proposed defensive model is robust.


2020 ◽  
Vol 34 (08) ◽  
pp. 13267-13272
Author(s):  
Alex Foo ◽  
Wynne Hsu ◽  
Mong Li Lee ◽  
Gilbert Lim ◽  
Tien Yin Wong

Although deep learning for Diabetic Retinopathy (DR) screening has shown great success in achieving clinically acceptable accuracy for referable versus non-referable DR, there remains a need to provide more fine-grained grading of the DR severity level as well as automated segmentation of lesions (if any) in the retina images. We observe that the DR severity level of an image is dependent on the presence of different types of lesions and their prevalence. In this work, we adopt a multi-task learning approach to perform the DR grading and lesion segmentation tasks. In light of the lack of lesion segmentation mask ground-truths, we further propose a semi-supervised learning process to obtain the segmentation masks for the various datasets. Experiments results on publicly available datasets and a real world dataset obtained from population screening demonstrate the effectiveness of the multi-task solution over state-of-the-art networks.


2014 ◽  
Vol 33 (3) ◽  
pp. 231 ◽  
Author(s):  
Etienne Decencière ◽  
Xiwei Zhang ◽  
Guy Cazuguel ◽  
Bruno Lay ◽  
Béatrice Cochener ◽  
...  

The Messidor database, which contains hundreds of eye fundus images, has been publicly distributed since 2008. It was created by the Messidor project in order to evaluate automatic lesion segmentation and diabetic retinopathy grading methods. Designing, producing and maintaining such a database entails significant costs. By publicly sharing it, one hopes to bring a valuable resource to the public research community. However, the real interest and benefit of the research community is not easy to quantify. We analyse here the feedback on the Messidor database, after more than 6 years of diffusion. This analysis should apply to other similar research databases.


Author(s):  
Alfiya Md. Shaikh

Abstract: Diabetic retinopathy (DR) is a medical condition that damages eye retinal tissues. Diabetic retinopathy leads to mild to complete blindness. It has been a leading cause of global blindness. The identification and categorization of DR take place through the segmentation of parts of the fundus image or the examination of the fundus image for the incidence of exudates, lesions, microaneurysms, and so on. This research aims to study and summarize various recent proposed techniques applied to automate the process of classification of diabetic retinopathy. In the current study, the researchers focused on the concept of classifying the DR fundus images based on their severity level. Emphasis is on studying papers that proposed models developed using transfer learning. Thus, it becomes vital to develop an automatic diagnosis system to support physicians in their work.


2020 ◽  
Vol 64 (2) ◽  
pp. 20502-1-20502-10
Author(s):  
Duygu Çelik Ertuğrul ◽  
Yıltan Bitirim ◽  
Basmah Yakoub Anber

Abstract Diabetic Retinopathy (DR) is a medical condition, also known as diabetic eye disease, which is vision-threatening damage to the retina of the eye caused by diabetes. As the technology advances, researchers are becoming more interested in intelligent medical diagnosis systems to assist screening of DR in earlier stages. In this study, variety of state-of-the-art procedures are used to extract the anatomic segments and lesions from the color fundus images. In addition, an automated system is proposed for the detection of anatomic segments and lesions by grading approach to help clinical diagnosis of the DR analysis. Four publicly available databases of color fundus images and various appropriate measurement techniques are used to compare quantitatively the performance of the proposed system. The experiments conducted on DIARETDB0, DIARETDB1, STARE, and HRF data sets have proved that accuracy, sensitivity, and specificity of the proposed system are comparable or superior to state-of-the-art methods.


Algorithms ◽  
2019 ◽  
Vol 12 (1) ◽  
pp. 14 ◽  
Author(s):  
Imran Qureshi ◽  
Jun Ma ◽  
Kashif Shaheed

Diabetic retinopathy (DR) is a complication of diabetes and is known as visual impairment, and is diagnosed in various ethnicities of the working-age population worldwide. Fundus angiography is a widely applicable modality used by ophthalmologists and computerized applications to detect DR-based clinical features such as microaneurysms (MAs), hemorrhages (HEMs), and exudates (EXs) for early screening of DR. Fundus images are usually acquired using funduscopic cameras in varied light conditions and angles. Therefore, these images are prone to non-uniform illumination, poor contrast, transmission error, low brightness, and noise problems. This paper presents a novel and real-time mechanism of fundus image enhancement used for early grading of diabetic retinopathy, macular degeneration, retinal neoplasms, and choroid disruptions. The proposed system is based on two folds: (i) An RGB fundus image is initially taken and converted into a color appearance module (called lightness and denoted as J) of the CIECAM02 color space model to obtain image information in grayscale with bright light. Afterwards, in step (ii), the achieved J component is processed using a nonlinear contrast enhancement approach to improve the textural and color features of the fundus image without any further extraction steps. To test and evaluate the strength of the proposed technique, several performance and quality parameters—namely peak signal-to-noise ratio (PSNR), contrast-to-noise ratio (CNR), entropy (content information), histograms (intensity variation), and a structure similarity index measure (SSIM)—were applied to 1240 fundus images comprised of two publicly available datasets, DRIVE and MESSIDOR. It was determined from the experiments that the proposed enhancement procedure outperformed histogram-based approaches in terms of contrast, sharpness of fundus features, and brightness. This further revealed that it can be a suitable preprocessing tool for segmentation and classification of DR-related features algorithms.


2020 ◽  
Author(s):  
Alejandro Noriega ◽  
Dalia Camacho ◽  
Daniela Meizner ◽  
Jennifer Enciso ◽  
Hugo Quiroz-Mercado ◽  
...  

Background: The automated screening of patients at risk of developing diabetic retinopathy (DR), represents an opportunity to improve their mid-term outcome and lower the public expenditure associated with direct and indirect costs of a common sight-threatening complication of diabetes. Objective: In the present study, we aim at developing and evaluating the performance of an automated deep learning-based system to classify retinal fundus images from international and Mexican patients, as referable and non-referable DR cases. In particular, we study the performance of the automated retina image analysis (ARIA) system under an independent scheme (i.e. only ARIA screening) and two assistive schemes (i.e., hybrid ARIA + ophthalmologist screening), using a web-based platform for remote image analysis. Methods: We ran a randomized controlled experiment where 17 ophthalmologists were asked to classify a series of retinal fundus images under three different conditions: 1) screening the fundus image by themselves (solo), 2) screening the fundus image after being exposed to the opinion of the ARIA system (ARIA answer), and 3) screening the fundus image after being exposed to the opinion of the ARIA system, as well as its level of confidence and an attention map highlighting the most important areas of interest in the image according to the ARIA system (ARIA explanation). The ophthalmologists' opinion in each condition and the opinion of the ARIA system were compared against a gold standard generated by consulting and aggregating the opinion of three retina specialists for each fundus image. Results: The ARIA system was able to classify referable vs. non-referable cases with an area under the Receiver Operating Characteristic curve (AUROC), sensitivity, and specificity of 98%, 95.1% and 91.5% respectively, for international patient-cases; and an AUROC, sensitivity, and specificity of 98.3%, 95.2%, 90% respectively for Mexican patient-cases. The results achieved on Mexican patient-cases outperformed the average performance of the 17 ophthalmologist participants of the study. We also find that the ARIA system can be useful as an assistive tool, as significant specificity improvements were observed in the experimental condition where participants were exposed to the answer of the ARIA system as a second opinion (93.3%), compared to the specificity of the condition where participants assessed the images independently (87.3%). Conclusions: These results demonstrate that both use cases of ARIA systems, independent and assistive, present a substantial opportunity for Latin American countries like Mexico towards an efficient expansion of monitoring capacity for the early detection of diabetes-related blindness.


2020 ◽  
Vol 13 (4) ◽  
Author(s):  
Huma Kayani

The term artificial intelligence (AI) was proposed in 1956 by Dartmouth scholar John McCarthy, which refers to hardware or software that exhibits behavior which appears intelligent.1  During recent times, AI gained immense popularity as new algorithms, specialized hardware, huge data and cloud-based services were developed. Machine learning (ML), a subset of AI, originated in 1980 and is defined as a set of methods that automatically detect patterns in data and then incorporate this information to predict future data under uncertain conditions. Another escalating technology of ML called Deep learning (DL), launched in 2000s, is an escalating technology of ML and has revolutionized the world of AI. These technologies are powerful tools utilized by modern society for objects' recognition in images, real-time languages' translation, device manipulation via speech (such as Apple's Siri®, Amazon’s Alexa®, Microsoft’s Cortana®, etc.). The steps for AI model include preprocessing image data, train, validate and test the model, and evaluate the trained model's performance. To increase AI prediction efficiency, raw data need to be preprocessed. Data collected from different sources needs to be integrated and the most relevant features selected and extracted to improve the learning process performance. Data set is randomly partitioned into two independent subsets, one is for modeling and the other is for testing. The test set is used to evaluate the final performance of the trained model. The area under receiver operating characteristic curves (AUC) is most used evaluation metrics for quantitative assessment of a model in AI diagnosis. The AUCs effective models range from 0.5 to 1; higher the value of AUC, better the performance of the model.2 In the medical field, AI gained popularity by visualization of input images of highly potential abnormal sites which can be reviewed and analyzed in future.           AI and DL algorithms or systems are also widely used in field of ophthalmology. More intensively studied fields are diabetic retinopathy, age related macular degeneration, and cataract and glaucoma. Various ophthalmic imaging modalities used for AI diagnosis include fundus image, optical coherence tomography (OCT), ocular ultrasound, slit-lamp image and visual field. Diabetic retinopathy (DR), a diabetic complication, is a vasculopathy that affects one-third of diabetic patients leading to irreversible blindness. AI has been in use to predict DR risk and its progression. Gulshan and colleague were the first to report the application of DL for DR identification.3 They used large fundus image data sets in supervised manner for DR detection. Other studies applied DL to identify and stage DR. DL-based computer-aided system was introduced to detect DR through OCT images, achieving a specificity of 0.98.4 A computer-aided diagnostic (CAD) system based on CML algorithms using optical coherence tomography angiography images to automatically diagnose non-proliferative DR (NPDR) also achieved high accuracy and AUC.5 Age-related macular degeneration (AMD) is the leading cause of irreversible blindness among old people in the developed world. ML algorithms are being used to identify AMD lesions and prompt early treatment with accuracy usually over 80%.6 Using ML to predict treatment of retinal neovascularity in AMD and DR by anti-vascular endothelial growth factor (Anti VEGF) injection requirements can manage patients' economic burden and resource management. ML algorithms have been applied to diagnose and grade cataract using fundus images, ultrasounds images, and visible wavelength eye images.7 Glaucoma is the third largest sight-threatening eye disease around the world. Glaucoma patients suffered from high intraocular pressure, damage of the optic nerve head, retina nerve fiber layer defect, and gradual vision loss. Studies using DL methods to diagnose glaucoma are few. So far, fundus images and wide-field OCT scans have all been used to construct DL-based glaucomatous diagnostic models. Mostly, the DL-based methods show excellent results.8           In this era of “evidence-based medicine,” clinicians and patients find it difficult to trust a mysterious machine to diagnose yet cannot provide explanations of why the patient has certain disease. In future, advanced AI interpreters will be launched which will contribute significantly to revolutionize current disease diagnostic pattern.


Author(s):  
Yanfei Guo ◽  
Yanjun Peng

AbstractDiabetic retinopathy is the leading cause of blindness in working population. Lesion segmentation from fundus images helps ophthalmologists accurately diagnose and grade of diabetic retinopathy. However, the task of lesion segmentation is full of challenges due to the complex structure, the various sizes and the interclass similarity with other fundus tissues. To address the issue, this paper proposes a cascade attentive RefineNet (CARNet) for automatic and accurate multi-lesion segmentation of diabetic retinopathy. It can make full use of the fine local details and coarse global information from the fundus image. CARNet is composed of global image encoder, local image encoder and attention refinement decoder. We take the whole image and the patch image as the dual input, and feed them to ResNet50 and ResNet101, respectively, for downsampling to extract lesion features. The high-level refinement decoder uses dual attention mechanism to integrate the same-level features in the two encoders with the output of the low-level attention refinement module for multiscale information fusion, which focus the model on the lesion area to generate accurate predictions. We evaluated the segmentation performance of the proposed CARNet on the IDRiD, E-ophtha and DDR data sets. Extensive comparison experiments and ablation studies on various data sets demonstrate the proposed framework outperforms the state-of-the-art approaches and has better accuracy and robustness. It not only overcomes the interference of similar tissues and noises to achieve accurate multi-lesion segmentation, but also preserves the contour details and shape features of small lesions without overloading GPU memory usage.


Sign in / Sign up

Export Citation Format

Share Document