Microscopy image segmentation tool: Robust image data analysis

2014 ◽  
Vol 85 (3) ◽  
pp. 033701 ◽  
Author(s):  
Ilya Valmianski ◽  
Carlos Monton ◽  
Ivan K. Schuller
2019 ◽  
Vol 8 (4) ◽  
pp. 9548-9551

Fuzzy c-means clustering is a popular image segmentation technique, in which a single pixel belongs to multiple clusters, with varying degree of membership. The main drawback of this method is it sensitive to noise. This method can be improved by incorporating multiresolution stationary wavelet analysis. In this paper we develop a robust image segmentation method using Fuzzy c-means clustering and wavelet transform. The experimental result shows that the proposed method is more accurate than the Fuzzy c-means clustering.


2012 ◽  
Vol 6 (4) ◽  
pp. 253-276 ◽  
Author(s):  
Daniel Baier ◽  
Ines Daniel ◽  
Sarah Frost ◽  
Robert Naundorf
Keyword(s):  

2021 ◽  
pp. 114811
Author(s):  
Aditi Joshi ◽  
Mohammed Saquib Khan ◽  
Asim Niaz ◽  
Farhan Akram ◽  
Hyun Chul Song ◽  
...  

2019 ◽  
Vol 20 (1) ◽  
Author(s):  
Fuyong Xing ◽  
Yuanpu Xie ◽  
Xiaoshuang Shi ◽  
Pingjun Chen ◽  
Zizhao Zhang ◽  
...  

Abstract Background Nucleus or cell detection is a fundamental task in microscopy image analysis and supports many other quantitative studies such as object counting, segmentation, tracking, etc. Deep neural networks are emerging as a powerful tool for biomedical image computing; in particular, convolutional neural networks have been widely applied to nucleus/cell detection in microscopy images. However, almost all models are tailored for specific datasets and their applicability to other microscopy image data remains unknown. Some existing studies casually learn and evaluate deep neural networks on multiple microscopy datasets, but there are still several critical, open questions to be addressed. Results We analyze the applicability of deep models specifically for nucleus detection across a wide variety of microscopy image data. More specifically, we present a fully convolutional network-based regression model and extensively evaluate it on large-scale digital pathology and microscopy image datasets, which consist of 23 organs (or cancer diseases) and come from multiple institutions. We demonstrate that for a specific target dataset, training with images from the same types of organs might be usually necessary for nucleus detection. Although the images can be visually similar due to the same staining technique and imaging protocol, deep models learned with images from different organs might not deliver desirable results and would require model fine-tuning to be on a par with those trained with target data. We also observe that training with a mixture of target and other/non-target data does not always mean a higher accuracy of nucleus detection, and it might require proper data manipulation during model training to achieve good performance. Conclusions We conduct a systematic case study on deep models for nucleus detection in a wide variety of microscopy images, aiming to address several important but previously understudied questions. We present and extensively evaluate an end-to-end, pixel-to-pixel fully convolutional regression network and report a few significant findings, some of which might have not been reported in previous studies. The model performance analysis and observations would be helpful to nucleus detection in microscopy images.


Plant Methods ◽  
2017 ◽  
Vol 13 (1) ◽  
Author(s):  
Xiong Xiong ◽  
Lingfeng Duan ◽  
Lingbo Liu ◽  
Haifu Tu ◽  
Peng Yang ◽  
...  

GigaScience ◽  
2017 ◽  
Vol 6 (11) ◽  
Author(s):  
Fernando Perez-Sanz ◽  
Pedro J Navarro ◽  
Marcos Egea-Cortines

Author(s):  
Hao Zheng ◽  
Lin Yang ◽  
Jianxu Chen ◽  
Jun Han ◽  
Yizhe Zhang ◽  
...  

Deep learning has been applied successfully to many biomedical image segmentation tasks. However, due to the diversity and complexity of biomedical image data, manual annotation for training common deep learning models is very timeconsuming and labor-intensive, especially because normally only biomedical experts can annotate image data well. Human experts are often involved in a long and iterative process of annotation, as in active learning type annotation schemes. In this paper, we propose representative annotation (RA), a new deep learning framework for reducing annotation effort in biomedical image segmentation. RA uses unsupervised networks for feature extraction and selects representative image patches for annotation in the latent space of learned feature descriptors, which implicitly characterizes the underlying data while minimizing redundancy. A fully convolutional network (FCN) is then trained using the annotated selected image patches for image segmentation. Our RA scheme offers three compelling advantages: (1) It leverages the ability of deep neural networks to learn better representations of image data; (2) it performs one-shot selection for manual annotation and frees annotators from the iterative process of common active learning based annotation schemes; (3) it can be deployed to 3D images with simple extensions. We evaluate our RA approach using three datasets (two 2D and one 3D) and show our framework yields competitive segmentation results comparing with state-of-the-art methods.


Author(s):  
Lars J. Isaksson ◽  
Paul Summers ◽  
Sara Raimondi ◽  
Sara Gandini ◽  
Abhir Bhalerao ◽  
...  

Abstract Researchers address the generalization problem of deep image processing networks mainly through extensive use of data augmentation techniques such as random flips, rotations, and deformations. A data augmentation technique called mixup, which constructs virtual training samples from convex combinations of inputs, was recently proposed for deep classification networks. The algorithm contributed to increased performance on classification in a variety of datasets, but so far has not been evaluated for image segmentation tasks. In this paper, we tested whether the mixup algorithm can improve the generalization performance of deep segmentation networks for medical image data. We trained a standard U-net architecture to segment the prostate in 100 T2-weighted 3D magnetic resonance images from prostate cancer patients, and compared the results with and without mixup in terms of Dice similarity coefficient and mean surface distance from a reference segmentation made by an experienced radiologist. Our results suggest that mixup offers a statistically significant boost in performance compared to non-mixup training, leading to up to 1.9% increase in Dice and a 10.9% decrease in surface distance. The mixup algorithm may thus offer an important aid for medical image segmentation applications, which are typically limited by severe data scarcity.


Sign in / Sign up

Export Citation Format

Share Document