scholarly journals Facial image denoising using AutoEncoder and UNET

2021 ◽  
Vol 3 (2) ◽  
pp. 89-96
Author(s):  
Milan Tripathi

Image denoising is a crucial topic in image processing. Noisy images are generated due to technical and environmental errors. Therefore, it is reasonable to consider image denoising an important topic to study, as it also helps to resolve other image processing issues. However, the challenge is that the classical techniques used are time-consuming and not flexible enough. This article compares the two major neural network architecture which looks promising to resolve this issues. The AutoEncoder and UNET is now the most researched subject in deep learning for image denoising. Multiple model architectures are designed, implement, and evaluated. The dataset is preprocessed and then it is used to train and test the model. It is clearly shown in this paper which model performs the best in this task by comparing both models using the most used parameters to evaluate image quality PSNR and SSIM.

2021 ◽  
Vol 3 (2) ◽  
pp. 102-111
Author(s):  
Milan Tripathi

Image denoising is an important aspect of image processing. Noisy images are produced as a result of technical and environmental flaws. As a result, it is reasonable to consider image denoising an important topic to research, as it also aids in the resolution of other image processing issues. The challenge, however, is that the traditional techniques used are time-consuming and inflexible. This article purposed a system of classifying and denoising noised images. A CNN and UNET based model architecture is designed, implement, and evaluated. The facial image dataset is processed and then it is used to train, valid and test the models. During preprocessing, the images are resized into 48*48, normalize, and various noises are added to the image. The preprocessing for each model is a bit different. The training and validation accuracy for the CNN model is 99.87% and 99.92% respectively. The UNET model is also able to get optimal PSNR and SSIM values for different noises.


2018 ◽  
Vol 8 (1) ◽  
Author(s):  
Zohreh Gholami Doborjeh ◽  
Nikola Kasabov ◽  
Maryam Gholami Doborjeh ◽  
Alexander Sumich

2021 ◽  
Author(s):  
◽  
Martin Mundt

Deep learning with neural networks seems to have largely replaced traditional design of computer vision systems. Automated methods to learn a plethora of parameters are now used in favor of previously practiced selection of explicit mathematical operators for a specific task. The entailed promise is that practitioners no longer need to take care of every individual step, but rather focus on gathering big amounts of data for neural network training. As a consequence, both a shift in mindset towards a focus on big datasets, as well as a wave of conceivable applications based exclusively on deep learning can be observed. This PhD dissertation aims to uncover some of the only implicitly mentioned or overlooked deep learning aspects, highlight unmentioned assumptions, and finally introduce methods to address respective immediate weaknesses. In the author’s humble opinion, these prevalent shortcomings can be tied to the fact that the involved steps in the machine learning workflow are frequently decoupled. Success is predominantly measured based on accuracy measures designed for evaluation with static benchmark test sets. Individual machine learning workflow components are assessed in isolation with respect to available data, choice of neural network architecture, and a particular learning algorithm, rather than viewing the machine learning system as a whole in context of a particular application. Correspondingly, in this dissertation, three key challenges have been identified: 1. Choice and flexibility of a neural network architecture. 2. Identification and rejection of unseen unknown data to avoid false predictions. 3. Continual learning without forgetting of already learned information. These latter challenges have already been crucial topics in older literature, alas, seem to require a renaissance in modern deep learning literature. Initially, it may appear that they pose independent research questions, however, the thesis posits that the aspects are intertwined and require a joint perspective in machine learning based systems. In summary, the essential question is thus how to pick a suitable neural network architecture for a specific task, how to recognize which data inputs belong to this context, which ones originate from potential other tasks, and ultimately how to continuously include such identified novel data in neural network training over time without overwriting existing knowledge. Thus, the central emphasis of this dissertation is to build on top of existing deep learning strengths, yet also acknowledge mentioned weaknesses, in an effort to establish a deeper understanding of interdependencies and synergies towards the development of unified solution mechanisms. For this purpose, the main portion of the thesis is in cumulative form. The respective publications can be grouped according to the three challenges outlined above. Correspondingly, chapter 1 is focused on choice and extendability of neural network architectures, analyzed in context of popular image classification tasks. An algorithm to automatically determine neural network layer width is introduced and is first contrasted with static architectures found in the literature. The importance of neural architecture design is then further showcased on a real-world application of defect detection in concrete bridges. Chapter 2 is comprised of the complementary ensuing questions of how to identify unknown concepts and subsequently incorporate them into continual learning. A joint central mechanism to distinguish unseen concepts from what is known in classification tasks, while enabling consecutive training without forgetting or revisiting older classes, is proposed. Once more, the role of the chosen neural network architecture is quantitatively reassessed. Finally, chapter 3 culminates in an overarching view, where developed parts are connected. Here, an extensive survey further serves the purpose to embed the gained insights in the broader literature landscape and emphasizes the importance of a common frame of thought. The ultimately presented approach thus reflects the overall thesis’ contribution to advance neural network based machine learning towards a unified solution that ties together choice of neural architecture with the ability to learn continually and the capability to automatically separate known from unknown data.


2020 ◽  
Vol 196 ◽  
pp. 02007
Author(s):  
Vladimir Mochalov ◽  
Anastasia Mochalova

In this paper, the previously obtained results on recognition of ionograms using deep learning are expanded to predict the parameters of the ionosphere. After the ionospheric parameters have been identified on the ionogram using deep learning in real time, we can predict the parameters for some time ahead on the basis of the new data obtained Examples of predicting the ionosphere parameters using an artificial recurrent neural network architecture long short-term memory are given. The place of the block for predicting the parameters of the ionosphere in the system for analyzing ionospheric data using deep learning methods is shown.


2021 ◽  
Vol 2137 (1) ◽  
pp. 012059
Author(s):  
Bowen Wei ◽  
Weixin Gao

Abstract At present, there are numerous losses caused by corrosion cracking of metal castings in engineering in China. In order to detect the possible defects of metal castings in engineering, the laser ultrasonic vision inspection technology is used to image the castings, and then the identification efficiency is low. In order to process these images efficiently and quickly, convolutional neural network image processing technology is introduced. According to the actual needs, a convolutional neural network architecture is designed to recognize images, and whether the architecture meets the requirements is verified. Experimental results show that the performance of the architecture meets the design requirements. Under the same conditions, this structure provides a solution for casting defect detection combined with artificial intelligence.


2020 ◽  
Vol 11 (18) ◽  
pp. 4618-4630 ◽  
Author(s):  
Jonathan A. Fine ◽  
Anand A. Rajasekar ◽  
Krupal P. Jethava ◽  
Gaurav Chopra

A new multi-label deep neural network architecture is used to combine Infrared and mass spectra, trained on single compounds to predict functional groups, and experimentally validated on complex mixtures.


Sign in / Sign up

Export Citation Format

Share Document