scholarly journals Computing 3D Phase-Type Holograms Based on Deep Learning Method

Photonics ◽  
2021 ◽  
Vol 8 (7) ◽  
pp. 280
Author(s):  
Huadong Zheng ◽  
Jianbin Hu ◽  
Chaojun Zhou ◽  
Xiaoxi Wang

Computer holography is a technology that use a mathematical model of optical holography to generate digital holograms. It has wide and promising applications in various areas, especially holographic display. However, traditional computational algorithms for generation of phase-type holograms based on iterative optimization have a built-in tradeoff between the calculating speed and accuracy, which severely limits the performance of computational holograms in advanced applications. Recently, several deep learning based computational methods for generating holograms have gained more and more attention. In this paper, a convolutional neural network for generation of multi-plane holograms and its training strategy is proposed using a multi-plane iterative angular spectrum algorithm (ASM). The well-trained network indicates an excellent ability to generate phase-only holograms for multi-plane input images and to reconstruct correct images in the corresponding depth plane. Numerical simulations and optical reconstructions show that the accuracy of this method is almost the same with traditional iterative methods but the computational time decreases dramatically. The result images show a high quality through analysis of the image performance indicators, e.g., peak signal-to-noise ratio (PSNR), structural similarity (SSIM) and contrast ratio. Finally, the effectiveness of the proposed method is verified through experimental investigations.

2020 ◽  
Vol 25 (2) ◽  
pp. 86-97
Author(s):  
Sandy Suryo Prayogo ◽  
Tubagus Maulana Kusuma

DVB merupakan standar transmisi televisi digital yang paling banyak digunakan saat ini. Unsur terpenting dari suatu proses transmisi adalah kualitas gambar dari video yang diterima setelah melalui proses transimisi tersebut. Banyak faktor yang dapat mempengaruhi kualitas dari suatu gambar, salah satunya adalah struktur frame dari video. Pada tulisan ini dilakukan pengujian sensitifitas video MPEG-4 berdasarkan struktur frame pada transmisi DVB-T. Pengujian dilakukan menggunakan simulasi matlab dan simulink. Digunakan juga ffmpeg untuk menyediakan format dan pengaturan video akan disimulasikan. Variabel yang diubah dari video adalah bitrate dan juga group-of-pictures (GOP), sedangkan variabel yang diubah dari transmisi DVB-T adalah signal-to-noise-ratio (SNR) pada kanal AWGN di antara pengirim (Tx) dan penerima (Rx). Hasil yang diperoleh dari percobaan berupa kualitas rata-rata gambar pada video yang diukur menggunakan metode pengukuran structural-similarity-index (SSIM). Dilakukan juga pengukuran terhadap jumlah bit-error-rate BER pada bitstream DVB-T. Percobaan yang dilakukan dapat menunjukkan seberapa besar sensitifitas bitrate dan GOP dari video pada transmisi DVB-T dengan kesimpulan semakin besar bitrate maka akan semakin buruk nilai kualitas gambarnya, dan semakin kecil nilai GOP maka akan semakin baik nilai kualitasnya. Penilitian diharapkan dapat dikembangkan menggunakan deep learning untuk memperoleh frame struktur yang tepat di kondisi-kondisi tertentu dalam proses transmisi televisi digital.


Author(s):  
S. Bash ◽  
B. Johnson ◽  
W. Gibbs ◽  
T. Zhang ◽  
A. Shankaranarayanan ◽  
...  

Abstract Objective This prospective multicenter multireader study evaluated the performance of 40% scan-time reduced spinal magnetic resonance imaging (MRI) reconstructed with deep learning (DL). Methods A total of 61 patients underwent standard of care (SOC) and accelerated (FAST) spine MRI. DL was used to enhance the accelerated set (FAST-DL). Three neuroradiologists were presented with paired side-by-side datasets (666 series). Datasets were blinded and randomized in sequence and left-right display order. Image features were preference rated. Structural similarity index (SSIM) and per pixel L1 was assessed for the image sets pre and post DL-enhancement as a quantitative assessment of image integrity impact. Results FAST-DL was qualitatively better than SOC for perceived signal-to-noise ratio (SNR) and artifacts and equivalent for other features. Quantitative SSIM was high, supporting the absence of image corruption by DL processing. Conclusion DL enables 40% spine MRI scan time reduction while maintaining diagnostic integrity and image quality with perceived benefits in SNR and artifact reduction, suggesting potential for clinical practice utility.


2021 ◽  
Vol 10 (9) ◽  
pp. 205846012110447
Author(s):  
Ryo Ogawa ◽  
Tomoyuki Kido ◽  
Masashi Nakamura ◽  
Atsushi Nozaki ◽  
R Marc Lebel ◽  
...  

Background Deep learning–based methods have been used to denoise magnetic resonance imaging. Purpose The purpose of this study was to evaluate a deep learning reconstruction (DL Recon) in cardiovascular black-blood T2-weighted images and compare with intensity filtered images. Material and Methods Forty-five DL Recon images were compared with intensity filtered and the original images. For quantitative image analysis, the signal to noise ratio (SNR) of the septum, contrast ratio (CR) of the septum to lumen, and sharpness of the endocardial border were calculated in each image. For qualitative image quality assessment, a 4-point subjective scale was assigned to each image (1 = poor, 2 = fair, 3 = good, 4 = excellent). Results The SNR and CR were significantly higher in the DL Recon images than in the intensity filtered and the original images ( p < .05 in each). Sharpness of the endocardial border was significantly higher in the DL Recon and intensity filtered images than in the original images ( p < .05 in each). The image quality of the DL Recon images was significantly better than that of intensity filtered and original images ( p < .001 in each). Conclusions DL Recon reduced image noise while improving image contrast and sharpness in the cardiovascular black-blood T2-weight sequence.


2021 ◽  
Vol 7 (3) ◽  
pp. 44
Author(s):  
Johannes Leuschner ◽  
Maximilian Schmidt ◽  
Poulami Somanya Ganguly ◽  
Vladyslav Andriiashen ◽  
Sophia Bethany Coban ◽  
...  

The reconstruction of computed tomography (CT) images is an active area of research. Following the rise of deep learning methods, many data-driven models have been proposed in recent years. In this work, we present the results of a data challenge that we organized, bringing together algorithm experts from different institutes to jointly work on quantitative evaluation of several data-driven methods on two large, public datasets during a ten day sprint. We focus on two applications of CT, namely, low-dose CT and sparse-angle CT. This enables us to fairly compare different methods using standardized settings. As a general result, we observe that the deep learning-based methods are able to improve the reconstruction quality metrics in both CT applications while the top performing methods show only minor differences in terms of peak signal-to-noise ratio (PSNR) and structural similarity (SSIM). We further discuss a number of other important criteria that should be taken into account when selecting a method, such as the availability of training data, the knowledge of the physical measurement model and the reconstruction speed.


2014 ◽  
Vol 26 (06) ◽  
pp. 1450074
Author(s):  
A. Sumaiya Begum ◽  
S. Poornachandra

In this paper a new ripplet-based shrinkage technique is used to suppress noise from Magnetic Resonance Imaging (MRI). The propitious properties of ripplet transform such as anisotropy, high directionality, good localization, and high-energy compaction make the proposed method efficient and feature preserving when compared to other transforms. Ripplet transform provides efficient representation of edges in images with a higher potential for image processing applications such as image restoration, compression, and de-noising. The proposed method implies a new nonlinear ripplet-based shrinkage technique to extract the spatial and frequency information from MRI corrupted by noise. The choice of this new shrinkage technique is due to its simplicity, versatility, and its efficiency in removing noise from homogenous regions and those regions with singularities, when compared to the existing filtering techniques. Experiments were conducted on several diffusion weighed images and anatomical images. The results show that the proposed de-noising technique shows competitive performance compared to the current state-of-art methods. Qualitative validation was performed based on several quality metrics and profound improvement over existing methods was obtained. Higher values of Peak Signal to Noise Ratio (PSNR), Correlation Coefficient (CC), mean structural similarity index (MSSIM), and lower values of Root Mean Square Error (RMSE) and computational time were obtained for the proposed ripplet-based shrinkage technique when compared to the existing ones.


2021 ◽  
Author(s):  
Kibo Ote ◽  
Fumio Hashimoto

Abstract Deep learning has attracted attention for positron emission tomography (PET) image reconstruction task, however, it remains necessary to further improve the image quality. In this study, we propose a novel CNN-based fast time-of-flight PET (TOF-PET) image reconstruction method to fully utilize the direction information of coincidence events. The proposed method inputs view-grouped histo-images into a 3D CNN as a multi-channel image to use the direction information of coincidence events. We evaluated the proposed method using Monte Carlo simulation data obtained from a digital brain phantom. Compared to the case without it, when using direction information, the peak signal-to-noise ratio and structural similarity were improved by 1.2 dB and 0.02, at a coincidence time resolution of 300 ps. The calculation times of the proposed method were significantly faster than the conventional iterative reconstruction. These results indicate that the proposed method improves both the speed and image quality of TOF-PET image reconstruction.


Symmetry ◽  
2020 ◽  
Vol 12 (2) ◽  
pp. 224
Author(s):  
Qing Yang ◽  
Ming Yu ◽  
Yan Xu ◽  
Shixin Cen

Rainy, as an inevitable weather condition, will affect the acquired image. To solve this problem, a single image rain removal algorithm based on deep learning and symmetric transformation is proposed. Because of the important characteristics of wavelet transform, such as symmetry, orthogonality, flexibility and limited support, wavelet transform is used to remove rain from a single image. The image is denoised by using wavelet decomposition, threshold value and wavelet reconstruction in wavelet transform, and the rain drop image is transformed from RGB space to YUV (luma chroma) space by using deep learning to obtain the brightness component and color component of the image. the brightness component and residual component of the raindrop source image and the ideal recovered image without raindrop are extracted. The residual image and brightness component are overlapped again, the reconstructed image is restored to RGB space by YUV inverse transformation, and the final color raindrop free image is obtained. After training the network, the optimal parameters of the network are obtained, and finally the convolution neural network which can effectively remove the rain line is obtained. Experimental results show that compared with other algorithms, the proposed algorithm achieves the highest value in both peak signal-to-noise ratio (PSNR) and structural similarity, which shows that the image effect of the algorithm is better after rain removal.


2020 ◽  
Vol 2 (2) ◽  
pp. 78-98 ◽  
Author(s):  
Sandra Aigner ◽  
Marco Körner

This paper analyzes in detail how different loss functions influence the generalization abilities of a deep learning-based next frame prediction model for traffic scenes. Our prediction model is a convolutional long-short term memory (ConvLSTM) network that generates the pixel values of the next frame after having observed the raw pixel values of a sequence of four past frames. We trained the model with 21 combinations of seven loss terms using the Cityscapes Sequences dataset and an identical hyper-parameter setting. The loss terms range from pixel-error based terms to adversarial terms. To assess the generalization abilities of the resulting models, we generated predictions up to 20 time-steps into the future for four datasets of increasing visual distance to the training dataset—KITTI Tracking, BDD100K, UA-DETRAC, and KIT AIS Vehicles. All predicted frames were evaluated quantitatively with both traditional pixel-based evaluation metrics, that is, mean squared error (MSE), peak signal-to-noise ratio (PSNR), and structural similarity index (SSIM), and recent, more advanced, feature-based evaluation metrics, that is, Fréchet inception distance (FID), and learned perceptual image patch similarity (LPIPS). The results show that solely by choosing a different combination of losses, we can boost the prediction performance on new datasets by up to 55%, and by up to 50% for long-term predictions.


2020 ◽  
Author(s):  
Hao Li ◽  
DeLiang Wang ◽  
Xueliang Zhang ◽  
Guanglai Gao

Entropy ◽  
2021 ◽  
Vol 23 (2) ◽  
pp. 223
Author(s):  
Yen-Ling Tai ◽  
Shin-Jhe Huang ◽  
Chien-Chang Chen ◽  
Henry Horng-Shing Lu

Nowadays, deep learning methods with high structural complexity and flexibility inevitably lean on the computational capability of the hardware. A platform with high-performance GPUs and large amounts of memory could support neural networks having large numbers of layers and kernels. However, naively pursuing high-cost hardware would probably drag the technical development of deep learning methods. In the article, we thus establish a new preprocessing method to reduce the computational complexity of the neural networks. Inspired by the band theory of solids in physics, we map the image space into a noninteraction physical system isomorphically and then treat image voxels as particle-like clusters. Then, we reconstruct the Fermi–Dirac distribution to be a correction function for the normalization of the voxel intensity and as a filter of insignificant cluster components. The filtered clusters at the circumstance can delineate the morphological heterogeneity of the image voxels. We used the BraTS 2019 datasets and the dimensional fusion U-net for the algorithmic validation, and the proposed Fermi–Dirac correction function exhibited comparable performance to other employed preprocessing methods. By comparing to the conventional z-score normalization function and the Gamma correction function, the proposed algorithm can save at least 38% of computational time cost under a low-cost hardware architecture. Even though the correction function of global histogram equalization has the lowest computational time among the employed correction functions, the proposed Fermi–Dirac correction function exhibits better capabilities of image augmentation and segmentation.


Sign in / Sign up

Export Citation Format

Share Document