scholarly journals Calibration of the EBT3 Gafchromic Film Using HNN Deep Learning

2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Liyun Chang ◽  
Shyh-An Yeh ◽  
Sheng-Yow Ho ◽  
Hueisch-Jy Ding ◽  
Pang-Yu Chen ◽  
...  

To achieve a dose distribution conformal to the target volume while sparing normal tissues, intensity modulation with steep dose gradient is used for treatment planning. To successfully deliver such treatment, high spatial and dosimetric accuracy are crucial and need to be verified. With high 2D dosimetry resolution and a self-development property, the Ashland Inc. product EBT3 Gafchromic film is a widely used quality assurance tool designed especially for this. However, the film should be recalibrated each quarter due to the “aging effect,” and calibration uncertainties always exist between individual films even in the same lot. Recently, artificial neural networks (ANN) are applied to many fields. If a physicist can collect the calibration data, it could be accumulated to be a substantial ANN data input used for film calibration. We therefore use the Keras functional Application Program Interface to build a hierarchical neural network (HNN), with the inputs of net optical densities, pixel values, and inverse transmittances to reveal the delivered dose and train the neural network with deep learning. For comparison, the film dose calculated using red-channel net optical density with power function fitting was performed and taken as a conventional method. The results show that the percentage error of the film dose using the HNN method is less than 4% for the aging effect verification test and less than 4.5% for the intralot variation test; in contrast, the conventional method could yield errors higher than 10% and 7%, respectively. This HNN method to calibrate the EBT film could be further improved by adding training data or adjusting the HNN structure. The model could help physicists spend less calibration time and reduce film usage.

2021 ◽  
Vol 11 (15) ◽  
pp. 7148
Author(s):  
Bedada Endale ◽  
Abera Tullu ◽  
Hayoung Shi ◽  
Beom-Soo Kang

Unmanned aerial vehicles (UAVs) are being widely utilized for various missions: in both civilian and military sectors. Many of these missions demand UAVs to acquire artificial intelligence about the environments they are navigating in. This perception can be realized by training a computing machine to classify objects in the environment. One of the well known machine training approaches is supervised deep learning, which enables a machine to classify objects. However, supervised deep learning comes with huge sacrifice in terms of time and computational resources. Collecting big input data, pre-training processes, such as labeling training data, and the need for a high performance computer for training are some of the challenges that supervised deep learning poses. To address these setbacks, this study proposes mission specific input data augmentation techniques and the design of light-weight deep neural network architecture that is capable of real-time object classification. Semi-direct visual odometry (SVO) data of augmented images are used to train the network for object classification. Ten classes of 10,000 different images in each class were used as input data where 80% were for training the network and the remaining 20% were used for network validation. For the optimization of the designed deep neural network, a sequential gradient descent algorithm was implemented. This algorithm has the advantage of handling redundancy in the data more efficiently than other algorithms.


Sensors ◽  
2020 ◽  
Vol 20 (22) ◽  
pp. 6460
Author(s):  
Dae-Yeon Kim ◽  
Dong-Sik Choi ◽  
Jaeyun Kim ◽  
Sung Wan Chun ◽  
Hyo-Wook Gil ◽  
...  

In this study, we propose a personalized glucose prediction model using deep learning for hospitalized patients who experience Type-2 diabetes. We aim for our model to assist the medical personnel who check the blood glucose and control the amount of insulin doses. Herein, we employed a deep learning algorithm, especially a recurrent neural network (RNN), that consists of a sequence processing layer and a classification layer for the glucose prediction. We tested a simple RNN, gated recurrent unit (GRU), and long-short term memory (LSTM) and varied the architectures to determine the one with the best performance. For that, we collected data for a week using a continuous glucose monitoring device. Type-2 inpatients are usually experiencing bad health conditions and have a high variability of glucose level. However, there are few studies on the Type-2 glucose prediction model while many studies performed on Type-1 glucose prediction. This work has a contribution in that the proposed model exhibits a comparative performance to previous works on Type-1 patients. For 20 in-hospital patients, we achieved an average root mean squared error (RMSE) of 21.5 and an Mean absolute percentage error (MAPE) of 11.1%. The GRU with a single RNN layer and two dense layers was found to be sufficient to predict the glucose level. Moreover, to build a personalized model, at most, 50% of data are required for training.


2021 ◽  
Vol 13 (19) ◽  
pp. 3859
Author(s):  
Joby M. Prince Czarnecki ◽  
Sathishkumar Samiappan ◽  
Meilun Zhou ◽  
Cary Daniel McCraine ◽  
Louis L. Wasson

The radiometric quality of remotely sensed imagery is crucial for precision agriculture applications because estimations of plant health rely on the underlying quality. Sky conditions, and specifically shadowing from clouds, are critical determinants in the quality of images that can be obtained from low-altitude sensing platforms. In this work, we first compare common deep learning approaches to classify sky conditions with regard to cloud shadows in agricultural fields using a visible spectrum camera. We then develop an artificial-intelligence-based edge computing system to fully automate the classification process. Training data consisting of 100 oblique angle images of the sky were provided to a convolutional neural network and two deep residual neural networks (ResNet18 and ResNet34) to facilitate learning two classes, namely (1) good image quality expected, and (2) degraded image quality expected. The expectation of quality stemmed from the sky condition (i.e., density, coverage, and thickness of clouds) present at the time of the image capture. These networks were tested using a set of 13,000 images. Our results demonstrated that ResNet18 and ResNet34 classifiers produced better classification accuracy when compared to a convolutional neural network classifier. The best overall accuracy was obtained by ResNet34, which was 92% accurate, with a Kappa statistic of 0.77. These results demonstrate a low-cost solution to quality control for future autonomous farming systems that will operate without human intervention and supervision.


2021 ◽  
Vol 11 (12) ◽  
pp. 3044-3053
Author(s):  
Rakesh Kumar Mahendran ◽  
V. Prabhu ◽  
V. Parthasarathy ◽  
A. Mary Judith

Myocardial infarction (MI) may precipitate severe health damage and lead to irreversible death of the heart muscle, the result of prolonged lack of oxygen if it is not treated in a timely manner. Lack of accurate and early detection techniques for this heart disease has reduced the efficiency of MI diagnosis. In this paper, the design, and implementation of an efficient deep learning algorithm called Adaptive Recurrent neural network (ARNN) is proposed for the MI detection. The main objective of the proposed work is the accurate identification of MI disease using ECG signals. ECG signal denoising has been performed using the Multi-Notch filter, which removes the specified noise frequency range. Discrete wavelet transform (DWT) is utilized for performing the feature extraction that decomposes the ECG signal into varied scales with waveletfiltering bank. After the extraction of specific QRS features, classification of the defected and normal ECG arrhythmic beat has been performed using the deep learning-based ARNN classifier. The MIT-BIH database has been used for testing and training data. The performance of the proposed algorithm is evaluated based on classification accuracy. Results that are attained include the classification accuracy of about 99.21%, 99% of sensitivity and 99.4% of specificity with PPV and NPV of about 99.4 and 99.01 values indicate the enhanced performance of our proposed work compared with the conventional LSTM-CAE and LSTM-CNN techniques.


Author(s):  
Uzma Batool ◽  
Mohd Ibrahim Shapiai ◽  
Nordinah Ismail ◽  
Hilman Fauzi ◽  
Syahrizal Salleh

Silicon wafer defect data collected from fabrication facilities is intrinsically imbalanced because of the variable frequencies of defect types. Frequently occurring types will have more influence on the classification predictions if a model gets trained on such skewed data. A fair classifier for such imbalanced data requires a mechanism to deal with type imbalance in order to avoid biased results. This study has proposed a convolutional neural network for wafer map defect classification, employing oversampling as an imbalance addressing technique. To have an equal participation of all classes in the classifier’s training, data augmentation has been employed, generating more samples in minor classes. The proposed deep learning method has been evaluated on a real wafer map defect dataset and its classification results on the test set returned a 97.91% accuracy. The results were compared with another deep learning based auto-encoder model demonstrating the proposed method, a potential approach for silicon wafer defect classification that needs to be investigated further for its robustness.


2022 ◽  
pp. 1559-1575
Author(s):  
Mário Pereira Véstias

Machine learning is the study of algorithms and models for computing systems to do tasks based on pattern identification and inference. When it is difficult or infeasible to develop an algorithm to do a particular task, machine learning algorithms can provide an output based on previous training data. A well-known machine learning model is deep learning. The most recent deep learning models are based on artificial neural networks (ANN). There exist several types of artificial neural networks including the feedforward neural network, the Kohonen self-organizing neural network, the recurrent neural network, the convolutional neural network, the modular neural network, among others. This article focuses on convolutional neural networks with a description of the model, the training and inference processes and its applicability. It will also give an overview of the most used CNN models and what to expect from the next generation of CNN models.


2020 ◽  
Vol 10 (21) ◽  
pp. 7817
Author(s):  
Ivana Marin ◽  
Ana Kuzmanic Skelin ◽  
Tamara Grujic

The main goal of any classification or regression task is to obtain a model that will generalize well on new, previously unseen data. Due to the recent rise of deep learning and many state-of-the-art results obtained with deep models, deep learning architectures have become one of the most used model architectures nowadays. To generalize well, a deep model needs to learn the training data well without overfitting. The latter implies a correlation of deep model optimization and regularization with generalization performance. In this work, we explore the effect of the used optimization algorithm and regularization techniques on the final generalization performance of the model with convolutional neural network (CNN) architecture widely used in the field of computer vision. We give a detailed overview of optimization and regularization techniques with a comparative analysis of their performance with three CNNs on the CIFAR-10 and Fashion-MNIST image datasets.


2020 ◽  
Vol 10 (5) ◽  
pp. 1657 ◽  
Author(s):  
Jieun Baek ◽  
Yosoon Choi

This paper proposes a deep neural network (DNN)-based method for predicting ore production by truck-haulage systems in open-pit mines. The proposed method utilizes two DNN models that are designed to predict ore production during the morning and afternoon haulage sessions, respectively. The configuration of the input nodes of the DNN models is based on truck-haulage conditions and corresponding operation times. To verify the efficacy of the proposed method, training data for the DNN models were generated by processing packet data collected over the two-month period December 2018 to January 2019. Subsequently, following training under different hidden-layer conditions, it was observed that the prediction accuracy of morning ore production was highest when the number of hidden layers and number of corresponding nodes were four and 50, respectively. The corresponding values of the determination coefficient and mean absolute percentage error (MAPE) were 0.99% and 4.78%, respectively. Further, the prediction accuracy of afternoon ore production was highest when the number of hidden layers was four and the corresponding number of nodes was 50. This yielded determination coefficient and MAPE values of 0.99% and 5.26%, respectively.


2020 ◽  
Vol 8 ◽  
Author(s):  
Adil Khadidos ◽  
Alaa O. Khadidos ◽  
Srihari Kannan ◽  
Yuvaraj Natarajan ◽  
Sachi Nandan Mohanty ◽  
...  

In this paper, a data mining model on a hybrid deep learning framework is designed to diagnose the medical conditions of patients infected with the coronavirus disease 2019 (COVID-19) virus. The hybrid deep learning model is designed as a combination of convolutional neural network (CNN) and recurrent neural network (RNN) and named as DeepSense method. It is designed as a series of layers to extract and classify the related features of COVID-19 infections from the lungs. The computerized tomography image is used as an input data, and hence, the classifier is designed to ease the process of classification on learning the multidimensional input data using the Expert Hidden layers. The validation of the model is conducted against the medical image datasets to predict the infections using deep learning classifiers. The results show that the DeepSense classifier offers accuracy in an improved manner than the conventional deep and machine learning classifiers. The proposed method is validated against three different datasets, where the training data are compared with 70%, 80%, and 90% training data. It specifically provides the quality of the diagnostic method adopted for the prediction of COVID-19 infections in a patient.


2019 ◽  
Vol 1 ◽  
pp. 1-1
Author(s):  
Tee-Ann Teo

<p><strong>Abstract.</strong> Deep Learning is a kind of Machine Learning technology which utilizing the deep neural network to learn a promising model from a large training data set. Convolutional Neural Network (CNN) has been successfully applied in image segmentation and classification with high accuracy results. The CNN applies multiple kernels (also called filters) to extract image features via image convolution. It is able to determine multiscale features through the multiple layers of convolution and pooling processes. The variety of training data plays an important role to determine a reliable CNN model. The benchmarking training data for road mark extraction is mainly focused on close-range imagery because it is easier to obtain a close-range image rather than an airborne image. For example, KITTI Vision Benchmark Suite. This study aims to transfer the road mark training data from mobile lidar system to aerial orthoimage in Fully Convolutional Networks (FCN). The transformation of the training data from ground-based system to airborne system may reduce the effort of producing a large training data set.</p><p>This study uses FCN technology and aerial orthoimage to localize road marks on the road regions. The road regions are first extracted from 2-D large-scale vector map. The input aerial orthoimage is 10&amp;thinsp;cm spatial resolution and the non-road regions are masked out before the road mark localization. The training data are road mark’s polygons, which are originally digitized from ground-based mobile lidar and prepared for the road mark extraction using mobile mapping system. This study reuses these training data and applies them for the road mark extraction using aerial orthoimage. The digitized training road marks are then transformed to road polygon based on mapping coordinates. As the detail of ground-based lidar is much better than the airborne system, the partially occulted parking lot in aerial orthoimage can also be obtained from the ground-based system. The labels (also called annotations) for FCN include road region, non-regions and road mark. The size of a training batch is 500&amp;thinsp;pixel by 500&amp;thinsp;pixel (50&amp;thinsp;m by 50&amp;thinsp;m on the ground), and the total number of training batches for training is 75 batches. After the FCN training stage, an independent aerial orthoimage (Figure 1a) is applied to predict the road marks. The results of FCN provide initial regions for road marks (Figure 1b). Usually, road marks show higher reflectance than road asphalts. Therefore, this study uses this characteristic to refine the road marks (Figure 1c) by a binary classification inside the initial road mark’s region.</p><p>To compare the automatically extracted road marks (Figure 1c) and manually digitized road marks (Figure 1d), most road marks can be extracted using the training set from ground-based system. This study also selects an area of 600&amp;thinsp;m&amp;thinsp;&amp;times;&amp;thinsp;200&amp;thinsp;m in quantitative analysis. Among the 371 reference road marks, 332 can be extracted from proposed scheme, and the completeness reached 89%. The preliminary experiment demonstrated that most road marks can be successfully extracted by the proposed scheme. Therefore, the training data from the ground-based mapping system can be utilized in airborne orthoimage in similar spatial resolution.</p>


Sign in / Sign up

Export Citation Format

Share Document