scholarly journals NLP-Based Approach for Predicting HMI State Sequences Towards Monitoring Operator Situational Awareness

Sensors ◽  
2020 ◽  
Vol 20 (11) ◽  
pp. 3228
Author(s):  
Harsh V. P. Singh ◽  
Qusay H. Mahmoud

A novel approach presented herein transforms the Human Machine Interface (HMI) states, as a pattern of visual feedback states that encompass both operator actions and process states, from a multi-variate time-series to a natural language processing (NLP) modeling domain. The goal of this approach is to predict operator response patterns for n − a h e a d time-step window given k − l a g g e d past HMI state patterns. The NLP approach offers the possibility of encoding (semantic) contextual relations within HMI state patterns. Towards which, a technique for framing raw HMI data for supervised training using sequence-to-sequence (seq2seq) deep-learning machine translation algorithms is presented. In addition, a custom Seq2Seq convolutional neural network (CNN) NLP model based on current state-of-the-art design elements such as attention, is compared against a standard recurrent neural network (RNN) based NLP model. Results demonstrate comparable effectiveness of both the designs of NLP models evaluated for modeling HMI states. RNN NLP models showed higher ( ≈ 26 % ) forecast accuracy, in general for both in-sample and out-of-sample test datasets. However, custom CNN NLP model showed higher ( ≈ 53 % ) validation accuracy indicative of less over-fitting with the same amount of available training data. The real-world application of the proposed NLP modeling of industrial HMIs, such as in power generating stations control rooms, aviation (cockpits), and so forth, is towards the realization of a non-intrusive operator situational awareness monitoring framework through prediction of HMI states.

2006 ◽  
Vol 16 (06) ◽  
pp. 423-434 ◽  
Author(s):  
MOHAMED ABDEL FATTAH ◽  
FUJI REN ◽  
SHINGO KUROIWA

Parallel corpora have become an essential resource for work in multi lingual natural language processing. However, sentence aligned parallel corpora are more efficient than non-aligned parallel corpora for cross language information retrieval and machine translation applications. In this paper, we present a new approach to align sentences in bilingual parallel corpora based on feed forward neural network classifier. A feature parameter vector is extracted from the text pair under consideration. This vector contains text features such as length, punctuate score, and cognate score values. A set of manually prepared training data has been assigned to train the feed forward neural network. Another set of data was used for testing. Using this new approach, we could achieve an error reduction of 60% over length based approach when applied on English–Arabic parallel documents. Moreover this new approach is valid for any language pair and it is quite flexible approach since the feature parameter vector may contain more/less or different features than that we used in our system such as lexical match feature.


2021 ◽  
Author(s):  
Harish Chandra ◽  
Xianwei Meng ◽  
Arman Margaryan

We propose and implement a novel approach to model the evolution of COVID-19 pandemic and predict the daily COVID-19 cases (infected, recovered and dead). Our model builds on the classical SEIR-based framework by adding additional compartments to capture recovered, dead and quarantined cases. Quarantine impacts are modeled using an Artificial Neural Network (ANN), leveraging alternative data sources such as the Google mobility reports. Since our model captures the impact of lockdown policies through the quarantine functions we designed, it is able to model and predict future waves of COVID-19 cases. We also benchmark out-of-sample predictions from our model versus those from other popular COVID-19 case projection models.


Author(s):  
C. Swetha Reddy Et.al

Surprisingly comprehensive learning methods are implemented in many large learning machine data, such as visual recognition and visual language processing. Much of the success of advanced training in recent years is due to leadership training, which requires a set of information for specific tasks, before such training. However, in reality, selected tasks related to personal study are gradually accumulated over time as it is difficult to collect and submit training data manually. It provides a way to continue learning some information columns and examples of steps that are specific to the new class and called additional learning. In this post, we recommend the best machine training method for further training for deep neural networks. The basic idea is to learn a deep system with strong connections that can be "activated" or "turned off" at different stages. The approach you suggest allows you to reduce the distribution of old services as you learn new for example new training, which increases the effectiveness of training in the additional training phase. Experiments with MNIST and CIFAR-100 show that our approach can be implemented in other long-term phases in deep neuron models and achieve better results from zero-base training.


Author(s):  
Changshun Du ◽  
Lei Huang

Text sentiment analysis is one of the most important tasks in the field of public opinion monitoring, service evaluation and satisfaction analysis under network environments. Compared with the traditional Natural Language Processing analysis tools, convolution neural networks can automatically learn useful features from sentences and improve the performance of the affective analysis model. However, the original convolution neural network model ignores sentence structure information which is very important for text sentiment analysis. In this paper, we add piece-wise pooling to the convolution neural network, which allows the model to obtain the sentence structure. And the main features of different sentences are extracted to analyze the emotional tendencies of the text. At the same time, the user’s feedback involves many different fields, and there is less labeled data. In order to alleviate the sparsity of the data, this paper also uses the generative adversarial network to make common feature extractions, so that the model can obtain the common features associated with emotions in different fields, and improves the model’s Generalization ability with less training data. Experiments on different datasets demonstrate the effectiveness of this method.


2021 ◽  
Vol 15 ◽  
Author(s):  
Jinhua Tian ◽  
Hailun Xie ◽  
Siyuan Hu ◽  
Jia Liu

The increasingly popular application of AI runs the risk of amplifying social bias, such as classifying non-white faces as animals. Recent research has largely attributed this bias to the training data implemented. However, the underlying mechanism is poorly understood; therefore, strategies to rectify the bias are unresolved. Here, we examined a typical deep convolutional neural network (DCNN), VGG-Face, which was trained with a face dataset consisting of more white faces than black and Asian faces. The transfer learning result showed significantly better performance in identifying white faces, similar to the well-known social bias in humans, the other-race effect (ORE). To test whether the effect resulted from the imbalance of face images, we retrained the VGG-Face with a dataset containing more Asian faces, and found a reverse ORE that the newly-trained VGG-Face preferred Asian faces over white faces in identification accuracy. Additionally, when the number of Asian faces and white faces were matched in the dataset, the DCNN did not show any bias. To further examine how imbalanced image input led to the ORE, we performed a representational similarity analysis on VGG-Face's activation. We found that when the dataset contained more white faces, the representation of white faces was more distinct, indexed by smaller in-group similarity and larger representational Euclidean distance. That is, white faces were scattered more sparsely in the representational face space of the VGG-Face than the other faces. Importantly, the distinctiveness of faces was positively correlated with identification accuracy, which explained the ORE observed in the VGG-Face. In summary, our study revealed the mechanism underlying the ORE in DCNNs, which provides a novel approach to studying AI ethics. In addition, the face multidimensional representation theory discovered in humans was also applicable to DCNNs, advocating for future studies to apply more cognitive theories to understand DCNNs' behavior.


Materials ◽  
2021 ◽  
Vol 14 (16) ◽  
pp. 4492
Author(s):  
Adam Kulawik ◽  
Joanna Wróbel ◽  
Alexey Mikhailovich Ikonnikov

The motivation of the presented paper is the desire to create a universal tool to analyse the process of austenite decomposition during the cooling process of various steel grades. The presented analysis concerns the application of Recurrent Artificial Neural Networks (RANN) of the Long Short-Term Memory (LSTM) type for the analysis of the transition path of the cooling curve. This type of network was selected due to its ability to predict events in time sequences. The proposed generalisation allows for the determination of the austenite transformation during the continuous cooling process for various cooling curves. As training data for the neural network, values determined from the macroscopic model based on the analysis of Continuous Cooling Transformation (CCT) diagrams were used. All relations and analyses used to build training/testing or validation sets are presented in the paper. The modelling with the use of LSTM network gives the possibility to determine the incremental changes of phase transformation (in a given time step) with the assumed changes of temperature resulting from the considered cooling rate.


2020 ◽  
Vol 10 (6) ◽  
pp. 1265-1273
Author(s):  
Lili Chen ◽  
Huoyao Xu

Sleep apnea (SA) is a common sleep disorders affecting the sleep quality. Therefore the automatic SA detection has far-reaching implications for patients and physicians. In this paper, a novel approach is developed based on deep neural network (DNN) for automatic diagnosis SA. To this end, five features are extracted from electrocardiogram (ECG) signals through wavelet decomposition and sample entropy. The deep neural network is constructed by two-layer stacked sparse autoencoder (SSAE) network and one softmax layer. The softmax layer is added at the top of the SSAE network for diagnosing SA. Afterwards, the SSAE network can get more effective high-level features from raw features. The experimental results reveal that the performance of deep neural network can accomplish an accuracy of 96.66%, a sensitivity of 96.25%, and a specificity of 97%. In addition, the performance of deep neural network outperforms the comparison models including support vector machine (SVM), random forest (RF), and extreme learning machine (ELM). Finally, the experimental results reveal that the proposed method can be valid applied to automatic SA event detection.


2020 ◽  
Vol 8 (2) ◽  
pp. 89 ◽  
Author(s):  
Bradford Knight ◽  
Kevin Maki

Accurate and efficient prediction of the forces on a propeller is critical for analyzing a maneuvering vessel with numerical methods. CFD methods like RANS, LES, or DES can accurately predict the propeller forces, but are computationally expensive due to the need for added mesh discretization around the propeller as well as the requisite small time-step size. One way of mitigating the expense of modeling a maneuvering vessel with CFD is to apply the propeller force as a body force term in the Navier–Stokes equations and to apply the force to the equations of motion. The applied propeller force should be determined with minimal expense and good accuracy. This paper examines and compares nonlinear regression and neural network predictions of the thrust, torque, and side force of a propeller both in open water and in the behind condition. The methods are trained and tested with RANS CFD simulations. The neural network approach is shown to be more accurate and requires less training data than the regression technique.


Author(s):  
Muhammad Ali Ramdhani ◽  
Dian Sa’adillah Maylawati ◽  
Teddy Mantoro

<span>Every language has unique characteristics, structures, and grammar. Thus, different styles will have different processes and result in processed in Natural Language Processing (NLP) research area. In the current NLP research area, Data Mining (DM) or Machine Learning (ML) technique is popular, especially for Deep Learning (DL) method. This research aims to classify text data in the Indonesian language using Convolutional Neural Network (CNN) as one of the DL algorithms. The CNN algorithm used modified following the Indonesian language characteristics. Thereby, in the text pre-processing phase, stopword removal and stemming are particularly suitable for the Indonesian language. The experiment conducted using 472 Indonesian News text data from various sources with four categories: ‘hiburan’ (entertainment), ‘olahraga’ (sport), ‘tajuk utama’ (headline news), and ‘teknologi’ (technology). Based on the experiment and evaluation using 377 training data and 95 testing data, producing five models with ten epoch for each model, CNN has the best percentage of accuracy around 90,74% and loss value around 29,05% for 300 hidden layers in classifying the Indonesian News data.</span>


2021 ◽  
Vol 45 (10) ◽  
Author(s):  
A. W. Olthof ◽  
P. M. A. van Ooijen ◽  
L. J. Cornelissen

AbstractIn radiology, natural language processing (NLP) allows the extraction of valuable information from radiology reports. It can be used for various downstream tasks such as quality improvement, epidemiological research, and monitoring guideline adherence. Class imbalance, variation in dataset size, variation in report complexity, and algorithm type all influence NLP performance but have not yet been systematically and interrelatedly evaluated. In this study, we investigate these factors on the performance of four types [a fully connected neural network (Dense), a long short-term memory recurrent neural network (LSTM), a convolutional neural network (CNN), and a Bidirectional Encoder Representations from Transformers (BERT)] of deep learning-based NLP. Two datasets consisting of radiologist-annotated reports of both trauma radiographs (n = 2469) and chest radiographs and computer tomography (CT) studies (n = 2255) were split into training sets (80%) and testing sets (20%). The training data was used as a source to train all four model types in 84 experiments (Fracture-data) and 45 experiments (Chest-data) with variation in size and prevalence. The performance was evaluated on sensitivity, specificity, positive predictive value, negative predictive value, area under the curve, and F score. After the NLP of radiology reports, all four model-architectures demonstrated high performance with metrics up to > 0.90. CNN, LSTM, and Dense were outperformed by the BERT algorithm because of its stable results despite variation in training size and prevalence. Awareness of variation in prevalence is warranted because it impacts sensitivity and specificity in opposite directions.


Sign in / Sign up

Export Citation Format

Share Document