scholarly journals Silent EEG-Speech Recognition Using Convolutional and Recurrent Neural Network with 85% Accuracy of 9 Words Classification

Sensors ◽  
2021 ◽  
Vol 21 (20) ◽  
pp. 6744
Author(s):  
Darya Vorontsova ◽  
Ivan Menshikov ◽  
Aleksandr Zubov ◽  
Kirill Orlov ◽  
Peter Rikunov ◽  
...  

In this work, we focus on silent speech recognition in electroencephalography (EEG) data of healthy individuals to advance brain–computer interface (BCI) development to include people with neurodegeneration and movement and communication difficulties in society. Our dataset was recorded from 270 healthy subjects during silent speech of eight different Russia words (commands): `forward’, `backward’, `up’, `down’, `help’, `take’, `stop’, and `release’, and one pseudoword. We began by demonstrating that silent word distributions can be very close statistically and that there are words describing directed movements that share similar patterns of brain activity. However, after training one individual, we achieved 85% accuracy performing 9 words (including pseudoword) classification and 88% accuracy on binary classification on average. We show that a smaller dataset collected on one participant allows for building a more accurate classifier for a given subject than a larger dataset collected on a group of people. At the same time, we show that the learning outcomes on a limited sample of EEG-data are transferable to the general population. Thus, we demonstrate the possibility of using selected command-words to create an EEG-based input device for people on whom the neural network classifier has not been trained, which is particularly important for people with disabilities.

Author(s):  
Lukas Hecker ◽  
Rebekka Rupprecht ◽  
Ludger Tebartz van Elst ◽  
Juergen Kornmeier

AbstractEEG and MEG are well-established non-invasive methods in neuroscientific research and clinical diagnostics. Both methods provide a high temporal but low spatial resolution of brain activity. In order to gain insight about the spatial dynamics of the M/EEG one has to solve the inverse problem, which means that more than one configuration of neural sources can evoke one and the same distribution of EEG activity on the scalp. Artificial neural networks have been previously used successfully to find either one or two dipoles sources. These approaches, however, have never solved the inverse problem in a distributed dipole model with more than two dipole sources. We present ConvDip, a novel convolutional neural network (CNN) architecture that solves the EEG inverse problem in a distributed dipole model based on simulated EEG data. We show that (1) ConvDip learned to produce inverse solutions from a single time point of EEG data and (2) outperforms state-of-the-art methods (eLORETA and LCMV beamforming) on all focused performance measures. (3) It is more flexible when dealing with varying number of sources, produces less ghost sources and misses less real sources than the comparison methods. (4) It produces plausible inverse solutions for real-world EEG recordings and needs less than 40 ms for a single forward pass. Our results qualify ConvDip as an efficient and easy-to-apply novel method for source localization in EEG and MEG data, with high relevance for clinical applications, e.g. in epileptology and real time applications.


Author(s):  
Soha Abd Mohamed El-Moamen ◽  
Marghany Hassan Mohamed ◽  
Mohammed F. Farghally

The need for tracking and evaluation of patients in real-time has contributed to an increase in knowing people’s actions to enhance care facilities. Deep learning is good at both a rapid pace in collecting frameworks of big data healthcare and good predictions for detection the lung cancer early. In this paper, we proposed a constructive deep neural network with Apache Spark to classify images and levels of lung cancer. We developed a binary classification model using threshold technique classifying nodules to benign or malignant. At the proposed framework, the neural network models training, defined using the Keras API, is performed using BigDL in a distributed Spark clusters. The proposed algorithm has metrics AUC-0.9810, a misclassifying rate from which it has been shown that our suggested classifiers perform better than other classifiers.


Author(s):  
Lenka Lhotská ◽  
Vladimír Krajca ◽  
Jitka Mohylová ◽  
Svojmil Petránek ◽  
Václav Gerla

This chapter deals with the application of principal components analysis (PCA) to the field of data mining in electroencephalogram (EEG) processing. The principal components are estimated from the signal by eigen decomposition of the covariance estimate of the input. Alternatively, they can be estimated by a neural network (NN) configured for extracting the first principal components. Instead of performing computationally complex operations for eigenvector estimation, the neural network can be trained to produce ordered first principal components. Possible applications include separation of different signal components for feature extraction in the field of EEG signal processing, adaptive segmentation, epileptic spike detection, and long-term EEG monitoring evaluation of patients in a coma.


2018 ◽  
Vol 119 (4) ◽  
pp. 1251-1253 ◽  
Author(s):  
Randolph F. Helfrich

Our continuous perception of the world could be the result of discrete sampling, where individual snapshots are seamlessly fused into a coherent stream. It has been argued that endogenous oscillatory brain activity could provide the functional substrate of cortical rhythmic sampling. A new study demonstrates that cortical rhythmic sampling is tightly linked to the oculomotor system, thus providing a novel perspective on the neural network underlying top-down guided visual perception.


2021 ◽  
Vol 12 ◽  
Author(s):  
Caroline Pinte ◽  
Mathis Fleury ◽  
Pierre Maurel

The simultaneous acquisition of electroencephalographic (EEG) signals and functional magnetic resonance images (fMRI) aims to measure brain activity with good spatial and temporal resolution. This bimodal neuroimaging can bring complementary and very relevant information in many cases and in particular for epilepsy. Indeed, it has been shown that it can facilitate the localization of epileptic networks. Regarding the EEG, source localization requires the resolution of a complex inverse problem that depends on several parameters, one of the most important of which is the position of the EEG electrodes on the scalp. These positions are often roughly estimated using fiducial points. In simultaneous EEG-fMRI acquisitions, specific MRI sequences can provide valuable spatial information. In this work, we propose a new fully automatic method based on neural networks to segment an ultra-short echo-time MR volume in order to retrieve the coordinates and labels of the EEG electrodes. It consists of two steps: a segmentation of the images by a neural network, followed by the registration of an EEG template on the obtained detections. We trained the neural network using 37 MR volumes and then we tested our method on 23 new volumes. The results show an average detection accuracy of 99.7% with an average position error of 2.24 mm, as well as 100% accuracy in the labeling.


2019 ◽  
Vol 9 (16) ◽  
pp. 3355 ◽  
Author(s):  
Min Seop Lee ◽  
Yun Kyu Lee ◽  
Dong Sung Pae ◽  
Myo Taeg Lim ◽  
Dong Won Kim ◽  
...  

Physiological signals contain considerable information regarding emotions. This paper investigated the ability of photoplethysmogram (PPG) signals to recognize emotion, adopting a two-dimensional emotion model based on valence and arousal to represent human feelings. The main purpose was to recognize short term emotion using a single PPG signal pulse. We used a one-dimensional convolutional neural network (1D CNN) to extract PPG signal features to classify the valence and arousal. We split the PPG signal into a single 1.1 s pulse and normalized it for input to the neural network based on the personal maximum and minimum values. We chose the dataset for emotion analysis using physiological (DEAP) signals for the experiment and tested the 1D CNN as a binary classification (high or low valence and arousal), achieving the short-term emotion recognition of 1.1 s with 75.3% and 76.2% valence and arousal accuracies, respectively, on the DEAP data.


2019 ◽  
Vol 1 (1) ◽  
pp. 87-95
Author(s):  
Bishon Lamichanne ◽  
Hari K.C.

Speech is one of the most natural ways to communicate between people. It plays an important role in our daily lives. To make machines able to talk with people is a challenging but very useful task. A crucial step is to enable machines to recognize and understand what people are saying. Hence, speech recognition becomes a key technique providing an interface for communication between machines and humans. There has been a long research history on speech recognition. Neural network is known as a technique that has ability to classify nonlinear problem. Today, lots of research are going in the field of speech recognition with the help of the Neural Network. Even though positive results have been obtained from continuous study, research on minimizing the error rate is still gaining lots attention. The English language offers a number of challenges for speech recognition. This paper implements the RNN to analyze and recognize speech from the set of spoken words.


Sensors ◽  
2019 ◽  
Vol 19 (18) ◽  
pp. 3930 ◽  
Author(s):  
Sunah Jung ◽  
Haesang Yang ◽  
Kiheum Park ◽  
Yutaek Seo ◽  
Woojae Seong

The use of accelerometer signals for early recognition of severe slugging is investigated in a pipeline-riser system conveying an air–water two-phase flow, where six accelerometers are installed from the bottom to the top of the riser. Twelve different environmental conditions are produced by changing water and gas superficial velocities, of which three conditions are stable states and the other conditions are related to severe slugging. For online recognition, simple parameters using statistics and linear prediction coefficients are employed to extract useful features. Binary classification to recognize stable flow and severe slugging is performed using a support vector machine and a neural network. In multiclass classification, the neural network is adopted to identify four flow patterns of stable state, two types of severe slugging, and an irregular transition state between severe slugging and dual-frequency severe slugging. The performance is compared and analyzed according to the signal length for three cases of sensor location: six accelerometers, one accelerometer at the riser base, and one accelerometer at the top of the riser.


Author(s):  
Yedilkhan Amirgaliyev ◽  
Kuanyshbay Kuanyshbay ◽  
Aisultan Shoiynbek

This paper evaluates and compares the performances of three well-known optimization algorithms (Adagrad, Adam, Momentum) for faster training the neural network of CTC algorithm for speech recognition. For CTC algorithms recurrent neural network has been used, specifically Long-Short-Term memory. LSTM is effective and often used model. Data has been downloaded from VCTK corpus of Edinburgh University. The results of optimization algorithms have been evaluated by the Label error rate and CTC loss.


Sign in / Sign up

Export Citation Format

Share Document