scholarly journals Assessment of Machine Learning Techniques in IoT-Based Architecture for the Monitoring and Prediction of COVID-19

Electronics ◽  
2021 ◽  
Vol 10 (15) ◽  
pp. 1834
Author(s):  
Abdullah Aljumah

From the end of 2019, the world has been facing the threat of COVID-19. It is predicted that, before herd immunity is achieved globally via vaccination, people around the world will have to tackle the COVID-19 pandemic using precautionary steps. This paper suggests a COVID-19 identification and control system that operates in real-time. The proposed system utilizes the Internet of Things (IoT) platform to capture users’ time-sensitive symptom information to detect potential cases of coronaviruses early on, to track the clinical measures adopted by survivors, and to gather and examine appropriate data to verify the existence of the virus. There are five key components in the framework: symptom data collection and uploading (via communication technology), a quarantine/isolation center, an information processing core (using artificial intelligent techniques), cloud computing, and visualization to healthcare doctors. This research utilizes eight machine/deep learning techniques—Neural Network, Decision Table, Support Vector Machine (SVM), Naive Bayes, OneR, K-Nearest Neighbor (K-NN), Dense Neural Network (DNN), and the Long Short-Term Memory technique—to detect coronavirus cases from time-sensitive information. A simulation was performed to verify the eight algorithms, after selecting the relevant symptoms, on real-world COVID-19 data values. The results showed that five of these eight algorithms obtained an accuracy of over 90%. Conclusively, it is shown that real-world symptomatic information would enable these three algorithms to identify potential COVID-19 cases effectively with enhanced accuracy. Additionally, the framework presents responses to treatment for COVID-19 patients.

Algorithms ◽  
2018 ◽  
Vol 11 (11) ◽  
pp. 170 ◽  
Author(s):  
Zhixi Li ◽  
Vincent Tam

Momentum and reversal effects are important phenomena in stock markets. In academia, relevant studies have been conducted for years. Researchers have attempted to analyze these phenomena using statistical methods and to give some plausible explanations. However, those explanations are sometimes unconvincing. Furthermore, it is very difficult to transfer the findings of these studies to real-world investment trading strategies due to the lack of predictive ability. This paper represents the first attempt to adopt machine learning techniques for investigating the momentum and reversal effects occurring in any stock market. In the study, various machine learning techniques, including the Decision Tree (DT), Support Vector Machine (SVM), Multilayer Perceptron Neural Network (MLP), and Long Short-Term Memory Neural Network (LSTM) were explored and compared carefully. Several models built on these machine learning approaches were used to predict the momentum or reversal effect on the stock market of mainland China, thus allowing investors to build corresponding trading strategies. The experimental results demonstrated that these machine learning approaches, especially the SVM, are beneficial for capturing the relevant momentum and reversal effects, and possibly building profitable trading strategies. Moreover, we propose the corresponding trading strategies in terms of market states to acquire the best investment returns.


Information ◽  
2021 ◽  
Vol 12 (12) ◽  
pp. 528
Author(s):  
David Opeoluwa Oyewola ◽  
Emmanuel Gbenga Dada ◽  
Sanjay Misra ◽  
Robertas Damaševičius

The application of machine learning techniques to the epidemiology of COVID-19 is a necessary measure that can be exploited to curtail the further spread of this endemic. Conventional techniques used to determine the epidemiology of COVID-19 are slow and costly, and data are scarce. We investigate the effects of noise filters on the performance of machine learning algorithms on the COVID-19 epidemiology dataset. Noise filter algorithms are used to remove noise from the datasets utilized in this study. We applied nine machine learning techniques to classify the epidemiology of COVID-19, which are bagging, boosting, support vector machine, bidirectional long short-term memory, decision tree, naïve Bayes, k-nearest neighbor, random forest, and multinomial logistic regression. Data from patients who contracted coronavirus disease were collected from the Kaggle database between 23 January 2020 and 24 June 2020. Noisy and filtered data were used in our experiments. As a result of denoising, machine learning models have produced high results for the prediction of COVID-19 cases in South Korea. For isolated cases after performing noise filtering operations, machine learning techniques achieved an accuracy between 98–100%. The results indicate that filtering noise from the dataset can improve the accuracy of COVID-19 case prediction algorithms.


2021 ◽  
Vol 9 ◽  
pp. 152-158
Author(s):  
Shubha Singh ◽  
Sreedevi Gutta ◽  
Ahmad Hadaegh

The Trend of stock price prediction is becoming more popular than ever. Share market is difficult to predict due to its volatile nature. There are no rules to follow to predict what will happen with the stock in the future. To predict accurately is a huge challenge since the market trend always keep changing depending on many factors. The objective is to apply machine learning techniques to predict stocks and maximize the profit. In this work, we have shown that with the help of artificial intelligence and machine learning, the process of prediction can be improved. While doing the literature review, we realized that the most effective machine learning tool for this research include: Artificial Neural Network (ANN), Support Vector Machine (SVM), and Genetic Algorithms (GA). All categories have common and unique findings and limitations. We collected data for about 10 years and used Long Short-Term Memory (LSTM) Neural Network-based machine learning models to analyze and predict the stock price. The Recurrent Neural Network (RNN) is useful to preserve the time-series features for improving profits. The financial data High and Close are used as input for the model.


2020 ◽  
Vol 12 (2) ◽  
pp. 84-99
Author(s):  
Li-Pang Chen

In this paper, we investigate analysis and prediction of the time-dependent data. We focus our attention on four different stocks are selected from Yahoo Finance historical database. To build up models and predict the future stock price, we consider three different machine learning techniques including Long Short-Term Memory (LSTM), Convolutional Neural Networks (CNN) and Support Vector Regression (SVR). By treating close price, open price, daily low, daily high, adjusted close price, and volume of trades as predictors in machine learning methods, it can be shown that the prediction accuracy is improved.


2021 ◽  
Vol 186 (Supplement_1) ◽  
pp. 445-451
Author(s):  
Yifei Sun ◽  
Navid Rashedi ◽  
Vikrant Vaze ◽  
Parikshit Shah ◽  
Ryan Halter ◽  
...  

ABSTRACT Introduction Early prediction of the acute hypotensive episode (AHE) in critically ill patients has the potential to improve outcomes. In this study, we apply different machine learning algorithms to the MIMIC III Physionet dataset, containing more than 60,000 real-world intensive care unit records, to test commonly used machine learning technologies and compare their performances. Materials and Methods Five classification methods including K-nearest neighbor, logistic regression, support vector machine, random forest, and a deep learning method called long short-term memory are applied to predict an AHE 30 minutes in advance. An analysis comparing model performance when including versus excluding invasive features was conducted. To further study the pattern of the underlying mean arterial pressure (MAP), we apply a regression method to predict the continuous MAP values using linear regression over the next 60 minutes. Results Support vector machine yields the best performance in terms of recall (84%). Including the invasive features in the classification improves the performance significantly with both recall and precision increasing by more than 20 percentage points. We were able to predict the MAP with a root mean square error (a frequently used measure of the differences between the predicted values and the observed values) of 10 mmHg 60 minutes in the future. After converting continuous MAP predictions into AHE binary predictions, we achieve a 91% recall and 68% precision. In addition to predicting AHE, the MAP predictions provide clinically useful information regarding the timing and severity of the AHE occurrence. Conclusion We were able to predict AHE with precision and recall above 80% 30 minutes in advance with the large real-world dataset. The prediction of regression model can provide a more fine-grained, interpretable signal to practitioners. Model performance is improved by the inclusion of invasive features in predicting AHE, when compared to predicting the AHE based on only the available, restricted set of noninvasive technologies. This demonstrates the importance of exploring more noninvasive technologies for AHE prediction.


2021 ◽  
Author(s):  
Praveeen Anandhanathan ◽  
Priyanka Gopalan

Abstract Coronavirus disease (COVID-19) is spreading across the world. Since at first it has appeared in Wuhan, China in December 2019, it has become a serious issue across the globe. There are no accurate resources to predict and find the disease. So, by knowing the past patients’ records, it could guide the clinicians to fight against the pandemic. Therefore, for the prediction of healthiness from symptoms Machine learning techniques can be implemented. From this we are going to analyse only the symptoms which occurs in every patient. These predictions can help clinicians in the easier manner to cure the patients. Already for prediction of many of the diseases, techniques like SVM (Support vector Machine), Fuzzy k-Means Clustering, Decision Tree algorithm, Random Forest Method, ANN (Artificial Neural Network), KNN (k-Nearest Neighbour), Naïve Bayes, Linear Regression model are used. As we haven’t faced this disease before, we can’t say which technique will give the maximum accuracy. So, we are going to provide an efficient result by comparing all the such algorithms in RStudio.


2018 ◽  
Vol 10 (1) ◽  
pp. 203 ◽  
Author(s):  
Xianming Dou ◽  
Yongguo Yang ◽  
Jinhui Luo

Approximating the complex nonlinear relationships that dominate the exchange of carbon dioxide fluxes between the biosphere and atmosphere is fundamentally important for addressing the issue of climate change. The progress of machine learning techniques has offered a number of useful tools for the scientific community aiming to gain new insights into the temporal and spatial variation of different carbon fluxes in terrestrial ecosystems. In this study, adaptive neuro-fuzzy inference system (ANFIS) and generalized regression neural network (GRNN) models were developed to predict the daily carbon fluxes in three boreal forest ecosystems based on eddy covariance (EC) measurements. Moreover, a comparison was made between the modeled values derived from these models and those of traditional artificial neural network (ANN) and support vector machine (SVM) models. These models were also compared with multiple linear regression (MLR). Several statistical indicators, including coefficient of determination (R2), Nash-Sutcliffe efficiency (NSE), bias error (Bias) and root mean square error (RMSE) were utilized to evaluate the performance of the applied models. The results showed that the developed machine learning models were able to account for the most variance in the carbon fluxes at both daily and hourly time scales in the three stands and they consistently and substantially outperformed the MLR model for both daily and hourly carbon flux estimates. It was demonstrated that the ANFIS and ANN models provided similar estimates in the testing period with an approximate value of R2 = 0.93, NSE = 0.91, Bias = 0.11 g C m−2 day−1 and RMSE = 1.04 g C m−2 day−1 for daily gross primary productivity, 0.94, 0.82, 0.24 g C m−2 day−1 and 0.72 g C m−2 day−1 for daily ecosystem respiration, and 0.79, 0.75, 0.14 g C m−2 day−1 and 0.89 g C m−2 day−1 for daily net ecosystem exchange, and slightly outperformed the GRNN and SVM models. In practical terms, however, the newly developed models (ANFIS and GRNN) are more robust and flexible, and have less parameters needed for selection and optimization in comparison with traditional ANN and SVM models. Consequently, they can be used as valuable tools to estimate forest carbon fluxes and fill the missing carbon flux data during the long-term EC measurements.


Author(s):  
Hossein Safarzadeh ◽  
Marco Leonesio ◽  
Giacomo Bianchi ◽  
Michele Monno

AbstractThis work proposes a model for suggesting optimal process configuration in plunge centreless grinding operations. Seven different approaches were implemented and compared: first principles model, neural network model with one hidden layer, support vector regression model with polynomial kernel function, Gaussian process regression model and hybrid versions of those three models. The first approach is based on an enhancement of the well-known numerical process simulation of geometrical instability. The model takes into account raw workpiece profile and possible wheel-workpiece loss of contact, which introduces an inherent limitation on the resulting profile waviness. Physical models, because of epistemic errors due to neglected or oversimplified functional relationships, can be too approximated for being considered in industrial applications. Moreover, in deterministic models, uncertainties affecting the various parameters are not explicitly considered. Complexity in centreless grinding models arises from phenomena like contact length dependency on local compliance, contact force and grinding wheel roughness, unpredicted material properties of the grinding wheel and workpiece, precision of the manual setup done by the operator, wheel wear and nature of wheel wear. In order to improve the overall model prediction accuracy and allow automated continuous learning, several machine learning techniques have been investigated: a Bayesian regularized neural network, an SVR model and a GPR model. To exploit the a priori knowledge embedded in physical models, hybrid models are proposed, where neural network, SVR and GPR models are fed by the nominal process parameters enriched with the roundness predicted by the first principle model. Those hybrid models result in an improved prediction capability.


2020 ◽  
Vol 17 (4) ◽  
pp. 1925-1930
Author(s):  
Ambeshwar Kumar ◽  
R. Manikandan ◽  
Robbi Rahim

It’s a new era technology in the field of medical engineering giving awareness about the various healthcare features. Deep learning is a part of machine learning, it is capable of handling high dimensional data and is efficient in concentrating on the right features. Tumor is an unbelievably complex disease: a multifaceted cell has more than hundred billion cells; each cell acquires mutation exclusively. Detection of tumor particles in experiment is easily done by MRI or CT. Brain tumors can also be detected by MRI, however, deep learning techniques give a better approach to segment the brain tumor images. Deep Learning models are imprecisely encouraged by information handling and communication designs in biological nervous system. Classification plays an significant role in brain tumor detection. Neural network is creating a well-organized rule for classification. To accomplish medical image data, neural network is trained to use the Convolution algorithm. Multilayer perceptron is intended for identification of a image. In this study article, the brain images are categorized into two types: normal and abnormal. This article emphasize the importance of classification and feature selection approach for predicting the brain tumor. This classification is done by machine learning techniques like Artificial Neural Networks, Support Vector Machine and Deep Neural Network. It could be noted that more than one technique can be applied for the segmentation of tumor. The several samples of brain tumor images are classified using deep learning algorithms, convolution neural network and multi-layer perceptron.


Computers ◽  
2019 ◽  
Vol 8 (1) ◽  
pp. 4 ◽  
Author(s):  
Jurgita Kapočiūtė-Dzikienė ◽  
Robertas Damaševičius ◽  
Marcin Woźniak

We describe the sentiment analysis experiments that were performed on the Lithuanian Internet comment dataset using traditional machine learning (Naïve Bayes Multinomial—NBM and Support Vector Machine—SVM) and deep learning (Long Short-Term Memory—LSTM and Convolutional Neural Network—CNN) approaches. The traditional machine learning techniques were used with the features based on the lexical, morphological, and character information. The deep learning approaches were applied on the top of two types of word embeddings (Vord2Vec continuous bag-of-words with negative sampling and FastText). Both traditional and deep learning approaches had to solve the positive/negative/neutral sentiment classification task on the balanced and full dataset versions. The best deep learning results (reaching 0.706 of accuracy) were achieved on the full dataset with CNN applied on top of the FastText embeddings, replaced emoticons, and eliminated diacritics. The traditional machine learning approaches demonstrated the best performance (0.735 of accuracy) on the full dataset with the NBM method, replaced emoticons, restored diacritics, and lemma unigrams as features. Although traditional machine learning approaches were superior when compared to the deep learning methods; deep learning demonstrated good results when applied on the small datasets.


Sign in / Sign up

Export Citation Format

Share Document