Tennessee Eastman Process Diagnosis Based on Dynamic Classification With SVDD

Author(s):  
Foued Theljani ◽  
Kaouther Laabidi ◽  
Salah Zidi ◽  
Moufida Ksouri

The support vector domain description (SVDD) is an efficient kernel method inspired from the SV machine (SVM) by Vapnik. It is commonly used for one-classification problems or novelty detection. The training algorithm solves a constrained convex quadratic programming (QP) problem. This assumes prior dense sampling (offline training) and it requires large memory and enormous amounts of training time. In this paper, we propose a fast SVDD dedicated for multiclassification problems. The proposed classifier deals with stationary as well as nonstationary (NS) data. The principle is based on the dynamic removal/insertion of informations according to adequate rules. To ensure the rapidity of convergence, the algorithm considers in each run a limited frame of samples for the training process. These samples are selected according to some approximations based on Karush–Kuhn–Tucker (KKT) conditions. An additional merge mechanism is proposed to avoid local optima drawbacks and improve performances. The developed method is assessed on some synthetic data to prove its effectiveness. Afterward, it is employed to solve a diagnosis problem and faults detection. We considered for this purpose a real industrial plant consisting in Tennessee Eastman process (TEP).

Author(s):  
Thế Cường Nguyễn ◽  
Thanh Vi Nguyen

In binary classification problems, two classes of data seem to be different from each other. It is expected to bemore complicated due to the number of data points of clusters in each class also be different. Traditional algorithmsas Support Vector Machine (SVM), Twin Support Vector Machine (TSVM), or Least Square Twin Support VectorMachine (LSTSVM) cannot sufficiently exploit information about the number of data points in each cluster of the data.Which may be effect to the accuracy of classification problems. In this paper, we propose a new Improvement LeastSquare - Support Vector Machine (called ILS-SVM) for binary classification problems with a class-vs-clusters strategy.Experimental results show that the ILS-SVM training time is faster than that of TSVM, and the ILS-SVM accuracy isbetter than LSTSVM and TSVM in most cases.


2021 ◽  
pp. 1-17
Author(s):  
Ming-Ai Li ◽  
Ruo-Tu Wang ◽  
Li-Na Wei

BACKGROUND: Motor imagery electroencephalogram (MI-EEG) play an important role in the field of neurorehabilitation, and a fuzzy support vector machine (FSVM) is one of the most used classifiers. Specifically, a fuzzy c-means (FCM) algorithm was used to membership calculation to deal with the classification problems with outliers or noises. However, FCM is sensitive to its initial value and easily falls into local optima. OBJECTIVE: The joint optimization of genetic algorithm (GA) and FCM is proposed to enhance robustness of fuzzy memberships to initial cluster centers, yielding an improved FSVM (GF-FSVM). METHOD: The features of each channel of MI-EEG are extracted by the improved refined composite multivariate multiscale fuzzy entropy and fused to form a feature vector for a trial. Then, GA is employed to optimize the initial cluster center of FCM, and the fuzzy membership degrees are calculated through an iterative process and further applied to classify two-class MI-EEGs. RESULTS: Extensive experiments are conducted on two publicly available datasets, the average recognition accuracies achieve 99.89% and 98.81% and the corresponding kappa values are 0.9978 and 0.9762, respectively. CONCLUSION: The optimized cluster centers of FCM via GA are almost overlapping, showing great stability, and GF-FSVM obtains higher classification accuracies and higher consistency as well.


2021 ◽  
Vol 37 (1) ◽  
pp. 43-56
Author(s):  
Nguyen The Cuong ◽  
Huynh The Phung

In binary classification problems, two classes of data seem to be different from each other. It is expected to be more complicated due to the clusters in each class also tend to be different. Traditional algorithms as Support Vector Machine (SVM) or Twin Support Vector Machine (TWSVM) cannot sufficiently exploit structural information with cluster granularity of the data, cause limitation on the capability of simulation of data trends. Structural Twin Support Vector Machine (S-TWSVM) sufficiently exploits structural information with cluster granularity for learning a represented hyperplane. Therefore, the capability of S-TWSVM’s data simulation is better than that of TWSVM. However, for the datasets where each class consists of clusters of different trends, the S-TWSVM’s data simulation capability seems restricted. Besides, the training time of S-TWSVM has not been improved compared to TWSVM. This paper proposes a new Weighted Structural - Support Vector Machine (called WS-SVM) for binary classification problems with a class-vs-clusters strategy. Experimental results show that WS-SVM could describe the tendency of the distribution of cluster information. Furthermore, both the theory and experiment show that the training time of the WS-SVM for classification problem has significantly improved compared to S-TWSVM.


Author(s):  
Thanh Vi Nguyen ◽  
Thế Cường Nguyễn

n binary classification problems, two classes of data seem tobe different from each other. It is expected to be more complicated dueto the number of data points of clusters in each class also be different.Traditional algorithms as Support Vector Machine (SVM), Twin Support Vector Machine (TSVM), or Least Square Twin Support VectorMachine (LSTSVM) cannot sufficiently exploit information about thenumber of data points in each cluster of the data. Which may be effectto the accuracy of classification problems. In this paper, we proposes anew Improved Least Square - Support Vector Machine (called ILS-SVM)for binary classification problems with a class-vs-clusters strategy. Experimental results show that the ILS-SVM training time is faster thanthat of TSVM, and the ILS-SVM accuracy is better than LSTSVM andTSVM in most cases.


2016 ◽  
Vol 25 (01) ◽  
pp. 1550026 ◽  
Author(s):  
Juan J. Carrasco ◽  
Mónica Millán-Giraldo ◽  
Juan Caravaca ◽  
Pablo Escandell-Montero ◽  
José M. Martínez-Martínez ◽  
...  

Extreme Learning Machine (ELM) is a recently proposed algorithm, efficient and fast for learning the parameters of single layer neural structures. One of the main problems of this algorithm is to choose the optimal architecture for a given problem solution. To solve this limitation several solutions have been proposed in the literature, including the regularization of the structure. However, to the best of our knowledge, there are no works where such adjustment is applied to classification problems in the presence of a non-linearity in the output; all published works tackle modelling or regression problems. Our proposal has been applied to a series of standard databases for the evaluation of machine learning techniques. Results obtained in terms of classification success rate and training time, are compared to the original ELM, to the well known Least Square Support Vector Machine (LS-SVM) algorithm and with two other methods based on the ELM regularization: Optimally Pruned Extreme Learning Machine (OP-ELM) and Bayesian Extreme Learning Machine (BELM). The obtained results clearly demonstrate the usefulness of the proposed method and its superiority over a classical approach.


2018 ◽  
Vol 41 (10) ◽  
pp. 2687-2698 ◽  
Author(s):  
Hajer Lahdhiri ◽  
Khaoula Ben Abdellafou ◽  
Okba Taouali ◽  
Majdi Mansouri ◽  
Ouajdi Korbaa

Process monitoring is an integral part of chemical process, required higher product quality and safety operation. Therefore, the objective of this paper is to ensure the suitable functioning and to improve the fault detection performance of conventional kernel Principal Components Analysis (KPCA). Thus, an online Reduced Rank KPCA (OnRR-KPCA) with adaptive model has been developed to monitor a dynamic nonlinear process. The developed method is proposed. Firstly, to extract the useful observations, from large amount of training data registered in normal operating conditions, in order to construct the reduced reference model. Secondly, to monitor the process online and update the reference model if a new useful observation is available and satisfies the condition of independencies between variables in feature space. To demonstrate the effectiveness of the OnRR-KPCA with adaptive model over the conventional KPCA and the RR-KPCA, the fault detection performances are illustrated through two examples: one using synthetic data, the second using a simulated Tennessee Eastman Process (TEP) data.


2017 ◽  
Vol 28 (02) ◽  
pp. 1750015 ◽  
Author(s):  
M. Andrecut

The least-squares support vector machine (LS-SVM) is a frequently used kernel method for non-linear regression and classification tasks. Here we discuss several approximation algorithms for the LS-SVM classifier. The proposed methods are based on randomized block kernel matrices, and we show that they provide good accuracy and reliable scaling for multi-class classification problems with relatively large data sets. Also, we present several numerical experiments that illustrate the practical applicability of the proposed methods.


Author(s):  
Anantvir Singh Romana

Accurate diagnostic detection of the disease in a patient is critical and may alter the subsequent treatment and increase the chances of survival rate. Machine learning techniques have been instrumental in disease detection and are currently being used in various classification problems due to their accurate prediction performance. Various techniques may provide different desired accuracies and it is therefore imperative to use the most suitable method which provides the best desired results. This research seeks to provide comparative analysis of Support Vector Machine, Naïve bayes, J48 Decision Tree and neural network classifiers breast cancer and diabetes datsets.


2021 ◽  
Vol 11 (9) ◽  
pp. 3863
Author(s):  
Ali Emre Öztürk ◽  
Ergun Erçelebi

A large amount of training image data is required for solving image classification problems using deep learning (DL) networks. In this study, we aimed to train DL networks with synthetic images generated by using a game engine and determine the effects of the networks on performance when solving real-image classification problems. The study presents the results of using corner detection and nearest three-point selection (CDNTS) layers to classify bird and rotary-wing unmanned aerial vehicle (RW-UAV) images, provides a comprehensive comparison of two different experimental setups, and emphasizes the significant improvements in the performance in deep learning-based networks due to the inclusion of a CDNTS layer. Experiment 1 corresponds to training the commonly used deep learning-based networks with synthetic data and an image classification test on real data. Experiment 2 corresponds to training the CDNTS layer and commonly used deep learning-based networks with synthetic data and an image classification test on real data. In experiment 1, the best area under the curve (AUC) value for the image classification test accuracy was measured as 72%. In experiment 2, using the CDNTS layer, the AUC value for the image classification test accuracy was measured as 88.9%. A total of 432 different combinations of trainings were investigated in the experimental setups. The experiments were trained with various DL networks using four different optimizers by considering all combinations of batch size, learning rate, and dropout hyperparameters. The test accuracy AUC values for networks in experiment 1 ranged from 55% to 74%, whereas the test accuracy AUC values in experiment 2 networks with a CDNTS layer ranged from 76% to 89.9%. It was observed that the CDNTS layer has considerable effects on the image classification accuracy performance of deep learning-based networks. AUC, F-score, and test accuracy measures were used to validate the success of the networks.


2021 ◽  
Vol 11 (9) ◽  
pp. 4280
Author(s):  
Iurii Katser ◽  
Viacheslav Kozitsin ◽  
Victor Lobachev ◽  
Ivan Maksimov

Offline changepoint detection (CPD) algorithms are used for signal segmentation in an optimal way. Generally, these algorithms are based on the assumption that signal’s changed statistical properties are known, and the appropriate models (metrics, cost functions) for changepoint detection are used. Otherwise, the process of proper model selection can become laborious and time-consuming with uncertain results. Although an ensemble approach is well known for increasing the robustness of the individual algorithms and dealing with mentioned challenges, it is weakly formalized and much less highlighted for CPD problems than for outlier detection or classification problems. This paper proposes an unsupervised CPD ensemble (CPDE) procedure with the pseudocode of the particular proposed ensemble algorithms and the link to their Python realization. The approach’s novelty is in aggregating several cost functions before the changepoint search procedure running during the offline analysis. The numerical experiment showed that the proposed CPDE outperforms non-ensemble CPD procedures. Additionally, we focused on analyzing common CPD algorithms, scaling, and aggregation functions, comparing them during the numerical experiment. The results were obtained on the two anomaly benchmarks that contain industrial faults and failures—Tennessee Eastman Process (TEP) and Skoltech Anomaly Benchmark (SKAB). One of the possible applications of our research is the estimation of the failure time for fault identification and isolation problems of the technical diagnostics.


Sign in / Sign up

Export Citation Format

Share Document