scholarly journals Correlation Kernels for Support Vector Machines Classification with Applications in Cancer Data

2012 ◽  
Vol 2012 ◽  
pp. 1-7 ◽  
Author(s):  
Hao Jiang ◽  
Wai-Ki Ching

High dimensional bioinformatics data sets provide an excellent and challenging research problem in machine learning area. In particular, DNA microarrays generated gene expression data are of high dimension with significant level of noise. Supervised kernel learning with an SVM classifier was successfully applied in biomedical diagnosis such as discriminating different kinds of tumor tissues. Correlation Kernel has been recently applied to classification problems with Support Vector Machines (SVMs). In this paper, we develop a novel and parsimonious positive semidefinite kernel. The proposed kernel is shown experimentally to have better performance when compared to the usual correlation kernel. In addition, we propose a new kernel based on the correlation matrix incorporating techniques dealing with indefinite kernel. The resulting kernel is shown to be positive semidefinite and it exhibits superior performance to the two kernels mentioned above. We then apply the proposed method to some cancer data in discriminating different tumor tissues, providing information for diagnosis of diseases. Numerical experiments indicate that our method outperforms the existing methods such as the decision tree method and KNN method.

2020 ◽  
Vol 2020 ◽  
pp. 1-7
Author(s):  
Nalindren Naicker ◽  
Timothy Adeliyi ◽  
Jeanette Wing

Educational Data Mining (EDM) is a rich research field in computer science. Tools and techniques in EDM are useful to predict student performance which gives practitioners useful insights to develop appropriate intervention strategies to improve pass rates and increase retention. The performance of the state-of-the-art machine learning classifiers is very much dependent on the task at hand. Investigating support vector machines has been used extensively in classification problems; however, the extant of literature shows a gap in the application of linear support vector machines as a predictor of student performance. The aim of this study was to compare the performance of linear support vector machines with the performance of the state-of-the-art classical machine learning algorithms in order to determine the algorithm that would improve prediction of student performance. In this quantitative study, an experimental research design was used. Experiments were set up using feature selection on a publicly available dataset of 1000 alpha-numeric student records. Linear support vector machines benchmarked with ten categorical machine learning algorithms showed superior performance in predicting student performance. The results of this research showed that features like race, gender, and lunch influence performance in mathematics whilst access to lunch was the primary factor which influences reading and writing performance.


2003 ◽  
Vol 15 (7) ◽  
pp. 1667-1689 ◽  
Author(s):  
S. Sathiya Keerthi ◽  
Chih-Jen Lin

Support vector machines (SVMs) with the gaussian (RBF) kernel have been popular for practical use. Model selection in this class of SVMs involves two hyper parameters: the penalty parameter C and the kernel width σ. This letter analyzes the behavior of the SVM classifier when these hyper parameters take very small or very large values. Our results help in understanding the hyperparameter space that leads to an efficient heuristic method of searching for hyperparameter values with small generalization errors. The analysis also indicates that if complete model selection using the gaussian kernel has been conducted, there is no need to consider linear SVM.


2013 ◽  
Vol 336-338 ◽  
pp. 277-280 ◽  
Author(s):  
Tian Lai Xu

The combination of Inertial Navigation System (INS) and Global Positioning System (GPS) provides superior performance in comparison with either a stand-alone INS or GPS. However, the positioning accuracy of INS/GPS deteriorates with time in the absence of GPS signals. A least squares support vector machines (LS-SVM) regression algorithm is applied to INS/GPS integrated navigation system to bridge the GPS outages to achieve seamless navigation. In this method, LS-SVM is trained to model the errors of INS when GPS is available. Once the LS-SVM is properly trained in the training phase, its prediction can be used to correct the INS errors during GPS outages. Simulations in INS/GPS integrated navigation showed improvements in positioning accuracy when GPS outages occur.


2013 ◽  
Vol 333-335 ◽  
pp. 1080-1084
Author(s):  
Zhang Fei ◽  
Ye Xi

In this paper, we will propose a novel classification method of high-resolution SAR using local autocorrelation and Support Vector Machines (SVM) classifier. The commonly applied spatial autocorrelation indexes, called Moran's Index; Geary's Index, Getis's Index, will be used to depict the feature of the land-cover. Then, the SVM based on these indexes will be applied as the high-resolution SAR classifier. A Cosmo-SkyMed scene in ChengDu city, China is used for our experiment. It is shown that the method proposed can lead to good classification accuracy.


2015 ◽  
Vol 24 (03) ◽  
pp. 1550010 ◽  
Author(s):  
Yassine Ben Ayed

In this paper, we propose an alternative keyword spotting method relying on confidence measures and support vector machines. Confidence measures are computed from phone information provided by a Hidden Markov Model based speech recognizer. We use three kinds of techniques, i.e., arithmetic, geometric and harmonic means to compute a confidence measure for each word. The acceptance/rejection decision of a word is based on the confidence vector processed by the SVM classifier for which we propose a new Beta kernel. The performance of the proposed SVM classifier is compared with spotting methods based on some confidence means. Experimental results presented in this paper show that the proposed SVM classifier method improves the performances of the keyword spotting system.


Author(s):  
KWANG IN KIM ◽  
JIN HYUNG KIM ◽  
KEECHUL JUNG

This paper presents a real-time face recognition system. For the system to be real time, no external time-consuming feature extraction method is used, rather the gray-level values of the raw pixels that make up the face pattern are fed directly to the recognizer. In order to absorb the resulting high dimensionality of the input space, support vector machines (SVMs), which are known to work well even in high-dimensional space, are used as the face recognizer. Furthermore, a modified form of polynomial kernel (local correlation kernel) is utilized to take account of prior knowledge about facial structures and is used as the alternative feature extractor. Since SVMs were originally developed for two-class classification, their basic scheme is extended for multiface recognition by adopting one-per-class decomposition. In order to make a final classification from several one-per-class SVM outputs, a neural network (NN) is used as the arbitrator. Experiments with ORL database show a recognition rate of 97.9% and speed of 0.22 seconds per face with 40 classes.


2021 ◽  
Vol 3 (3) ◽  
pp. 482-499
Author(s):  
Roberto Leporini ◽  
Davide Pastorello

We analyze possible connections between quantum-inspired classifications and support vector machines. Quantum state discrimination and optimal quantum measurement are useful tools for classification problems. In order to use these tools, feature vectors have to be encoded in quantum states represented by density operators. Classification algorithms inspired by quantum state discrimination and implemented on classic computers have been recently proposed. We focus on the implementation of a known quantum-inspired classifier based on Helstrom state discrimination showing its connection with support vector machines and how to make the classification more efficient in terms of space and time acting on quantum encoding. In some cases, traditional methods provide better results. Moreover, we discuss the quantum-inspired nearest mean classification.


Author(s):  
Cecilio Angulo ◽  
Luis Gonzalez-Abril

Support Vector Machines -- SVMs -- are learning machines, originally designed for bi-classification problems, implementing the well-known Structural Risk Minimization (SRM) inductive principle to obtain good generalization on a limited number of learning patterns (Vapnik, 1998). The optimization criterion for these machines is maximizing the margin between two classes, i.e. the distance between two parallel hyperplanes that split the vectors of each one of the two classes, since larger is the margin separating classes, smaller is the VC dimension of the learning machine, which theoretically ensures a good generalization performance (Vapnik, 1998), as it has been demonstrated in a number of real applications (Cristianini, 2000). In its formulation is applicable the kernel trick, which improves the capacity of these algorithms, learning not being directly performed in the original space of data but in a new space called feature space; for this reason this algorithm is one of the most representative of the called Kernel Machines (KMs). Main theory was originally developed on the sixties and seventies by V. Vapnik and A. Chervonenkis (Vapnik et al., 1963, Vapnik et al., 1971, Vapnik, 1995, Vapnik, 1998), on the basis of a separable binary classification problem, however generalization in the use of these learning algorithms did not take place until the nineties (Boser et al., 1992). SVMs has been used thoroughly in any kind of learning problems, mainly in classification problems, although also in other problems like regression (Schölkopf et al., 2004) or clustering (Ben-Hur et al., 2001). The fields of Optic Character Recognition (Cortes et al., 1995) and Text Categorization (Sebastiani, 2002) were the most important initial applications where SVMs were used. With the extended application of new kernels, novel applications have taken place in the field of Bioinformatics, concretely many works are related with the classification of data in Genetic Expression (Microarray Gene Expression) (Brown et al., 1997) and detecting structures between proteins and their relationship with the chains of DNA (Jaakkola et al., 2000). Other applications include image identification, voice recognition, prediction in time series, etc. A more extensive list of applications can be found in (Guyon, 2006).


Sign in / Sign up

Export Citation Format

Share Document