scholarly journals Variational Information Bottleneck for Semi-Supervised Classification

Entropy ◽  
2020 ◽  
Vol 22 (9) ◽  
pp. 943 ◽  
Author(s):  
Slava Voloshynovskiy ◽  
Olga Taran ◽  
Mouad Kondah ◽  
Taras Holotyak ◽  
Danilo Rezende

In this paper, we consider an information bottleneck (IB) framework for semi-supervised classification with several families of priors on latent space representation. We apply a variational decomposition of mutual information terms of IB. Using this decomposition we perform an analysis of several regularizers and practically demonstrate an impact of different components of variational model on the classification accuracy. We propose a new formulation of semi-supervised IB with hand crafted and learnable priors and link it to the previous methods such as semi-supervised versions of VAE (M1 + M2), AAE, CatGAN, etc. We show that the resulting model allows better understand the role of various previously proposed regularizers in semi-supervised classification task in the light of IB framework. The proposed IB semi-supervised model with hand-crafted and learnable priors is experimentally validated on MNIST under different amount of labeled data.

Informatics ◽  
2019 ◽  
Vol 6 (2) ◽  
pp. 17
Author(s):  
Athanasios Davvetas ◽  
Iraklis A. Klampanos ◽  
Spiros Skiadopoulos ◽  
Vangelis Karkaletsis

Evidence transfer for clustering is a deep learning method that manipulates the latent representations of an autoencoder according to external categorical evidence with the effect of improving a clustering outcome. Evidence transfer’s application on clustering is designed to be robust when introduced with a low quality of evidence, while increasing the effectiveness of the clustering accuracy during relevant corresponding evidence. We interpret the effects of evidence transfer on the latent representation of an autoencoder by comparing our method to the information bottleneck method. Information bottleneck is an optimisation problem of finding the best tradeoff between maximising the mutual information of data representations and a task outcome while at the same time being effective in compressing the original data source. We posit that the evidence transfer method has essentially the same objective regarding the latent representations produced by an autoencoder. We verify our hypothesis using information theoretic metrics from feature selection in order to perform an empirical analysis over the information that is carried through the bottleneck of the latent space. We use the relevance metric to compare the overall mutual information between the latent representations and the ground truth labels before and after their incremental manipulation, as well as, to study the effects of evidence transfer regarding the significance of each latent feature.


2018 ◽  
Vol 7 (4.5) ◽  
pp. 248 ◽  
Author(s):  
Syed Muzamil Basha ◽  
Dharmendra Singh Rajput ◽  
Ravi Kumar Poluru ◽  
S. Bharath Bhushan ◽  
Shaik Abdul Khalandar Basha

The classification task is to predict the value of the target variable from the values of the input variables. If a target is provided as part of the dataset, then classification is a supervised task. It is important to analysis the performance of supervised classification models before using them in classification task. In our research we would like to propose a novel way to evaluated the performance of supervised     classification models like Decision Tree and Naïve Bayes using KNIME Analytics platform. Experiments are conducted on Multi variant dataset consisting 58000 instances, 9 columns associated specially for classification, collected from UCI Machine learning repositories  (http://archive.ics.uci.edu/ml/datasets/statlog+(shuttle)) and compared the performance of both the models in terms of Classification  Accuracy (CA) and Error Rate. Finally, validated both the models using Metric precision, recall and F-measure. In our finding, we found that  Decision tree acquires CA (99.465%) where as Naïve Bayes attain CA (90.358%). The F-measure of Decision tree is 0.984, whereas Naïve Bayes acquire 0.7045.  


2021 ◽  
Vol 2021 (7) ◽  
Author(s):  
Dipankar Barman ◽  
Subhajit Barman ◽  
Bibhas Ranjan Majhi

Abstract We investigate the effects of field temperature T(f) on the entanglement harvesting between two uniformly accelerated detectors. For their parallel motion, the thermal nature of fields does not produce any entanglement, and therefore, the outcome is the same as the non-thermal situation. On the contrary, T(f) affects entanglement harvesting when the detectors are in anti-parallel motion, i.e., when detectors A and B are in the right and left Rindler wedges, respectively. While for T(f) = 0 entanglement harvesting is possible for all values of A’s acceleration aA, in the presence of temperature, it is possible only within a narrow range of aA. In (1 + 1) dimensions, the range starts from specific values and extends to infinity, and as we increase T(f), the minimum required value of aA for entanglement harvesting increases. Moreover, above a critical value aA = ac harvesting increases as we increase T(f), which is just opposite to the accelerations below it. There are several critical values in (1 + 3) dimensions when they are in different accelerations. Contrary to the single range in (1 + 1) dimensions, here harvesting is possible within several discrete ranges of aA. Interestingly, for equal accelerations, one has a single critical point, with nature quite similar to (1 + 1) dimensional results. We also discuss the dependence of mutual information among these detectors on aA and T(f).


2021 ◽  
Vol 15 ◽  
pp. 174830262110249
Author(s):  
Cong-Zhe You ◽  
Zhen-Qiu Shu ◽  
Hong-Hui Fan

Recently, in the area of artificial intelligence and machine learning, subspace clustering of multi-view data is a research hotspot. The goal is to divide data samples from different sources into different groups. We proposed a new subspace clustering method for multi-view data which termed as Non-negative Sparse Laplacian regularized Latent Multi-view Subspace Clustering (NSL2MSC) in this paper. The method proposed in this paper learns the latent space representation of multi view data samples, and performs the data reconstruction on the latent space. The algorithm can cluster data in the latent representation space and use the relationship of different views. However, the traditional representation-based method does not consider the non-linear geometry inside the data, and may lose the local and similar information between the data in the learning process. By using the graph regularization method, we can not only capture the global low dimensional structural features of data, but also fully capture the nonlinear geometric structure information of data. The experimental results show that the proposed method is effective and its performance is better than most of the existing alternatives.


MIS Quarterly ◽  
2021 ◽  
Vol 45 (3) ◽  
pp. 1113-1148
Author(s):  
Angela Xia Liu ◽  
◽  
Yilin Li ◽  
Sean Xu ◽  
◽  
...  

This work examines the question of who is more likely to provide future helpful reviews in the context of online product reviews by synergistically using personality theories and data analytics. It trains a deep learning model to infer a reviewer’s personality traits. This enables analyses to reveal the role of personality traits in review helpfulness among a large population of reviewers. We develop hypotheses on how personality traits are associated with review helpfulness, followed by hypotheses testing that confirms that higher review helpfulness is related to higher openness, conscientiousness, extraversion, and agreeableness and to lower emotional stability. These results suggest the appropriateness of using these five personality traits as inputs for developing a model for predicting future review helpfulness. Based on an ensemble model using supervised classification algorithms, we develop a predictive model and demonstrate its superior performance. Theoretical and practical implications are discussed.


Author(s):  
Gang Liu ◽  
Chunlei Yang ◽  
Sen Liu ◽  
Chunbao Xiao ◽  
Bin Song

A feature selection method based on mutual information and support vector machine (SVM) is proposed in order to eliminate redundant feature and improve classification accuracy. First, local correlation between features and overall correlation is calculated by mutual information. The correlation reflects the information inclusion relationship between features, so the features are evaluated and redundant features are eliminated with analyzing the correlation. Subsequently, the concept of mean impact value (MIV) is defined and the influence degree of input variables on output variables for SVM network based on MIV is calculated. The importance weights of the features described with MIV are sorted by descending order. Finally, the SVM classifier is used to implement feature selection according to the classification accuracy of feature combination which takes MIV order of feature as a reference. The simulation experiments are carried out with three standard data sets of UCI, and the results show that this method can not only effectively reduce the feature dimension and high classification accuracy, but also ensure good robustness.


1994 ◽  
Vol 09 (18) ◽  
pp. 3245-3282 ◽  
Author(s):  
B.A. HARRIS ◽  
G.C. JOSHI

Recent developments in quantum gravity suggest that wormholes may influence the observed values of the constants of nature. The Euclidean formulation of quantum gravity predicts that wormholes induce a probability distribution in the space of possible fundamental constants. In particular, the effective action on a large spherical space may lead to the vanishing of the cosmological constant and possibly determine the values of other constants of nature. The ability to perform calculations involving interacting quantum fields, particularly non-Abelian models, on a four-sphere is vital if one is to investigate this possibility. In this paper we present a new formulation of field theory on a four-sphere using the angular momentum space representation of SO(5). We give a review of field theory on a sphere and then show how a matrix element prescription in angular momentum space and a new summation technique based on the complex l plane, overcome previous limitations in calculational techniques. The standard one-loop graphs of QED are given as examples.


Author(s):  
Wenqi Zhao ◽  
Satoshi Oyama ◽  
Masahito Kurihara

Counterfactual explanations help users to understand the behaviors of machine learning models by changing the inputs for the existing outputs. For an image classification task, an example counterfactual visual explanation explains: "for an example that belongs to class A, what changes do we need to make to the input so that the output is more inclined to class B." Our research considers changing the attribute description text of class A on the basis of the attributes of class B and generating counterfactual images on the basis of the modified text. We can use the prediction results of the model on counterfactual images to find the attributes that have the greatest effect when the model is predicting classes A and B. We applied our method to a fine-grained image classification dataset and used the generative adversarial network to generate natural counterfactual visual explanations. To evaluate these explanations, we used them to assist crowdsourcing workers in an image classification task. We found that, within a specific range, they improved classification accuracy.


Sign in / Sign up

Export Citation Format

Share Document