High Dimensional Brushing for Interactive Exploration of Multivariate Data

Author(s):  
A.R. Martin ◽  
M.O. Ward
Author(s):  
Lavanya K ◽  
L.S.S. Reddy ◽  
B. Eswara Reddy

Multiple imputations (MI) are predominantly applied in such processes that are involved in the transaction of huge chunks of missing data. Multivariate data that follow traditional statistical models undergoes great suffering for the inadequate availability of pertinent data. The field of distributed computing research faces the biggest hurdle in the form of insufficient high dimensional multivariate data. It mainly deals with the analysis of parallel input problems found in the cloud computing network in general and evaluation of high-performance computing in particular. In fact, it is a tough task to utilize parallel multiple input methods for accomplishing remarkable performance as well as allowing huge datasets achieves scale. In this regard, it is essential that a credible data system is developed and a decomposition strategy is used to partition workload in the entire process for minimum data dependence. Subsequently, a moderate synchronization and/or meager communication liability is followed for placing parallel impute methods for achieving scale as well as more processes. The present article proposes many novel applications for better efficiency. As the first step, this article suggests distributed-oriented serial regression multiple imputation for enhancing the efficiency of imputation task in high dimensional multivariate normal data. As the next step, the processes done in three diverse with parallel back ends viz. Multiple imputation that used the socket method to serve serial regression and the Fork Method to distribute work over workers, and also same work experiments in dynamic structure with a load balance mechanism. In the end, the set of distributed MI methods are used to experimentally analyze amplitude of imputation scores spanning across three probable scenarios in the range of 1:500. Further, the study makes an important observation that due to the efficiency of numerous imputation methods, the data is arranged proportionately in a missing range of 10% to 50%, low to high, while dealing with data between 1000 and 100,000 samples. The experiments are done in a cloud environment and demonstrate that it is possible to generate a decent speed by lessening the repetitive communication between processors.


2020 ◽  
Vol 34 (04) ◽  
pp. 6786-6794
Author(s):  
Lifeng Zhang

Detecting relationships among multivariate data is often of great importance in the analysis of high-dimensional data sets, and has received growing attention for decades from both academic and industrial fields. In this study, we propose a statistical tool named the neighbor correlation coefficient (nCor), which is based on a new idea that measures the local continuity of the reordered data points to quantify the strength of the global association between variables. With sufficient sample size, the new method is able to capture a wide range of functional relationship, whether it is linear or nonlinear, bivariate or multivariate, main effect or interaction. The score of nCor roughly approximates the coefficient of determination (R2) of the data which implies the proportion of variance in one variable that is predictable from one or more other variables. On this basis, three nCor based statistics are also proposed here to further characterize the intra and inter structures of the associations from the aspects of nonlinearity, interaction effect, and variable redundancy. The mechanisms of these measures are proved in theory and demonstrated with numerical analyses.


2021 ◽  
pp. 1471082X2110410
Author(s):  
Elena Tuzhilina ◽  
Leonardo Tozzi ◽  
Trevor Hastie

Canonical correlation analysis (CCA) is a technique for measuring the association between two multivariate data matrices. A regularized modification of canonical correlation analysis (RCCA) which imposes an [Formula: see text] penalty on the CCA coefficients is widely used in applications with high-dimensional data. One limitation of such regularization is that it ignores any data structure, treating all the features equally, which can be ill-suited for some applications. In this article we introduce several approaches to regularizing CCA that take the underlying data structure into account. In particular, the proposed group regularized canonical correlation analysis (GRCCA) is useful when the variables are correlated in groups. We illustrate some computational strategies to avoid excessive computations with regularized CCA in high dimensions. We demonstrate the application of these methods in our motivating application from neuroscience, as well as in a small simulation example.


Methodology ◽  
2011 ◽  
Vol 7 (4) ◽  
pp. 134-144 ◽  
Author(s):  
Ali Ünlü ◽  
Waqas Ahmed Malik

Gauguin (Grouping And Using Glyphs Uncovering Individual Nuances) is statistical data visualization software for the interactive graphical exploration of multivariate data using glyph representations. Glyphs are defined as geometric shapes scaled by the values of multivariate data. Each glyph represents one high-dimensional data point or the prototype (average) of a group or cluster of data points. This paper reviews the capabilities, functionality, and interactive properties of this software package. Key features of Gauguin are illustrated with data from the Programme for International Student Assessment.


2012 ◽  
Vol 67 (1) ◽  
pp. 81-99 ◽  
Author(s):  
M. Rauf Ahmad ◽  
Dietrich von Rosen ◽  
Martin Singull

2014 ◽  
Vol 33 (3) ◽  
pp. 101-110 ◽  
Author(s):  
S. Liu ◽  
B. Wang ◽  
P.-T. Bremer ◽  
V. Pascucci

Sign in / Sign up

Export Citation Format

Share Document