Multinomial Data

2005 ◽  
pp. 106-123
Keyword(s):  
1990 ◽  
Vol 29 (03) ◽  
pp. 200-204 ◽  
Author(s):  
J. A. Koziol

AbstractA basic problem of cluster analysis is the determination or selection of the number of clusters evinced in any set of data. We address this issue with multinomial data using Akaike’s information criterion and demonstrate its utility in identifying an appropriate number of clusters of tumor types with similar profiles of cell surface antigens.


2019 ◽  
Author(s):  
Mark Andrews

A Gibbs sampler for the hierarchical Dirichlet process mixture model (HDPMM) when used with multinomial data.


Algorithms ◽  
2021 ◽  
Vol 14 (10) ◽  
pp. 296
Author(s):  
Lucy Blondell ◽  
Mark Z. Kos ◽  
John Blangero ◽  
Harald H. H. Göring

Statistical analysis of multinomial data in complex datasets often requires estimation of the multivariate normal (mvn) distribution for models in which the dimensionality can easily reach 10–1000 and higher. Few algorithms for estimating the mvn distribution can offer robust and efficient performance over such a range of dimensions. We report a simulation-based comparison of two algorithms for the mvn that are widely used in statistical genetic applications. The venerable Mendell-Elston approximation is fast but execution time increases rapidly with the number of dimensions, estimates are generally biased, and an error bound is lacking. The correlation between variables significantly affects absolute error but not overall execution time. The Monte Carlo-based approach described by Genz returns unbiased and error-bounded estimates, but execution time is more sensitive to the correlation between variables. For ultra-high-dimensional problems, however, the Genz algorithm exhibits better scale characteristics and greater time-weighted efficiency of estimation.


2013 ◽  
Vol 41 (4) ◽  
pp. 701-715 ◽  
Author(s):  
Andrew Cron ◽  
Liang Zhang ◽  
Deepak Agarwal

Author(s):  
Shijia Wang ◽  
Liangliang Wang ◽  
Tim B. Swartz
Keyword(s):  

2019 ◽  
Vol 38 (25) ◽  
pp. 4963-4976
Author(s):  
V. Landsman ◽  
D. Landsman ◽  
C.S. Li ◽  
H. Bang
Keyword(s):  

1997 ◽  
Vol 33 (1) ◽  
pp. 41-48 ◽  
Author(s):  
Marc Aerts ◽  
Ilse Augustyns ◽  
Paul Janssen
Keyword(s):  

Biometrics ◽  
2019 ◽  
Vol 76 (3) ◽  
pp. 834-842
Author(s):  
Farzana Afroz ◽  
Matt Parry ◽  
David Fletcher
Keyword(s):  

1989 ◽  
Vol 6 (1) ◽  
pp. 73-95 ◽  
Author(s):  
Pascale Rousseau ◽  
David Sankoff
Keyword(s):  

2019 ◽  
Vol 09 (03) ◽  
pp. 2050008
Author(s):  
Xiaona Yang ◽  
Zhaojun Wang ◽  
Xuemin Zi

This paper develops an outlier detection procedure for multinomial data when the number of categories tends to infinity. Most of the outlier detection methods are based on the assumption that the observations follow multivariate normal distribution, while in many modern applications, the observations either are measured on a discrete scale or naturally have some categorical structures. For such multinomial observations, there are rather limited approaches for outlier detection. To overcome the main obstacle, the least trimmed distances estimator for multinomial data and a fast algorithm to identify the clean subset are introduced in this work. Also, a threshold rule is considered through the asymptotic distribution of measure distance to identify outliers. Furthermore, a one-step reweighting scheme is proposed to improve the efficiency of the procedure. Finally, the finite sample performance of our method is evaluated through simulations and is compared with that of available outlier detection methods.


Sign in / Sign up

Export Citation Format

Share Document