Missing phenotype data imputation in pedigree data analysis

2008 ◽  
Vol 32 (1) ◽  
pp. 52-60 ◽  
Author(s):  
Brooke L. Fridley ◽  
Mariza de Andrade
Genetics ◽  
2003 ◽  
Vol 164 (4) ◽  
pp. 1561-1566
Author(s):  
Sharon Browning

AbstractWe propose a new method for calculating probabilities for pedigree genetic data that incorporates crossover interference using the chi-square models. Applications include relationship inference, genetic map construction, and linkage analysis. The method is based on importance sampling of unobserved inheritance patterns conditional on the observed genotype data and takes advantage of fast algorithms for no-interference models while using reweighting to allow for interference. We show that the method is effective for arbitrarily many markers with small pedigrees.


2002 ◽  
Vol 22 (3) ◽  
pp. 221-232 ◽  
Author(s):  
Mariza de Andrade ◽  
René Guéguen ◽  
Sophie Visvikis ◽  
Catherine Sass ◽  
Gérard Siest ◽  
...  

Genome ◽  
2020 ◽  
Vol 63 (11) ◽  
pp. 577-581
Author(s):  
Davoud Torkamaneh ◽  
Jérôme Laroche ◽  
François Belzile

Genotyping-by-sequencing (GBS) is a rapid, flexible, low-cost, and robust genotyping method that simultaneously discovers variants and calls genotypes within a broad range of samples. These characteristics make GBS an excellent tool for many applications and research questions from conservation biology to functional genomics in both model and non-model species. Continued improvement of GBS relies on a more comprehensive understanding of data analysis, development of fast and efficient bioinformatics pipelines, accurate missing data imputation, and active post-release support. Here, we present the second generation of Fast-GBS (v2.0) that offers several new options (e.g., processing paired-end reads and imputation of missing data) and features (e.g., summary statistics of genotypes) to improve the GBS data analysis process. The performance assessment analysis showed that Fast-GBS v2.0 outperformed other available analytical pipelines, such as GBS-SNP-CROP and Gb-eaSy. Fast-GBS v2.0 provides an analysis platform that can be run with different types of sequencing data, modest computational resources, and allows for missing-data imputation for various species in different contexts.


2015 ◽  
Vol 9s3 ◽  
pp. BBI.S29473 ◽  
Author(s):  
William Seffens ◽  
Chad Evans ◽  

Health-care initiatives are pushing the development and utilization of clinical data for medical discovery and translational research studies. Machine learning tools implemented for Big Data have been applied to detect patterns in complex diseases. This study focuses on hypertension and examines phenotype data across a major clinical study called Minority Health Genomics and Translational Research Repository Database composed of self-reported African American (AA) participants combined with related cohorts. Prior genome-wide association studies for hypertension in AAs presumed that an increase of disease burden in susceptible populations is due to rare variants. But genomic analysis of hypertension, even those designed to focus on rare variants, has yielded marginal genome-wide results over many studies. Machine learning and other nonparametric statistical methods have recently been shown to uncover relationships in complex phenotypes, genotypes, and clinical data. We trained neural networks with phenotype data for missing-data imputation to increase the usable size of a clinical data set. Validity was established by showing performance effects using the expanded data set for the association of phenotype variables with case/control status of patients. Data mining classification tools were used to generate association rules.


2019 ◽  
Vol 12 (5) ◽  
pp. 404-411
Author(s):  
Mahdi Akbarzadeh ◽  
Abbas Moghimbeigi ◽  
Nathan Morris ◽  
Maryam S. Daneshpour ◽  
Hossein Mahjub ◽  
...  

2015 ◽  
Vol 6 ◽  
Author(s):  
Md. Matiur Rahaman ◽  
Dijun Chen ◽  
Zeeshan Gillani ◽  
Christian Klukas ◽  
Ming Chen

2001 ◽  
Vol 28 (2) ◽  
pp. 313 ◽  
Author(s):  
Douglas A. Wolf

Microsimulation is well known as a tool for static analysis of tax and transfer policies, for the generation of programmatic cost estimates, and dynamic analyses of socio-economic and demographic systems. However, microsimulation also has the potential to contribute to longitudinal data analysis in several ways, including extending the range of outputs generated by a model, addressing several defective-data problems, and serving as a vehicle for missing-data imputation. This paper discusses microsimulation procedures suitable for several commonly-used statistical models applied to longitudinal data. It also addresses the unique role that can be played by microsimulation in longitudinal data analysis, and the problem of accounting for the several sources of variability associated with microsimulation procedures.


2013 ◽  
Vol 1 (2) ◽  
Author(s):  
Alma Molytė ◽  
Vaidutis Kučinskas ◽  
Aušra Matulevičienė ◽  
Eglė Preikšaitienė
Keyword(s):  

2005 ◽  
Vol 29 (1) ◽  
pp. 12-22 ◽  
Author(s):  
Tao Wang ◽  
Robert C. Elston

Sign in / Sign up

Export Citation Format

Share Document