scholarly journals Comparative Pathway Integrator: a framework of meta-analytic integration of multiple transcriptomic studies for consensual and differential pathway analysis

2018 ◽  
Author(s):  
Xiangrui Zeng ◽  
Zhou Fang ◽  
Tianzhou Ma ◽  
Chien-Wei Lin ◽  
George C. Tseng

AbstractMotivationPathway analysis provides a knowledge-driven approach to interpret differentially expressed genes associated with disease status. Many tools have been developed to analyze a single study. When multiple studies of different conditions are jointly analyzed, novel integrative tools are needed. In addition, pathway redundancy issue introduced by combining public pathway databases hinders knowledge discovery.Methods and ResultsWe present a meta-analytic integration tool, Comparative Pathway Integrator (CPI), to address these issues using adaptively weighted Fisher’s method to discover consensual and differential enrichment patterns, consensus clustering to reduce pathway redundancy, and a novel text mining algorithm to assist interpretation of the pathway clusters. We applied CPI to jointly analyze six psychiatric disorder transcriptomic studies to demonstrate its effectiveness, and found functions confirmed by previous biological studies as well novel enrichment patterns.AvailabilityCPI is accessible online: http://tsenglab.biostat.pitt.edu/[email protected]

Genes ◽  
2020 ◽  
Vol 11 (6) ◽  
pp. 696
Author(s):  
Xiangrui Zeng ◽  
Wei Zong ◽  
Chien-Wei Lin ◽  
Zhou Fang ◽  
Tianzhou Ma ◽  
...  

Pathway enrichment analysis provides a knowledge-driven approach to interpret differentially expressed genes associated with disease status. Many tools have been developed to analyze a single study. However, when multiple studies of different conditions are jointly analyzed, novel integrative tools are needed. In addition, pathway redundancy introduced by combining multiple public pathway databases hinders interpretation and knowledge discovery. We present a meta-analytic integration tool, Comparative Pathway Integrator (CPI), to address these issues using adaptively weighted Fisher’s method to discover consensual and differential enrichment patterns, a tight clustering algorithm to reduce pathway redundancy, and a text mining algorithm to assist interpretation of the pathway clusters. We applied CPI to jointly analyze six psychiatric disorder transcriptomic studies to demonstrate its effectiveness, and found functions confirmed by previous biological studies as well as novel enrichment patterns. CPI’s R package is accessible online on Github metaOmics/MetaPath.


2019 ◽  
Vol 20 (1) ◽  
Author(s):  
Marcelo P. Segura-Lepe ◽  
Hector C. Keun ◽  
Timothy M. D. Ebbels

Abstract Background Transcriptomic data is often used to build statistical models which are predictive of a given phenotype, such as disease status. Genes work together in pathways and it is widely thought that pathway representations will be more robust to noise in the gene expression levels. We aimed to test this hypothesis by constructing models based on either genes alone, or based on sample specific scores for each pathway, thus transforming the data to a ‘pathway space’. We progressively degraded the raw data by addition of noise and examined the ability of the models to maintain predictivity. Results Models in the pathway space indeed had higher predictive robustness than models in the gene space. This result was independent of the workflow, parameters, classifier and data set used. Surprisingly, randomised pathway mappings produced models of similar accuracy and robustness to true mappings, suggesting that the success of pathway space models is not conferred by the specific definitions of the pathway. Instead, predictive models built on the true pathway mappings led to prediction rules with fewer influential pathways than those built on randomised pathways. The extent of this effect was used to differentiate pathway collections coming from a variety of widely used pathway databases. Conclusions Prediction models based on pathway scores are more robust to degradation of gene expression information than the equivalent models based on ungrouped genes. While models based on true pathway scores are not more robust or accurate than those based on randomised pathways, true pathways produced simpler prediction rules, emphasizing a smaller number of pathways.


2019 ◽  
Author(s):  
Nicholas Franzese ◽  
Adam Groce ◽  
T. M. Murali ◽  
Anna Ritz

AbstractCharacterizing cellular responses to different extrinsic signals is an active area of research, and curated pathway databases describe these complex signaling reactions. Here, we revisit a fundamental question in signaling pathway analysis: are two molecules “connected” in a network? This question is the first step towards understanding the potential influence of molecules in a pathway, and the answer depends on the choice of modeling framework. We examined the connectivity of Reactome signaling pathways using four different pathway representations. We find that Reactome is very well connected as a graph, moderately well connected as a compound graph or bipartite graph, and poorly connected as a hypergraph (which captures many-to-many relationships in reaction networks). We present a novel relaxation of hypergraph connectivity that iteratively increases connectivity from a node while preserving the hypergraph topology. This measure, B-relaxation distance, provides a parameterized transition between hypergraph connectivity and graph connectivity. B-relaxation distance is sensitive to the presence of small molecules that participate in many functionally unrelated reactions in the network. We also define a score that quantifies one pathway’s downstream influence on another, which can be calculated as B-relaxation distance gradually relaxes the connectivity constraint in hypergraphs. Computing this score across all pairs of 34 Reactome pathways reveals pairs of pathways statistically significant influence. We present two such case studies, and we describe the specific reactions that contribute to the large influence score. Finally, we investigate the ability for connectivity measures to capture functional relationships among proteins, and use the evidence channels in the STRING database as a benchmark dataset. STRING interactions whose proteins are B-connected in Reactome have statistically significantly higher scores than interactions connected in the bipartite graph representation. Our method lays the groundwork for other generalizations of graph-theoretic concepts to hypergraphs in order to facilitate signaling pathway analysis.Author summarySignaling pathways describe how cells respond to external signals through molecular interactions. As we gain a deeper understanding of these signaling reactions, it is important to understand how molecules may influence downstream responses and how pathways may affect each other. As the amount of information in signaling pathway databases continues to grow, we have the opportunity to analyze properties about pathway structure. We pose an intuitive question about signaling pathways: when are two molecules “connected” in a pathway? This answer varies dramatically based on the assumptions we make about how reactions link molecules. Here, examine four approaches for modeling the structural topology of signaling pathways, and present methods to quantify whether two molecules are “connected” in a pathway database. We find that existing approaches are either too permissive (molecules are connected to many others) or restrictive (molecules are connected to a handful of others), and we present a new measure that offers a continuum between these two extremes. We then expand our question to ask when an entire signaling pathway is “downstream” of another pathway, and show two case studies from the Reactome pathway database that uncovers pathway influence. Finally, we show that the strict notion of connectivity can capture functional relationships among proteins using an independent benchmark dataset. Our approach to quantify connectivity in pathways considers a biologically-motivated definition of connectivity, laying the foundation for more sophisticated analyses that leverage the detailed information in pathway databases.


Author(s):  
Enchong Zhang ◽  
Fujisawa Shiori ◽  
Oscar YongNan Mu ◽  
Jieqian He ◽  
Yuntian Ge ◽  
...  

Prostate cancer (PCa) is the most common malignant tumor affecting males worldwide. The substantial heterogeneity in PCa presents a major challenge with respect to molecular analyses, patient stratification, and treatment. Least absolute shrinkage and selection operator was used to select eight risk-CpG sites. Using an unsupervised clustering analysis, called consensus clustering, we found that patients with PCa could be divided into two subtypes (Methylation_H and Methylation_L) based on the DNA methylation status at these CpG sites. Differences in the epigenome, genome, transcriptome, disease status, immune cell composition, and function between the identified subtypes were explored using The Cancer Genome Atlas database. This analysis clearly revealed the risk characteristics of the Methylation_H subtype. Using a weighted correlation network analysis to select risk-related genes and least absolute shrinkage and selection operator, we constructed a prediction signature for prognosis based on the subtype classification. We further validated its effectiveness using four public datasets. The two novel PCa subtypes and risk predictive signature developed in this study may be effective indicators of prognosis.


2020 ◽  
Vol 15 (5) ◽  
pp. 379-395
Author(s):  
Ali Ghulam ◽  
Xiujuan Lei ◽  
Min Guo ◽  
Chen Bian

Pathway analysis integrates most of the computational tools for the investigation of high-level and complex human diseases. In the field of bioinformatics research, biological pathways analysis is an important part of systems biology. The molecular complexities of biological pathways are difficult to understand in human diseases, which can be explored through pathway analysis. In this review, we describe essential information related to pathway databases and their mechanisms, algorithms and methods. In the pathway database analysis, we present a brief introduction on how to gain knowledge from fundamental pathway data in regard to specific human pathways and how to use pathway databases and pathway analysis to predict diseases during an experiment. We also provide detailed information related to computational tools that are used in complex pathway data analysis, the roles of these tools in the bioinformatics field and how to store the pathway data. We illustrate various methodological difficulties that are faced during pathway analysis. The main ideas and techniques for the pathway-based examination approaches are presented. We provide the list of pathway databases and analytical tools. This review will serve as a helpful manual for pathway analysis databases.


2021 ◽  
Author(s):  
Xiaoxi Dong ◽  
Kovidh Vegesna ◽  
Cory Brouwer ◽  
Weijun Luo

AbstractPathway analysis is widely used in genomics and omics research, but the data visualization has been highly limited in function, pathway coverage and data format. Here we develop SBGNview a comprehensive solution to address these needs. By adopting the standard SBGN format, SBGNview greatly extend the coverage of pathway based analysis and data visualization to essentially all major pathway databases beyond KEGG, including 5200 reference pathways and over 3000 species. In addition, SBGNview substantially extends current tools in both design and function, including standard coherent input/output formats, high quality graphics convenient for both computational and manual analysis, and flexible and open-end workflow. In addition to pathway analysis and data visualization, SBGNview provides essential infrastructure for SBGN data manipulation and processing. SBGNview is available online: https://github.com/datapplab/SBGNview.


Author(s):  
R. C. Moretz ◽  
D. F. Parsons

Short lifetime or total absence of electron diffraction of ordered biological specimens is an indication that the specimen undergoes extensive molecular structural damage in the electron microscope. The specimen damage is due to the interaction of the electron beam (40-100 kV) with the specimen and the total removal of water from the structure by vacuum drying. The lower percentage of inelastic scattering at 1 MeV makes it possible to minimize the beam damage to the specimen. The elimination of vacuum drying by modification of the electron microscope is expected to allow more meaningful investigations of biological specimens at 100 kV until 1 MeV electron microscopes become more readily available. One modification, two-film microchambers, has been explored for both biological and non-biological studies.


Author(s):  
Peter Rez

In high resolution microscopy the image amplitude is given by the convolution of the specimen exit surface wave function and the microscope objective lens transfer function. This is usually done by multiplying the wave function and the transfer function in reciprocal space and integrating over the effective aperture. For very thin specimens the scattering can be represented by a weak phase object and the amplitude observed in the image plane is1where fe (Θ) is the electron scattering factor, r is a postition variable, Θ a scattering angle and x(Θ) the lens transfer function. x(Θ) is given by2where Cs is the objective lens spherical aberration coefficient, the wavelength, and f the defocus.We shall consider one dimensional scattering that might arise from a cross sectional specimen containing disordered planes of a heavy element stacked in a regular sequence among planes of lighter elements. In a direction parallel to the disordered planes there will be a continuous distribution of scattering angle.


Author(s):  
Murray Vernon King ◽  
Donald F. Parsons

Effective application of the high-voltage electron microscope to a wide variety of biological studies has been restricted by the radiation sensitivity of biological systems. The problem of radiation damage has been recognized as a serious factor influencing the amount of information attainable from biological specimens in electron microscopy at conventional voltages around 100 kV. The problem proves to be even more severe at higher voltages around 1 MV. In this range, the problem is the relatively low sensitivity of the existing recording media, which entails inordinately long exposures that give rise to severe radiation damage. This low sensitivity arises from the small linear energy transfer for fast electrons. Few developable grains are created in the emulsion per electron, while most of the energy of the electrons is wasted in the film base.


Author(s):  
Jane K. Rosenthal ◽  
Dianne L. Atkins ◽  
William J. Marvin ◽  
Penny A. Krumm

To comprehend structural changes in cardiac myocytes accompanying adrenergic innervation, it is essential that a three dimensional analysis be performed. To date, biological studies which utilize stereological methods have been limited to cells in tissue and in organs. Our laboratory has utilized current stereological techniques for measuring absolute volumes of individual myocytes in primary culture. Cell volumes are calculated for two distinct groups of cells at 96 hours in culture: isolated myocytes and myocytes innervated with adrenergic neurons (Figure 1).Cardiac myocytes are cultured from the ventricular apices of newborn rats. Cells are plated directly onto tissue culture dishes with or without preplated explants from the paravertebral thoracolumbar sympathetic chain. On day four cultures are photographed and marked for one-to-one cell location. Following conventional fixation and embeddment in eponate-12, the cells are relocated and mounted for microtomy. The cells are completely sectioned at 120nm in their parallel orientation to the surface of the dish (Figure 2). Serial sections are collected on formvar coated slotted grids and are recorded in sequence.


Sign in / Sign up

Export Citation Format

Share Document