In RNA-seq differential expression analysis, investigators aim to detect those genes with changes in expression level across conditions, despite technical and biological variability in the ...observations. A common task is to accurately estimate the effect size, often in terms of a logarithmic fold change (LFC).
When the read counts are low or highly variable, the maximum likelihood estimates for the LFCs has high variance, leading to large estimates not representative of true differences, and poor ranking of genes by effect size. One approach is to introduce filtering thresholds and pseudocounts to exclude or moderate estimated LFCs. Filtering may result in a loss of genes from the analysis with true differences in expression, while pseudocounts provide a limited solution that must be adapted per dataset. Here, we propose the use of a heavy-tailed Cauchy prior distribution for effect sizes, which avoids the use of filter thresholds or pseudocounts. The proposed method, Approximate Posterior Estimation for generalized linear model, apeglm, has lower bias than previously proposed shrinkage estimators, while still reducing variance for those genes with little information for statistical inference.
The apeglm package is available as an R/Bioconductor package at https://bioconductor.org/packages/apeglm, and the methods can be called from within the DESeq2 software.
Supplementary data are available at Bioinformatics online.
The power prior: theory and applications Ibrahim, Joseph G.; Chen, Ming-Hui; Gwon, Yeongjin ...
Statistics in medicine,
10 December 2015, Letnik:
34, Številka:
28
Journal Article
On the normalized power prior Carvalho, Luiz Max; Ibrahim, Joseph G.
Statistics in medicine,
30 October 2021, Letnik:
40, Številka:
24
Journal Article
Recenzirano
Odprti dostop
The power prior is a popular tool for constructing informative prior distributions based on historical data. The method consists of raising the likelihood to a discounting factor in order to control ...the amount of information borrowed from the historical data. However, one often wishes to assign this discounting factor a prior distribution and estimate it jointly with the parameters, which in turn necessitates the computation of a normalizing constant. In this article, we are concerned with how to approximately sample from joint posterior of the parameters and the discounting factor. We first show a few important properties of the normalizing constant and then use these results to motivate a bisection‐type algorithm for computing it on a fixed budget of evaluations. We give a large array of illustrations and discuss cases where the normalizing constant is known in closed‐form and where it is not. We show that the proposed method produces approximate posteriors that are very close to the exact distributions and also produces posteriors that cover the data‐generating parameters with higher probability in the intractable case. Our results suggest that the proposed method is an accurate and easy to implement technique to include this normalization, being applicable to a large class of models. They also reinforce the notion that proper inclusion of the normalizing constant is crucial to the drawing of correct inferences and appropriate quantification of uncertainty.
Joint models for longitudinal and survival data are particularly relevant to many cancer clinical trials and observational studies in which longitudinal biomarkers (eg, circulating tumor cells, ...immune response to a vaccine, and quality-of-life measurements) may be highly associated with time to event, such as relapse-free survival or overall survival. In this article, we give an introductory overview on joint modeling and present a general discussion of a broad range of issues that arise in the design and analysis of clinical trials using joint models. To demonstrate our points throughout, we present an analysis from the Eastern Cooperative Oncology Group trial E1193, as well as examine some operating characteristics of joint models through simulation studies.
Medication errors (MEs) result in preventable harm to nursing home (NH) residents and pose a significant financial burden. Institutionalized older people are particularly vulnerable because of ...various organizational and individual factors. This systematic review reports the prevalence of MEs leading to hospitalization and death in NH residents and the factors associated with risk of death and hospitalization. A systematic search was conducted of the relevant peer‐reviewed research published between January 1, 2000, and October 1, 2015, in English, French, German, or Spanish examining serious outcomes of MEs in NHs residents. Eleven studies met the inclusion criteria and examined three types of MEs: all MEs (n = 5), transfer‐related MEs (n = 5), and potentially inappropriate medications (PIMs) (n = 1). MEs were common, involving 16–27% of residents in studies examining all types of MEs and 13–31% of residents in studies examining transfer‐related MEs, and 75% of residents were prescribed at least one PIM. That said, serious effects of MEs were surprisingly low and were reported in only a small proportion of errors (0–1% of MEs), with death being rare. Whether MEs resulting in serious outcomes are truly infrequent, or are underreported because of the difficulty in ascertaining them, remains to be elucidated to assist in designing safer systems.
Expression quantitative trait loci (eQTL) studies are used to understand the regulatory function of non-coding genome-wide association study (GWAS) risk loci, but colocalization alone does not ...demonstrate a causal relationship of gene expression affecting a trait. Evidence for mediation, that perturbation of gene expression in a given tissue or developmental context will induce a change in the downstream GWAS trait, can be provided by two-sample Mendelian Randomization (MR). Here, we introduce a new statistical method, MRLocus, for Bayesian estimation of the gene-to-trait effect from eQTL and GWAS summary data for loci with evidence of allelic heterogeneity, that is, containing multiple causal variants. MRLocus makes use of a colocalization step applied to each nearly-LD-independent eQTL, followed by an MR analysis step across eQTLs. Additionally, our method involves estimation of the extent of allelic heterogeneity through a dispersion parameter, indicating variable mediation effects from each individual eQTL on the downstream trait. Our method is evaluated against other state-of-the-art methods for estimation of the gene-to-trait mediation effect, using an existing simulation framework. In simulation, MRLocus often has the highest accuracy among competing methods, and in each case provides more accurate estimation of uncertainty as assessed through interval coverage. MRLocus is then applied to five candidate causal genes for mediation of particular GWAS traits, where gene-to-trait effects are concordant with those previously reported. We find that MRLocus's estimation of the causal effect across eQTLs within a locus provides useful information for determining how perturbation of gene expression or individual regulatory elements will affect downstream traits. The MRLocus method is implemented as an R package available at https://mikelove.github.io/mrlocus.
Adherence to medication is vital for disease management while simultaneously reducing healthcare expenditure. Older persons with cognitive impairment (CI) are at risk for non-adherence as cognitive ...processes are needed to manage medications. This systematic review focuses on the relationship between medication non-adherence and specific cognitive domains in persons with CI, and explores determinants of medication non-adherence. When available, relationships and factors are compared with cognitively intact populations.
A seven database systematic search of studies published between 1 January 1949-31 December 2015 examining medication non-adherence in community dwelling persons with CI or dementia was conducted. Articles reporting medication non-adherence in people with CI or dementia in the community, with or without caregiver supports were eligible for inclusion. Papers reporting adherence to treatments in cognitively intact populations, populations from hospital or institutional settings, for non-prescribed medication or those describing dementia as a factor predicting medication non-adherence were excluded. Data on study and population characteristics, research design, data sources and analysis, specific cognitive domains, non-adherence prevalence, measurement of adherence, salient findings, factors associated with adherence and strategies to improve medication adherence were extracted. Study limitations included inconsistencies between data sources and definitions, resulting in a loss of fidelity in the value and comprehensiveness of data, as well as exclusion of non-pharmacological treatments and regimens.
Fifteen studies met inclusion criteria. Adherence among CI subjects ranged from 10.7%-38% with better rates of adherence in non-CI individuals. Medication non-adherence definitions varied considerably. New-learning, memory and executive functioning were associated with improved adherence and formed the focus of most studies. Multiple factors were identified as modulators of non-adherence.
This review highlights a gap in knowledge on how specific cognitive domains contribute to medication non-adherence amongst CI populations, and demonstrates the current focus is limited to two domains: memory and executive functioning.
Missing data are a prevailing problem in any type of data analyses. A participant variable is considered missing if the value of the variable (outcome or covariate) for the participant is not ...observed. In this article, various issues in analyzing studies with missing data are discussed. Particularly, we focus on missing response and/or covariate data for studies with discrete, continuous, or time-to-event end points in which generalized linear models, models for longitudinal data such as generalized linear mixed effects models, or Cox regression models are used. We discuss various classifications of missing data that may arise in a study and demonstrate in several situations that the commonly used method of throwing out all participants with any missing data may lead to incorrect results and conclusions. The methods described are applied to data from an Eastern Cooperative Oncology Group phase II clinical trial of liver cancer and a phase III clinical trial of advanced non-small-cell lung cancer. Although the main area of application discussed here is cancer, the issues and methods we discuss apply to any type of study.
We consider the problem of Bayesian sample size determination for a clinical trial in the presence of historical data that inform the treatment effect. Our broadly applicable, simulation-based ...methodology provides a framework for calibrating the informativeness of a prior while simultaneously identifying the minimum sample size required for a new trial such that the overall design has appropriate power to detect a non-null treatment effect and reasonable type I error control. We develop a comprehensive strategy for eliciting null and alternative sampling prior distributions which are used to define Bayesian generalizations of the traditional notions of type I error control and power. Bayesian type I error control requires that a weighted-average type I error rate not exceed a prespecified threshold. We develop a procedure for generating an appropriately sized Bayesian hypothesis test using a simple partial-borrowing power prior which summarizes the fraction of information borrowed from the historical trial. We present results from simulation studies that demonstrate that a hypothesis test procedure based on this simple power prior is as efficient as those based on more complicated meta-analytic priors, such as normalized power priors or robust mixture priors, when all are held to precise type I error control requirements. We demonstrate our methodology using a real data set to design a follow-up clinical trial with time-to-event endpoint for an investigational treatment in high-risk melanoma.