Abstract
apeglm provides empirical Bayes shrinkage estimators for effect sizes for a variety of GLM models; apeglm stands for “Approximate Posterior Estimation for GLM”. apeglm package version: 1.20.0Note: the typical RNA-seq workflow for users would be to call apeglm estimation from within the lfcShrink
function from the DESeq2 package. The unevaluated code chunk shows how to obtain apeglm shrinkage estimates after running DESeq
. See the DESeq2 vignette for more details. The lfcShrink
wrapper function takes care of many details below, and unifies the interface for multiple shrinkage estimators. The coefficient to shrink can be specified either by name or by number (following the order in resultsNames(dds)
). Be aware that DESeq2’s lfcShrink
interface provides LFCs on the log2 scale, while apeglm provides coefficients on the natural log scale.
Joshua Zitovsky contributed fast C++ code for the beta-binomial likelihood, demonstrated in a later section of the vignette.
We have benefited in the development of apeglm
from feedback or contributions from the following individuals:
Wolfgang Huber, Cecile Le Sueur, Charlotte Soneson
Here we show example code which mimics what will happen inside the lfcShrink
function when using the apeglm method (Zhu, Ibrahim, and Love 2018).
Load a prepared SummarizedExperiment
:
## SRR1039508 SRR1039509 SRR1039512 SRR1039513 SRR1039516
## ENSG00000000003 679 448 873 408 1138
## ENSG00000000005 0 0 0 0 0
## ENSG00000000419 467 515 621 365 587
## ENSG00000000457 260 211 263 164 245
## ENSG00000000460 60 55 40 35 78
## ENSG00000000938 0 0 2 0 1
## SRR1039517 SRR1039520 SRR1039521
## ENSG00000000003 1047 770 572
## ENSG00000000005 0 0 0
## ENSG00000000419 799 417 508
## ENSG00000000457 331 233 229
## ENSG00000000460 63 76 60
## ENSG00000000938 0 0 0
For demonstration, we will use 3000 genes of the airway
dataset, the first from those genes with at least 10 counts across all samples.
First run a DESeq2 differential expression analysis (Love, Huber, and Anders 2014) (size factors, and dispersion estimates could similarly be estimated using edgeR):
library(DESeq2)
dds <- DESeqDataSet(airway, ~cell + dex)
dds$dex <- relevel(dds$dex, "untrt")
dds <- DESeq(dds)
res <- results(dds)
Defining data and parameter objects necessary for apeglm
. We must multiply the coefficients from DESeq2 by a factor, because apeglm provides natural log coefficients. Again, this would be handled inside of lfcShrink
in DESeq2 for a typical RNA-seq analysis.
Here apeglm
on 3000 genes takes a few seconds on a laptop. It scales with number of genes, the number of samples and the number of variables in the design formula, where here we have 5 coefficients (one for the four cell cultures and one for the difference due to dexamethasone treatment).
We provide apeglm
with the SummarizedExperiment although the function can also run on a matrix of counts or other observed data. We specify a coef
as well as a threshold
which we discuss below. Note that we multiple the threshold
by log(2)
to convert from log2 scale to natural log scale.
The original interface to apeglm looked as follows, but was slow, we therefore recommend using the faster interfaces shown below.
# original interface, also may give Lapack error
fit <- apeglm(Y=airway, x=x, log.lik=logLikNB, param=param, coef=ncol(x),
threshold=log(2) * 1, mle=mle, offset=offset)
There are better, faster implementations of apeglm specifically for negative binomial likelihoods. The version nbinomR
is ~5 times faster than the default method="general"
.
We will use this run for downstream analysis. Here we pass the SummarizedExperiment, which will pass back the ranges. For faster apeglm
runs, provide only the counts matrix (as seen in subsequent chunks).
library(apeglm)
system.time({
fitR <- apeglm(Y=airway, x=x, log.lik=NULL, param=param, coef=ncol(x),
threshold=log(2) * 1, mle=mle, offset=offset, method="nbinomR")
})
## user system elapsed
## 2.622 0.332 2.955
## [1] "map" "sd" "prior.control" "fsr"
## [5] "svalue" "interval" "thresh" "diag"
## [9] "ranges"
## List of 7
## $ no.shrink : int [1:4] 1 2 3 4
## $ prior.mean : num 0
## $ prior.scale : num 0.304
## $ prior.df : num 1
## $ prior.no.shrink.mean : num 0
## $ prior.no.shrink.scale: num 15
## $ prior.var : num 0.0921
The version nbinomCR
is ~10 times faster than the default general
.
system.time({
fitCR <- apeglm(Y=assay(airway), x=x, log.lik=NULL, param=param, coef=ncol(x),
threshold=log(2) * 1, mle=mle, offset=offset, method="nbinomCR")
})
## user system elapsed
## 1.264 0.004 1.268
The version nbinomC
returns only the MAP coefficients and can be ~50-100 times faster than the default general
. The MAP coefficients are the same as returned by nbinomCR
above, we just skip the calculation of posterior SD. A variant of nbinomC
is nbinomC*
which includes random starts.
system.time({
fitC <- apeglm(Y=assay(airway), x=x, log.lik=NULL, param=param, coef=ncol(x),
threshold=log(2) * 1, mle=mle, offset=offset, method="nbinomC")
})
## user system elapsed
## 0.088 0.004 0.093
Among other output, we have the estimated coefficients attached to the ranges of the SummarizedExperiment used as input:
## [1] "CompressedGRangesList"
## attr(,"package")
## [1] "GenomicRanges"
## DataFrame with 3000 rows and 5 columns
## X.Intercept. cellN061011 cellN080611 cellN61311 dextrt
## <numeric> <numeric> <numeric> <numeric> <numeric>
## ENSG00000000003 6.64788 0.0546137 0.2242529 -0.1570180 -0.2602199
## ENSG00000000419 6.23638 -0.0912014 -0.0206726 -0.0526312 0.1203700
## ENSG00000000457 5.45524 0.0540325 -0.0611456 0.0435641 0.0133132
## ENSG00000000460 3.77417 0.5394114 0.2564594 0.3450027 -0.0420874
## ENSG00000000971 8.53432 0.1733404 0.1364808 -0.4750196 0.2732967
## ... ... ... ... ... ...
## ENSG00000107731 6.63160 0.00129427 -0.3588497 0.2327120 -0.6946087
## ENSG00000107736 2.06827 -0.47986419 -0.5995966 -0.4171680 -0.0224223
## ENSG00000107738 8.62310 -0.14041395 -0.8643571 -0.2631779 0.2137555
## ENSG00000107742 2.06485 0.10029416 -0.4432328 0.0207201 -0.0116752
## ENSG00000107745 7.45809 0.01255731 0.0417601 -0.1318218 -0.0729778
We can compare the coefficients from apeglm with the "normal"
shrinkage type from the original DESeq2 paper (2014). This method, which makes use of a Normal-based prior, is no longer the default shrinkage estimator for lfcShrink
. apeglm provides coefficients on the natural log scale, so we must convert to log2 scale by multiplying by log2(exp(1))
. Note that DESeq2’s lfcShrink
function converts apeglm coefficients to the log2 scale internally.
## user system elapsed
## 1.296 0.036 1.332
Here we plot apeglm estimators against DESeq2:
Here we plot MLE, DESeq2 and apeglm estimators against the mean of normalized counts:
par(mfrow=c(1,3))
lims <- c(-8,8)
hline <- function() abline(h=c(-4:4 * 2),col=rgb(0,0,0,.2))
xlab <- "mean of normalized counts"
plot(res$baseMean, res$log2FoldChange, log="x",
ylim=lims, main="MLE", xlab=xlab)
hline()
plot(res$baseMean, DESeq2.lfc, log="x",
ylim=lims, main="DESeq2", xlab=xlab)
hline()
plot(res$baseMean, apeglm.lfc, log="x",
ylim=lims, main="apeglm", xlab=xlab)
hline()
Note that p-values and FSR define different events, and are not on the same scale. An FSR of 0.5 means that the estimated sign is as bad as random guess.
par(mfrow=c(1,2),mar=c(5,5,1,1))
plot(res$pvalue, fit$fsr, col="blue",
xlab="DESeq2 pvalue", ylab="apeglm local FSR",
xlim=c(0,1), ylim=c(0,.5))
abline(0,1)
plot(-log10(res$pvalue), -log10(fit$fsr),
xlab="-log10 DESeq2 pvalue", ylab="-log10 apeglm local FSR",
col="blue")
abline(0,1)
The s-value was proposed by Stephens (2016), as a statistic giving the aggregate false sign rate for tests with equal or lower s-value than the one considered. We recommend using a lower threshold on s-values than typically used for adjusted p-values, for example one might be interested in sets with 0.01 or 0.005 aggregate FSR.
plot(res$padj, fit$svalue, col="blue",
xlab="DESeq2 padj", ylab="apeglm svalue",
xlim=c(0,.2), ylim=c(0,.02))
More scrutiny can be applied by using an LFC threshold greater than zero, and asking for the probability of a “false-sign-or-small” (FSOS) event: that the effect size is not further from zero in distance than the threshold amount. We can run the svalue
function on these per-gene probabilities to produce s-values that bound the FSOS rate for sets of genes. By specifying threshold=log(2) * 1
above, apeglm
will then output a vector thresh
in the results list that gives the per-gene probabilities of false-sign-or-small events.
We have created a separate GitHub repository giving an example of how the apeglm estimator can be used for Zero-Inflated Negative Binomial data. This approach uses the zinbwave method and Bioconductor package to estimate the probability of each zero belonging to the excess zero component. We compare using a Negative Binomial likelihood with the excess zeros down-weighted and using a Zero-Inflated Negative Binomial likelihood. These two approaches with apeglm perform similarly but we note that the first approach involves less additional code and is faster to compute.
We also show an short example using an alternative likelihood to the negative binomial. Suppose we have allele-specific counts for n=20 vs 20 samples across 5000 genes. We can define a binomial model and test for the allelic balance across groups of samples.
Here we will simulate allele counts from our existing dataset for demonstration. We spike in 10 genes with strong allelic imbalance (instead of an allelic ratio close to 0.5, these will have a ratio of 0.75).
library(emdbook)
n <- 20
f <- factor(rep(1:2,each=n))
mu <- ifelse(res$baseMean > 50, res$baseMean, 50)
set.seed(1)
cts <- matrix(rnbinom(nrow(dds)*2*n,
mu=mu,
size=1/dispersions(dds)),
ncol=2*n)
theta <- runif(nrow(cts),1,1000)
prob <- rnorm(nrow(cts),.5,.05) # close to 0.5
ase.cts <- matrix(rbetabinom(prod(dim(cts)), prob=prob,
size=cts, theta=rep(theta,ncol(cts))),
nrow=nrow(cts))
idx <- 1:10
idx2 <- which(f == 2)
theta[idx] <- 1000
prob[idx] <- 0.75
# the spiked in genes have an allelic ratio of 0.75
ase.cts[idx,idx2] <- matrix(rbetabinom(length(idx)*length(idx2), prob=prob[idx],
size=cts[idx,idx2], theta=theta[idx]),
nrow=length(idx))
We first need to estimate MLE coefficients and standard errors.
One option to run apeglm
would be to define a beta-binomial likelihood function which uses the total counts as a parameter, and the logit function as a link. And then this function could be provided to the log.lik
argument.
betabinom.log.lik <- function(y, x, beta, param, offset) {
xbeta <- x %*% beta
p.hat <- (1+exp(-xbeta))^-1
dbetabinom(y, prob=p.hat, size=param[-1], theta=param[1], log=TRUE)
}
However, apeglm
has faster C++ implementations for the beta-binomial, which were implemented and tested by Joshua Zitovsky. Here, using method="betabinCR"
is 3 times faster than using the R-defined log.lik
, and the "betabinCR"
method also scales significantly better with more samples and more coefficients. As with the negative binomial, method="betabinC"
can be used if the standard errors are not needed, and this will be 5 times faster than the R-defined log.lik
approach on this dataset.
The following code performs two iterations of estimating the MLE coefficients, and then estimating the beta-binomial dispersion.
theta.hat <- 100 # rough initial estimate of dispersion
x <- model.matrix(~f)
niter <- 3
system.time({
for (i in 1:niter) {
param <- cbind(theta.hat, cts)
fit.mle <- apeglm(Y=ase.cts, x=x, log.lik=NULL, param=param,
no.shrink=TRUE, log.link=FALSE, method="betabinCR")
theta.hat <- bbEstDisp(success=ase.cts, size=cts,
x=x, beta=fit.mle$map,
minDisp=.01, maxDisp=5000)
}
})
## user system elapsed
## 8.123 0.116 8.240
We can then plot the MLE estimates over the mean:
coef <- 2
xlab <- "mean of normalized counts"
plot(res$baseMean, fit.mle$map[,coef], log="x", xlab=xlab, ylab="log odds")
points(res$baseMean[idx], fit.mle$map[idx,coef], col="dodgerblue", cex=3)
Now we run the posterior estimation, including a prior on the second coefficient:
mle <- cbind(fit.mle$map[,coef], fit.mle$sd[,coef])
param <- cbind(theta.hat, cts)
system.time({
fit2 <- apeglm(Y=ase.cts, x=x, log.lik=NULL, param=param,
coef=coef, mle=mle, threshold=0.5,
log.link=FALSE, method="betabinCR")
})
## user system elapsed
## 1.498 0.004 1.502
In the apeglm
plot, we color in red the genes with a low aggregate probability of false-sign-or-small (FSOS) events (s-value < .01), where we’ve again defined “small” on the log odds scale using the threshold
argument above.
par(mfrow=c(1,2))
ylim <- c(-1,1.5)
s.val <- svalue(fit2$thresh) # small-or-false-sign value
plot(res$baseMean, fit.mle$map[,coef], main="MLE",
log="x", xlab=xlab, ylab="log odds", ylim=ylim)
points(res$baseMean[idx], fit.mle$map[idx,coef], col="dodgerblue", cex=3)
abline(h=0,col=rgb(1,0,0,.5))
cols <- ifelse(s.val < .01, "red", "black")
plot(res$baseMean, fit2$map[,coef], main="apeglm",
log="x", xlab=xlab, ylab="log odds", col=cols, ylim=ylim)
points(res$baseMean[idx], fit2$map[idx,coef], col="dodgerblue", cex=3)
abline(h=0,col=rgb(1,0,0,.5))
## [1] 1.098612
##
## TRUE
## 10
## R version 4.2.1 (2022-06-23)
## Platform: x86_64-pc-linux-gnu (64-bit)
## Running under: Ubuntu 20.04.5 LTS
##
## Matrix products: default
## BLAS: /home/biocbuild/bbs-3.16-bioc/R/lib/libRblas.so
## LAPACK: /home/biocbuild/bbs-3.16-bioc/R/lib/libRlapack.so
##
## locale:
## [1] LC_CTYPE=en_US.UTF-8 LC_NUMERIC=C
## [3] LC_TIME=en_GB LC_COLLATE=C
## [5] LC_MONETARY=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8
## [7] LC_PAPER=en_US.UTF-8 LC_NAME=C
## [9] LC_ADDRESS=C LC_TELEPHONE=C
## [11] LC_MEASUREMENT=en_US.UTF-8 LC_IDENTIFICATION=C
##
## attached base packages:
## [1] stats4 stats graphics grDevices utils datasets methods
## [8] base
##
## other attached packages:
## [1] emdbook_1.3.12 apeglm_1.20.0
## [3] DESeq2_1.38.0 airway_1.17.0
## [5] SummarizedExperiment_1.28.0 Biobase_2.58.0
## [7] GenomicRanges_1.50.0 GenomeInfoDb_1.34.0
## [9] IRanges_2.32.0 S4Vectors_0.36.0
## [11] BiocGenerics_0.44.0 MatrixGenerics_1.10.0
## [13] matrixStats_0.62.0
##
## loaded via a namespace (and not attached):
## [1] httr_1.4.4 sass_0.4.2 bit64_4.0.5
## [4] jsonlite_1.8.3 splines_4.2.1 bslib_0.4.0
## [7] assertthat_0.2.1 highr_0.9 blob_1.2.3
## [10] GenomeInfoDbData_1.2.9 yaml_2.3.6 numDeriv_2016.8-1.1
## [13] pillar_1.8.1 RSQLite_2.2.18 lattice_0.20-45
## [16] glue_1.6.2 bbmle_1.0.25 digest_0.6.30
## [19] RColorBrewer_1.1-3 XVector_0.38.0 colorspace_2.0-3
## [22] plyr_1.8.7 htmltools_0.5.3 Matrix_1.5-1
## [25] XML_3.99-0.12 pkgconfig_2.0.3 genefilter_1.80.0
## [28] zlibbioc_1.44.0 mvtnorm_1.1-3 xtable_1.8-4
## [31] scales_1.2.1 BiocParallel_1.32.0 tibble_3.1.8
## [34] annotate_1.76.0 KEGGREST_1.38.0 generics_0.1.3
## [37] ggplot2_3.3.6 cachem_1.0.6 cli_3.4.1
## [40] survival_3.4-0 magrittr_2.0.3 crayon_1.5.2
## [43] memoise_2.0.1 evaluate_0.17 fansi_1.0.3
## [46] MASS_7.3-58.1 tools_4.2.1 lifecycle_1.0.3
## [49] stringr_1.4.1 locfit_1.5-9.6 munsell_0.5.0
## [52] DelayedArray_0.24.0 AnnotationDbi_1.60.0 Biostrings_2.66.0
## [55] compiler_4.2.1 jquerylib_0.1.4 rlang_1.0.6
## [58] grid_4.2.1 RCurl_1.98-1.9 bitops_1.0-7
## [61] rmarkdown_2.17 gtable_0.3.1 codetools_0.2-18
## [64] DBI_1.1.3 R6_2.5.1 bdsmatrix_1.3-6
## [67] knitr_1.40 dplyr_1.0.10 fastmap_1.1.0
## [70] bit_4.0.4 utf8_1.2.2 stringi_1.7.8
## [73] parallel_4.2.1 Rcpp_1.0.9 vctrs_0.5.0
## [76] geneplotter_1.76.0 png_0.1-7 coda_0.19-4
## [79] tidyselect_1.2.0 xfun_0.34
Love, Michael I., Wolfgang Huber, and Simon Anders. 2014. “Moderated estimation of fold change and dispersion for RNA-seq data with DESeq2.” Genome Biology 15 (12): 550. https://doi.org/10.1186/s13059-014-0550-8.
Stephens, Matthew. 2016. “False Discovery Rates: A New Deal.” Biostatistics 18 (2). https://doi.org/10.1093/biostatistics/kxw041.
Zhu, Anqi, Joseph G. Ibrahim, and Michael I. Love. 2018. “Heavy-Tailed Prior Distributions for Sequence Count Data: Removing the Noise and Preserving Large Differences.” Bioinformatics. https://doi.org/10.1093/bioinformatics/bty895.