importData_soGGi
(recommended)We first set global chunk options and load the packages that are necessary to run the vignette.
library(DChIPRep)
## Warning: Package 'DChIPRep' is deprecated and will be removed from Bioconductor version 3.12
library(ggplot2)
library(DESeq2)
The package implements the analysis strategy of (Chabbert et al. 2015). Along with an experimental protocol to multiplex ChIP–Seq experiments, (Chabbert et al. 2015) developped a methodology to assess differences between chromatin modification profiles in replicated ChIP–Seq studies that builds on the packages DESeq2 and fdrtool. The package also includes a preprocessing script in Python that allows the user to import sam files into data structures than can be used with the package.
Note that the Python script is no longer tested / updated. It
remains in the package in order to provide a data import that is similar
to Chabbert et al. (2015). The function importData_soGGi
provides a Bioconductor–only
import–solution that is an integral part of the package.
Together with the release of the R package DChIPRep, we provide a python framework that should allow users to generate count tables. These may then be directly used to evaluate metagene enrichment profiles. This framework is entirely contained in a script that only requires the installation of Python 2.7 and of the packages HTSeq and Numpy, which are free to download. This section describes briefly the various possibilities of analysis offered by this tool together with the various input options and parameters. Additional information may be found directly in the source code or via the usual help framework offered by Python.
The script comes with the package and can be located on your system by the following commands:
pythonScriptsDir <- system.file( "exec" , package = "DChIPRep" )
pythonScriptsDir
## [1] "/tmp/RtmpIarzaa/Rinst41c17921421f/DChIPRep/exec"
list.files(pythonScriptsDir)
## [1] "DChIPRep.py"
This script will process alignments of paired-end reads, filter out divergent pairs and low quality alignments and ultimately identify potential PCR duplicates. For the pairs of read retained after filtering, the center of the genomic loci determined by each pair is estimated using mapping coordinates. This center constitutes an approximation of the center of each observed nucleosome. The provided annotation file is then used to estimate the nucleosome counts around the features of interest. These counts are ultimately used to provide a count table in which each row corresponds to one feature and each column to a genomic position around the feature start. This table may then be imported in R using the importData function of the DChIPRep package.
Here we show how to quickly generate a count table from an alignment file. The DChIPRep.py script only requires 4 arguments (all other options have a default setting, cf below):
An alignment file in the SAM format (some HTSeq functions are not currently supported for BAM files) that may be zipped. It is specified by the ‘-i’ option on the command line.
An annotation file in the gff format (specifications for the gff format may be found here) specified in the ‘-a’ option
…
It is crucial to ensure that the chromosome names used in the alignment file and this information file are identical. This information is used by the program to pre–allocate memory to store the counts. A valid path to this text file should be specified under the ‘-g’ option.
A valid path for an output file specified under the option ‘-o’. A new file will be created automatically - if it already exists, its content will be erased and replaced by the script output.
Here is an example of the command line to be run
python DChIPRep.py -i alignment.sam -a my_gff.gff -g chromosome_sizes.txt -o my_count_table.txt
In order to provide greater flexibility in data pre-processing, we have added a panel of additional options that may be specified when running the script. The script is executable, With the –help option, one can get an overview of the available options. They are given below.
usage: DChIPRep.py [-h] -i SAM/BAM -a GFF -g Chromosome Sizes File -o Count Table [-v] [-q TH_QC] [-l TH_LOW] [-L TH_HIGH] [-d TH_PCR] [-f FEATURETYPE]
[-w DOWNSTREAM] [-u UPSTREAM]
optional arguments:
-h, --help show this help message and exit
-i SAM/BAM, --input SAM/BAM
The alignment file to be used for to generate the count table. The file may be in the sam (zipped or not)format. The extension of
the file should contain either the '.sam' indication. Bam files are not supported at the moment due to soome instability in the BAM
reader regarding certain aligner formats.
-a GFF, --annotation GFF
The annotation file that will be used to generate the counts. The file should be in the gff format (see
https://www.sanger.ac.uk/resources/software/gff/spec.html for details).
-g Chromosome Sizes File, --genome_details Chromosome Sizes File
A tabulated file containing the names and sizes of each chromosome. !!! The the chromosome names should be identical to the ones
used to generate the alignment file !!! The file should look like this example (no header): chromI 1304563 chromII 6536634 ...
-o Count Table, --output_file Count Table
The output file where the count table should be stored. If the specified file does not already exist it will be created
automatically. Otherwise, it will be overwritten
-v, --verbose When specified, the option will result in the printing of informations on the alignments and the filtering steps (number of
ambiguous alignments...etc). Default: OFF
-q TH_QC, --quality_threshold TH_QC
The quality threshold below which alignments will be discarded. The alignment quality index typically ranges between 1 and 41.
Default: 30.
-l TH_LOW, --lowest_size TH_LOW
The lowest possible size accepted for DNA fragments. Any pair of reads with an insert size below that value will be discarded.
Default: 130.
-L TH_HIGH, --longest_size TH_HIGH
The longest possible size accepted for DNA fragments. Any pair of reads with an insert size above this value will be discarded.
Default: 180.
-d TH_PCR, --duplicate_filter TH_PCR
The number of estimated PCR duplicates for accepted for a given genomic location. Default: 1.
-f FEATURETYPE, --feature_type FEATURETYPE
The feature types to be used when generating the count table. The feature type will be matched 3rd column of the GFF file. Default:
'Transcript
-w DOWNSTREAM, --downstream_window DOWNSTREAM
The window size used to obtain the counts downstream of the TSS. Default: 1000bp
-u UPSTREAM, --upstream_window UPSTREAM
The window size used to obtain the counts upstream of the TSS. Default: 1500bp
As an example, let us assume that we want to process an alignment file with the following criteria:
Remove reads with an alignment quality below 20
Focus on the pairs with a spanning size ranging between 120 and 160 bp
Get the counts information around the TSS of the coding sequences (CDS) in a symmetrical window of 500bp
Only tolerate two copies on the same molecules upstream
The command line would then be:
python DChIPRep.py -i alignment.sam -a my_gff.gff
-g chromosome_sizes.txt -o my_count_table.txt -v -q 20 -l 120
-L 160 -d 2 -f CDS -u 500 -w 500
The default number of upstream position is 1500bp, the default number of downstream
positions is 1000bp. This results in count tables with 2502 columns in total,
corresponding to the basepairs up– and downstream as well as one column for
for the 0 position and one column for the feature IDs.
In what follows, we assume that we count nucleosomes around transcription start sites (TSS), that is the start of transcripts, which is also the default option in the preprocessing script.
Note that the Python script is no longer tested / updated. It
remains in the package in order to provide a data import that is similar
to Chabbert et al. (2015). The function importData_soGGi
provides a Bioconductor–only
import–solution that is an integral part of the package.
After the data has been preproccesed, we first need an annotation table for our samples that looks like this:
data(exampleSampleTable)
exampleSampleTable
## ChIP Input upstream downstream condition
## 1 BY_conf_K4me3_1__ORF_count.txt BY_conf_WCE_1__ORF_count.txt 1000 1500 K4me3_WT
## 2 BY_conf_K4me3_2__ORF_count.txt BY_conf_WCE_2__ORF_count.txt 1000 1500 K4me3_WT
## 3 BY_conf_K4me3_3__ORF_count.txt BY_conf_WCE_3__ORF_count.txt 1000 1500 K4me3_WT
## 4 SET2_conf_K4me3_1__ORF_count.txt SET2_conf_WCE_1__ORF_count.txt 1000 1500 K4me3_SET2
## 5 SET2_conf_K4me3_2__ORF_count.txt SET2_conf_WCE_2__ORF_count.txt 1000 1500 K4me3_SET2
## 6 SET2_conf_K4me3_3__ORF_count.txt SET2_conf_WCE_3__ORF_count.txt 1000 1500 K4me3_SET2
## sampleID
## 1 K4me3_WT_1
## 2 K4me3_WT_2
## 3 K4me3_WT_3
## 4 K4me3_SET2_1
## 5 K4me3_SET2_2
## 6 K4me3_SET2_3
It gives the names for the input count files in the columns ChIP and Input respectively and the number of base pairs used for analysis upstream and downstream of the TSS in the columns upstream and downstream. Note that the number of upstream and downstream positions needs to be the same for all samples.
The input files must be tab separated files with genomic features in the rows and the positions around the TSS in the columns. In addition to the position columns, a column containing a feature ID must be present.
As mentioned above, the tables have upstream + downstream + 2 columns in total, the two extra columns correspond to the center of the profile (e.g. a TSS) and a column containing the feature IDs.
The sampleID column contains unique sample IDs.
We can then import the data as follows using the function importData which needs
the sample annotation table, a data.frame
and the directory where the raw data files
are stored.
We use the down–sampled raw data that come with the package to illustrate
the function use.
By default the feature ID column is assumed to have the name name, however this can specified in the call to importData via the ID argument.
directory <- file.path(system.file("extdata", package="DChIPRep"))
importedData <- importData(exampleSampleTable, directory)
The imported data is a DChIPRepResults object that contains the data as a DESeqDataSet. It can be accessed via the function DESeq2Data. The DESeqDataSet also contains normalization factors that are equal to the counts from the chromatin inputs, after being multiplied by the coverage ratio between the input and the ChIP data sets.
importedData
## Summary of the DESeqDataSet:
## ============================
## class: DESeqDataSet
## dim: 2501 6
## metadata(1): version
## assays(2): counts normalizationFactors
## rownames(2501): Pos_-1000 Pos_-999 ... Pos_1499 Pos_1500
## rowData names(0):
## colnames(6): K4me3_WT_1 K4me3_WT_2 ... K4me3_SET2_2 K4me3_SET2_3
## colData names(6): ChIP Input ... condition sampleID
##
##
##
## Results:
## ============================
## No results available yet, call runTesting first
DESeq2Data(importedData)
## class: DESeqDataSet
## dim: 2501 6
## metadata(1): version
## assays(2): counts normalizationFactors
## rownames(2501): Pos_-1000 Pos_-999 ... Pos_1499 Pos_1500
## rowData names(0):
## colnames(6): K4me3_WT_1 K4me3_WT_2 ... K4me3_SET2_2 K4me3_SET2_3
## colData names(6): ChIP Input ... condition sampleID
head(normalizationFactors(DESeq2Data(importedData)))
## K4me3_WT_1 K4me3_WT_2 K4me3_WT_3 K4me3_SET2_1 K4me3_SET2_2 K4me3_SET2_3
## Pos_-1000 4.069 0.8868 2.8974 0.3647 2.58 0.2305
## Pos_-999 5.426 3.1038 1.0865 0.2084 1.29 0.1536
## Pos_-998 1.356 0.8868 0.7244 0.1042 1.29 0.3073
## Pos_-997 8.139 0.4434 1.8109 0.2084 1.72 0.2305
## Pos_-996 2.713 0.8868 3.2596 0.1042 1.29 0.6146
## Pos_-995 2.713 2.2170 2.8974 0.2084 1.72 0.2305
If your data is already available within R, you can import it directly via importDataFromMatrices from an input and a ChIP Matrix. The rows of the matrices should correspond to the positions and the columns to the samples. You furthermore need a sample table as described above, however the columns Input and ChIP are not needed.
If you have data that is still on the level of the individual features, you can use the helper function summarizeCountsPerPosition to summarize the counts for individual features per position.
The package comes with example data sets that correspond to the example sample table shown above.
We first show 10 random rows from the two data matrices and then use the function importDataFromMatrices to import them to a DChIPRepResults object.
As you can see the rows should be named according to the positions up– and downstream of the TSS that they contain and the columns should be named after the samples.
Note however, that a correct ordering is not checked, it is assumed that the rows are ordered according to the position they represent (from upstream to downstream)
data(exampleInputData)
data(exampleChipData)
exampleSampleTable
## ChIP Input upstream downstream condition
## 1 BY_conf_K4me3_1__ORF_count.txt BY_conf_WCE_1__ORF_count.txt 1000 1500 K4me3_WT
## 2 BY_conf_K4me3_2__ORF_count.txt BY_conf_WCE_2__ORF_count.txt 1000 1500 K4me3_WT
## 3 BY_conf_K4me3_3__ORF_count.txt BY_conf_WCE_3__ORF_count.txt 1000 1500 K4me3_WT
## 4 SET2_conf_K4me3_1__ORF_count.txt SET2_conf_WCE_1__ORF_count.txt 1000 1500 K4me3_SET2
## 5 SET2_conf_K4me3_2__ORF_count.txt SET2_conf_WCE_2__ORF_count.txt 1000 1500 K4me3_SET2
## 6 SET2_conf_K4me3_3__ORF_count.txt SET2_conf_WCE_3__ORF_count.txt 1000 1500 K4me3_SET2
## sampleID
## 1 K4me3_WT_1
## 2 K4me3_WT_2
## 3 K4me3_WT_3
## 4 K4me3_SET2_1
## 5 K4me3_SET2_2
## 6 K4me3_SET2_3
exampleInputData[1:10, ]
## K4me3_WT_1 K4me3_WT_2 K4me3_WT_3 K4me3_SET2_1 K4me3_SET2_2 K4me3_SET2_3
## X.1000 1417 2036 2957 1729 1741 1842
## X.999 1442 2075 2889 1756 1703 1796
## X.998 1408 2031 2784 1761 1803 1795
## X.997 1433 1962 2853 1740 1702 1783
## X.996 1465 2109 2905 1759 1701 1845
## X.995 1466 2077 2778 1703 1690 1852
## X.994 1504 2074 2839 1706 1729 1867
## X.993 1433 2028 2787 1725 1670 1845
## X.992 1416 1930 2739 1668 1729 1719
## X.991 1372 2005 2743 1693 1666 1785
exampleChipData[1:10, ]
## K4me3_WT_1 K4me3_WT_2 K4me3_WT_3 K4me3_SET2_1 K4me3_SET2_2 K4me3_SET2_3
## X.1000 4374 1928 2235 198 1595 334
## X.999 4139 1951 2200 182 1591 306
## X.998 4348 1907 2209 211 1581 281
## X.997 4326 1939 2265 225 1555 343
## X.996 4256 1883 2214 211 1609 340
## X.995 4227 1888 2198 198 1565 332
## X.994 4219 1884 2203 195 1584 313
## X.993 4222 1911 2230 198 1599 321
## X.992 4192 1866 2182 191 1549 330
## X.991 4126 1892 2157 199 1546 314
imDataFromMatrices <- importDataFromMatrices(inputData = exampleInputData,
chipData = exampleChipData,
sampleTable = exampleSampleTable)
The imported data is again a DChIPRepResults object that contains the data as a DESeqDataSet.
importData_soGGi
(recommended)It is also possible to use the function importData_soGGi
, which is
based on the function regionPlot
from the package soGGi
to import data from .bam files directly.
It will return a matrix with one column per .bam file and the respective counts per postion in the rows.
In the example below, we use a subsampled .bam file (0.1 % of the reads)
from the Galonska et. al. WCE (whole cell extract) H3Kme3 data and associated
TSS near identified peaks. For additional details on the data, see the help pages
on input_galonska
and TSS_galonska
. The code below is not evaluated,
since it takes some time to compute.
data(sample_table_galonska)
data(TSS_galonska)
bam_dir <- file.path(system.file("extdata", package="DChIPRep"))
wce_bam <- "subsampled_0001_pc_SRR2144628_WCE_bowtie2_mapped-only_XS-filt_no-dups.bam"
mat_wce <- importData_soGGi(bam_paths = file.path(bam_dir, wce_bam),
TSS = TSS_galonska,
fragment_lengths = sample_table_galonska$input_fragment_length[1],
sample_ids = sample_table_galonska$input[1],
paired = FALSE,
removeDup=FALSE
)
head(mat_wce)
After the data import, we are ready to perform the tests for differential enrichment. The tests are performed position–wise and wrap DESeq2 and fdrtool. Briefly the DChIPRep testing workflow is as follows:
The function estimateDispersions from DESeq2 is called and the dispersions are estimated.
Then the position–wise tests to compare the experimental conditions are performed. A minimum fold change is used for the null hypothesis, the default value used is 0.05 on a log2 scale.
A possible strategy to infer this threshold from the data is to look a the average fold–change between technical replicates.
imDataFromMatrices <- runTesting(imDataFromMatrices, plotFDR = TRUE)
## Step 1... determine cutoff point
## Step 2... estimate parameters of null distribution and eta0
## Step 3... compute p-values and estimate empirical PDF/CDF
## Step 4... compute q-values and local fdr
## Step 5... prepare for plotting
The results can now be accessed via the function resultsDChIPRep.
res <- resultsDChIPRep(imDataFromMatrices)
head(res)
## baseMean log2FoldChange lfcSE stat pvalue padj lfdr
## Pos_-1000 1013.4 -0.02983 0.05319 0.00000 1.0000 1.0000 1
## Pos_-999 978.5 -0.03271 0.05592 0.00000 1.0000 1.0000 1
## Pos_-998 996.3 0.05379 0.05360 0.07077 0.9436 1.0000 1
## Pos_-997 1056.4 -0.04933 0.05193 0.00000 1.0000 1.0000 1
## Pos_-996 1011.3 -0.11783 0.05190 -1.30675 0.1913 0.8983 1
## Pos_-995 1005.2 -0.06681 0.05287 -0.31789 0.7506 1.0000 1
table( res$lfdr < 0.2)
##
## FALSE TRUE
## 2452 49
At an lfdr of 0.2 we identify 49 significant positions.
We can first of all plot the TSS profiles by coloring the individual points by significance.
Points corresponding to significant positions are colored black in both of the conditions. The replicate–samples are sumarized by using a positionwise robust Huber estimator for the mean (Hampel, Hennig, and Ronchetti 2011).
The function returns the plot as a ggplot2 object that can be modified afterwards.
sigPlot <- plotSignificance(imDataFromMatrices)
sigPlot
This plot is similar to Figure S17B of (Chabbert et al. 2015). We see an enrichment for significant position near the end of the downstream window considered.
We can produce the typical, smoothed plots of the TSS profiles as well. Here we use again the smoothed Huber estimator for the mean to compute a summary per experimental group.
profilePlot <- plotProfiles(imDataFromMatrices)
profilePlot
This plot is similar to Figure 5B of (Chabbert et al. 2015).
sessionInfo()
## R version 4.0.0 (2020-04-24)
## Platform: x86_64-pc-linux-gnu (64-bit)
## Running under: Ubuntu 18.04.4 LTS
##
## Matrix products: default
## BLAS: /home/biocbuild/bbs-3.11-bioc/R/lib/libRblas.so
## LAPACK: /home/biocbuild/bbs-3.11-bioc/R/lib/libRlapack.so
##
## locale:
## [1] LC_CTYPE=en_US.UTF-8 LC_NUMERIC=C LC_TIME=en_US.UTF-8
## [4] LC_COLLATE=C LC_MONETARY=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8
## [7] LC_PAPER=en_US.UTF-8 LC_NAME=C LC_ADDRESS=C
## [10] LC_TELEPHONE=C LC_MEASUREMENT=en_US.UTF-8 LC_IDENTIFICATION=C
##
## attached base packages:
## [1] parallel stats4 stats graphics grDevices utils datasets methods base
##
## other attached packages:
## [1] ggplot2_3.3.0 DChIPRep_1.18.0 DESeq2_1.28.0
## [4] SummarizedExperiment_1.18.0 DelayedArray_0.14.0 matrixStats_0.56.0
## [7] Biobase_2.48.0 GenomicRanges_1.40.0 GenomeInfoDb_1.24.0
## [10] IRanges_2.22.0 S4Vectors_0.26.0 BiocGenerics_0.34.0
## [13] rmarkdown_2.1 knitr_1.28 BiocStyle_2.16.0
##
## loaded via a namespace (and not attached):
## [1] colorspace_1.4-1 hwriter_1.3.2 seqinr_3.6-1
## [4] ellipsis_0.3.0 futile.logger_1.4.3 XVector_0.28.0
## [7] farver_2.0.3 ChIPpeakAnno_3.22.0 bit64_0.9-7
## [10] AnnotationDbi_1.50.0 codetools_0.2-16 splines_4.0.0
## [13] geneplotter_1.66.0 ade4_1.7-15 Rsamtools_2.4.0
## [16] annotate_1.66.0 GO.db_3.10.0 dbplyr_1.4.3
## [19] png_0.1-7 graph_1.66.0 BiocManager_1.30.10
## [22] compiler_4.0.0 httr_1.4.1 assertthat_0.2.1
## [25] RcppZiggurat_0.1.5 Matrix_1.2-18 lazyeval_0.2.2
## [28] limma_3.44.0 formatR_1.7 htmltools_0.4.0
## [31] prettyunits_1.1.1 tools_4.0.0 smoothmest_0.1-2
## [34] gtable_0.3.0 glue_1.4.0 GenomeInfoDbData_1.2.3
## [37] reshape2_1.4.4 dplyr_0.8.5 rappdirs_0.3.1
## [40] ShortRead_1.46.0 Rcpp_1.0.4.6 vctrs_0.2.4
## [43] Biostrings_2.56.0 multtest_2.44.0 nlme_3.1-147
## [46] preprocessCore_1.50.0 rtracklayer_1.48.0 xfun_0.13
## [49] stringr_1.4.0 lifecycle_0.2.0 ensembldb_2.12.0
## [52] XML_3.99-0.3 idr_1.2 zlibbioc_1.34.0
## [55] MASS_7.3-51.6 scales_1.1.0 BSgenome_1.56.0
## [58] hms_0.5.3 ProtGenerics_1.20.0 RBGL_1.64.0
## [61] AnnotationFilter_1.12.0 lambda.r_1.2.4 RColorBrewer_1.1-2
## [64] yaml_2.2.1 curl_4.3 memoise_1.1.0
## [67] biomaRt_2.44.0 latticeExtra_0.6-29 stringi_1.4.6
## [70] RSQLite_2.2.0 genefilter_1.70.0 GenomicFeatures_1.40.0
## [73] BiocParallel_1.22.0 chipseq_1.38.0 rlang_0.4.5
## [76] pkgconfig_2.0.3 bitops_1.0-6 evaluate_0.14
## [79] lattice_0.20-41 purrr_0.3.4 labeling_0.3
## [82] GenomicAlignments_1.24.0 Rfast_1.9.9 bit_1.1-15.2
## [85] tidyselect_1.0.0 plyr_1.8.6 magrittr_1.5
## [88] bookdown_0.18 R6_2.4.1 magick_2.3
## [91] DBI_1.1.0 mgcv_1.8-31 withr_2.2.0
## [94] pillar_1.4.3 survival_3.1-12 RCurl_1.98-1.2
## [97] tibble_3.0.1 crayon_1.3.4 futile.options_1.0.1
## [100] fdrtool_1.2.15 soGGi_1.20.0 BiocFileCache_1.12.0
## [103] jpeg_0.1-8.1 progress_1.2.2 locfit_1.5-9.4
## [106] grid_4.0.0 blob_1.2.1 digest_0.6.25
## [109] xtable_1.8-4 VennDiagram_1.6.20 tidyr_1.0.2
## [112] regioneR_1.20.0 openssl_1.4.1 munsell_0.5.0
## [115] askpass_1.1
Chabbert, C. D., S. H. Adjalley, B. Klaus, E. S. Fritsch, I. Gupta, V. Pelechano, and L. M. Steinmetz. 2015. “A High-Throughput ChIP-Seq for Large-Scale Chromatin Studies.” Molecular Systems Biology 11 (1). EMBO:777–77. https://doi.org/10.15252/msb.20145776.
Hampel, Frank, Christian Hennig, and Elvezio Ronchetti. 2011. “A Smoothing Principle for the Huber and Other Location M-Estimators.” Computational Statistics & Data Analysis 55 (1). Elsevier BV:324–37. https://doi.org/10.1016/j.csda.2010.05.001.