This repository contains codes for processing scATAC-seq data produced by the plate-based scATAC-seq method. The method was originally published in Nature Communications 2018. A more detailed step-by-step protocol was published in Nature Protocols 2021.
In this pipeline, we have:
- Added
config.json
file to make the processing more flexible and easier to modify. - Used the
BED
file for the MACS2 peak calling. See this tweet for the reason. - Produced
10x Genomics
like output files in the finalouts
directory so that they can be easily put into downwstream analysis package like Signac. See below for more details.
Install the following packages either using conda or pip or directly download the pre-compiled binary from the website:
python3
snakemake 5.3.0
numpy v1.18.5
scipy v1.5.0
pandas v1.0.5
hisat2 v2.1.0
samtools v1.9
bedtools v2.27.1
fastp v0.20.1
macs2 v2.2.7.1
tabix 0.2.5
Get the picard tool picard.jar
from https://github.com/broadinstitute/picard/releases
Get calc
, addCols
, bedClip
, bedGraphToBigWig
and fetchChromSizes
from UCSC utilities, and make sure they are executable and in your $PATH
.
Get bdg2bw to convert the macs2 generated begraph to bigwig for visualisation.
There are a few things you need to change in the config.json
file according to your computing environment:
genome: This is the prefix of hisat2
index for the genome you are working on. It is basically passed to the -x
flag of hisat2
during alignment.
hisat2_X: This is the -X
flag of hisat2
, which suggests the maximum frament in bp allowed during the alignment. People normally use 2000
for ATAC-seq.
picard_jar: The location of the picard.jar
file.
blacklist: The ENCODE blacklist region to exclude for analysis. Check this publication for more details. There are a few pre-compiled blacklists for different genome builds that can be found here.
gsize: This is the genome size -g
option for macs2
during the peak calling.
bpk: This contains the macs2
flags for broadPeak calling. In most cases, ATAC-seq signals are sharp, we normally leave this empty here.
chromsize: The location pointing to the tab-delimited file that contains the length of each chromosome. Use the UCSC fetchChromSizes
program to get this file. For example, to get the file for hg38, simply run fetchChromSizes hg38 > hg38.chrom.sizes
.
macs2_format: This is the file format duing macs2
peak calling. We use BED
in this pipeline.
macs2_shift: The flags used for calling narrowPeak. Use --nomodel --shift -100 --extsize 200
to centre the reads on the Tn5 cutting sites.
The starting point of the pipeline is the fastq
files. Put your fastq
files inside each plate directory. Also put the Snakefile
, config.json
and the scripts
folder from this repository to your experiment directory. The structure will be like this:
Experiment
│
├── config.json
│
├── cluster.json
│
├── submit_snake.sh
│
├── plate1
│ │
│ ├── fastq
│ │ │
│ │ ├── scATAC_p1_cell_001_r1.fq.gz
│ │ ├── scATAC_p1_cell_001_r2.fq.gz
│ │ ├── scATAC_p1_cell_002_r1.fq.gz
│ │ ├── scATAC_p1_cell_002_r2.fq.gz
│ │ ├── scATAC_p1_cell_003_r1.fq.gz
│ │ ├── scATAC_p1_cell_003_r2.fq.gz
. . .
. . .
├── plate2
│ │
│ ├── fastq
│ │ │
│ │ ├── scATAC_p2_cell_001_r1.fq.gz
│ │ ├── scATAC_p2_cell_001_r2.fq.gz
│ │ ├── scATAC_p2_cell_002_r1.fq.gz
│ │ ├── scATAC_p2_cell_002_r2.fq.gz
│ │ ├── scATAC_p2_cell_003_r1.fq.gz
│ │ ├── scATAC_p2_cell_003_r2.fq.gz
. . .
. . .
├── scripts
│ │
│ ├── collect_metadata.py
│ ├── generate_count_csc_mtx.py
│ ├── generate_fragments_file.sh
│ ├── get_depth_mr.sh
│ ├── get_dup_level.sh
│ ├── get_frac_open.sh
│ ├── get_frip.sh
│ ├── get_lib_size.sh
│ ├── get_ufrags_mt.sh
│ └── list_bam.sh
│
└── Snakefile
To use all available cores to run the pipeline, simply type snakemake --cores
under the Experiment
directory.
You can also run the pipeline using bsub
, using the command and setting provided in the submit_snake.sh
and cluster.json
files. These two files can be ignored if you are not using bsub
.
There will be quite a few useful intermediate files generated during the process of the pipeline. They are organised into each directory, and the name of the file is self-explanatory. The most important output files are in the outs
directory. If the the pipeline runs successfully, you should expect an outs
directory under the Experiment
directory. Inside the outs
directory, there will be six files:
aggregate_fragments.tsv.gz: this is a tab-delimited file that contains the ATAC fragments of all cells after deduplication, with the following specification:
column | meaning |
---|---|
1st column | chromosome of the fragment |
2nd column | 0-based start coordinate of the fragment |
3rd column | 1-based end corrdinate of the fragment |
4th column | the cell name from where the fragment comes |
5th column | just '1' (might change) |
aggregate_fragments.tsv.gz.tbi: the index of the fragment file, created by the indexFrag
rule from the Snakefile
.
count_matrix_over_aggregate.mtx: the peak-by-cell count matrix in matrix market format
. This is basically sparse.csc_matrix
if you use python
; or this can be treated as dgCMatrix
if you use R
.
count_matrix_over_aggregate.cols: the name of each cell in plain text format.
count_matrix_over_aggregate.rows: the peak location in a 3-column bed
format.
sample_info.csv: a csv file containing the basic quality metrics of each cell. The meaning of each column is described as follows:
column | value | typical range for a successful cell |
---|---|---|
cell | the name of the cell | N/A |
frac_open | percentage (%) of all peaks detected (at least one read) in the cell | 1 - 20 |
mapping_rate | overall alignment rate (%) from hisat2 | 70 - 99 |
mt_content | percentage (%) of reads mapped to the reference genome | 0.1 - 90 |
uniq_nuc_frags | number of read mapped to the nuclear genome after deduplication | 10,000 - 100,000 |
dup_level | duplication level estimated by the picard tool, indicating the fraction of duplicate reads | 0.4 - 0.9 |
frip | percentage (%) of reads that come from the peak region | 20 - 80 |
sequencing_depth | total number of reads sequenced per cell | 10,000 - 1,000,000 |
library_size | library complexity (number of unique fragments) estimated by the picard tool | 10,000 - 1,000,000 |
5. Load the output files into Signac
If you use python
, use mmread
from scipy
to load the mtx
file and conduct analysis using differen packages from scikit-learn
. You can also try EpiScanpy. If you use R
, you have many choices for the analysis. To load data into Signac
, use the following lines of code:
library(Signac)
library(Seurat)
library(ggplot2)
library(patchwork)
library(hdf5r)
library(dplyr)
library(readr)
# read the content from the 'outs' directory
setwd("/your/working/directory")
mex_dir_path <- "/path/to/mtx"
mtx_path <- paste(mex_dir_path, "count_matrix_over_aggregate.mtx", sep = '/')
feature_path <- paste(mex_dir_path, "count_matrix_over_aggregate.rows", sep = '/')
barcode_path <- paste(mex_dir_path, "count_matrix_over_aggregate.cols", sep = '/')
features <- readr::read_tsv(feature_path, col_names = F) %>% tidyr::unite(feature)
barcodes <- readr::read_tsv(barcode_path, col_names = F) %>% tidyr::unite(barcode)
metadata <- read.csv(
file = "/path/to/outs/sample_info.csv",
header = TRUE,
row.names = 1
)
# create a Signac chromatin assay and a Seurat object
mtx <- Matrix::readMM(mtx_path) %>%
magrittr::set_rownames(features$feature) %>%
magrittr::set_colnames(barcodes$barcode)
chrom_assay <- CreateChromatinAssay(
counts = mtx,
sep = c("_", "_"),
genome = 'hg38',
min.cells = 10,
min.features = 200
)
atac <- CreateSeuratObject(
counts = chrom_assay,
assay = 'peaks',
project = 'scATAC-seq_is_cool',
meta.data = metadata
)
Xi Chen
[email protected]