Install an environment with snakemake:
conda install -y -c conda-forge mamba
mamba create -y -c conda-forge -c bioconda -n snakemake python=3.9.1 snakemake=6.9.1 numpy=1.21.2 pandas=1.3.3 matplotlib=3.4.3 seaborn=0.11.2
This may take a while.
If snakemake says mamba is not installed, run
conda install -y -n snakemake -c conda-forge mamba
while snakemake environment is activated.
You would need to install a slurm profile if you want to submit jobs on computer clusters. If you have a workstation with enough resources, this step could be skipped.
Please install the profile from Ben Parks:
git clone https://github.com/bnprks/snakemake-slurm-profile.git
bash install_snakemake_profile.sh
Ben's defaults are:
"time" : "01:00:00",
"cores" : "{threads}",
"partition" : "wjg,biochem,sfgf",
"priority" : "normal",
"memory" : "8G"
Follow the instructions here to set up.
To use the slurm profile in the Snakefile with custom parameters, add in the params:
section:
rule my_job:
input: ...
output: ...
threads: 8
params:
cluster_memory = "12G",
cluster_time = "4:00:00",
num_cores = "6"
shell: ...
I put mine at $GROUP_HOME
git clone https://github.com/keyuxi/array_analysis.git
For aligning paired end reads. The path should be entered in the snakemake configuration file.
To run, copy a configuration yaml file in the config
directory for the experiment and make modifications. Usually, make one config file that corresponds to one working directory on $SCRATCH
or $GROUP_SCRATCH
(recommended) for each MiSeq chip. An example yaml file is pasted below:
# === Experiment-dependent configurations === #
experimentName:
NNNlib2b
processingType:
# either 'pre-array' or 'post-array'
# 'pre-array' for sequencing processing only
pre-array
datadir:
/scratch/groups/wjg/kyx/NNNlib2b_Nov11/data/
fastq:
read1:
/oak/stanford/groups/wjg/kyx/data/NNNlib2b_processing_20211111/NNNlib2b_S1_L001_R1_001.fastq
read2:
/oak/stanford/groups/wjg/kyx/data/NNNlib2b_processing_20211111/NNNlib2b_S1_L001_R2_001.fastq
sequencingResult:
# csv (CPseq) file holding sequencing data aligned to the library
# ocation: {datadir}/aligned/
ConsensusReads_20211115_exact.CPseq
referenceLibrary:
# RefSeqs of the library design for alignment
# should have a column named 'RefSeq'
/scratch/groups/wjg/kyx/NNNlib2b_Nov11/data/reference/NNNlib2b_RefSeqs.csv
tifdir:
# directory holding all tif images, with a trailing slash
/scratch/groups/wjg/kyx/NNNlib2b_Nov11/data/images_20211022/
fluordir:
# output directory for CPfluor files, with a trailing slash
# recommemded: {datadir}/fluor/
/scratch/groups/wjg/kyx/NNNlib2b_Nov11/data/fluor/
seriesdir:
# output directory for CPseries files, with a trailing slash
# recommended: {datadir}/series/
/scratch/groups/wjg/kyx/NNNlib2b_Nov11/data/series/
mapfile:
# a csv file with column `condition`
config/nnnlib2b_map.csv
FIDfilter:
data/reference/FID.CPfilter
LibRegionFilter:
data/reference/LibRegion.CPfilter
# ====== Environment Configurations =====
FLASHdir:
/home/groups/wjg/kyx
An example directory structure of the data processing directory after running this snakemake workflow is shown below, where different subfolders in data
are different steps of processing. Note that you may include array data from multiple experiments but the same chip here. This directory is usually kept on $OAK or $GROUP_SCRATCH. Note that if it's on $GROUP_SCRATCH, you will need to copy the needed output files to $OAK later.
I also currently keep the folders of images for registration here and run a bash script stored in bash_scripts
on array day. This is fast and it works, but may be incorporated in the future.
{name-of-your-chip}
├── bash_scripts
├── data
│ ├── aligned
│ ├── fastq
│ ├── fiducial_images
│ ├── filtered_tiles
│ ├── filtered_tiles_libregion
│ ├── fluor
│ ├── {your-array-experiment-name}
│ ├── {another-array-experiment-name}
│ ├── {registration-images}
│ ├── library
│ ├── reference
│ ├── roff
│ ├── tiles
│ └── tmp
├── fig
│ └── fiducial
└── out
You should prepare this working directory (datadir
in the config file) before running the pipeline. An example directory after sequencing but before array is:
{name-of-your-chip}
├── bash_scripts
├── data
└── fastq
An example workflow after sequencing but before array experiment may look like:
an example workflow after sequencing but before array experiment
Before running the workflow,
conda activate snakemake
snakemake -np | less
for a dry run. Snakemake will print out the commands to execute.
To execute, run
snakemake --profile slurm --use-conda
The jobs will be automatically submitted in the right order.
At the beginning of Snakefile
, modify here
####### SELECT CONFIG FILE HERE #######
configfile: "config/config_NNNlib2b_Nov11.yaml"
#######################################
-
At the end of the configuration, change the
FLASH
location to where you installed it# ====== Environment Configurations ===== FLASHdir: /home/groups/wjg/kyx
-
The experiment name will show up in some of the processed data file names and is quite flexible:
experimentName: NNNlib2b
-
Depending on whether you are only processing the sequencing data or the array data, set
processingType: # either 'pre-array' or 'post-array' # 'pre-array' for sequencing processing only pre-array
-
Add the locations of your main data processing directory. Point to where the fastq files and your reference library are located
datadir: /scratch/groups/wjg/kyx/NNNlib2b_Nov11/data/ fastq: read1: /oak/stanford/groups/wjg/kyx/data/NNNlib2b_processing_20211111/NNNlib2b_S1_L001_R1_001.fastq read2: /oak/stanford/groups/wjg/kyx/data/NNNlib2b_processing_20211111/NNNlib2b_S1_L001_R2_001.fastq referenceLibrary: # RefSeqs of the library design for alignment # should have a column named 'RefSeq' /scratch/groups/wjg/kyx/NNNlib2b_Nov11/data/reference/NNNlib2b_RefSeqs.csv
Your reference library should be what you designed and in read1 orientation. The sequences should be in a column named
RefSeq
(case sensitive).I left the name of the output of the alignment customizable in case one needs to play with alignment parameters.
The fiducial filter file is included in the github repository. If you used a different fiducial (unlikely), you may need to point to another fiducial file.
You will get:
- Fiducial filtered CPseq files for registration on the array day and plots of them in
fig/fiducial
- CPseq file aligned to the designed library and an aggregated STATS report
- TODO: Add QC plots of the sequencing results to the workflow
Note: array data tif files should have the same prefix, i.e. experiment name like NNNlib2b_DNA
, for snakemake to match the filenames successfully.
-
In the config file, modify
processingType
topost-array
. -
If you need to tune the global variables for registration, the file is located at
scripts/array_tools/GlobalVars.m
. Common parameters to play with includescorrelationSuccessPeakHeight
,maxOffsetGlobalRegistration
, andscalingFactor
.
Run all snakemake
commands right at the array_analysis
directory, where Snakefile
lives. snakemake
will try to find Snakefile
here.
Run with -np
for a dry run. Snakemake will print out the commands to execute.
conda activate snakemake
snakemake -np | less
Optionally, plot the DAG with
snakemake --dag | dot -Tsvg > dag.svg
To execute, run
snakemake --profile slurm --use-conda
The jobs will be automatically submitted in the right order.
The first time you run the pipeline, snakemake will use the yaml files in envs/
to install conda
environments. Version numbers of packages are included in these files. After the first run, these installed conda environments will be used and the future runs will be faster.
If you shut down you computer after the first jobs were submitted and walk away, it should be fine even if the future jobs were not yet submitted.
Yuxi Ke ([email protected])