-
Notifications
You must be signed in to change notification settings - Fork 14
/
base.config
executable file
·123 lines (106 loc) · 3.68 KB
/
base.config
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
params {
wf {
basecaller_container = "ontresearch/dorado"
container_sha_basecalling = "sha8bc19cc3c78c1ce847fbf4a879cad997b7e8c430"
bonito_container = "ontresearch/bonito"
bonito_sha = "shaea43ca2333f91fa78a823f640ba158e4268f1f98"
common_sha = "shad28e55140f75a68f59bbecc74e880aeab16ab158"
}
}
// used by default for "standard" (docker) and singularity profiles,
// other profiles may override.
process {
withLabel:wf_basecalling {
container = "${params.wf.basecaller_container}:${params.wf.container_sha_basecalling}"
}
withLabel:wf_common {
container = "ontresearch/wf-common:${params.wf.common_sha}"
}
shell = ['/bin/bash', '-euo', 'pipefail']
// by default GPU tasks will run in serial to avoid GPU management.
// cluster and cloud users can remove this with -profile discrete_gpus.
// we use profiles to handle this as maxForks cannot be set dynamically
// see https://github.com/nextflow-io/nextflow/discussions/3806 and CW-1857
withLabel:gpu {
maxForks = 1
}
}
profiles {
// the "standard" profile is used implicitely by nextflow
// if no other profile is given on the CLI
standard {
docker {
enabled = true
// this ensures container is run as host user and group, but
// also adds host user to the within-container group
runOptions = "--user \$(id -u):\$(id -g) --group-add 100"
}
process."withLabel:gpu".containerOptions = "--gpus all"
}
// using singularity instead of docker
singularity {
singularity {
enabled = true
autoMounts = true
//envWhitelist = "" // if your cluster sets a variable to indicate which GPU has been assigned you will want to allow it here
}
process."withLabel:gpu".containerOptions = "--nv"
}
// keep stub conda profile to prevent unknown profile warning so users get a better error
conda {
conda.enabled = true
}
// Using AWS batch.
// May need to set aws.region and aws.batch.cliPath
awsbatch {
process {
executor = 'awsbatch'
queue = "${params.aws_queue}"
memory = "16 GB" // likely not enough!
withLabel:wf_common {
container = "${params.aws_image_prefix}-wf-common:${params.wf.common_sha}"
}
shell = ['/bin/bash', '-euo', 'pipefail']
// lift limit on simultaneous gpu jobs for cloud
// and ensure that the host mounts relevant driver bobbins inside the container
withLabel:gpu {
maxForks = null
containerOptions = "-e NVIDIA_DRIVER_CAPABILITIES=compute,utility --gpus all"
}
withLabel:wf_basecalling {
container = "${params.aws_image_prefix}-dorado:${params.wf.container_sha_basecalling}"
}
withLabel:wf_bonito {
container = "${params.aws_image_prefix}-bonito:${params.wf.bonito_sha}"
}
}
}
// local profile for simplified development testing
local {
process.executor = 'local'
}
// lift limit on simultaneous gpu jobs
discrete_gpus {
process."withLabel:gpu".maxForks = null
}
}
timeline {
enabled = true
file = "${params.out_dir}/execution/timeline.html"
overwrite = true
}
report {
enabled = true
file = "${params.out_dir}/execution/report.html"
overwrite = true
}
trace {
enabled = true
file = "${params.out_dir}/execution/trace.txt"
overwrite = true
}
env {
PYTHONNOUSERSITE = 1
JAVA_TOOL_OPTIONS = "-Xlog:disable -Xlog:all=warning:stderr"
}
cleanup = true