Author: Barry Digby
Contact Info: [email protected]
System Administrator: Chris Duke
To use the lugh configuration profile with your pipeline, add -profile lugh
to your nextflow run
command:
nextflow -bg run nf-core/rnaseq -profile test,nuig
Please take care to use the -bg
flag, or run the job on a compute node.
The configuration file will load prerequisite modules for users (Java
& Singularity
), however it is up to the user to have a functional version of nextflow
installed in their path. Follow nextflow
installation instructions at the following link.
Queue | Hostnames | Max Memory | Max CPUS | Max Time |
---|---|---|---|---|
MSC | compute[01-03] | 32GB | 16 | 336.h |
Normal | compute[10-29] | 64GB | 16 | 240.h |
Highmem | compute[04-09] | 128GB | 32 | 2880.h |
The configuration profile design is very simple. If your process exceeds 64GB memory or 16 cpus, it is sent to the highmem
queue. If not, it is sent to the normal
queue. Please do not use the MSC
queue, this is reserved for Masters students.
Take others into consideration when deploying your workflow (do not hog the cluster 🐷). If you need to hammer the cluster with a pipeline, please reach out to me and we can tweak the configuration profile to dispatch jobs to only a handful of compute nodes via hostnames.
Your workflow containers are stored under /data/containers/
which is accessible to all users on lugh.