3 | | Nextflow is a workflow system for creating scalable, portable, and reproducible workflows. It allows you to run your analysis pipeline on a large-scale dataset in a streamlined and parallel manner. Nextflow can deploy workflows on a variety of execution platforms, including your local machine, HPC schedulers, and cloud. Additionally, Nextflow supports a range of compute environments, software container runtimes, and package managers, allowing workflows to be executed in reproducible and isolated environments. |
| 3 | Nextflow is a workflow system for creating scalable, portable, and reproducible workflows. It has been developed specifically to ease the creation and execution of bioinformatics pipelines. It allows you to run your analysis pipeline on a large-scale dataset in a streamlined and parallel manner. Nextflow can deploy workflows on a variety of execution platforms, including your local machine, HPC schedulers, and cloud. Additionally, Nextflow supports a range of compute environments, software container runtimes, and package managers, allowing workflows to be executed in reproducible and isolated environments. |
| 79 | The simplest way to run is with -profile docker (or singularity). This instructs Nextflow to execute jobs locally, with Docker (or Singularity) to fulfill software dependencies. |
| 80 | |
| 81 | Please note that if you are running the pipeline on the slurm cluster, you can only use -profile singularity, because you don't have the permission to run docker on it. |
| 82 | |
| 83 | Conda is also supported with -profile conda. However, this option is not recommended, as reproducibility of results can’t be guaranteed without containerization. |
| 84 | |
| 85 | |
| 86 | 3. Run the tests for your pipeline in the terminal to confirm everything is working: |
| 87 | |
| 88 | |
| 89 | {{{ |
| 90 | nextflow run nf-core/<pipeline_name> -profile test,singularity --outdir <OUTDIR> |
| 91 | }}} |
| 92 | |
| 93 | Replace <pipeline_name> with the name of an nf-core pipeline. |
| 94 | |
| 95 | Nextflow will pull the code from the GitHub repository and fetch the software requirements automatically, so there is no need to download anything first. |
| 96 | |
| 97 | |
| 98 | 4. Read the pipeline documentation to see which command-line parameters are required. This will be specific to your data type and usage. |
| 99 | |
| 100 | 5. To launch the pipeline with real data, omit the test config profile and provide the required pipeline-specific parameters. For example, to run the CUTandRun pipeline, your command will be similar to this: |
| 101 | |
| 102 | |
| 103 | {{{ |
| 104 | nextflow run nf-core/cutandrun |
| 105 | -profile singularity |
| 106 | —input samplesheet.csv |
| 107 | —peakcaller ‘seacr,MACS2’ |
| 108 | —genome GRCh38 |
| 109 | —outdir nextflow_cutandrun |
| 110 | }}} |
| 111 | |
| 112 | 6. Once complete, check the pipeline execution and quality control reports (such as multiqc_report.html files for MultiQC reports). Each pipeline’s documentation describes the outputs to expect. |
| 113 | |
| 114 | Please refer to nf-core documentation for more details (https://nf-co.re/docs/usage/getting_started/introduction). |