Part 1: Run a demo pipeline¶
In this first part of the Hello nf-core training course, we show you how to find and try out an nf-core pipeline, configure and customize it for your needs, and understand how input validation protects against common errors.
We are going to use a pipeline called nf-core/demo that is maintained by the nf-core project as part of its inventory of pipelines for demonstration and training purposes.
Make sure your working directory is set to hello-nf-core/ as instructed on the Getting started page.
1. Find and retrieve the nf-core/demo pipeline¶
Let's start by locating the nf-core/demo pipeline on the project website at nf-co.re, which centralizes all information such as: general documentation and help articles, documentation for each of the pipelines, blog posts, event announcements and so forth.
1.1. Find the pipeline on the website¶
In your web browser, go to https://nf-co.re/pipelines/ and type demo in the search bar.

Click on the pipeline name, demo, to access the pipeline documentation page.
Each released pipeline has a dedicated page that includes the following documentation sections:
- Introduction: An introduction and overview of the pipeline
- Usage: Descriptions of how to execute the pipeline
- Parameters: Grouped pipeline parameters with descriptions
- Output: Descriptions and examples of the expected output files
- Results: Example output files generated from the full test dataset
- Releases & Statistics: Pipeline version history and statistics
Whenever you are considering adopting a new pipeline, you should read the pipeline documentation carefully first to understand what it does and how it should be configured before attempting to run it.
Have a look now and see if you can find out:
- Which tools the pipeline will run (Check the tab:
Introduction) - Which inputs and parameters the pipeline accepts or requires (Check the tab:
Parameters) - What are the outputs produced by the pipeline (Check the tab:
Output)
1.1.1. Pipeline overview¶
The Introduction tab provides an overview of the pipeline, including a visual representation (called a subway map) and a list of tools that are run as part of the pipeline.

- Read QC (FASTQC)
- Adapter and quality trimming (SEQTK_TRIM)
- Present QC for raw reads (MULTIQC)
1.1.2. Example command line¶
The documentation also provides an example input file (discussed further below) and an example command line.
nextflow run nf-core/demo \
-profile <docker/singularity/.../institute> \
--input samplesheet.csv \
--outdir <OUTDIR>
You'll notice that the example command does NOT specify a workflow file, just the reference to the pipeline repository, nf-core/demo.
When invoked this way, Nextflow will assume that the code is organized in a certain way. Let's retrieve the code so we can examine this structure.
1.2. Retrieve the pipeline code¶
Once we've determined that the pipeline appears to be suitable for our purposes, let's try it out. Fortunately Nextflow makes it easy to retrieve pipelines from correctly-formatted repositories without having to download anything manually.
Let's return to the terminal and run the following:
Command output
Nextflow does a pull of the pipeline code, meaning it downloads the full repository to your local drive.
To be clear, you can do this with any Nextflow pipeline that is appropriately set up in GitHub, not just nf-core pipelines. However nf-core is the largest open-source collection of Nextflow pipelines.
You can get Nextflow to give you a list of what pipelines you have retrieved in this way:
You'll notice that the files are not in your current work directory.
By default, Nextflow saves them to $NXF_HOME/assets.
Note
The full path may differ on your system if you're not using our training environment.
Nextflow keeps the downloaded source code intentionally 'out of the way' on the principle that these pipelines should be used more like libraries than code that you would directly interact with.
Optional: Create a symlink to explore the source code
To make it easier to browse the pipeline source code later, create a symbolic link to the assets directory:
This creates a shortcut so you can explore the code with tree -L 2 pipelines or open files directly.
We will use this in Part 2 when examining how nf-core pipelines are structured internally.
Tip
You can also browse any nf-core pipeline's source code on GitHub, e.g. github.com/nf-core/demo. Every nf-core pipeline follows the same directory layout, so once you know the structure, you can find configuration files, modules, and workflows for any pipeline the same way.
Now, on to running the pipeline!
Takeaway¶
You now know how to find a pipeline via the nf-core website and retrieve a local copy of the source code.
What's next?¶
Learn how to try out an nf-core pipeline with minimal effort.
2. Try out the pipeline with its test profile¶
Conveniently, every nf-core pipeline comes with a test profile. This is a minimal set of configuration settings for the pipeline to run using a small test dataset hosted in the nf-core/test-datasets repository. It's a great way to quickly try out a pipeline at small scale.
Note
Nextflow's configuration profile system allows you to easily switch between different container engines or execution environments. For more details, see Hello Nextflow Part 6: Configuration.
2.1. Examine the test profile¶
It's good practice to check what a pipeline's test profile specifies before running it.
The test profile for nf-core/demo lives in the configuration file conf/test.config.
You can find it locally inside the pipeline source that nextflow pull downloaded:
Here is the content of that file:
You'll notice right away that the comment block at the top includes a usage example showing how to run the pipeline with this test profile.
| conf/test.config | |
|---|---|
The only things we need to supply are what's shown between carets in the example command: <docker/singularity> and <OUTDIR>.
As a reminder, <docker/singularity> refers to the choice of container system. All nf-core pipelines are designed to be usable with containers (Docker, Singularity, etc.) to ensure reproducibility and eliminate software installation issues.
So we'll need to specify whether we want to use Docker or Singularity to test the pipeline.
The --outdir <OUTDIR> part refers to the directory where Nextflow will write the pipeline's outputs.
We need to provide a name for it, which we can just make up.
If it does not exist already, Nextflow will create it for us at runtime.
Moving on to the section after the comment block, the test profile shows us what has been pre-configured for testing: most notably, the input parameter is already set to point to a test dataset, so we don't need to provide our own data.
If you follow the link to the pre-configured input, you'll see it is a csv file containing sample identifiers and file paths for several experimental samples.
sample,fastq_1,fastq_2
SAMPLE1_PE,https://raw.githubusercontent.com/nf-core/test-datasets/viralrecon/illumina/amplicon/sample1_R1.fastq.gz,https://raw.githubusercontent.com/nf-core/test-datasets/viralrecon/illumina/amplicon/sample1_R2.fastq.gz
SAMPLE2_PE,https://raw.githubusercontent.com/nf-core/test-datasets/viralrecon/illumina/amplicon/sample2_R1.fastq.gz,https://raw.githubusercontent.com/nf-core/test-datasets/viralrecon/illumina/amplicon/sample2_R2.fastq.gz
SAMPLE3_SE,https://raw.githubusercontent.com/nf-core/test-datasets/viralrecon/illumina/amplicon/sample1_R1.fastq.gz,
SAMPLE3_SE,https://raw.githubusercontent.com/nf-core/test-datasets/viralrecon/illumina/amplicon/sample2_R1.fastq.gz,
This is called a samplesheet, and is the most common form of input to nf-core pipelines.
Note
Don't worry if you're not familiar with the data formats and types, it's not important for what follows.
So this confirms that we have everything we need to try out the pipeline.
2.2. Run the pipeline¶
Let's decide to use Docker for the container system and demo-results as the output directory, and we're ready to run the test command:
Command output
N E X T F L O W ~ version 25.04.3
Launching `https://github.com/nf-core/demo` [magical_pauling] DSL2 - revision: db7f526ce1 [master]
------------------------------------------------------
,--./,-.
___ __ __ __ ___ /,-._.--~'
|\ | |__ __ / ` / \ |__) |__ } {
| \| | \__, \__/ | \ |___ \`-._,-`-,
`._,._,'
nf-core/demo 1.0.2
------------------------------------------------------
Input/output options
input : https://raw.githubusercontent.com/nf-core/test-datasets/viralrecon/samplesheet/samplesheet_test_illumina_amplicon.csv
outdir : demo-results
Institutional config options
config_profile_name : Test profile
config_profile_description: Minimal test dataset to check pipeline function
Generic options
trace_report_suffix : 2025-11-21_04-57-41
Core Nextflow options
revision : master
runName : magical_pauling
containerEngine : docker
launchDir : /workspaces/training/hello-nf-core
workDir : /workspaces/training/hello-nf-core/work
projectDir : /workspaces/.nextflow/assets/nf-core/demo
userName : root
profile : docker,test
configFiles : /workspaces/.nextflow/assets/nf-core/demo/nextflow.config
!! Only displaying parameters that differ from the pipeline defaults !!
------------------------------------------------------
* The pipeline
https://doi.org/10.5281/zenodo.12192442
* The nf-core framework
https://doi.org/10.1038/s41587-020-0439-x
* Software dependencies
https://github.com/nf-core/demo/blob/master/CITATIONS.md
executor > local (7)
[ff/a6976b] NFCORE_DEMO:DEMO:FASTQC (SAMPLE3_SE) | 3 of 3 ✔
[39/731ab7] NFCORE_DEMO:DEMO:SEQTK_TRIM (SAMPLE3_SE) | 3 of 3 ✔
[7c/78d96e] NFCORE_DEMO:DEMO:MULTIQC | 1 of 1 ✔
-[nf-core/demo] Pipeline completed successfully-
If your output matches that, congratulations! You've just run your first nf-core pipeline.
You'll notice that there is a lot more console output than when you run a basic Nextflow pipeline. There's a header that includes a summary of the pipeline's version, inputs and outputs, and a few elements of configuration.
Note
Your output will show different timestamps, execution names, and file paths, but the overall structure and process execution should be similar.
Notice the line near the top of the output:
This tells you which revision of the pipeline was used.
Because we did not specify a version, Nextflow used the latest commit on master.
For reproducible runs, you should pin a specific release using the -r flag:
This ensures that the same pipeline code is used every time, regardless of new commits or releases.
For this training we omit -r for simplicity, but in production you should always specify it.
Moving on to the execution output, let's have a look at the lines that tell us what processes were run:
[ff/a6976b] NFCORE_DEMO:DEMO:FASTQC (SAMPLE3_SE) | 3 of 3 ✔
[39/731ab7] NFCORE_DEMO:DEMO:SEQTK_TRIM (SAMPLE3_SE) | 3 of 3 ✔
[7c/78d96e] NFCORE_DEMO:DEMO:MULTIQC | 1 of 1 ✔
This tells us that three processes were run, corresponding to the three tools shown in the pipeline documentation page on the nf-core website: FASTQC, SEQTK_TRIM and MULTIQC.
The full process names as shown here, such as NFCORE_DEMO:DEMO:MULTIQC, are longer than what you may have seen in the introductory Hello Nextflow material.
These include the names of their parent workflows and reflect the modularity of the pipeline code.
We'll go into more detail about that in Part 2 of this course.
2.3. Examine the pipeline's outputs¶
Finally, let's have a look at the demo-results directory produced by the pipeline.
Directory contents
demo-results
├── fastqc
│ ├── SAMPLE1_PE
│ ├── SAMPLE2_PE
│ └── SAMPLE3_SE
├── fq
│ ├── SAMPLE1_PE
│ ├── SAMPLE2_PE
│ └── SAMPLE3_SE
├── multiqc
│ ├── multiqc_data
│ ├── multiqc_plots
│ └── multiqc_report.html
└── pipeline_info
├── execution_report_2025-11-21_04-57-41.html
├── execution_timeline_2025-11-21_04-57-41.html
├── execution_trace_2025-11-21_04-57-41.txt
├── nf_core_demo_software_mqc_versions.yml
├── params_2025-11-21_04-57-46.json
└── pipeline_dag_2025-11-21_04-57-41.html
That might seem like a lot.
To learn more about the nf-core/demo pipeline's outputs, check out its documentation page.
At this stage, what's important to observe is that the results are organized by module, and there is additionally a directory called pipeline_info containing various timestamped reports about the pipeline execution.
For example, the execution_timeline_* file shows you what processes were run, in what order and how long they took to run:

Note
Here the tasks were not run in parallel because we are running on a minimalist machine in Github Codespaces. To see these run in parallel, try increasing the CPU allocation of your codespace and the resource limits in the test configuration.
These reports are generated automatically for all nf-core pipelines.
Takeaway¶
You know how to run an nf-core pipeline using its built-in test profile and where to find its outputs.
What's next?¶
Learn how to configure and customize the pipeline.
3. Configure and customize the pipeline¶
Nextflow distinguishes between parameters (what a pipeline processes) and configuration (how it runs).
nf-core builds on this by requiring all parameters to be formally defined in a JSON schema file (nextflow_schema.json).
This schema records each parameter's type, description, default value, and grouping.
It powers the --help output, automated validation at startup, the Parameters tab on the nf-core website that you explored in section 1, and launch forms in tools like Seqera Platform.
3.1. Get help¶
nf-core pipelines use the nf-schema plugin to generate structured, grouped --help output automatically from the schema file.
Run the help command for the demo pipeline:
The output groups parameters into categories (Input/output options, Reference genome options, etc.) with types and descriptions for each one.
In plain Nextflow pipelines, --help only works if the developer implemented it manually.
Tip
Use --help --show_hidden to see additional parameters that are hidden by default, such as --publish_dir_mode or --monochrome_logs.
3.2. Set parameters¶
As covered in Hello Config, you can override parameters on the command line with --param_name or collect them in a YAML file and pass it with -params-file.
Both approaches work the same way with nf-core pipelines.
For example, to skip the trimming step:
The SEQTK_TRIM process no longer appears in the output.
nf-core discourages setting parameters inside config files using params { } blocks.
Parameters defined in a custom config file passed with -c may not override defaults already set in the pipeline's own nextflow.config, depending on Nextflow's configuration precedence rules.
Using --param_name on the command line or -params-file is more reliable, as these always take precedence.
Warning
As a rule of thumb: if it appears in the --help output, set it via the command line or a params file. Everything else belongs in a config file.
3.3. Input validation¶
nf-core pipelines use the nf-schema plugin to validate inputs at startup, before any processes run.
Because all parameters are defined in the nextflow_schema.json schema, nf-schema can check that every parameter you pass is recognized and has the correct value.
3.3.1. Unrecognized parameters¶
Try passing a parameter that does not exist:
The console output includes a warning:
The pipeline still runs, but the warning alerts you that --foobar is not a recognized parameter.
This catches typos like --outDir instead of --outdir before you waste compute time wondering why the output went to the wrong place.
3.3.2. Invalid parameter values¶
Validation also checks parameter values.
The --skip_trim parameter is a boolean flag, so passing a string value causes the pipeline to fail immediately:
ERROR ~ Validation of pipeline parameters failed!
The following invalid input values have been detected:
* --skip_trim (yes): Value is [string] but should be [boolean]
The pipeline stops before any processes run, saving you from a failed or incorrect execution.
Boolean parameters should be passed as flags (--skip_trim) without a value, or set to true/false in a params file.
3.3.3. Samplesheet validation¶
nf-schema also validates the input samplesheet against a schema.
The nf-core/demo pipeline expects a CSV file with columns sample, fastq_1, and fastq_2.
This is defined in a schema file (assets/schema_input.json) that specifies the expected structure, column types, and constraints.
To see this in action, create a samplesheet with a missing column and a non-existent file path:
nextflow run nf-core/demo -profile docker,test --outdir demo-results --input malformed_samplesheet.csv
The pipeline fails immediately with a clear error message:
ERROR ~ Validation of pipeline parameters failed!
The following invalid input values have been detected:
* --input (malformed_samplesheet.csv): Validation of file failed:
-> Entry 1: Error for field 'fastq_1' (/not/a/real/file.fastq.gz): the file or directory
'/not/a/real/file.fastq.gz' does not exist
The error identifies the exact entry and field that caused the problem, so you can fix your samplesheet without having to debug a mid-pipeline failure.
Tip
We cover how to define these schemas from the developer side in Part 5 of this course.
3.4. Customize configuration¶
Configuration controls how the pipeline runs: resource allocation, tool-specific arguments, where jobs execute, and which software packaging system to use.
nf-core pipelines include default configuration in nextflow.config and the conf/ directory.
Before overriding anything, it helps to know where the defaults live.
You already saw in section 2.1 that the pipeline source code lives in $NXF_HOME/assets.
List the config files to see what's available:
The three most important configuration files are:
conf/base.config: Defines resource labels (process_low,process_medium,process_high) that assign CPUs, memory, and time to processes. When you see a process using more resources than expected, this is where those defaults come from.conf/modules.config: Sets per-process tool arguments (ext.args) and output publishing settings (publishDir). Open this file to see what arguments each tool receives by default.conf/test.config: The test profile you used in section 2.1, which caps resources viaresourceLimitsand sets a test samplesheet. Activated with-profile test.
There is also a conf/test_full.config for running with a full-sized test dataset, useful for benchmarking.
The central nextflow.config loads all of the above and sets default parameter values.
You do not need to modify any of these files directly.
Instead, create your own config file and pass it with -c. Your values override the pipeline defaults.
3.4.1. Change resource allocation for a process¶
The demo pipeline assigns resources using labels defined in base.config.
For example, FASTQC uses the process_medium label, which allocates 6 CPUs and 36 GB of memory.
The test profile caps resources via resourceLimits, but you can also override resources for specific processes.
Create a file called custom.config:
Run the pipeline with your custom config:
The -c flag adds your config on top of the pipeline's built-in configuration.
Your values take precedence for the processes you specify.
3.4.2. Override tool arguments with ext.args¶
Some tool options are exposed as pipeline parameters. For example, --skip_trim controls whether SEQTK_TRIM runs at all.
For tool options that are not exposed as parameters, nf-core modules use a convention called ext.args to pass arguments to the underlying tool through configuration.
For example, the FASTQC module is configured with ext.args = '--quiet' by default (defined in conf/modules.config).
The SEQTK_TRIM module has no default ext.args.
To add a trimming argument to SEQTK_TRIM, update your custom.config:
| custom.config | |
|---|---|
This tells seqtk trimfq to trim 5 bases from the beginning of each read in addition to quality trimming.
Run the pipeline again with this config to see the effect:
To verify the argument was applied, find the SEQTK_TRIM work directory hash from the run output (e.g. work/ab/cd1234...) and check the .command.sh file inside it:
You should see -b 5 in the seqtk trimfq command, confirming your ext.args override took effect.
Note
The ext.args value completely replaces the default for that process.
If the default was '--quiet' and you set ext.args = '--kmers 8', the --quiet flag is no longer applied.
To keep both, set ext.args = '--quiet --kmers 8'.
Takeaway¶
You know how to get help from an nf-core pipeline, set parameters and understand how they are validated, and customize configuration through config files.
What's next?¶
Take a break! When you're ready, move on to Part 2, where you will create your own nf-core compatible pipeline from scratch.
Tip
If you would like to learn how to compose workflows with subworkflows before moving on to the next part, check out the Workflows of Workflows Side Quest.