Nextflow for reproducible and automated data analysis
Are you a research staff member working with bioinformatics data? This hands-on training is designed to help you build scalable, reproducible data analysis pipelines using Nextflow. As data volumes grow and analyses become more complex, mastering workflow management tools is essential for ensuring reproducibility and efficiency. In this course, you’ll learn how to develop and run pipelines on your local machine, HPC clusters, and cloud platforms. Starting from the basics, we’ll guide you through building a simple pipeline and progressively enhance it using Nextflow’s modern DSL2 syntax. By the end, you’ll be equipped with practical skills to streamline your own analyses and collaborate more effectively across computational environments.
Objectives
- Describe the core components of Nextflow (e.g., channels, processes, modules, workflows) and explain their roles in pipeline development.
- Construct a basic Nextflow pipeline using DSL2 syntax and execute it on a local machine.
- Modify configuration files to adapt pipeline parameters for different computing environments (e.g., local, cluster, cloud).
- Apply best practices for reproducibility and scalability in workflow design.
- Debug and refactor a simple pipeline to improve clarity, modularity, and performance.
Organized by VIB Training and Conferences and ELIXIR Belgium