Skip to content

Commit

Permalink
Merge pull request #4 from crickbabs/master
Browse files Browse the repository at this point in the history
Put input files into channels
  • Loading branch information
drpatelh committed Dec 11, 2018
2 parents 4516f13 + e1039bb commit 31bb2c0
Show file tree
Hide file tree
Showing 9 changed files with 95 additions and 162 deletions.
4 changes: 2 additions & 2 deletions Dockerfile
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
FROM nfcore/base

MAINTAINER Harshil Patel <harshil.patel@crick.ac.uk>
LABEL authors="harshil.patel@crick.ac.uk" \
description="Docker image containing all requirements for the nfcore/atacseq pipeline"
maintainer="harshil.patel@crick.ac.uk" \
description="Docker image containing all requirements for the nfcore/atacseq pipeline" \

COPY environment.yml /
RUN conda env create -f /environment.yml && conda clean -a
Expand Down
23 changes: 0 additions & 23 deletions conf/crick.config

This file was deleted.

6 changes: 3 additions & 3 deletions conf/test.config
Original file line number Diff line number Diff line change
Expand Up @@ -13,11 +13,11 @@ params {
max_time = 12.h

// Input data
design = 'https://github.com/nf-core/test-datasets/raw/atacseq/design.csv'
design = 'https://raw.githubusercontent.com/nf-core/test-datasets/atacseq/design.csv'

// Genome references - use existing files from nf-core/chipseq
fasta = 'https://github.com/nf-core/test-datasets/raw/chipseq/reference/genome.fa'
gtf = 'https://github.com/nf-core/test-datasets/raw/chipseq/reference/genes.gtf'
fasta = 'https://raw.githubusercontent.com/nf-core/test-datasets/chipseq/reference/genome.fa'
gtf = 'https://raw.githubusercontent.com/nf-core/test-datasets/chipseq/reference/genes.gtf'

// Not mandatory but permits the pipeline to run through peak-calling steps
macs_gsize = 1.2e7
Expand Down
4 changes: 2 additions & 2 deletions docs/configuration/adding_your_own.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,11 +2,11 @@

It is entirely possible to run this pipeline on other clusters, though you will need to set up your own config file so that the pipeline knows how to work with your cluster.

> If you think that there are other people using the pipeline who would benefit from your configuration (eg. other common cluster setups), please let us know. We can add a new configuration and profile which can used by specifying `-profile <name>` when running the pipeline.
> If you think that there are other people using the pipeline who would benefit from your configuration (eg. other common cluster setups), please let us know. We can add a new configuration and profile which can used by specifying `-profile <name>` when running the pipeline. The config file will then be hosted at `nf-core/configs` and will be pulled automatically before the pipeline is executed.
If you are the only person to be running this pipeline, you can create your config file as `~/.nextflow/config` and it will be applied every time you run Nextflow. Alternatively, save the file anywhere and reference it when running the pipeline with `-c path/to/config` (see the [Nextflow documentation](https://www.nextflow.io/docs/latest/config.html) for more).

A basic configuration comes with the pipeline, which runs by default (the `standard` config profile - see [`conf/base.config`](../../conf/base.config)). This means that you only need to configure the specifics for your system and overwrite any defaults that you want to change.
A basic configuration comes with the pipeline, which loads the [`conf/base.config`](../../conf/base.config) by default. This means that you only need to configure the specifics for your system and overwrite any defaults that you want to change.

## Cluster Environment
By default, pipeline uses the `local` Nextflow executor - in other words, all jobs are run in the login session. If you're using a simple server, this may be fine. If you're using a compute cluster, this is bad as all jobs will run on the head node.
Expand Down
38 changes: 0 additions & 38 deletions docs/configuration/crick.md

This file was deleted.

5 changes: 0 additions & 5 deletions docs/installation.md
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,6 @@ To start using the nf-core/atacseq pipeline, follow the steps below:
4. [Reference genomes](#4-reference-genomes)
5. [Appendices](#5-appendices)
* [Running on UPPMAX](#running-on-uppmax)
* [Running at the Crick](#running-at-the-crick)

## 1) Install NextFlow
Nextflow runs on most POSIX systems (Linux, Mac OSX etc). It can be installed by running the following commands:
Expand Down Expand Up @@ -119,7 +118,3 @@ Note that you will need to specify your UPPMAX project ID when running a pipelin
```nextflow
params.project = 'project_ID' // eg. b2017123
```

#### Running at the Crick

See [`docs/configuration/crick.md`](configuration/crick.md)
52 changes: 28 additions & 24 deletions docs/usage.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,14 +8,11 @@
* [Reproducibility](#reproducibility)
* [Main arguments](#main-arguments)
* [`-profile`](#-profile-single-dash)
* [`standard`](#standard)
* [`awsbatch`](#awsbatch)
* [`conda`](#conda)
* [`docker`](#docker)
* [`singularity`](#singularity)
* [`conda`](#conda)
* [`awsbatch`](#awsbatch)
* [`crick`](#crick)
* [`test`](#test)
* [`none`](#none)
* [`--design`](#--design)
* [Generic arguments](#generic-arguments)
* [`--singleEnd`](#--singleEnd)
Expand All @@ -31,6 +28,7 @@
* [`--macs_gsize`](#--macs_gsize)
* [`--blacklist`](#--blacklist)
* [`--saveReference`](#--saveReference)
* [`--igenomesIgnore`](#--igenomesignore)
* [Adapter trimming](#adapter-trimming)
* [`--skipTrimming`](#--skipTrimming)
* [`--saveTrimmed`](#--saveTrimmed)
Expand All @@ -53,10 +51,11 @@
* [`-name`](#-name-single-dash)
* [`-resume`](#-resume-single-dash)
* [`-c`](#-c-single-dash)
* [`--custom_config_version`](#--custom_config_version)
* [`--max_memory`](#--max_memory)
* [`--max_time`](#--max_time)
* [`--max_cpus`](#--max_cpus)
* [`--plaintext_emails`](#--plaintext_emails)
* [`--plaintext_email`](#--plaintext_email)
* [`--multiqc_config`](#--multiqc_config)

## General Nextflow info
Expand Down Expand Up @@ -104,28 +103,22 @@ This version number will be logged in reports when you run the pipeline, so that
### `-profile`
Use this parameter to choose a configuration profile. Profiles can give configuration presets for different compute environments. Note that multiple profiles can be loaded, for example: `-profile standard,docker` - the order of arguments is important!

* `standard`
* The default profile, used if `-profile` is not specified at all.
* Runs locally and expects all software to be installed and available on the `PATH`.
If `-profile` is not specified at all the pipeline will be run locally and expects all software to be installed and available on the `PATH`.

* `awsbatch`
* A generic configuration profile to be used with AWS Batch.
* `conda`
* A generic configuration profile to be used with [conda](https://conda.io/docs/)
* Pulls most software from [Bioconda](https://bioconda.github.io/)
* `docker`
* A generic configuration profile to be used with [Docker](http://docker.com/)
* Pulls software from dockerhub: [`nfcore/atacseq`](http://hub.docker.com/r/nfcore/atacseq/)
* `singularity`
* A generic configuration profile to be used with [Singularity](http://singularity.lbl.gov/)
* Pulls software from singularity-hub
* `conda`
* A generic configuration profile to be used with [conda](https://conda.io/docs/)
* Pulls most software from [Bioconda](https://bioconda.github.io/)
* `awsbatch`
* A generic configuration profile to be used with AWS Batch.
* `crick`
* Designed to use Singularity on the CAMP HPC system at [The Francis Crick Institute](https://www.crick.ac.uk/)
* See [`docs/configuration/crick.md`](configuration/crick.md)
* `test`
* A profile with a complete configuration for automated testing
* Includes links to test data so needs no other parameters
* `none`
* No configuration at all. Useful if you want to build your own config from scratch and want to avoid loading in the default `base` config profile (not recommended).

### `--design`
You will need to create a design file with information about the samples in your experiment before running the pipeline. Use this parameter to specify its location.
Expand Down Expand Up @@ -178,7 +171,7 @@ The pipeline config files come bundled with paths to the illumina iGenomes refer
### `--genome` (using iGenomes)
There are 31 different species supported in the iGenomes references. To run the pipeline, you must specify which to use with the `--genome` flag.

You can find the keys to specify the genomes in the [`igenomes.config`](../conf/igenomes.config). Common genomes that are supported are:
You can find the keys to specify the genomes in the [`iGenomes config file`](../conf/igenomes.config). Common genomes that are supported are:

* Human
* `--genome GRCh37`
Expand Down Expand Up @@ -251,6 +244,9 @@ If provided, alignments that overlap with the regions in this file will be filte
### `--saveReference`
Supply this parameter to save any generated reference genome files such as the BWA index to your results folder. These can then be used for future pipeline runs, reducing processing times.

### `--igenomesIgnore`
Do not load `igenomes.config` when running the pipeline. You may choose this option if you observe clashes between custom parameters and those supplied in `igenomes.config`.

## Adapter trimming
The pipeline accepts a number of parameters to change how the trimming is done, according to your data type.
You can specify custom trimming parameters as follows:
Expand Down Expand Up @@ -292,7 +288,11 @@ By default, intermediate BAM files will not be saved. The final BAM files create
Each step in the pipeline has a default set of requirements for number of CPUs, memory and time. For most of the steps in the pipeline, if the job exits with an error code of `143` (exceeded requested resources) it will automatically resubmit with higher requests (2 x original, then 3 x original). If it still fails after three times then the pipeline is stopped.

### Custom resource requests
Wherever process-specific requirements are set in the pipeline, the default value can be changed by creating a custom config file. See the files in [`conf`](../conf) for examples.
Wherever process-specific requirements are set in the pipeline, the default value can be changed by creating a custom config file. See the files hosted at [`nf-core/configs`](https://github.com/nf-core/configs/tree/master/conf) for examples.

If you are likely to be running `nf-core` pipelines regularly it may be a good idea to request that your custom config file is uploaded to the `nf-core/configs` git repository. Before you do this please can you test that the config file works with your pipeline of choice using the `-c` parameter (see definition below). You can then create a pull request to the `nf-core/configs` repository with the addition of your config file, associated documentation file (see examples in [`nf-core/configs/docs`](https://github.com/nf-core/configs/tree/master/docs)), and amending [`nfcore_custom.config`](https://github.com/nf-core/configs/blob/master/nfcore_custom.config) to include your custom profile.

If you have any questions or issues please send us a message on [`Slack`](https://nf-core-invite.herokuapp.com/).

## AWS Batch specific parameters
Running the pipeline on AWS Batch requires a couple of specific parameters to be set according to your AWS Batch configuration. Please use the `-awsbatch` profile and then specify all of the following parameters.
Expand Down Expand Up @@ -330,10 +330,14 @@ Specify the path to a specific config file (this is a core NextFlow command).

**NB:** Single hyphen (core Nextflow option)

Note - you can use this to override defaults. For example, you can specify a config file using `-c` that contains the following:
Note - you can use this to override pipeline defaults.

```nextflow
process.$multiqc.module = []
### `--custom_config_version`
Provide git commit id for custom Institutional configs hosted at `nf-core/configs`. This was implemented for reproducibility purposes. Default is set to `master`.

```bash
## Download and use config file with following git commid id
--custom_config_version d52db660777c4bf36546ddb188ec530c3ada1b96
```

### `--max_memory`
Expand Down
Loading

0 comments on commit 31bb2c0

Please sign in to comment.