Count Reads¶
The properly rendered version of this document can be found at Read The Docs. If you are reading this on github, you should instead click here. |
This simple pipeline counts reads and can be run either against a BAM file in Google Cloud Storage or against data accessed via the Google Genomics Reads API. It demonstrates the decoupling of reads data processing from ways of getting the read data and shows how to use common classes for getting reads from BAM or API data sources.
The pipeline produces a small text file with the number of reads counted.
The pipeline is implemented on Google Cloud Dataflow.
Setup Dataflow¶
Most users launch Dataflow jobs from their local machine. This is unrelated to where the job itself actually runs (which is controlled by the --runner
parameter). Either way, Java 8 is needed to run the Jar that kicks off the job.
- If you have not already done so, follow the Genomics Quickstart.
- If you have not already done so, follow the Dataflow Quickstart including installing gcloud and running
gcloud init
.
If you do not have Java on your local machine, the following setup instructions will allow you to launch Dataflow jobs using the Google Cloud Shell:
- If you have not already done so, follow the Genomics Quickstart.
- If you have not already done so, follow the Dataflow Quickstart.
- Use the Cloud Console to activate the Google Cloud Shell.
- Run the following commands in the Cloud Shell to install Java 8.
sudo apt-get update
sudo apt-get install --assume-yes openjdk-8-jdk maven
sudo update-alternatives --config java
sudo update-alternatives --config javac
Note
Depending on the pipeline, Cloud Shell may not not have sufficient memory to run the pipeline locally (e.g., without the --runner
command line flag). If you get error java.lang.OutOfMemoryError: Java heap space
, follow the instructions to run the pipeline using Compute Engine Dataflow workers instead of locally (e.g. use --runner=DataflowPipelineRunner
).
If you want to run a small pipeline on your machine before running it in parallel on Compute Engine, you will need ALPN since many of these pipelines require it. When running locally, this must be provided on the boot classpath but when running on Compute Engine Dataflow workers this is already configured for you. You can download it from here. For example:
wget -O alpn-boot.jar \
http://central.maven.org/maven2/org/mortbay/jetty/alpn/alpn-boot/8.1.8.v20160420/alpn-boot-8.1.8.v20160420.jar
Download the latest GoogleGenomics dataflow runnable jar from the Maven Central Repository. For example:
wget -O google-genomics-dataflow-runnable.jar \
https://search.maven.org/remotecontent?filepath=com/google/cloud/genomics/google-genomics-dataflow/v1-0.1/google-genomics-dataflow-v1-0.1-runnable.jar
Run the pipeline¶
The following command will count reads from a BAM in Google Cloud Storage, specifically those in the BRCA1 region for sample NA12877 within the Illumina Platinum Genomes dataset:
java -Xbootclasspath/p:alpn-boot.jar \
-cp google-genomics-dataflow-runnable.jar \
com.google.cloud.genomics.dataflow.pipelines.CountReads \
--references=chr17:41196311:41277499 \
--BAMFilePath=gs://genomics-public-data/platinum-genomes/bam/NA12877_S1.bam \
--output=gs://YOUR-BUCKET/dataflow-output/NA12877-BAM-reads.tsv
The following command will count those same reads but from the Google Genomics Reads API:
java -Xbootclasspath/p:alpn-boot.jar \
-cp google-genomics-dataflow-runnable.jar \
com.google.cloud.genomics.dataflow.pipelines.CountReads \
--references=chr17:41196311:41277499 \
--readGroupSetId=CMvnhpKTFhD3he72j4KZuyc \
--output=gs://YOUR-BUCKET/dataflow-output/NA12877-API-reads.tsv
You can check your results by ensuring that both of these examples return the answer 45,081 in their output files.
The above command line runs the pipeline locally over a small portion of the genome, only taking a few minutes. If modified to run over a larger portion of the genome or the entire genome, it may take a few hours depending upon how many virtual machines are configured to run concurrently via --numWorkers
. Add the following additional command line parameters to run the pipeline on Google Cloud instead of locally:
--runner=DataflowPipelineRunner \
--project=YOUR-GOOGLE-CLOUD-PLATFORM-PROJECT-ID \
--stagingLocation=gs://YOUR-BUCKET/dataflow-staging \
--numWorkers=#
Use a comma-separated list to run over multiple disjoint regions. For example to run over BRCA1 and BRCA2 --references=chr13:32889610:32973808,chr17:41196311:41277499
.
To run this pipeline over the entire genome, use --allReferences
instead of --references=chr17:41196311:41277499
.
To run the pipeline on a different read group set:
- Change the
--readGroupSetId
id parameter. - Update the
--references
as appropriate (e.g., add/remove the ‘chr’ prefix on reference names).
To run the pipeline over a different BAM file:
- Change
--BAMFilePath
parameter. Set--shardBAMReading=false
if no BAM index file is available. - Update the
--references
as appropriate (e.g., add/remove the ‘chr’ prefix on reference names).
Additional details¶
If the Application Default Credentials are not sufficient, use --client-secrets PATH/TO/YOUR/client_secrets.json
. If you do not already have this file, see the authentication instructions to obtain it.
Use --help
to get more information about the command line options. Change
the pipeline class name below to match the one you would like to run.
java -cp google-genomics-dataflow*runnable.jar \
com.google.cloud.genomics.dataflow.pipelines.VariantSimilarity --help
See the source code for implementation details: https://github.com/googlegenomics/dataflow-java