Understanding Read Formats and Quality Controlling Data
=======================================================
Note: there are generic instructions for doing quality control at the
`khmer-protocols Web site
`__. These should work for
most Illumina data sets, even those consisting of multiple files.
Below, we've done a bit of a shorthand because we only have a small
data set to filter.
The fastq Format
````````````````
After spending weeks, nay, months of time on designing your study and planning your
bioinformatics goals (right?), you finally get the email from you sequencing
center: they have your data! You get a link to an ftp server and some login information,
and are presented with a list of files. But what are these formats?
Most commonly, you'll get your data in fastq format. fastq is a really simple
way of representing sequence in plain text which is understood by pretty much
every piece of bioinformatics software. A fastq file can contain anywhere from
one to billions of sequences, and is usually used for reads before they have been
assembled. A faux example of the format is::
@read1
+
ATCGTAGCTAGCTAGCT
+
DHread1
ATCGTAGGTAGGATATA
fasta is usually output by assembly programs, and can be used if data has already
been quality controlled and needs to be a more manageable size. However, if you're
not sure what preprocessing steps your data has been through, but you have fasta
instead of fastq, you'd be well-off to make sure of what those steps were.
Getting the Data
````````````````
Now that you know a little about the format, let's get some data. The data set
we'll be using is from everyone's favorite bacterium, ecoli. We'll use the
command line tool curl to download it to our Amazon machines::
cd /mnt
mkdir ecoli
cd ecoli
curl -O ftp://ftp.sra.ebi.ac.uk/vol1/fastq/SRR390/SRR390202/SRR390202_1.fastq.gz
curl -O ftp://ftp.sra.ebi.ac.uk/vol1/fastq/SRR390/SRR390202/SRR390202_2.fastq.gz
The data came from `this `__ study,
if you're interested.
To take a quick look at the files, use less::
less SRR390202_1.fastq.gz
Hit 'q' to quit less.
These reads are compressed with gzip to save some space, and are in two files,
because they are paired -- the first read in ``SRR390202_1.fastq.gz``
is paired with the first read in ``SRR390202_2.fastq.gz`` and so on. Some programs
prefer paired reads to be interleaved, that is, in the same file alternating between
the first and second read pair. Many programs also require the name field in
the fasta/q to explicitly state which part of a pair a read is with a /1 or /2; for example,
in an interleaved file, you might have::
@SRR390202.1 M10_0139:1:2:18915:1321/1
ATCAAGAAAGATTTTAACAGCATTGAC
+
ECCFFFDDHGHFDHJJJJIGIDIJJJJ
@SRR390202.1 M10_0139:1:2:18915:1321/2
GTTCATAGTGACAAGGTAATATTTGTC
+
FDFFFFHHGGIJIF?CIGJJGI@FEFH
Naturally, because this is a standard, almost every program has a different way of
doing it. So, be sure to double check the pairing format in your data!
Getting the Dependencies
````````````````````````
Before we can work with our data, we need to grab a few more dependencies. We'll download screed, which is
a simple python module for parsing fasta files developed by our lab at MSU::
pip install screed
And then khmer, which is a Python interface to a really fast (and awesome) piece
of software for counting k-mers, also developed by our lab (more on that later)::
cd /usr/local/share
git clone https://github.com/ged-lab/khmer.git
cd khmer
git checkout 2013-caltech-cemi
make
Now that we've downloaded and built khmer, we'll add it to the system's python
PATH so that our scripts know where to find it::
echo 'export PYTHONPATH=/usr/local/share/khmer/python' >> ~/.bashrc
source ~/.bashrc
Get Trimmomatic, which is used for adapter removal and quality filtering::
curl -O http://www.usadellab.org/cms/uploads/supplementary/Trimmomatic/Trimmomatic-0.30.zip
unzip Trimmomatic-0.30.zip
cd Trimmomatic-0.30/
cp trimmomatic-0.30.jar /usr/local/bin
cp -r adapters /usr/local/share/adapters
And then fastx, and its dependencies, which is another tool for quality filtering::
cd /root
curl -O http://hannonlab.cshl.edu/fastx_toolkit/libgtextutils-0.6.1.tar.bz2
tar xjf libgtextutils-0.6.1.tar.bz2
cd libgtextutils-0.6.1/
./configure && make && make install
cd /root
curl -O http://hannonlab.cshl.edu/fastx_toolkit/fastx_toolkit-0.0.13.2.tar.bz2
tar xjf fastx_toolkit-0.0.13.2.tar.bz2
cd fastx_toolkit-0.0.13.2/
./configure && make && make install
And finally, FastQC, which is a program for assessing quality, finding contaminants,
and generally producing nice plots::
apt-get -y install lighttpd
Now, configure::
cd /etc/lighttpd/conf-enabled
ln -fs ../conf-available/10-cgi.conf ./
echo 'cgi.assign = ( ".cgi" => "" )' >> 10-cgi.conf
echo 'index-file.names += ( "index.cgi" ) ' >> 10-cgi.conf
/etc/init.d/lighttpd restart
And install FastQC::
cd /usr/local/share
curl -O http://www.bioinformatics.babraham.ac.uk/projects/fastqc/fastqc_v0.10.1.zip
unzip fastqc_v0.10.1.zip
chmod +x FastQC/fastqc
Assessing your Data with FastQC
```````````````````````````````
First we're going to separate out a subset of the reads for demonstrative
purposes; otherwise, things take way too long::
cd /mnt/ecoli
gunzip -c SRR390202_1.fastq.gz | head -n 400000 > SRR390202_1.head.fastq
gunzip -c SRR390202_2.fastq.gz | head -n 400000 > SRR390202_2.head.fastq
Before you go wildly charging at your data with trimmers and filters, it's always
a good idea to know what your data looks like ahead of time. The program we will
use for this is FastQC, which parses the quality information from all the reads
and produces handy charts and statistics::
mkdir /var/www/ecoli_fastqc
/usr/local/share/FastQC/fastqc SRR390202_1.head.fastq SRR390202_2.head.fastq -o /var/www/ecoli_fastqc
In the previous step, we actually also installed a very lightweight http server.
This allows you to host things publicly on your instance and view it through a
browser, which in some cases avoids having to download the data to your computer.
In the last command, we put our output in the web server's folder, so let's go
ahead and access it. In a new browser tab, go to::
http://ec2-???????????.compute-1.amazonaws.com/ecoli_fastqc/
replacing the question marks with your EC2 URL. You should be presented with
something like `this `__.
There is a folder for each of your sequence files, each of which contains a file
called ``fastqc_report.html``. Clicking on that file will render the report
in your browser.
Trimming Your Data
``````````````````
Based on the FastQC report for the reads, we should probably quality trim them.
Although there aren't any flagged overrepresented sequences, it's good practice
to filter for adapters as well, which can confound assemblers. Trimmomatic can
both filter adapters and quality trim, though we'll only use it for adapter removal
here::
mkdir trim
cd trim
java -jar /usr/local/bin/trimmomatic-0.30.jar PE ../SRR390202_1.head.fastq ../SRR390202_2.head.fastq s1_pe s1_se s2_pe s2_se ILLUMINACLIP:/usr/local/share/adapters/TruSeq3-PE.fa:2:30:10
fastx is an alternative which performs many of the same functions as Trimmomatic.
We'll use it for quality filtering; the following flags direct its fastq_quality_filter
module to keep reads if 50% of the bases have a quality score over 30::
/usr/local/share/khmer/scripts/interleave-reads.py s1_pe s2_pe > combined.fq
fastq_quality_filter -Q33 -q 30 -p 50 -i combined.fq > combined-trim.fq
fastq_quality_filter -Q33 -q 30 -p 50 -i s1_pe > s1_se.filt
We also interleaved the reads in the previous block, as fastx requires it. We'll
now separate out the reads which had their pair thrown out into their own file,
combine them with the output of Trimmomatic, and compress the results::
/usr/local/share/khmer/scripts/extract-paired-reads.py combined-trim.fq
cat combined-trim.fq.se s1_se.filt | gzip -9c > ../SRR390202.head.se.qc.fq.gz
gzip -9c combined-trim.fq.pe > ../SRR390202.head.pe.qc.fq.gz
Finally, move up a directory and get rid of all the unneeded intermediate files::
cd ../
rm -fr trim
Reassess Data Quality
`````````````````````
Once the reads have been quality-controlled, we should check to make sure that our
measures were actually helpful::
mkdir /var/www/ecoli_qc_fastqc
/usr/local/share/FastQC/fastqc SRR390202.head.pe.qc.fq.gz SRR390202.head.se.qc.fq.gz -o /var/www/ecoli_qc_fastqc
Check the output the same way as above.
The Trimmed Data
````````````````
We only quality-controlled a subset of the reads, but we'll want all of them later
on. To that end, we've run the programs on the full data, which you can download with::
curl -O https://s3.amazonaws.com/public.ged.msu.edu/SRR390202.pe.qc.fq.gz
curl -O https://s3.amazonaws.com/public.ged.msu.edu/SRR390202.se.qc.fq.gz