How Perl Saved the Human Genome Project

by Lincoln Stein

Reprinted courtesy of the Perl Journal, http://www.tpj.com

Lincoln Stein's website is http://stein.cshl.org

DATE: Early February, 1996

LOCATION: Cambridge, England, in the conference room of the largest DNA 
sequencing center in Europe.

OCCASION: A high level meeting between the computer scientists of this 
center and the largest DNA sequencing center in the United States.

THE PROBLEM: Although the two centers use almost identical laboratory 
techniques, almost identical databases, and almost identical data 
analysis tools, they still can't interchange data or meaningfully 
compare results.

THE SOLUTION: Perl.

The human genome project was inaugurated at the beginning of the decade 
as an ambitious international effort to determine the complete DNA 
sequence of human beings and several experimental animals. The 
justification for this undertaking is both scientific and medical. By 
understanding the genetic makeup of an organism in excruciating detail, 
it is hoped that we will be better able to understand how organisms 
develop from single eggs into complex multicellular beings, how food is 
metabolized and transformed into the constituents of the body, how the 
nervous system assembles itself into a smoothly functioning ensemble. 
From the medical point of view, the wealth of knowledge that will come 
from knowing the complete DNA sequence will greatly accelerate the 
process of finding the causes of (and potential cures for) human 
diseases.

Six years after its birth, the genome project is ahead of schedule. 
Detailed maps of the human and all the experimental animals have been 
completed (mapping out the DNA using a series of landmarks is an 
obligatory first step before determining the complete DNA sequence). 
The sequence of the smallest model organism, yeast, is nearly 
completed, and the sequence of the next smallest, a tiny soil-dwelling 
worm, isn't far behind. Large scale sequencing efforts for human DNA 
started at several centers a number of months ago and will be in full 
swing within the year.

The scale of the human DNA sequencing project is enough to send your 
average Unix system administrator running for cover. From the 
information-handling point of view, DNA is a very long string 
consisting of the four letters G, A, T and C (the letters are 
abbreviations for the four chemical units that form the "rungs" of the 
DNA double helix ladder). The goal of the project is to determine the 
order of letters in the string. The size of the string is impressive 
but not particularly mind-boggling: 3 x 10^9 letters long, or some 3 
gigabytes of storage space if you use 1 byte to store each letter and 
don't use any compression techniques.

Three gigabytes is substantial but certainly manageable by today's 
standards. Unfortunately, this is just the storage space required to 
store the finished data. The storage requirements for the experimental 
data needed to determine this sequence is far more vast. The essential 
problem is that DNA sequencing technology is currently limited to 
reading stretches of at most 500 contiguous letters. In order to 
determine sequences longer than that, the DNA must be sequenced as 
small overlapping fragments called "reads" and the jigsaw puzzle 
reassembled by algorithms that look for areas where the sequences 
match. Because the DNA sequence is nonrandom (similar but 
not-entirely-identical motifs appear many times throughout the genome), 
and because DNA sequencing technology is noisy and error-prone, one 
ends up having to sequence each region of DNA five to 10 times in order 
to reliably assemble the reads into the true sequence. This increases 
the amount of data to manage by an order of magnitude. On top of this 
is all the associated information that goes along with laboratory work: 
who performed the experiment, when it was performed, the section of the 
genome was sequenced, the identity and version of the software used to 
assemble the sequence, any comments someone wants to attach to the 
experiment, and so forth. In addition, one generally wants to store the 
raw output from the machine that performs the sequencing itself. Each 
500 letters of sequence generates a data file that's 20-30 kilobytes 
long!

That's not the whole of it. It's not enough just to determine the 
sequence of the DNA. Within the sequence are functional areas scattered 
among long stretches of nonfunctional areas. There are genes, control 
regions, structural regions, and even a few viruses that got entangled 
in human DNA long ago and persist as fossilized remnants. Because the 
genes and control regions are responsible for health and disease, one 
wants to identify and mark them as the DNA sequence is assembled. This 
type of annotation generates yet more data that needs to be tracked and 
stored.

Altogether, people estimate that some 1 to 10 terabytes of information 
will need to be stored in order to see the human genome project to its 
conclusion.

So what's Perl got to do with it? From the beginning researchers 
realized that informatics would have to play a large role in the genome 
project. An informatics core formed an integral part of every genome 
center that was created. The mission of these cores was two-fold: to 
provide computer support and databasing services for their affiliated 
laboratories, and to develop data analysis and management software for 
use by the genome community as a whole.

It's fair to say that the initial results of the informatics groups 
efforts were mixed. Things were slightly better on the laboratory 
management side of the coin. Some groups attempted to build large 
monolithic systems on top of complex relational databases; they were 
thwarted time and again by the highly dynamic nature of biological 
research. By the time a system that could deal with the ins and outs of 
a complex laboratory protocol had been designed, implemented and 
debugged, the protocol had been superseded by new technology and the 
software engineers had to go back to the drawing board.

Most groups, however, learned to build modular, loosely-coupled systems 
whose parts could be swapped in and out without retooling the whole 
system. In my group, for example, we discovered that many data analysis 
tasks involve a sequence of semi-independent steps. Consider the steps 
that one may want to perform on a bit of DNA that has just been 
sequenced (Figure 1). First there's a basic quality check on the 
sequence: is it long enough? Are the number of ambiguous letters below 
the maximum limit? Then there's the "vector check". For technical 
reasons, the human DNA must be passed through a bacterium before it can 
be sequenced (this is the process of "cloning"). Not infrequently, the 
human DNA gets lost somewhere in the process and the sequence that's 
read consists entirely of the bacterial vector. The vector check 
ensures that only human DNA gets into the database. Next there's a 
check for repetitive sequences. Human DNA is full of repetitive 
elements that make fit! ting the sequencing jigsaw puzzle together 
challenging. The repetitive sequence check tries to match the new 
sequence against a library of known repetitive elements. A penultimate 
step is to attempt to match the new sequence against other sequences in 
a large community database of DNA sequences. Often a match at this 
point will provide a clue to the function of the new DNA sequence. 
After performing all these checks, the sequence along with the 
information that's been gathered about it along the way is loaded into 
the local laboratory database.

The process of passing a DNA sequence through these independent 
analytic steps looks kind of like a pipeline, and it didn't take us 
long to realize that a Unix pipe could handle the job. We developed a 
simple Perl-based data exchange format called "boulderio" that allowed 
loosely coupled programs to add information to a pipe-based I/O stream. 
Boulderio is based on tag/value pairs. A Perl module makes it easy for 
programs to reach into the input stream, pull out only the tags they're 
interested in, do something with them, and drop new tags into output 
the stream. Any tags that the program isn't interested in are just 
passed through to standard output so that other programs in the 
pipeline can get to them.

Using this type of scheme, the process of analyzing a new DNA sequence 
looks something like this (this is not exactly the set of scripts that 
we use, but it's close enough):

 

name_sequence.pl < new.dna |

quality_check.pl |

vector_check.pl |

find_repeats.pl |

search_big_database.pl |

load_lab_database.pl

 

A file containing the new DNA sequence is processed by a perl script 
named "name_sequence.pl", whose only job is to give the sequence a new 
unique name and to put it into boulder format. Its output looks like 
this:

 

NAME=L26P93.2

SEQUENCE=GATTTCAGAGTCCCAGATTTCCCCCAGGGGGTTTCCAGAGAGCCC......

 

The output from name_sequence.pl is next passed to the quality checking 
program, which looks for the SEQUENCE tag, runs the quality checking 
algorithm, and writes its conclusion to the data stream. The data 
stream now looks like this:

NAME=L26P93.2

SEQUENCE=GATTTCAGAGTCCCAGATTTCCCCCAGGGGGTTTCCAGAGAGCCC......

QUALITY_CHECK=OK

Now the data stream enters the vector checker. It pulls the SEQUENCE 
tag out of the stream and runs the vector checking algorithm. The data 
stream now looks like this:

NAME=L26P93.2

SEQUENCE=GATTTCAGAGTCCCAGATTTCCCCCAGGGGGTTTCCAGAGAGCCC......

QUALITY_CHECK=OK

VECTOR_CHECK=OK

VECTOR_START=10

VECTOR_LENGTH=300

This continues down the pipeline, until at last the 
"load_lab_database.pl" script collates all the data collected, makes 
some final conclusions about whether the sequence is suitable for 
further use, and enters all the results into the laboratory database.

One of the nice features of the boulderio format is that multiple 
sequence records can be processed sequentially in the same Unix 
pipeline. An "=" sign marks the end of one record and the beginning of 
the next:

 

NAME=L26P93.2

SEQUENCE=GATTTCAGAGTCCCAGATTTCCCCCAGGGGGTTTCCAGAGAGCCC......

=

NAME=L26P93.3

SEQUENCE=CCCCTAGAGAGAGAGAGCCGAGTTCAAAGTCAAAACCCATTCTCTCTCCTC...

=

There's also a way to create subrecords within records, allowing for 
structured data types.

Here's an example of a script that processes boulderio format. It uses 
an object-oriented style, in which records are pulled out of the input 
stream, modified, and dropped back in:

use Boulder::Stream;

$stream = new Boulder::Stream;

while ($record = $stream->read_record('NAME','SEQUENCE')) {

$name = $record->get('NAME');

$sequence = $record->get('SEQUENCE');

...[continue processing]...

$record->add(QUALITY_CHECK=>"OK");

$stream->write_record($record);

}

(If you're interested, more information about the boulderio format and 
the perl libraries to manipulate it can be found at 
http://stein.cshl.org/software/boulder/).

The interesting thing is that multiple informatics groups independently 
converged on solutions that were similar to the boulderio idea. For 
example, several groups involved in the worm sequencing project began 
using a data exchange format called ".ace". Although this format was 
initially designed as the data dump and reload format for the ACE 
database (a database specialized for biological data), it happens to 
use a tag/value format that's very similar to boulderio. Soon .ace 
files were being processed by Perl script pipelines and loaded into the 
ACE database at the very last step.

Perl found uses in other aspects of laboratory management. For example, 
many centers, including my own, use Web based interfaces for displaying 
the status of projects and allowing researchers to take actions. Perl 
scripts are the perfect engine for Web CGI scripts. Similarly, Perl 
scripts run e-mail database query servers, supervise cron jobs, prepare 
nightly reports summarizing laboratory activity, create instruction 
files to control robots, and handle almost every other information 
management task that a busy genome center needs.

So as far as laboratory management went, the informatics cores were 
reasonably successful. As far as development and distributing generally 
useful, however, things were not so rosy.

The problem will be familiar to anyone who has worked in a large, 
loosely organized software project. Despite best intentions, the 
project begins to drift. Programmers go off to work on ideas that 
interest them, modules that need to interface with one another are 
designed independently, and the same problems get solved several times 
in different, mutually incompatible ways. When the time comes to put 
all the parts together, nothing works.

This is what happened in the genome project. Despite the fact that 
everyone was working on the same problems, no two groups took exactly 
the same approach. Programs to solve a given problem were written and 
rewritten multiple times. While a given piece of software wasn't 
guaranteed to work better than its counterpart developed elsewhere, you 
could always count on it to sport its own idiosyncratic user interface 
and data format. A typical example is the central algorithm that 
assembles thousands of short DNA reads into an ordered set of overlaps. 
At last count there were at least six different programs in widespread 
use, and no two of them use the same data input or output formats.

This lack of interchangeability presents terrible dilemma for the 
genome centers. Without interchangeability, an informatics group is 
locked into using the software that it developed in-house. If another 
genome center has come up with a better software tool to attack the 
same problem, a tremendous effort is required by the first center to 
retool their system in order to use that tool.

The long range solution to this problem is to come up with uniform data 
interchange standards that genome software must adhere to. This would 
allow common modules to be swapped in and out easily. However, 
standards require time to agree on, and while the various groups are 
involved in discussion and negotiation, there is still an urgent need 
to adapt existing software to the immediate needs of the genome centers.

Here is where Perl again came to the rescue. The Cambridge summit 
meeting that introduced this article was called in part to deal with 
the data interchange problem. Despite the fact that the two groups 
involved were close collaborators and superficially seemed to be using 
the same tools to solve the same problems, on closer inspection nothing 
they were doing was exactly the same.

The main software components in a DNA sequencing projects are:

    a trace editor to analyze, display and the short DNA read 
chromatograms from DNA sequencing machines.
    a read assembler, to find overlaps between the reads and assemble 
them together into long contiguous sections.
    an assembly editor, to view the assemblies and make changes in 
places where the assembler went wrong.
    a database to keep track of it all. 

Over the course of a few years, the two groups had developed suites of 
software that worked well in their hands. Following the familiar genome 
center model, some of the components were developed in-house while 
others were imported from outside. As shown in Figure 2, Perl was used 
for the glue to make these pieces of software fit together. Between 
each pair of interacting modules were one or more Perl scripts 
responsible for massaging the output of one module into the expected 
input for another.

When the time came to interchange data, however, the two groups hit a 
snag. Between them they were now using two trace editors, three 
assemblers, two assembly editors and (thankfully) one database. If two 
Perl scripts were required for each pair of components (one for each 
direction), one would need as many as 62 different scripts to handle 
all the possible interconversion tasks. Every time the input or ouput 
format of one of these modules changed, 14 scripts might need to be 
examined and fixed.

The solution that was worked out during this meeting is the obvious one 
shown in Figure 3. The two groups decided to adopt a common data 
exchange format known as CAF (an acronym whose exact meaning was 
forgotten during the course of the meeting). CAF would contain a 
superset of the data that each of the analysis and editing tools 
needed. For each module, two Perl scripts would be responsible for 
converting from CAF into whatever format Module A expects 
("CAF2ModuleA") and converting Module A's output back into CAF 
("ModuleA2CAF"). This simplified the programming and maintenance task 
considerably. Now there were only 16 Perl scripts to write; when one 
module changed, only two scripts would need to be examined.

This episode is not unique. Perl has been the solution of choice for 
genome centers whenever they need to exchange data, or to retrofit one 
center's software module to work with another center's system.

So Perl has become the software mainstay for computation within genome 
centers as well as the glue that binds them together. Although genome 
informatics groups are constantly tinkering with other "high level" 
languages such as Python, Tcl and recently Java, nothing comes close to 
Perl's popularity. How has Perl achieved this remarkable position?

I think several factors are responsible:

1. Perl is remarkably good for slicing, dicing, twisting, wringing, 
smoothing, summarizing and otherwise mangling text. Although the 
biological sciences do involve a good deal of numeric analysis now, 
most of the primary data is still text: clone names, annotations, 
comments, bibliographic references. Even DNA sequences are textlike. 
Interconverting incompatible data formats is a matter of text mangling 
combined with some creative guesswork. Perl's powerful regular 
expression matching and string manipulation operators simplify this job 
in a way that isn't equalled by any other modern language.

2. Perl is forgiving. Biological data is often incomplete, fields can 
be missing, or a field that is expected to be present once occurs 
several times (because, for example, an experiment was run in 
duplicate), or the data was entered by hand and doesn't quite fit the 
expected format. Perl doesn't particularly mind if a value is empty or 
contains odd characters. Regular expressions can be written to pick up 
and correct a variety of common errors in data entry. Of course this 
flexibility can be also be a curse. I talk more about the problems with 
Perl below.

3. Perl is component-oriented. Perl encourages people to write their 
software in small modules, either using Perl library modules or with 
the classic Unix tool-oriented approach. External programs can easily 
be incorporated into a Perl script using a pipe, system call or socket. 
The dynamic loader introduced with Perl5 allows people to extend the 
Perl language with C routines or to make entire compiled libraries 
available for the Perl interpreter. An effort is currently under way to 
gather all the world's collected wisdom about biological data into a 
set of modules called "bioPerl" (discussed at length in an article to 
be published later in the Perl Journal).

4. Perl is easy to write and fast to develop in. The interpreter 
doesn't require you to declare all your function prototypes and data 
types in advance, new variables spring into existence as needed, calls 
to undefined functions only cause an error when the function is needed. 
The debugger works well with Emacs and allows a comfortable interactive 
style of development.

5. Perl is a good prototyping language. Because Perl is quick and 
dirty, it often makes sense to prototype new algorithms in Perl before 
moving them to a fast compiled language. Sometimes it turns out that 
Perl is fast enough so that of the algorithm doesn't have to be ported; 
more frequently one can write a small core of the algorithm in C, 
compile it as a dynamically loaded module or external executable, and 
leave the rest of the application in Perl (for an example of a complex 
genome mapping application implemented in this way, see 
http://waldo.wi.mit.edu/ftp/distribution/software/rhmapper/).

6. Perl is a good language for Web CGI scripting, and is growing in 
importance as more labs turn to the Web for publishing their data.

My experience in using Perl in a genome center environment has been 
extremely favorable overall. However I find that Perl has its problems 
too. Its relaxed programming style leads to many errors that more 
uptight languages would catch. For example, Perl lets you use a 
variable before its been assigned to, a useful feature when that's what 
you intend but a disaster when you've simply mistyped a variable name. 
Similarly, it's easy to forget to declare make a variable used in a 
subroutine local, inadvertently modifying a global variable.

If one uses the -w switch religiously and turn on the "use strict vars" 
pragma, these Perl will catch these problems and others. However there 
are more subtle gotchas in the language that are not so easy to fix. A 
major one is Perl's lack of type checking. Strings, floats and integers 
all interchange easily. While this greatly speeds up development, it 
can cause major headaches. Consider a typical genome center Perl script 
that's responsible for recording the information of short named 
subsequences within a larger DNA sequence. When the script was written, 
the data format was expected to consist of tab-delimited fields: a 
string followed by two integers representing the name, starting 
position and length of a DNA subsequence within a larger sequence. An 
easy way to parse this would to split() into an list like this:

 

($name,$start_position,$length) = split("\t");

Later on in this script some arithmetic is performed with the two 
integer values and the result written to a database or to standard 
output for further processing.

Then one day the input file's format changes without warning. Someone 
bumps the field count up by one by sticking a comment field between the 
name and the first integer. Now the unknowing script assigns a string 
to a variable that's expected to be numeric and silently discards the 
last field on the line. Rather than crashing or returning an error 
code, the script merrily performs integer arithmetic on a string, 
assuming a value of zero for the string (unless it happens to start 
with a digit). Although the calculation is meaningless, the output may 
look perfectly good, and the error may not be caught until some point 
well downstream in the processing.

The final Perl deficiency has been a way to create graphical user 
interfaces. Although Unix True Believers know that anything worth doing 
can be done on the command line, most end-users don't agree. Windows, 
menus and bouncing icons have become de rigueur for programs intended 
for use by mere mortals.

Until recently, GUI development in Perl was awkward to impossible. 
However the work of Nick Ing-Simmons and associates on perlTK (pTK) has 
made Perl-driven GUIs possible on X-windows systems. My associates and 
I have written several pTK-based applications for internal use at the 
MIT genome center, and it's been a satisfying experience overall. Other 
genome centers make much more extensive use of pTK, and in some places 
its become a mainstay of production.

Unfortunately, I'm sad to confess that a few months ago when I needed 
to put a graphical front end on a C++ image analysis program I'd 
written, I turned to the standard Tcl/Tk library rather than to pTK. I 
made this choice because I intended the application for widespread 
distribution. I find pTK still too unstable for export: new releases of 
pTK discover lurking bugs in Perl, and vice-versa. Further, I find that 
even seasoned system administrators run into glitches when compiling 
and installing Perl modules, and I worried that users would hit some 
sort of road block while installing either pTK or the modules needed to 
support my application, and owuld give up. In contrast, many systems 
have the Tcl/Tk libraries preinstalled; if they don't, installation is 
quick and painless.

In short, when the genome project was foundering in a sea of 
incompatible data formats, rapidly-changing techniques, and monolithic 
data analysis programs that were already antiquated on the day of their 
release, Perl saved the day. Although it's not perfect, Perl seems to 
fill the needs of the genome centers remarkably well, and is usually 
the first tool we turn to when we have a problem to solve.