A team of researchers at Washington State University (WSU) wanted to know the minimum proteins that were required for gram-negative microbes called Proteobacteria to survive. The team compiled a dataset of 2,300 bacterial genomes, containing sequences for nearly nine million proteins, which were grouped together when similar. As they began to assess those sequences, they started finding errors in genomic data that is publicly available and used by scientists, including them. Their study, which will probably have major implications for future research, has been reported in Frontiers in Microbiology.
"Just in the last two years, researchers have sequenced more than twice the number of bacterial genomes as they did in the twenty years before that," said Shira Broschat, a professor in the School of Electrical Engineering and Computer Science at WSU.
A tool for genetic sequencing was pioneered in 1977 by Frederick Sanger and colleagues, commercialized, and brought into the research lab, where it had huge impacts. It took over ten years and $2.7 billion to sequence the human genome that way. Recent years have seen the advent of a new tool: next-gen sequencing, which is described in the video, and is able to deliver a human genome sequence in under an hour for $1500. One can imagine that this new technology immediately created a staggering amount of new genetic data.
Unfortunately, the WSU researchers have found that this massive amount of data introduces a type of error that the researchers found to be a serious problem, noted lead author Svetlana Lockwood, a graduate student in computer science. "A single annotation error can propagate rapidly because scientists build on previous annotation when they sequence new genomes," she explained. Gene annotations describe the coding regions of genes, which create a protein, and where they are located; they can help scientists ascertain gene function, for example.
"We found that for each of the proteins, there were mistakes in annotation of their genes, which resulted in truncated or missing sequences," Broschat added.
Other studies have identified annotation errors, but in this work, the WSU team listed and explained the different types of errors they found in currently available genetic data.
"With the scale of misannotation we found, researchers have to reevaluate the reliability of publicly available genome data for use in big data applications," Broschat noted.
These errors have both human and technological causes, said Kelly Brayton, a professor in the Department of Veterinary Microbiology and Pathology. DNA sequencing tools are good but they aren’t perfect, so they sometimes make the wrong determination when identifying nucleotide bases. A lack of understanding of proteins can also create problems.
This team still had to use computational tools to perform their analysis; it’s simply too much data for people to comb through by hand. They used information contained in the databases of the National Center for Biotechnology Information. Their efforts continue, as Brayton and Broschat seek to develop a tool that can find annotation errors in datasets.
Sources: AAAS/Eurekalert! via WSU, Frontiers in Microbiology