Research Article | Open Access
Pamela Vinitha Eric, Gopakumar Gopalakrishnan, Muralikrishnan Karunakaran, "An Optimal Seed Based Compression Algorithm for DNA Sequences", Advances in Bioinformatics, vol. 2016, Article ID 3528406, 7 pages, 2016. https://doi.org/10.1155/2016/3528406
An Optimal Seed Based Compression Algorithm for DNA Sequences
This paper proposes a seed based lossless compression algorithm to compress a DNA sequence which uses a substitution method that is similar to the LempelZiv compression scheme. The proposed method exploits the repetition structures that are inherent in DNA sequences by creating an offline dictionary which contains all such repeats along with the details of mismatches. By ensuring that only promising mismatches are allowed, the method achieves a compression ratio that is at par or better than the existing lossless DNA sequence compression algorithms.
There is an exponential increase in the amount of DNA being sequenced, thus leading to problems in storage, comprehension, and transmission. The cost of storage has been reducing dramatically in the past few years, but the exponential growth in the amount of DNA being sequenced leads to tremendous increase in the amount of data that needs to be stored online thereby making storage one of the biggest cost elements. Another challenge faced is how to make sense out of this huge mass of data. With whole genomes, we now have to deal with millions or billions of base pairs. When we have a database of such genomes, as is typically the case, the problem becomes even more compounding. Thus, new and more effective techniques are needed for the compression of biological sequence data, particularly DNA sequences.
DNA sequences are expected to be nonrandom and hence it is possible to remove redundancy, resulting in compression. It is estimated that more than 50% of the human genome is repeat DNA . Compression will solve the issues related to storage and also improve the understanding of these sequences. Chen et al.  showed that compressibility is a good measurement of relatedness between sequences and can be effectively used in sequence alignment and evolutionary tree construction. According to Allison et al.  compression of DNA sequences also results in the intelligent analysis of these sequences. Compression also plays an important role in efficient sequence classification .
DNA sequences consist of four nucleotide bases, A (adenine), C (cytosine), G (guanine), and T (thymine), and two bits are sufficient to represent each of these nucleotide bases. Moreover the repeats found in DNA sequences are not always exact; they can be of different types like approximate, reverse, complementary, reverse complementary, and tandem. Also these repeats are long and less frequent. Traditional text compression algorithms are only effective in capturing short and frequent repeats; hence using them to compress DNA sequences often results in expansion of the same. Therefore finding all the different types of repeats in a DNA sequence and encoding them in order to achieve a good compression ratio is a challenging task.
This paper proposes a seed based algorithm which uses a substitution method that is in line with the LempelZiv [5, 6] compression scheme to compress DNA sequences. The proposed algorithm captures all the various types of repeats like exact, tandem, approximate, reverse, complemented, and reverse complemented and stores them onto an offline dictionary. These repeats are then removed from the original sequence to form the final parsed sequence. The offline dictionary along with the final parsed sequence forms the compressed sequence. Mismatches that give good compression gain are tolerated and recorded along with the repeat substrings in the offline dictionary.
This paper is organized as follows. Section 2 reviews the various DNA compression algorithms. Section 3 describes the proposed method and Section 4 analyzes the results obtained. This is followed by conclusion in Section 5.
2. Related Work
Compression of biological sequences can be either horizontal or vertical as proposed by Grumbach and Tahi [7, 8]. Horizontal mode compresses a biological sequence by making use of information contained within it, like references to the substrings, whereas vertical mode takes a set of biological sequences and compresses each sequence based on the information derived from this set. Horizontal mode is of interest for the reduction of storage and transmission costs  and uses compression techniques like substitution, statistical, or a combination of these two . Statistical compression uses a statistical model of the data, comprised of variable sized codes, and the quality of compression obtained depends on the data model . Substitution or dictionary based method selects several strings of symbols that occur frequently and encodes each string as a token which is a pointer to the string in a dictionary. The dictionary itself can be static or dynamic. Compression algorithms based on LZ method use online dictionary whereas in case of methods using offline dictionary compression occurs in two passes: the first pass identifies all repeats and stores them in a dictionary and the second pass encodes these repeats as pointers to the dictionary [5, 6]. A third category of compression is the hybrid technique which makes use of a combination of substitution and statistical techniques to compress data.
Most of the compression methods available for compressing biological sequences like [2, 7, 8, 12] use substitution methods. The earliest special purpose DNA compression algorithm found in the literature is Biocompress developed by Grumbach and Tahi [7, 8]. They proposed Biocompress and Biocompress 2 which detects repeats of substrings that occurred earlier in the sequence and encodes them as length of repeat and position of previous occurrence. They also employ order 2 arithmetic coding to encode nonrepeat regions. Chen et al.  developed DNACompress that uses the software utility Pattern Hunter  to identify significant approximate repeat regions and then encodes these repeat regions by a pointer to their earlier occurrence. The nonrepeat regions are also encoded using arithmetic coding. The offline approach by Apostolico and Lonardi  iteratively selects repeated substrings for which encoding would gain maximum compression. A similar substitution approach is used in GenCompress by Chen et al.  where they concentrate on finding an optimal prefix that can be encoded economically. Here approximate repeats are exploited. Adjeroh et al. [15, 16] create an offline dictionary of short repeats and code all occurrences of a given repeat with reference to the position of that repeat in the dictionary. Cfact developed by Rivals et al.  constructs a suffix tree in the first pass and uses this data structure to search for the longest exact matching repeat in the second pass.
A few methods like XM, CDNA, and ARM employ statistical techniques. Expert model (XM) proposed by Cao et al.  uses an order 2 Markov expert and a copy expert to predict the probability of occurrence of a symbol. It also employs adaptive coding for correct or incorrect predictions. The CDNA algorithm by Loewenstern and Yianilos  is a pure statistical algorithm, where the probability distribution of each symbol is obtained by approximate partial matches from history. Each approximate match is with a previous subsequence having a small Hamming distance to the context preceding the symbol to be encoded. The latter ARM algorithm by Allison et al.  is also a pure statistical algorithm that forms the probability of a subsequence by summing the probabilities over all explanations as to how the subsequence is generated.
A method that employs hybrid technique was introduced by Korodi and Tabus [20, 21] where encoding is done by using a simple normalized maximum likelihood model for discrete regression, through reference to preceding approximate matching blocks and encoding them by a first-order context coding. In its improvement, GeNML by Korodi and Tabus [20, 21], the DNA sequence is split into fixed size blocks. The bit mask is encoded using a probability distribution estimated by the normalized maximum likelihood of similarity between the regressor and the block. Matsumoto et al.  use a combination of LZ [5, 6] and CTW . They first identify approximate repeat regions using hash and dynamic programming and then replace these repeat regions with an offset and length. Edit operations are encoded using arithmetic coding and nonrepeat areas by an order 32 context tree weighting.
3. Optimal Seed Based Compression Algorithm for DNA Sequences
The proposed method consists of a seed based algorithm that identifies potentially good matches. The matching substrings so identified are later extended in both the directions, that is, to the left and right.
Let be the DNA sequence to be compressed and the length of the DNA sequence. represents the th symbol of the given DNA sequence, where and is a substring of of length where . The seed is also a substring of of length . The initial seed is and the first substring to be matched is . The values of are incremented until a repeat substring is identified such that . If no such matching substring is encountered, are incremented and the search is continued until for some , where ranges from to and from to . Now the length of the match is initialized to and , the number of mismatches, is initialized to 0.
The repeat substring and the seed are extended and compared. The extension is done first to the left and then to the right. The length of the match is incremented for each symbol matched. If a mismatch occurs while extending the repeat substring and the seed, decision regarding permitting this mismatch is made, based on the total number of mismatches until then and whether permitting this mismatch would result in a compression gain. If is greater than the threshold, repeat extension in the direction in which the mismatch occurred is temporarily terminated. Extension to the left is also stopped whenever there is an overlap between the extended seed and the extended repeat substring.
Assume that the substrings and are the extended repeat and the extended seed so obtained. An offline dictionary, as shown in Table 1, stores the extended seed , position of occurrence of repeat , length of the repeat , type of repeat, and the details of mismatches that have occurred if any. is then removed from and the remaining symbols of are concatenated to form the next sequence . The process is repeated on sequence until all approximate repeats of are identified and stored in the offline dictionary. This offline dictionary is similar to the one created by Adjeroh et al. [15, 16, 25].
Finally the extended seed is removed from and the remaining symbols of are concatenated to form the new sequence . The position of the extended seed , , is recorded as the last entry under that seed.
The above process is repeated on the new sequence until all the exact and approximate repeats are identified and removed from the sequence and the remaining nonrepeat regions of the sequence are concatenated to form the final parsed sequence. This process along with an example is depicted in Figure 1. The offline dictionary along with the final parsed sequence (the original sequence from which all seeds and repeats have been removed) forms the compressed sequence. The final parsed sequence is further compressed using adaptive arithmetic coding. The extended seed entries in the offline dictionary are also encoded using arithmetic coding.
3.1. Encoding of Mismatches
The proposed method identifies all significant approximate repeats in the DNA sequence and stores them onto an offline dictionary. All approximate repeats have mismatches which are encoded and written onto the offline dictionary. Mismatches occur due to mutation and can be defined by any of the edit operations like insertion, deletion, or substitution of some base. The mismatch details are recorded in the table as a triple , where is the position of mismatch within the extended seed, the type of edit operation, and the symbol to be inserted or substituted. When the edit operation is deletion, the last field of the triple may be omitted. The same representation is also used in DNACompress  and GenCompress . The edit operations are encoded as 00 insertion, 01 deletion, and 10 substitution and the bases A, C, G, and T are encoded as 00, 01, 10, and 11, respectively. Suppose that the seed is the substring GCACTTACT and the approximate repeat found is GCACTTTCT. Here the symbol A which occurs at the th position in the seed has been substituted with the symbol T in the repeat. This is represented by a triple of 7 bits as (1111011).
3.2. Determining the Threshold
Mismatches are allowed while extending the repeats in both directions but the number of such mismatches should not exceed a predetermined threshold. When the predetermined threshold is exceeded temporarily suspend extension of seed and repeat in that direction until the length of the repeat and seed has increased to such an extent that extension of repeat in the suspended direction becomes feasible again. But any mismatch is tolerated if and only if allowing such a mismatch results in a compression gain. The threshold value is determined dynamically with respect to the length of the extended repeat. Experimental results show that the total number of mismatches allowed at any instance should never exceed ; here is the length of the extended repeat.
3.3. Calculation of Compression Ratio
It takes at the most 2 bits to encode each symbol of a DNA sequence. The objective of DNA compression is to bring down the bits needed to represent each base to less than 2. In the proposed method the output comprises the offline dictionary along with the final parsed sequence and the compression ratio specifies the bits per symbol (bps) and can be calculated by the following formula :
The cost of a variable is the number of bits required to represent it. The term vocabulary refers to the identified repeats without reference to their specific locations in the sequence. The size of the dictionary denoted as gives the number of distinct repetitions in the dictionary. The length of the th seed is denoted by and the number of repetitions of th seed by . If the position of the th occurrence of repeat pattern is given as and mismatch in the th occurrence of repeat pattern as , then cost of positions and mismatch details can be given as
Here is the number of bits required to represent and the last term gives the number of bits needed to represent the type of repetition of each repeat.
The cost of dictionary is the sum of the cost of vocabulary and cost of positions and mismatch details:
The final parsed sequence is the sequence from which all the repeats have been removed. The number of bits to represent is . Therefore
The seed based compression algorithm was experimentally verified on a set of DNA sequences in FASTA format as the input. The method was tested on the same standard benchmark data used in [2, 7, 12, 18, 22]. These standard sequences include human growth hormone (HUMGHCSA), human DNA sequence (HUMHDABCD), vaccinia virus Copenhagen complete genome (VACCG), Marchantia polymorpha mitochondrion complete genome (MPOMTCG), human beta globin region on chromosome 11 (HUMHBB), Homo sapiens dystrophin gene (HUMDYSTROP), and human hypoxanthine phosphoribosyltransferase gene (HUMHPRTB).
The algorithm was tested on this data set for various seed lengths “,” to decide upon an optimum “” value for the compression. The seed length “” was varied from 5 to 11 on various runs of the data set and the best “” value was inferred to be 8 as this gives better compression ratio than smaller “” values. BLAST , being another bioinformatics local alignment search tool, also uses 11 as the standard seed length whereas SENSEI  uses 8 as the seed length. Also, it was noticed that even though time complexity increases when was incremented further a substantial improvement in compression ratio does not occur to warrant such an increase. Thus the “” value was inferred to be 8. A graph comparing compression ratios against varying values for different sequences is shown in Figure 2.
Any mismatch is tolerated if and only if allowing such a mismatch results in a compression gain. To ensure compression gain a mismatch is allowed only if the next few characters are an exact match. The results of testing for different values showed that ensuring the next three characters are exact matches gives good compression ratio. The permitted threshold for the number of mismatches allowed was varied from to and it was found that gives better compression. The graph depicting the compression ratio achieved when the threshold for the number of mismatches allowed is and is shown in Figure 3.
The tabulated result of comparison of compression ratios of the proposed seed based method against other existing algorithms is shown in Table 2.
The execution time taken by few of the reviewed methods for the benchmark sequences was determined after executing them on a machine with a quad core processor having a clock speed of 2.60 GHz, 8 GB RAM, and 64-bit operating system. The execution time taken by the reviewed methods is given in Table 3.
Decompression. Algorithm 1 was implemented and the result of decompression was verified to ensure that the compression method proposed is indeed lossless.
A substitutional compression algorithm for DNA sequence is proposed. On extensive testing, the optimum seed length for the method was decided to be 8. As seen from the results, it is observed that the proposed method performs with compression ratios comparable to the existing algorithms and even better for a few standard sequences. Further the speed of execution can be improved by incorporating any known methods of string comparison like suffix trees or bitwise XOR operation as used in SENSEI  in the initial phase while looking for exact seed matches.
The authors declare that they have no competing interests.
- E. S. Lander, L. M. Linton, B. Birren et al., “Initial sequencing and analysis of the human genome,” Nature, vol. 409, no. 6822, pp. 860–921, 2001.
- X. Chen, S. Kwong, and M. Li, “Compression algorithm for DNA sequences and its applications in genome comparison,” in Proceedings of the 4th Annual International Conference on Computational Molecular Biology (RECOMB '00), p. 107, ACM, Tokyo, Japan, April 2000.
- L. Allison, L. Stern, T. Edgoose, and T. I. Dix, “Sequence complexity for biological sequence analysis,” Computers and Chemistry, vol. 24, no. 1, pp. 43–55, 2000.
- E. Keogh, S. Lonardi, and C. A. Ratanamahatana, “Towards parameter-free data mining,” in Proceedings of the 10th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 206–215, August 2004.
- J. Ziv and A. Lempel, “A universal algorithm for sequential data compression,” IEEE Transactions on Information Theory, vol. 23, no. 3, pp. 337–343, 1977.
- J. Ziv and A. Lempel, “Compression of individual sequences via variable-rate coding,” IEEE Transactions on Information Theory, vol. 24, no. 5, pp. 530–536, 1978.
- S. Grumbach and F. Tahi, “Compression of DNA sequences,” in Proceedings of the IEEE Symposium on Data Compression, pp. 340–350, Snowbird, Utah, USA, 1993.
- S. Grumbach and F. Tahi, “A new challenge for compression algorithms: genetic sequences,” Information Processing and Management, vol. 30, no. 6, pp. 875–886, 1994.
- R. Giancarlo, D. Scaturro, and F. Utro, “Textual data compression in computational biology: a synopsis,” Bioinformatics, vol. 25, no. 13, pp. 1575–1586, 2009.
- T. M. Cover and J. A. Thomas, Elements of Information Theory, John Wiley & Sons, New York, NY, USA, 2012.
- D. Salomon, Data Compression: The Complete Reference, Springer Science and Business Media, 2004.
- X. Chen, M. Li, B. Ma, and J. Tromp, “DNACompress: fast and effective DNA sequence compression,” Bioinformatics, vol. 18, no. 12, pp. 1696–1698, 2002.
- B. Ma, J. Tromp, and M. Li, “PatternHunter: faster and more sensitive homology search,” Bioinformatics, vol. 18, no. 3, pp. 440–445, 2002.
- A. Apostolico and S. Lonardi, “Compression of biological sequences by greedy off-line textual substitution,” in Proceedings of the Data Compression Conference (DDC '00), pp. 143–152, March 2000.
- D. Adjeroh and F. Nan, “On compressibility of protein sequences,” in Proceedings of the Data Compression Conference (DCC '06), p. 10, Snowbird, Utah, USA, March 2006.
- D. Adjeroh, Y. Zhang, A. Mukherjee, M. Powell, and T. Bell, “DNA sequence compression using the Burrows-Wheeler Transform,” in Proceedings of the IEEE Computer Society Bioinformatics Conference, Computer Society, vol. 1, pp. 303–313, 2002.
- É. Rivals, M. Dauchet, J. P. Delahaye, and O. Delgrange, “Compression and genetic sequence analysis,” Biochimie, vol. 78, no. 5, pp. 315–322, 1996.
- M. D. Cao, T. I. Dix, L. Allison, and C. Mears, “A simple statistical algorithm for biological sequence compression,” in Proceedings of the Data Compression Conference (DCC '07), pp. 43–52, IEEE, Snowbird, Utah, USA, March 2007.
- D. Loewenstern and P. N. Yianilos, “Significantly lower entropy estimates for natural DNA sequences,” Journal of Computational Biology, vol. 6, no. 1, pp. 125–142, 1999.
- G. Korodi and I. Tabus, “An efficient normalized maximum likelihood algorithm for DMA sequence compression,” ACM Transactions on Information Systems, vol. 23, no. 1, pp. 3–34, 2005.
- J. I. Myung, D. J. Navarro, and M. A. Pitt, “Model selection by normalized maximum likelihood,” Journal of Mathematical Psychology, vol. 50, no. 2, pp. 167–179, 2006.
- T. Matsumoto, K. Sadakane, and H. Imai, “Biological sequence compression algorithms,” Genome Informatics, vol. 11, pp. 43–52, 2000.
- F. M. J. Willems, Y. M. Shtarkov, and T. J. Tjalkens, “The context-tree weighting method: basic properties,” IEEE Transactions on Information Theory, vol. 41, no. 3, pp. 653–664, 1995.
- B. Behzadi and F. Le Fessant, “DNA compression challenge revisited: a dynamic programming approach,” in Proceedings of the Annual Symposium on Combinatorial Pattern Matching, pp. 190–200, Springer, Berlin, Germany, 2005.
- D. Adjeroh and J. Feng, “The SCP and compressed domain analysis of biological sequences,” in Proceedings of the IEEE Bioinformatics Conference (CSB '03), pp. 587–592, Stanford, Calif, USA, August 2003.
- S. F. Altschul, W. Gish, W. Miller, E. W. Myers, and D. J. Lipman, “Basic local alignment search tool,” Journal of Molecular Biology, vol. 215, no. 3, pp. 403–410, 1990.
- P. Agarwal, “Compact encoding strategies for DNA sequence similarity search,” in Proceedings of the International Conference on Intelligent Systems for Molecular Biology (ISMB '95), vol. 4, pp. 211–217, 1995.
Copyright © 2016 Pamela Vinitha Eric et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.