Background

RNA performs many diverse functions in the cell in addition to its role as a messenger of genetic information. It can form enzymes, for example for cleavage of itself or of other RNA, or to create peptide bonds as a fundamental constituent of the ribosome [1]. It can act as a signalling molecule for regulation of gene expression, for protein export, or for guiding post-translational modifications [25].

As for proteins, RNA function depends on its folding to a well-defined three-dimensional shape. In contrast to proteins, the folding of RNA is hierarchical [6]. Secondary structure, or the particular set of contacts between pairs of complementary bases mediated by hydrogen bonding and stacking of bases, provides a significant amount of information. This can be helpful in predicting function or accessibility to ligands [710]. Computational prediction of the secondary structure of RNA from its sequence is therefore of great interest. The most widely-used automated prediction methods attempt to estimate the thermodynamic stability of RNA, using empirical parameters determined from experiments on oligonucleotides [11, 12].

CUDA is a programming interface developed by the company NVIDIA to facilitate general-purpose, parallel high-performance computing on multiprocessor graphics processing units (GPUs) [13]. In recent years, many scientific computing applications have been implemented on GPUs using CUDA, in many cases yielding speed-ups of several orders of magnitude [14, 15]. However, to our knowledge, only a handful of publications have appeared describing GPU implementations of codes for RNA secondary structure prediction. Rizk and Lavenier described a CUDA implementation of structure prediction by free energy minimization [16]. Their work was limited to relatively short sequences (up to 120 bases) and a simplified energy model that neglects coaxial stacking. The GPU implementation was faster than the serial implementation by a factor of 10 and 17 depending on the particular hardware. More recently, Lei et al. [17] also reported a parallelized implementation of free energy minimization using CUDA. They used only a coarse-grained parallelization scheme where the minimum-free-energy structures for subsequences of a given length are calculated in parallel, but the search over structures for each subsequence is done serially. Their work was also limited to relatively short sequences (up to 221 bases). They reported speedups of up to a factor of 16. It should be noted that these parallelized implementations neither use the latest thermodynamic parameters for loops [18], nor include coaxial stacking interactions [19].

In addition, recent work demonstrates that calculations of base-pairing probabilities calculated with partition functions can provide additional useful information [20]. Structures composed of highly probable pairs are more accurate than lowest free energy structures [20, 21]. The base pair probabilities provide confidence intervals for prediction of individual pairs. Base pairing probabilities can also be used to predict pseudoknots [22]. We have additionally extended this work to predictions using multiple, homologous sequences, where the same principles hold true [2326].

In this paper, we present the calculation of base-pair probabilities for 44 sequences containing up to 10,000 bases, using an optimized and parallelized version of the “partition” code in the RNAstructure package [27]. Our test set contained both random sequences of varying lengths (up to 10,000 bases) and actual sequences of biological importance (see Table 1), the longest being the HIV-1 NL43 genome (GenBank: AF324493) containing 9709 bases.

Table 1 Biological RNA sequences examined *

We employed a sophisticated and accurate energy calculation that includes coaxial stacking [19]. Before attempting a parallel implementation, we first wrote an optimized serial version of the original code (the “partition” program in RNAstructure), implementing only a subset of its functionality, while improving efficiency and reducing memory usage. Subsequently, we parallelized the optimized code for GPU hardware, using CUDA. Here we made use of a fine-grained parallelization scheme, in which the calculation of the restricted partition function for each subsequence of a given length is parallelized, as well as the calculation of the restricted partition function for each subsequence. We found that this fine-grained parallelization resulted in greater speedups than a simpler coarse-grained-only parallelization (up to factors of ∼60 compared to the optimized serial version and ∼116 compared to the original code).

Implementation

Calculating base pair probabilities

The probability pi,j of a canonical pair between bases i and j related to the standard free energy change Δ G i , j 0 of the restricted ensemble of structures containing the pair, and the standard free energy change Δ G0 of the unrestricted ensemble of all possible structures, both relative to the state in which all bases are unpaired:

p i , j = w Δ G i , j 0 w Δ G 0
(1)

where

w(g)exp g / RT
(2)

is the Boltzmann weight corresponding to standard free energy change g at temperature T.

The standard free energy changes Δ G i , j 0 and Δ G0 are estimated using the Turner nearest-neighbor rules [11, 18, 37], and pseudoknots are excluded. In that case, Δ G i , j 0 depends on two independent contributions, one for the interior fragment containing the pair i,j and bases in between (but excluding bases from the 5’ end to i−1 and from j+1 to the 3’ end), and one for the exterior fragment containing the pair and bases from the two ends (but excluding bases from i+1 to j−1). We define Vi,j to be the relative standard free energy change for the interior fragment in the case that i<j and for the exterior fragment otherwise, following the convention used in the mfold prediction software [38]. In that case,

Δ G i , j 0 =RTlog exp V i , j / RT + exp V j , i / RT
(3)

or more succinctly

w Δ G i , j 0 =w V i , j +w V j , i
(4)

The standard free energy changes Vi,j are calculated using the following set of recursions:

w V i , j = Q i , j hairpin + Q i , j stack + Q i , j internal + Q i , j multibranch for i < j Q i , j exterior + Q i , j stack + Q i , j internal + Q i , j multibranch for i > j
(5)

where

Q i , j hairpin = w Δ G i , j hairpin
(6)
Q i , j stack = w V i + 1 , j 1 + Δ G i , j , i + 1 , j 1 stack
(7)
Q i , j internal = i < i < j < j w V i , j + Δ G i , j , i , j internal
(8)
Q i , j multibranch = w W i + 1 , j 1 MB + a + c + w W i + 2 , j 1 MB + Δ G i , j , i + 1 3 dangle + a + b + c + w W i + 1 , j 2 MB + Δ G i , j , j 1 5 dangle + a + b + c + w W i + 2 , j 2 MB + Δ G i , j , i + 1 , j 1 terminal mismatch + a + 2 b + c + i < k < j w V i + 1 , k + Y k + 1 , j 1 + Δ G j , i , i + 1 , k coaxial flush + a + 2 c + i < k < j w V i + 2 , k + Y k + 2 , j 1 + Δ G j , i , i + 2 , k coaxial mismatch(2) + a + 2 b + 2 c + i < k < j w V i + 2 , k + Y k + 1 , j 2 + Δ G j , i , i + 2 , k coaxial mismatch(1) + a + 2 b + 2 c + i < k < j w V k , j 1 + Y i + 1 , k 1 + Δ G k , j 1 , j , i coaxial flush + a + 2 c + i < k < j w V k , j 2 + Y i + 1 , k 2 + Δ G k , j 2 , j , i coaxial mismatch(1) + a + 2 b + 2 c + i < k < j w V k , j 2 + Y i + 2 , k 1 + Δ G k , j 2 , j , i coaxial mismatch(2) + a + 2 b + 2 c
(9)
Q i , j exterior = w W i + 1 3 + W j 1 5 + w W i + 2 3 + W j 1 5 + Δ G i , j , i + 1 3 dangle + + w W i + 1 3 + W j 2 5 + Δ G i , j , j 1 5 dangle + + w W i + 2 3 + W j 2 5 + Δ G i , j , i + 1 , j 1 terminal mismatch + i < k < j w V i + 1 , k + W k + 1 3 + W j 1 5 + Δ G j , i , i + 1 , k coaxial flush + i < k < j w V i + 2 , k + W k + 2 3 + W j 1 5 + Δ G j , i , i + 2 , k coaxial mismatch(2) + i < k < j w V i + 2 , k + W k + 1 3 + W j 2 5 + Δ G j , i , i + 2 , k coaxial mismatch(1) + i < k < j w V k , j 1 + W i + 1 3 + W k 1 5 + Δ G k , j 1 , j , i coaxial flush + i < k < j w V k , j 2 + W i + 1 3 + W k 2 5 + Δ G k , j 2 , j , i coaxial mismatch(1) + i < k < j w V k , j 2 + W i + 2 3 + W k 1 5 + Δ G k , j 2 , j , i coaxial mismatch(2)
(10)
w W i , j L = w W i + 1 , j L + b + w V i , j + c + w V i , j 1 + Δ G j 1 , i , j 3 dangle + b + c + w V i + 1 , j + Δ G j , i + 1 , i 5 dangle + b + c + w V i + 1 , j 1 + Δ G j 1 , i + 1 , j , i terminal mismatch + 2 b + c
(11)
w W i , j Q = w V i , j + w V i , j 1 + Δ G j 1 , i , j 3 dangle + w V i + 1 , j + Δ G j , i + 1 , i 5 dangle + w V i + 1 , j 1 + Δ G j 1 , i + 1 , j , i terminal mismatch
(12)
w W i , j = w W i , j 1 + b + w W i , j L
(13)
w W i , j coax = i < k < j w V i , k + V k + 1 , j + Δ G i , k , k + 1 , j coaxial flush + 2 c + i < k < j w V i + 1 , k + V k + 2 , j + Δ G i + 1 , k , kj + 2 , j coaxial mismatch(1) + 2 b + 2 c + i < k < j w V i , k + V k + 2 , j 1 + Δ G i , k , k + 2 , j 1 coaxial mismatch(2) + 2 b + 2 c
(14)
w Z i , j = w W i , j coax + w V i , j + c + w V i , j 1 + Δ G j 1 , i , j 3 dangle + b + c + w V i + 1 , j + Δ G j , i + 1 , i 5 dangle + b + c + w V i + 1 , j 1 + Δ G j 1 , i + 1 , j , i terminal mismatch + 2 b + c
(15)
w W i , j MBL = w W i + 1 , j MBL + w W i , j coax + i < k < j w Z i , k + Y k + 1 , j L
(16)
w W i , j MB = w W i , j 1 MB + b + w W i , j MBL
(17)
w Y i , j = w W i , j + w W i , j MB
(18)
w Y i , j L = w W i , j L + w W i , j MBL
(19)
w W i 5 = w W i 1 5 + j < i w W j 1 5 + W j , i Q
(20)
w W i 3 = w W i + 1 3 + j > i w W j + 1 3 + W i , j Q
(21)

These recursions are slightly different from—but equivalent to—those presented in reference [20] and used in the previous code. It should be noted that there was an error in equation 15 of reference [20]: in the second line, WMBL(k+1,j) should be replaced by [WMBL(k+1,j) + WL(k+1,j)].

The quantities V, W, WMB, WL, Wcoax, W5′, and W3′ are simply −R T times the logarithms of the quantities in reference [20]. In addition, four new arrays are introduced:

  1. 1.

    W Q is the standard free energy change corresponding to the sum of terms on the right hand side of equation 11 of reference [20] not including the scaling by W5(k).

  2. 2.

    Elements of Y are −R T times the logarithm of the sum of the Boltzmann weights of corresponding elements of W and W MB.

  3. 3.

    Elements of Y L are −R T times the logarithm of the sum of the Boltzmann weights of corresponding elements of W L and W MBL.

  4. 4.

    Elements of Z are −R T times the logarithm of the sum of the Boltzmann weights of corresponding elements of W L (except for the term depending on the next-smallest fragment) and W coax.

Reorganizing the recursions in this way might appear to use more memory because of the additional arrays. In fact, the modified version requires less memory, because several of the arrays do not need to be stored in their entirety. Specifically, using the modified recursions, storage is only required for two diagonals of W, WL, and WMBL; for five diagonals of WMB; and for a half-triangle of WQ. Reducing memory usage is important as the size of the full arrays scales as O(N2) and the available GPU memory on our hardware was limited to ∼2.5 GB. The modified recursions use four full N×N arrays and one half-triangle, rather than the six full N×N arrays used in the original recursions, and therefore reduce memory usage by about 25%. In addition, the calculation of W5′ and W3′ is simplified (compare equations 20 and 21 above with equation 11 of reference [20]).

As in the previous work, the various Δ G parameters above are from the Turner nearest-neighbor rules [11], while a,b, and c are from the following estimate of the standard free energy change for multibranch loop initiation:

Δ G 0 =a+bn+ch
(22)

Here n is the number of unpaired nucleotides and h is the number of branching helices [21]. By convention, the size of internal loops is limited to thirty unpaired nucleotides, so the number of terms in equation 8 and the overall computational expense scales as O(N3) where N is the size of the sequence.

The largest difference in the new implementation is that logarithms of probabilities and partition functions (i.e., standard free energy changes) are used rather than probabilities themselves, which is convenient when working in single precision in order to avoid overflow or underflow errors. This requires that exponentials and logarithms are calculated at each step of the calculation where sums are performed. This approach is a departure from the previous implementation, which used scaling factors. Having to compute logarithms and exponentials does entail some additional computational expense, but this does not appear to be exhorbitant on the GPU, because optimized intrinsic mathematical functions are used. (i.e., the code was compiled with the NVIDIA compiler using the -use_fast_math option). The function required for the sum of two free energies a and b (expressed in units such that R T=1) is

f(a,b)=log e a + e b
(23)

We calculated this in the following way:

f(a,b)= a log 1 + e a b a < b b log 1 + e b a a > b a log ( 2 ) a = b
(24)

This was done for two reasons: it requires at most a single call to exp, rather than two; and it can make use of the log1p function from the standard math library, which calculates log(1+x) accurately even for small x. This is important because often ea and eb will differ by several orders of magnitude, and simply adding them and then taking the logarithm can lead to significant roundoff error.

In order to determine how much additional computational overhead was imposed by the calculation of exp and log1p we performed a comparison with an artificial reference calculation, which was identical except that calls to these functions were omitted. We found that for a 1,000-mer, the actual GPU calculation is only ∼20% more expensive than this reference calculation.

For a serial calculation on the CPU, there is a larger performance hit; the actual calculation is about a factor of two more expensive than the reference without exp or log1p. However, it should be noted that this is not the entire story, because overall, the new optimized serial code, which uses logarithms, is still faster than the original code, which does not. Running the calculation in log space results in simplifications such as not requiring checking for overflow and not having to multiply by scaling factors, which reduces computational expense.

Parallelization of the partition function calculation using CUDA

In the CUDA programming model, overall execution of the program is still governed by the CPU. Compute-intensive portions of the program are then delegated to subroutines executed by the GPU, or kernels. In general, the GPU has its own memory, so data must be copied to and from the GPU before and after kernel execution. Many copies of a kernel, or threads, run in parallel, each of which belongs to a block. During kernel execution, threads belonging to the same block can share data and synchronize, whereas threads belonging to different blocks cannot. A program can contain many kernels, which can execute either serially or in parallel [13].

The algorithm for calculating partition functions is recursive: partition functions for larger fragments depend on those for smaller fragments. As such, the overall calculation proceeds serially, in order of fragment size. We used two levels of parallelization, a block level and a thread level. Calculations for all fragments of a given size may be done in parallel, with no communication. This was implemented in CUDA at the level of blocks of threads. The partition function for a given fragment depends on sums, with the number of terms on the order of the fragment size (e.g., equation 9). These sums were parallelized at the level of threads within a block, since calculating a sum in parallel relies on communication between the threads. In our experience, a greater speedup was obtained from this “inner loop” parallelization, even though it requires more communication between threads. Most likely, this is because optimal efficiency on GPU hardware is obtained when identical mathematical operations are performed in lockstep on different data [13]. We stress that these two different levels of parallelization are not mutually exclusive and optimal performance was obtained from including both. A separate block of threads was run for each fragment, while 256 threads were run within a block. The number of threads per block was chosen by trial and error and was optimal for our hardware (the simple sum reduction scheme we chose requires it to be a power of two). In our code, this value is set at compile time (but this is not required by CUDA—it could be set at run time if desired).

Results and discussion

Accuracy

Peak floating point performance for NVIDIA Tesla GPUs are faster by a factor of two when working in single compared with working in double precision (http://www.nvidia.com), but single precision introduces greater roundoff error. In order to examine accuracy, we calculated base-pair probabilities for the same sequences using both the parallel CUDA/GPU implementation in single precision, and the serial implementation in double precision. We also calculated probabilities in double precision using a set of nearest-neighbor parameters slightly modified by adding a random variate chosen from a Gaussian distribution with mean 0 and standard deviation 0.01 kcal/mol, which is comparable to or smaller than their experimental uncertainty (0.1 kcal/mol for parameters describing helical stacking and 0.5 kcal/mol for parameters describing loops [18, 37]).

Figure 1 shows the root-mean-square deviation (RMSD) between base-pair probabilities calculated using double precision and either single precision, or the modified parameters. The RMSD of the calculation using modified parameters decreases as the size of the sequences increases (because the fraction of pairs with very small probabilities increases). In contrast, the roundoff error due to working in single precision increases as the cube of the number of bases in the sequence (i.e., the number of operations involved in the calculation). However, it remains relatively small for sequences of up to 10,000 bases, and negligible compared with the differences resulting from the modified parameters. Our conclusion is that working in single precision does not introduce unacceptable roundoff error for RNA secondary structure prediction for sequences of this size and most likely substantially larger.

Figure 1
Figure 1
Full size image

Accuracy of base-pair probabilities. Root-mean-square deviation in base-pair probabilities calculated using the CUDA single-precision implementation (red plus signs) and the serial double-precision implementation with slightly modified parameters (green crosses), compared with the serial double-precision implementation using the original parameters.

We also used the calculated base pair probabilities to determine a single consensus structure for each sequences, using the ProbKnot algorithm [22]. In this case, working in single precision led to differences with double precision only for one sequence (a random sequence of 6000 bases) out of the 44 we examined, and these were very small (only three bases out of the 6000 were matched with a different partner). Working with the modified parameters led to larger discrepancies although these were still fairly small (for sequences containing more than 100 bases, at most 6% of bases were matched with different partners). This is consistent with a previous report that predictions using base pair probabilities are significantly less sensitive to errors in thermodynamic parameters than using only lowest free energy structures [39].

Computational expense

We compared overall execution time for the original serial code developed in our laboratory, the optimized serial code, and the parallel CUDA/GPU implementation. Calculations were performed for sequences containing from 10 to 10,000 bases, on compute nodes containing dual hex-core Intel Xeon 2.67 GHz CPUs and dual NVIDIA Tesla M2050 GPUs, each of which contains 448 multiprocessor cores (only a single GPU was used). Figure 2 shows the execution time (from the user time reported by the UNIX “time” command) of the original and optimized serial codes as well as the parallel CUDA/GPU implementation. The execution time for all codes scales as the cube of the sequence length for large sequences. The CUDA version was able to calculate base-pair probabilities for the sequence for the full HIV-1 NL43 genome (9709 nucleotides) in 27 minutes. We note that there is a small overhead (a few hundredths of a second) involved in running calculations with either the original serial code or the CUDA code, which is not present for the optimized serial code. This overhead has different origins: for the original code, it is probably due to reading the parameter files from disk, while for the CUDA code, it is most likely due to copying parameters and other data between the GPU and CPU.

Figure 2
Figure 2
Full size image

Computational expense. Execution time in seconds for the original serial implementation (red plus signs), optimized serial implementation (green crosses), and parallel GPU implementation (blue asterisks). The red, green, and blue lines are the cube of the number of bases n in the sequences, multiplied by a constant chosen to best fit the corresponding execution times.

Conclusions

In this work, we introduced a modified set of recursions for calculating RNA secondary structure partition functions and base-pairing probabilities using a dynamic programming algorithm, and implemented these in parallel using the CUDA framework for multiprocessor GPUs. For large sequences, the GPU implementation reduces execution time by a factor of close to 60 compared with an optimized serial implementation, and by a factor of 116 compared with the original code. It is clear from our work that using GPUs can greatly accelerate computation of RNA secondary structure partition functions, allowing calculation of base-pair probabilities for large sequences in a reasonable amount of time, with a negligible compromise in accuracy due to working in single precision. It is expected that parallelization using CUDA should be applicable to other implementations of dynamic programming algorithms [12] besides ours, and result in similar speedups.

Two levels of parallelization were implemented. Calculations for all fragments of a given size were done in parallel, with no communication between threads. This was implemented in CUDA at the level of blocks of threads. In addition, the sums contributing to the partition function for a given fragment were calculated in parallel, with communication required between threads. These sums were parallelized at the level of thread within a block. We found that this “inner loop” parallelization resulted in a significantly greater speedup than the “outer loop” parallelization alone.

Availability and requirements

● Project name: partition-cuda; part of RNAstructure, version 5.5 and later

● Project home page: http://rna.urmc.rochester.edu/RNAstructure.html

● Operating system(s): Unix

● Programming languages: C and CUDA

● Other requirements: CUDA compiler, available from

● http://www.nvidia.com/object/cuda\_home\_new.html

● License: GNU GPL

● Any restrictions to use by non-academics: None.