Selected Papers from the Midwest Symposium on Circuits and Systems
View this Special IssueResearch Article  Open Access
JunKyu Lee, Gregory D. Peterson, Robert J. Harrison, Robert J. Hinde, "Implementation of HardwareAccelerated Scalable Parallel Random Number Generators", VLSI Design, vol. 2010, Article ID 930821, 11 pages, 2010. https://doi.org/10.1155/2010/930821
Implementation of HardwareAccelerated Scalable Parallel Random Number Generators
Abstract
The Scalable Parallel Random Number Generators (SPRNGs) library is widely used in computational science applications such as Monte Carlo simulations since SPRNG supports fast, parallel, and scalable random number generation with good statistical properties. In order to accelerate SPRNG, we develop a HardwareAccelerated version of SPRNG (HASPRNG) on the Xilinx XC2VP50 Field Programmable Gate Arrays (FPGAs) in the Cray XD1 that produces identical results. HASPRNG includes the reconfigurable logic for FPGAs along with a programming interface which performs integer random number generation. To demonstrate HASPRNG for Reconfigurable Computing (RC) applications, we also develop a Monte Carlo estimator for the Cray XD1. The RC Monte Carlo estimator shows a 19.1 speedup over the 2.2 GHz AMD Opteron processor in the Cray XD1. In this paper we describe the FPGA implementation for HASPRNG and a estimator example application exploiting the finegrained parallelism and mathematical properties of the SPRNG algorithm.
1. Introduction
Random numbers are required in a wide variety of applications such as circuit testing, system simulation, gameplaying, cryptography, evaluation of multiple integrals, and computational science Monte Carlo (MC) applications [1].
In particular, MC applications require a huge quantity of highquality random numbers in order to obtain a highquality solution [2–4]. To support MC applications effectively, a random number generator should have certain characteristics [5]. First, the random numbers must maintain good statistical properties (e.g., no biases) to guarantee valid results. Second, the generator should have a long period. Third, the random numbers should be reproducible. Fourth, the random number generation should be fast since generating a huge quantity of random numbers requires substantial execution time. Finally, the generator should require little storage to allow the MC application to use the rest of the storage resources.
Many MC applications are embarrassingly parallel [6]. To exploit the parallelism, Parallel Pseudorandom Number Generators (PPRNGs) are required for such applications to achieve fast random number generation [2, 6, 7]. Random numbers from a PPRNG should be statistically independent each other to guarantee a highquality solution. Many common Pseudorandom Number Generators (PRNGs) fail statistical tests of their randomness [8, 9]; so computational scientists are cautious in selecting PRNG algorithms.
The Scalable Parallel Random Number Generators (SPRNGs) library is one of the best candidates for parallel random number generation satisfying the five characteristics, since it supports fast, scalable, and parallel random number generation with good randomness [2, 7]. SPRNG consists of 6 types of random number generators: Modified Lagged Fibonacci Generator (Modified LFG), 48bit Linear Congruential Generator with prime addend (48bit LCG), 64bit Linear Congruential Generator with prime addend (64bit LCG), Combined Multiple Recursive Generator (CMRG), Multiplicative Lagged Fibonacci Generator (MLFG), and Prime Modulus Linear Congruential Generator (PMLCG).
We desire to improve the generation speed by implementing a hardware version of random number generators for simulation applications, since generating random numbers takes a considerable amount of the execution time for applications which require huge quantities of random numbers [10]. FPGAs have several advantages in terms of speedup, energy, power, and flexibility for implementation of the random number generators [11, 12].
HighPerformance Reconfigurable Computing (HPRC) platforms employ FPGAs to execute the computationally intensive portion of an application [13, 14]. The Cray XD1 is an HPRC platform providing a flexible interface between the microprocessors and FPGAs (Xilinx XC2VP50 or XC4VLX160). FPGAs are able to communicate with a microprocessor directly through the interface [15].
Therefore, we explore the use of reconfigurable computing to achieve faster random number generation. In order to provide the highquality, scalable random number generation associated with SPRNG combined with the capabilities of HPRC, we developed the HardwareAccelerated Scalable Parallel Random Number Generators library (HASPRNGs) that provides bitequivalent results to operate on a coprocessor FPGA [16–18]. HASPRNG can be used to target computational science applications on arbitrarily large supercomputing systems (e.g., Cray XD1, XT5h), subject to FPGA resource availability [15].
Although the computational science application could be executed on the node microprocessors with the FPGAs accelerating the PPRNGs, the HASPRNG implementation could also be colocated with the MC application on the FPGA. The latter approach avoids internal bandwidth constraints and enables more aggressive parallel processing. This presents significant potential benefit from tightly coupling HASPRNG with Reconfigurable Computing Monte Carlo (RC MC) applications. For example, the RC MC estimation application can employ a huge quantity of random numbers using as many parallel generators as the hardware resources can support.
In this paper we describe the implementation of the HASPRNG library for the Cray XD1 for the full set of integer random number generators in SPRNG and demonstrate the potential of HASPRNG to accelerate RC MC applications by exploring a estimator on the Cray XD1.
2. Implementation
SPRNG includes 6 different types of generators and a number of default parameters [2, 7]. Each of the SPRNG library random number generators and its associated parameter sets are implemented in HASPRNG. The VHSICs (Very HighSpeed Integrated Circuits) Hardware Description Language (VHDL) is used for designing the reconfigurable logic of HASPRNG and the RC MC estimator. To provide a flexible interface to RC MC applications, a onebit control input is employed to start and stop HASPRNG operation. When HASPRNG is paused, all the state information inside HASPRNG is kept to enable the resumption of random number generation when needed. Similarly, a onebit control output signals the availability of a valid random number. These two ports are sufficient for the control interface to a RC MC application developer.
To provide high performance, eight generators are implemented for HASPRNG on the Cray XD1. The modified lagged Fibonacci and multiplicative lagged Fibonacci generators have two implementations to exploit the potential concurrency associated with different parameter sets. We call the eight generators in HASPRNG as follows: HardwareAccelerated Modified Lagged Fibonacci Generator for OddOdd type seed (1: HALFGOO), HardwareAccelerated Modified Lagged Fibonacci Generator for OddEven type seed (2: HALFGOE), HardwareAccelerated Linear Congruential Generator for 48 bits (3: HALCG48), HardwareAccelerated Linear Congruential Generator for 64 bits (4: HALCG64), HardwareAccelerated Combined Multiple Recursive Generator (5: HACMRG), HardwareAccelerated Multiplicative Lagged Fibonacci Generator for Short lag seed (6: HAMLFGS), HardwareAccelerated Multiplicative Lagged Fibonacci Generator for Long lag seed (7: HAMLFGL), and HardwareAccelerated Prime Modulus Linear Congruential Generator (8: HAPMLCG). Since HASPRNG implements the same integer random number generation as SPRNG, every HASPRNG generator returns 32bit positive integer values (most significant bit equals “0”). In consequence, HASPRNG converts different bitwidth random numbers from different generators (e.g., 32 bits for LFG, 48 bits for LCG48, 64 bits for LCG64) to positive 32bit integer random numbers and generates random numbers from 0 to (). For the conversion, HASPRNG masks the upper 31 bits of the different bitwidth random numbers and prepends “0” as the most significant bit of the 31bit data in order to produce positive 32bit random numbers for all generators in HASPRNG. The techniques to produce 32bit random numbers are the same as in SPRNG [2, 7]. The random numbers before the conversion are fed back to the generators to produce future random numbers in HASPRNG.
HASPRNG can be used not only for improving the speed of software applications by providing a programming interface but also for accelerating RC MC applications by colocating the application and the hardware accelerated generator(s) on the FPGAs (see Sections 2.6, 2.7, 3.4, and 4).
2.1. HardwareAccelerated Modifed Lagged Fibonacci Generator (HALFG)
The Modified LFG in SPRNG is computed by the XOR bit operation of two additive LFG products [2, 7]. The Modified LFG is expressed by (1), (2), and (3): where and are called the lags of the generator. The generator follows the convention of . and are 32bit random numbers generated by the two additive LFGs. is represented by setting the least significant bit of to 0. is represented by shifting right by one bit. is the th random number of the generator [2, 7]. In SPRNG the generator produces a random number every two stepoperations. To provide bitequivalent results and improve performance, two types of designs are used depending on whether is odd or even [19, 20]. We employ the design for HALFG as in [19]. The two types of designs are represented by HALFGOO (both and are odd numbers in (1), (2), and (3)) and HALFGOE ( is odd and is even). The HALFGOO and HALFGOE employ block memory modules in order to store initial seeds and state. Note that small values of result in data hazards that complicate pipelining. Hence, these designs are optimized for the specific memory access patterns dictated by and . Figure 1 shows the HALFGOO and HALFGOE architectures. The block memories are represented by , , and and the results from the two additions are represented by and in Figure 1. HALFG requires an initialization process to store previous values in the three block memories, since and in (2) and (3) require the previous values. Therefore HALFG requires separate buffers of previous values to generate the different random number sequences for and .
(a)
(b)
After the initialization, the seeds are accessed to produce random numbers. For example, represents a random number, or , and at the same time is stored to and to produce future random numbers. This method is able to generate a random number every two step operations. The HALFGOO and HALFGOE generators produce a random number every clock cycle.
2.2. HardwareAccelerated 48bit and 64bit Linear Congruential Generators (HALCGs)
The two LCGs use the same algorithm with different data sizes, 48 bits and 64 bits. The LCG characteristic equation is represented by
where is a prime number, is a multiplier, is 2^{48} for the 48bit LCG and 2^{64} for the 64bit LCG, and is the th random number [2, 7]. The HALCGs must use the previous random number as in (4). Consequently, the HALCGs face data feedback (hazards) in the multiplication which reduces performance. Thanks to the simple recursion relation of (4) we are able to generate a future random number by unrolling the recursion in (4) based on pipeline depths. Equations (5) and (6) represent the modified equations which produce identical results as (4) in the HALCGs. HALCG implementations employing (5) and (6) generate two random numbers every clock cycle using two sevenstage pipelined multipliers: The architecture for the HALCGs is shown in Figure 2. Two generation engines are employed to produce two random numbers. One generation engine (Generators 1 in Figure 2) produces the odd indexed random numbers (e.g., (17), (19), (21), …) and the other generator (Generator 2 in Figure 2) produces the even indexed random numbers (e.g., (16), (18), (20), …). For the multiplications in (5), we employ builtin multipliers inside the generators (Generator 1 and 2 in Figure 2) (see Table 3).
Instead of having internal logic modules to obtain the and in (6), the coefficients are precalculated in software to save hardware resources. Software also calculates 15 initial random numbers (()–Z(15)) to provide the initial state to the HALCGs. The pregenerated 15 random numbers and an initial seed are stored in the register file (Register File in Figure 2: (Initial seed), ) having 16 48/64bit registers during the initialization process. The HALCGs produce 15 random numbers during initialization before they generate random number outputs. In consequence, the generator generates two 32bit random numbers every clock cycle. We provide two random numbers every clock cycle to exploit the Cray XD1 bandwidth since the Cray XD1 can support a 64 bit data transfer between SDRAM and FPGAs every clock cycle. A microprocessor can access the SDRAM directly (see Section 2.6).
2.3. HardwareAccelerated Combined Multiple Recursive Generator (HACMRG)
The SPRNG CMRG employs two generators. One is a 64bit LCG and the other is a recursive generator [7]. The recursive generator is expressed by (8). The CMRG combines the two generators as follows: where is generated by a 64bit LCG, represents a 31bit random number, and is the resulting random number [2, 7]. The implementation equation is represented by: where is the upper 32 bits of . Equation (9) produces identical results as (7).
Figure 3 shows the HACMRG hardware architecture. The architecture has two parts, each generating a partial result. The first part is an HALCG64, and the second part is a generator having two lag factors as in (8).
The left part in Figure 3 is the HALCG64 employing a twostaged multiplier. The HACMRG HALCG64 produces one random number every other clock cycle to synchronize with the generator composed of two twostaged multipliers, fourdeep FIFO registers, and some combinational logic. In the generator, the multiplexor controlling the left multiplier inputs switches between an input value of “1” during the initial random number and the previous value thereafter.
For the “” modulo operator implementation, the 62bit value summed from the two multiplier’s outputs is shifted right by 31bits and added to the lower 31bits of the value before the shifting. The shifted value represents the modulo value for the higher 31 bits of the 62 bit value and the lower 31 bits of the 62 bit value represents the modulo value itself. The summed value represents the total modulo value. The total modulo value is reexamined to represent the final modulo value. If the total modulo value is larger or equal than 2^{31}, the final modulo value is represented by adding “1” to the lower 31bit data of the total modulo value since the () modulo value of 2^{31} is “1”. The value obtained from the modulus operation fans out three ways. The first one goes to the left multiplier input port in Figure 3 in order to save one clock cycle latency, the second one goes to the FIFO, and the third one goes to the final adder to add the value to the result generated by the HALCG64. The resulting upper 31 bits represent the next random number. The HACMRG generates a random number every other clock cycle.
2.4. HardwareAccelerated Multiplicative Lagged Fibonnaci Generator (HAMLFG)
The SPRNG MLFG characteristic equation is given by where and are time lags and is the resulting random number [2, 7].
The HAMLFGS and HAMLFGL employ two generators to produce two random numbers every clock cycle based on (10). The HAMLFGS and HAMLFGL require threeport RAMs to consistently keep previous random numbers since two ports are required to store two random numbers and one port is required to read a random number at the same time. However, only two ports are supported for the DPRAMs. Fortunately, (10) reveals data access patterns such that one port is enough for storing two random numbers. Tables 1 and 2 show the data accessing patterns in the case of the and parameter sets. In these tables, we reference the “oddodd” case when the longer lag factor is odd and the shorter lag factor is odd and the “oddeven” case when the longer lag factor is odd and the shorter lag factor is even. Note that is always odd for the SPRNG parameter sets [7].



In the case of oddodd parameter sets, the odd part always accesses the even part and the even part always accesses the odd part in Table 1. In the case of oddeven parameter sets, the odd part always accesses the odd part and the even part always accesses the even part in Table 2. The data access patterns are described in bold face in Table 2. Figure 4 shows the HAMLFGS and HAMLFGL architecture exploiting these data access patterns. HAMLFGS and HAMLFGL employ four DPRAMs to store random numbers and two multipliers to produce two random numbers every clock cycle, since the generator can access four values every clock cycle (see Tables 1 and 2). The two upper DPRAMs (DPRAM 1 and DPRAM 2 in Figure 4) are used to generate odd indexed random numbers and the two lower DPRAMs (DPRAM 3 and DPRAM 4) are used to generate even indexed random numbers.
HAMLFGS and HAMLFGL also employ one readcontroller to read data from DPRAMs and one writecontroller to write data to DPRAMs from the multipliers. The two multiplexers (MUX in Figure 4) exploit the data access patterns. They select the appropriate values for the oddodd and oddeven parameter sets. Twostage multipliers are employed for HAMLFGS for the and parameter sets and sevenstage pipelined multipliers are employed for HAMLFGL for the other nine parameter sets to avoid data hazards (SPRNG provides eleven parameter sets for MLFG) [7]. Even though twostaged multipliers are employed for the HAMLFGS implementation, data hazards still exist in Tables 1 and 2 since a two clock cycle latency is required due to the twostage multipliers along with an additional two clock cycles for the DPRAM access, one for writing and one for reading. In order to avoid the two clock cycle delay from the DPRAM access, we employ forwarding techniques when the parameter set is or . Instead of storing the data into DPRAMs, the data is fed into the multipliers directly from the multiplier’s outputs through the multiplexers. The forwarding is shown as dottedlines in Figure 4. If the required latency is three clock cycles, such as for the odd part in and both parts in , one stall is inserted to synchronize the data access. HAMLFGS and HAMLFGL produce two random numbers every clock cycle.
2.5. HardwareAccelerated Prime Modulus Linear Congruential Generator (HAPLMCG)
The SPRNG PMLCG generates random numbers as follows: where is a multiplier and is the resulting random number [2, 7]. Figure 5 shows the HAPMLCG architecture.
HAPMLCG employs four twostage 32bit multipliers that execute in parallel. The initial seed and the initial multiplier coefficient are divided by the lower 31 bits and upper 30 bits and are stored into the gray registers in Figure 5. In order to make those data 32 bits for 32bit multiplications, the value of “0” is attached to the lower 31bit data at the most significant bit position and the value of “00” is attached to the higher 30bit data at the most significant bit position. All the shift and mask operations before the IFcondition black box are needed to make two 61bit data multiplication (one is a multiplier coefficient and the other is a random number) as in (11). The last part of the implementation is described by the IFcondition black box in Figure 5. The IFcondition black box performs the modulus and data check operations. When the 62nd bit of the data is “1”, the data is modified by adding “1” after the 61bit mask operation. If the modified data is 2^{61}, the data is changed to the value “1” in order to prevent the generator from producing “0” as a random number [7, 19, 20]. The final data from the IFcondition black box fans out two directions. One is fed into one of the inputs of the multiplexer in order to generate the next random number. The other represents a resultant random number. After the first operation, the multiplexer always selects the resultant random numbers instead of the initial seed. The HAPMLCG generates a random number every other clock cycle.
The HAPMLCG did not use the HALCG technique for generating future random numbers, since it was extremely hard to derive an equation to generate future random numbers due to the complicated modulo value while preventing the generation of a zero random number.
2.6. Programming Interface for HASPRNG
The programming interface for HASPRNG employs the C language and allows users to use HASPRNG the same way as SPRNG 2.0. The programming interface requires two buffers in main memory on the Cray XD1 to transfer data between the FPGAs and microprocessors. The main memory plays a role as a bridge to communicate between the microprocessor and the FPGA directly through HyperTransport (HT)/RapidArrayTransport (RT) [15].
HASPRNG provides the initialization function, the generation function, and the free function for the programming interface as in SPRNG 2.0 [7]. To the programmer, the functions for HASPRNG are identical to those for SPRNG, but with “sprng” replaced by “hasprng.” The initialization function initializes HASPRNG. The generation function returns a random number per function call. The free function releases the memory resources when the random number generation is completely done. The three functions are implemented using Application Programming Interface (API) commands in C. Figure 6 describes the hardware architecture to support the programming interface of the Cray XD1.
The HASPRNG initialization function is responsible for reconfiguring the FPGA logic (FPGA in Figure 6) and registering buffers (Buffer 1 and Buffer 2 in Figure 6) in the Cray XD1 main memory. The initialization function requires five integer parameters to initialize HASPRNG. The five parameters represent the generator type, the stream identification number, the number of streams, an initial seed, and an initial parameter as with the initialization of SPRNG generators [7]. Refer to [7] for further explanation of the five initialization parameters. The logic for the specified generator type parameter is used to program the FPGAs. Once the generator core (Generator Core in Figure 6) is programmed in the FPGAs, the initialization function lets the generator core generate random numbers based on the five integer parameters and send them to a buffer until the two buffers are full of random numbers. A FIFO is employed to hide the latency for random numbers transfer and control the generator core operation. When the FIFO is full of data, the full flag signal is generated and it makes the generator core stop random number generation. When the full flag signal is released, the generator core starts generating random numbers again. The FIFO sends data directly to a buffer in the memory every clock cycle through HT/RT unless the buffer is full of data.
When the initialization function is complete, the random numbers are stored in two buffers so that a microprocessor can read a random number by a generation function call. The initialization function is called only once. Hence, the time initialization, including function call overhead, is negligible to overall performance.
The generation function allows a microprocessor to read random numbers from one buffer while the FPGA fills the other buffer. Once the buffer currently used by the processor is empty, the other buffer becomes active and the FPGA fills the empty buffer. In consequence the latency of random number generation can be hidden.
The free function stops random number generation and releases hardware resources. The free function makes the generator core inactive by sending a signal to the FPGAs and releases the two buffers and the pointer containing HASPRNG information such as the generator type and the buffer size.
The performance of random number generation depends on the buffer size. We trade off the reduced overhead of swapping buffers by enlarging buffer size with fitting the buffer in cache. We optimized HASPRNG buffer size empirically, choosing 1MB as the most appropriate.
2.7. FPGA Estimator Using HASPRNG
We demonstrate an FPGA estimator implementation for the Cray XD1. The implementation of the estimation can be described by the following formula:
Based on the Law of Large Numbers (LLNs), the formula converges to the expected value of as the number of samples increases [21]. Equation (14) describes the relation between the LLN and the MC integration for estimation:
where and represent samples and is the number of sample trials of and . For example, if the estimation consumes two random numbers, one for and one for , then the value of is “1”.
The FPGA estimator implementation employs eight HALCG48 generators. Hence, the FPGA estimator is able to consume 16 random numbers per cycle (8 samples). Figure 7 describes the architecture of the FPGA estimator. Each “A” represents a HALCG48, each “B” represents a 32bit multiplier, each “C” is an accumulator module, and the “D” is logic which produces the complete signal when the required number of iterations is done. A random number from the HASPRNG ranges from “0” to “”. Each pair of random numbers is interpreted as values in the unit square.
The multipliers compute the square of these numbers which are added as follows:
where and are random numbers generated by HALCG48s in the FPGA estimator. If the resultant falls inside the unit circle, the accumulator adds “1” to the count value. When done, the FPGA estimator returns the countvalue and the complete signal. When the complete signal is “1”, the microprocessor reads the count values and computes the value of based on (18):
where = =
3. Results
Each HASPRNG generator occupies a small portion of the FPGA, leaving plenty of room for RC MC candidate applications. Each of the HASPRNG generators is verified with over 1 million random numbers. Through the estimator results we demonstrate that HASPRNG is able to improve the performance significantly for RC MC applications. We use Xilinx 8.1 ISE Place And Route (PAR) tools to get hardware resource usage.
3.1. HASPRNG Hardware Resources
Table 3 shows the hardware resource usage for the XC2VP50 FPGA in each Cray XD1 node and the maximum allowable number of generators on each FPGA
Each HASPRNG LFG consumes an average of about 5% of the slices, 8% of the builtin multipliers, and 16% of the BRAMs. Multipliers are used when SPRNG characteristic equations contain multiplication operations and BRAMs are used when the initial seed data and random numbers are needed to be stored to generator random numbers. In consequence, the LCG type generators (HALCG48/64, HACMRG, and HAPMLCG) do not need BRAMs and the HALFGOO/OE do not need multipliers. HASPRNG performance can be linearly improved by adding extra random number generator copies.
3.2. HASPRNG Performance
Each generator has a different clock rate. Table 4 shows applied clock rates and performance. The difference between actual performance and theoretical performance mainly comes from the data transfer overhead caused by transferring data from FPGAs to the main memory because the bus is saturated.

Table 5 shows performance evaluation for a single generator from HASPRNG compared to SPRNG. The gcc optimization level is set to O2. Because there are eleven parameter sets for the MLFG in SPRNG, we compute the HASPRNG and SPRNG performance in Table 5 using the average speed of each parameter set.

HASPRNG shows 1.5 overall performance improvement for a single HASPRNG hardware generator over SPRNG running on a 2.2 GHz AMD Opteron processor on the Cray XD1. Note that this speedup is limited by the bandwidth between the FPGA and microprocessor. Because SPRNG is designed to support additional random streams, we can easily add HASPRNG generators in the FPGA (up to the maximum copies in the last column of Table 3). This will not help applications on the Opteron processor because the link between the FPGA and Opteron’s main memory is already saturated.
Other reconfigurable computing systems with faster links will obtain higher performance (with a speedup of up to 13.7 as shown in Table 5), or applications can be partially or entirely mapped to the FPGA for additional performance as seen in Section 2.7 with the estimator. Moreover, the Opteron microprocessor is now free to perform other portions of the computational science application.
3.3. HASPRNG Verification
For the verification platforms, each generator in HASPRNG was compared to its SPRNG counterpart to ensure bitequivalent behavior [16, 18]. SPRNG was installed on the Cray XD1 to compare the results from SPRNG with the ones from HASPRNG. We observe that HASPRNG produces identical results with SPRNG for each type of generator and for each parameter set given in SPRNG. We verified over 1 million random numbers on the verification platform for each of these configurations.
3.4. Reconfigurable Computing Estimator
HASPRNG can improve the performance of RC MC applications as shown with the estimator. The estimator runs at 150 MHz and can consume 2.4 billion random numbers per second. The HASRPNG estimator shows 19.1 speedup over the software estimator employing SPRNG when the random samples are sufficiently large for Monte Carlo applications as shown in Table 6. We would expect the significant speedup for such computational science applications. It is worth noting that these results are for eight pairs of points generated per clock cycle. It takes less than 7 minutes to estimate generating 1 trillion random numbers for the estimator on a single FPGA node.

Table 7 shows the numerical results for the absolute errors between the true value of and the estimated based on five different experiments for each sample size. Mean values and the standard deviations are computed assuming that the five sample errors follow a normal distribution [5, 21]. We seek 95% confidence intervals according to different random sample sizes. The error in the estimate of decreases as increases, proportional to [3]. When similar MC integration applications consume an additional 100 times more random samples, they are able to obtain a solution having one more decimal digit of accuracy. Table 8 shows the hardware resource usage for the FPGA estimator.


4. Conclusions
Random number generation for Monte Carlo methods requires high performance, scalability, and good statistical properties. HASPRNG satisfies these requirements by providing a highperformance implementation of random number generators using FPGAs that produce bitequivalent results to those provided by SPRNG. The bandwidth between the processor and FPGA is saturated with random number values, which limits the speedup on the Cray XD1. The reconfigurable computing Monte Carlo estimation application using HASPRNG shows good performance and numerical results with a speedup of 19.1 over the software estimator employing SPRNG. Hence, HASPRNG promises to help computational scientists to accelerate their applications.
Acknowledgments
This work was partially supported by the National Science Foundation, Grant NSF CHE0625598. The authors would like to thank Oak Ridge National Laboratory for access to the Cray XD1. They thank the reviewers for their helpful comments.
References
 T. Warnock, “Randomnumber generators,” Los Alamos Science, vol. 15, pp. 137–141, 1987. View at: Google Scholar
 M. Mascagni and A. Srinivasan, “Algorithm 806: SPRNG: a scalable library for pseudorandom number generation,” ACM Transactions on Mathematical Software, vol. 26, no. 3, pp. 436–461, 2000. View at: Google Scholar
 S. Weinzierl, “Introduction to Monte Carlo methods,” Tech. Rep. NIKHEF00012, NIKHEF Theory Group, Amsterdam, The Netherlands, June 2000. View at: Google Scholar
 N. Metropolis and S. Ulam, “The Monte Carlo method,” Journal of the American Statistica Association, vol. 44, no. 247, pp. 335–341, 1949. View at: Google Scholar
 R. L. Scheaffer, Introduction to Probability and Its Applications, Duxbury Press, Belmont, Calif, USA, 1995.
 M. Mascagni, “Parallel pseudorandom number generation,” SIAM News, vol. 32, no. 5, pp. 221–251, 1999. View at: Google Scholar
 Scalable parallel pseudo random number generators library, June 2007, http://sprng.fsu.edu.
 P. L'Ecuyer and R. Simard, “TestU01: a C library for empirical testing of random number generators,” ACM Transactions on Mathematical Software, vol. 33, no. 4, article 22, 2007. View at: Publisher Site  Google Scholar
 G. Marsaglia, “DIEHARD: a battery of tests of randomness,” 1996, http://stat.fsu.edu/pub/diehard/l. View at: Google Scholar
 J. M. McCollum, J. M. Lancaster, D. W. Bouldin, and G. D. Peterson, “Hardware acceleration of pseudorandom number generation for simulation applications,” in Proceedings of the 35th IEEE Southeastern Symposium on System Theory (SSST '03), 2003. View at: Google Scholar
 K. Underwood, “FPGAs vs. CPUs: trends in peak floatingpoint performance,” in Proceedings of the ACM International Symposium on Field Programmable Gate Arrays (FPGA '04), vol. 12, pp. 171–180, Monterrey, Calif, USA, February 2004. View at: Google Scholar
 P. P. Chu and R. E. Jones, “Design techniques of FPGA based random number generator (extended abstract),” in Proceedings of the Military and Aerospace Applications of Programmable Devices and Technologies Conference (MAPLD '99), 1999. View at: Google Scholar
 D. Buell, T. ElGhazawi, K. Gaj, and V. Kindratenko, “Highperformance reconfigurable computing,” Computer, vol. 40, no. 3, pp. 23–27, 2007. View at: Publisher Site  Google Scholar
 M. C. Smith and G. D. Peterson, “Parallel application performance on shared high performance reconfigurable computing resources,” Performance Evaluation, vol. 60, no. 1–4, pp. 107–125, 2005. View at: Publisher Site  Google Scholar
 Cray Inc., “Cray XD1 FPGA development,” 2005. View at: Google Scholar
 J. Lee, G. D. Peterson, and R. J. Harrison, “Hardware accelerated scalable parallel random number generators,” in Proceedings of the 3rd Annual Reconfigurable Systems Summer Institute, July 2007. View at: Google Scholar
 J. Lee, Y. Bi, G. D. Peterson, R. J. Hinde, and R. J. Harrison, “HASPRNG: hardware accelerated scalable parallel random number generators,” Computer Physics Communications, vol. 180, no. 12, pp. 2574–2581, 2009. View at: Publisher Site  Google Scholar
 J. Lee, G. D. Peterson, R. J. Harrison, and R. J. Hinde, “Hardware accelerated scalable parallel random number generators for Monte Carlo methods,” in Proceedings of the Midwest Symposium on Circuits and Systems, pp. 177–180, August 2008. View at: Publisher Site  Google Scholar
 Y. Bi, A reconfigurable supercomputing library for accelerated parallel laggedFibonacci pseudorandom number generation, M.S. thesis, Computer Engineering, University of Tennessee, December 2006.
 Y. Bi, G. D. Peterson, G. L. Warren, and R. J. Harrison, “Hardware acceleration of parallel lagged Fibonacci pseudo random number generation,” in Proceedings of the International Conference on Engineering of Reconfigurable Systems and Algorithms, June 2006. View at: Google Scholar
 J. Jacob and P. Protter, Probability Essentials, Springer, Berlin, Germany, 2004.
Copyright
Copyright © 2010 JunKyu Lee et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.