Abstract

Universal filtered multicarrier (UFMC) is a low complexity promising waveform that provides quasi-orthogonal property among subcarriers. In addition, it can achieve much better out-of-band emission performance than orthogonal frequency division multiplexing (OFDM) system. Authors have proposed a hardware platform to implement a UFMC transmitter in this paper. Highly reduced complexity schemes for IFFT, filtering, and spectrum shifting are realized on actual hardware. This helps to achieve overall architecture of the transmitter at the cost of minimal FPGA resource usage. Hence, the overall design uses only 1038 slice registers, 1154 slice LUTs, and 64 multipliers of Xilinx Virtex-7 XC7VX330t device. A throughput of 773.5 Msamples/sec at an operational frequency of 364 MHz is achieved. This throughput is adequate for processing 50 Physical Resource Blocks (PRB) of LTE 10 MHz channelization in required time. The presented architecture provides a latency of only 2% of one LTE 10MHz channelization symbol due to the implementation of pipelining at different levels. Although the presented hardware design in its current form meets LTE 10MHz channelization throughput requirements, further increase in throughput is possible due to the scalable nature of the architecture. To the best of our knowledge, this work is first ever FPGA solution for UFMC transmitter presented in the literature.

1. Introduction

ADVENT of 5G mobile telecommunication technology has sparked the start of new era of research in the field of telecommunications. 5G standardization aims to address requirements related to three main communication scenarios: enhanced mobile broadband (eMBB), massive machine type communications (mMTC), and ultrareliable low latency communications (URLLC) [1, 2]. To achieve such challenging performance requirements, numerous waveforms have been proposed [3, 4], which include cyclic prefix-orthogonal frequency division multiplexing (CP-OFDM), filter bank multicarrier (FBMC) [5, 6], generalized frequency division multiplexing (GFDM) [7], universal filtered multicarrier (UFMC) [8], filtered OFDM (FOFDM) [9], windowed OFDM (WOLA-OFDM) [10], and their variants [11, 12]. Among these waveforms, UFMC is a promising choice as per subband filtering which is performed to trade off the out-of-band emission and complexity performance. When comparing with FBMC, UFMC is better in case of short packet length [13] and has only slightly worth out-of-band emission for large packet size [9, 14]. In addition, it has much better compatibility to multiantenna system than FBMC system. In addition, as a subband filtered multicarrier (SFMC) system, UFMC is flexible to support multiservice radio access network (RAN) slicing in physical layer [15, 16]. For example, based on UFMC, [17] proposes a comprehensive framework for multiservice system to support multiple types of services/slices in both generalized and nongeneralized synchronized systems. This aspect addresses the flexibility requirement perspective of 5G.

Apart from performance, hardware complexity is another key factor associated with acceptability of any proposed waveform. In this regard, a reduced complexity UFMC transmitter architecture is proposed in [18] that uses frequency domain approach for generating UFMC waveform. Such simplified idea is also mentioned in [9] to reduce the implementation complexity of the algorithm. The main idea behind is application of filters on each PRB in frequency domain and then converting it back to time domain using IFFT. If classical scheme of UFMC transmitter [8] has complexity of 150 times that of CP-OFDM, the frequency domain UFMC implementation has complexity equals 120 times. Another scheme to reduce UFMC transmitter complexity is presented by Knopp et al. in [19]. This modified approach reduces complexity to around 25 times of CP-OFDM by reducing size of IFFT and then implementing upsampling, filtering, and spectrum shifting. Further reduction of this model is proposed in [20] which will be used as baseline in this paper. However, the complexity reduction claims of above stated UFMC schemes are based on counting the arithmetic operations involved but not on actual hardware implementation. This is due to the unavailability of actual hardware implementation results for UFMC. On the other hand, actual FPGA implementation results are available for FBMC transmitter in the work of Nadal et al. [21] and Robin et al. [22].

In this paper, we have filled this gap of unavailability of actual hardware implementation of a UFMC transmitter. Hence, first real time FPGA implementation of UFMC transmitter complying with the timing requirements of 10MHz channelization of LTE is presented here. To achieve this real time hardware, the following is the summary of contributions:(i)Proposal of reduced complexity implementation solutions for all constituent building blocks of the transmitter while avoiding computation/storage of redundant information(ii)Hardware design of the constituent building blocks while achieving highest possible operational frequency at low area overhead(iii)Resource allocation to constituent blocks and their scheduling in order to meet the established timing requirements for LTE 10MHz channelization(iv)Finally, UFMC FPGA implementation results comparison to OFDM transmitter implementation results.

The rest of the paper is organized as follows: the next section presents the system model. Section 3 is dedicated to architectural choices for our transmitter implementation. Design of transmitter is discussed in Section 4. In Section 5, implementation and performance results are tabulated to compare with the state of the art. Finally, Section 6 concludes the paper.

2. System Model

UFMC is filtered variant of zero padding based OFDM (ZP-OFDM). The symbols obtained from mapper are divided into groups of carriers. These groups of carriers, i.e., PRBs, go through the process of IFFT and filtering after which they are summed up to generate the final UFMC waveform. UFMC model shown in Figure 1 as proposed in [8] is the basic UFMC model. On the other hand, a low complexity UFMC transmitter model proposed in [19] is shown in Figure 2. Both implementations are explained below in detail.

2.1. Classical UFMC Transmitter

As shown in Figure 1, the input to this classical scheme is a set of constellation mapped symbols. These symbols are divided into number of PRBs in LTE standard. Each of these PRBs contain data carriers in LTE. Let be the number of PRBs. Each of these PRBs is zero padded to make a stream of N symbols. Each stream undergoes N-point IFFT which is then passed through Dolph-Chebyshev FIR filter. At the end, all streams are added up to form final UFMC waveform. UFMC waveform is expressed using following equation [20]:where and is the filter length.

In case of transforming an OFDM based waveform into UFMC waveform with the same system overhead in time domain, filter length will be equal to -1 where is the cyclic prefix length. For 10MHz channelization of LTE, there are 600 data carriers which are divided into 50 PRBs and (). Filter length depends on long or short cyclic prefix of LTE and hence value varies from 73 to 80 taps [19].

2.2. Modified UFMC Transmitter

Unlike the classic model, the modified UFMC model proposed in [19] uses smaller IFFT size, i.e., . Moreover, subcarriers of each PRB are placed on same initial positions as that of first PRB subcarriers positions. Each data block is then zero padded with zeros. -IFFT instead of -IFFT is applied [19]. The time domain data is then upsampled by the factor of . The upsampled data from each PRB stream is passed through same low pass filter to remove out-of-band emissions. Each PRB related data is then shifted to its allocated frequency position. Finally data related to all PRBs is added to form final UFMC waveform. This scheme provides complexity reduction through reducing the size of IFFT and avoiding multiplication with zero during filtering of upsampled data. The expression for overall complexity for this scheme of implementation in terms of arithmetic operations is given in [19]. Using this expression, the complexity of UFMC transmitter is almost 23.5 times as compared to CP-OFDM scheme for 10MHz channelization. Hence, as compared to classical UFMC scheme of Figure 1 that has 150 times complexity of CP-OFDM, more than 6 times complexity is reduced.

3. Architectural Choices

We have taken the simplest model of UFMC proposed in [19] as baseline for our work. For scheme in [19], reduced complexity solution for IFFT, filtering, and spectrum shifting are proposed in [20]. These simplifications are discussed briefly below whereas their actual hardware implementation with timing constraints is presented later in this paper.

3.1. IFFT

To obtain low computational complexity in IFFT, Radix-2 algorithm is selected as there are opportunities to avoid redundant computation in case of UFMC implementation through UFMC transmission scheme of Figure 2.

Consider the portion of Radix-2 decimation in time (DIT) with bit reversed input format implementation of 64-point IFFT as show in Figure 3. In first stage, only upper inputs of used butterflies (BF) have nonzero value, i.e., values from 12 subcarriers of a PRB, whereas the second inputs are zero. Hence, the portion of data path shown as bold lines in Figure 3 needs computations. Since the multiplication with twiddle factor is present in lower path, shown as dotted lines, no multiplication is required. Hence, the input value will appear on both outputs without any processing. It is, therefore, the computations in first stage that are not required. Same is the case with the second stage. In third stage, only half of the BFs will be executed as shown in Figure 3. In subsequent stages all BFs will be fully used. Hence, out of 192 BFs only 112 BFs will be computed. As the computation of one Radix-2 BF needs 4 real multiplications (RM) and 6 real additions/subtractions (RA/S), the overall complexity of 112 BF will be 448 RM and 672 RA/S.

In order to exploit this simplification, one needs to copy 12 pieces of data at multiple locations before starting the computation of useful butterflies, e.g., first sample of PRB; i.e., X(0) shall be copied at locations 32, 16, and 48 of input memory of third stage of butterflies as shown in Figure 3. In the same way 12 carriers will be copied at required locations.

To implement 64-point IFFT, Radix-4 architecture is also considered. The partial data flow of 64-point IFFT using Radix-4 implementation is shown in Figure 4. In total there are 48 butterflies, i.e., 3 stages with each stage having 16 butterflies. The first-stage BF computations are redundant as only first input is nonzero which appears on all 4 outputs of each BF (shown as solid line). In the second stage, the last inputs of all 16 BF (shown as dotted lines) are zero; this removes the use of 1 complex multiplication and 4 complex additions in each butterfly of second stage. Finally, in third stage all inputs are fully utilized. Placement of hardware for one complete Radix-4 BF, i.e., hardware comprised of 3 complex multipliers and 12 complex adders/subtractors, will be fully utilized in third stage; however, for second stage BFs computation of one complex multiplier and 4 complex adders will not be utilized. On the other hand, if resources in shape of one complex multiplication and one complex addition/subtraction (as used in Radix-2 BF hardware architecture) are provided to compute Radix-4 BFs, additional hardware for control signal generation and registers to store intermediate results will be required. Moreover, there will a penalty of extra clock cycles while performing read/write operations on intermediate data during Radix-4 BF computation. This will slow down the whole process.

While comparing the utilization of Radix-2 and Radix-4 implementation schemes for simplified IFFT computation, Radix-2 has a clear advantage over Radix-4 scheme in terms of lesser hardware placement, 100 % utilization rate of placed hardware, and simplification of the implementation. In addition, Radix-2 solution encourages the scalable architecture where more Radix-2 BFs can be placed in parallel to achieve higher throughput. Hence, placement of fewer hardware resources, higher utilization rate of placed hardware, less complexity of architecture, and support for scalability are the key elements for selecting Radix-2 implementation in order to exploit the simplified solution of computing a 64-point IFFT operation during UFMC waveform synthesis.

3.2. Filtering

As shown in Figure 2, upsampling by a factor of is performed by insertion of zeros between each output sample of IFFT. However, multiplication of only nonzero samples with filter coefficients is useful. The scenario of upsampling with a factor of 16 is shown in Figure 5. Here, out of sixteen, only one element is nonzero which is multiplied with filtering coefficients. Hence, multiplier for zero element is not required. Moreover, during the computation of different filter outputs, one nonzero value is multiplied with different filtering coefficients . Hence, we can multiplex the filtering coefficients. The circuit diagram shown in Figure 6 implements this idea as given in [23] for poly-phase filter applications.

Once a sample enters the filter and shift of memory elements is performed, then in next 16 cycles the filter coefficients are multiplexed one by one to generate 16 outputs of the filter. Hence, for our case study, on one side only 64 samples from IFFT operations will be required; i.e., no actual zero padding is required and secondly only 5 multipliers, 4 shift registers, 4 adders, and 5 16-to-1 multiplexers will be used in place of 73 to 80 multipliers and 72 to 79 adders (depending on filter taps) and shift registers. Moreover, the same hardware can be used for considered tap lengths of 73 or 80 by changing the values of filter coefficients at the input of multiplexers.

3.3. Spectrum Shifting

In the scheme of [19], the filtered data of each PRB except first one is shifted to its original spectrum. For this, each filtered data element is multiplied with a complex spectrum shifting coefficient which is computed asIn order to elaborate the idea, consider the case where , , is filtered data sample number ranging from 0 to 1095, i.e., total of 1096 elements. Finally, represents PRB number. Hence, for 50 PRBs, 49 sets of filtering coefficients are required. While keeping 16 bits for each real and imaginary component of one filtering coefficient, total memory of Mbits will be required to store spectrum shifting coefficients in Look-Up Table (LUT) for our case study. On the other hand, if these coefficients are computed on the fly, the processing delay will occur. Hence, in [20] a simpler solution is proposed by simplifying the ratio of and in (2) as under

In expression (3), we need 256 discrete values, i.e., 0 to 255 for complete circle, and values are repeated beyond it. Thus, if we make a ROM containing 256 sine and cosine values in the step of , address generated through will provide the sine and cosine value for a particular value of and in (2). Moreover, can be achieved easily by taking 8 least significant bits from the result of . Hence, here we need only Kbits and two multipliers to generate filtering coefficients for 50 PRBs as shown in Figure 7.

There is another possibility of reduction in the size of LUT which is shown in Figure 8. Here we store 128 values related to one-half of cosine wave in a dual port ROM. This ROM has two address lines, i.e., Address 1 and Address 2. Address 1 lines are connected with 7 least significant bits (LSbs) of output of operation whereas the most significant bit (MSb) is connected to a multiplexer MUX-1. The function of MUX-1 is to output the value of . In case the output of operation is within 0-127 the associated cosine value read from Data 1 lines is transferred to the output; otherwise the 2’s complement of read value (shown as multiplication with -1) is sent to output. In case negative of sine wave is required, which leads cosine by an angle of , 64 is added in the output of operation. The 7 LSBs of the result of this addition are connected to Address 2 lines of 128-location ROM. The data output mechanism to generate from MUX-2 is same as described for cosine values.

Using this scheme, the LUT size reduces 4 times as compared to scheme presented in Figure 7 at a cost of few combinational logic elements and provision of second port of the ROM. In our implementation, we have used this scheme for spectrum shifting.

4. UFMC Transmitter Design for LTE 10MHz Channelization

In order to achieve a hardware design for LTE 10MHz channelization, we need to find the required time to process all PRBs. Based on this information we can scale and schedule hardware resources.

4.1. Timing Requirements

The useful OFDM symbol time is 66.7μsec whereas there are two types of CP lengths. The long CP is 5.2μsec and short one has a duration of 4.7μsec. Keeping in view the worst case timing constraint, we consider short symbol time, i.e., 71.3μsec, as maximum time to process 50 PRBs to form a UFMC waveform. In order to scale hardware resources, maximum achievable frequency of a hardware architecture must be known. Hence, we separately modeled data path of two-stage pipelined architecture of IFFT butterfly and 3-stage pipelined architecture for Multiply and Accumulate (MAC) operators of FIR filter (detailed in subsequent sections). We then went through place and route process on Virtex-7 FPGA to find the maximum achievable frequency. It was found that the maximum achievable frequency is 364MHz. Based on the available processing time () and the frequency () at which our hardware can work, following calculation is made to find the number of clock cycles available for UFMC processing:Hence, target is to process single PRB in 519 clock cycles.

4.2. Hardware Architecture of Building Blocks

The UFMC waveform synthesis is performed in 5 processes as shown in Figure 2. In our hardware, to eliminate redundant butterfly computations we have split IFFT process into two processes, namely, data fetching and IFFT computation. The upsampling is not actually performed in hardware due to adopted architecture of filtering. FIR filtering, spectrum shifting, and addition of processed PRBs are the next three processes in our implementation. The details related to hardware for each process are given in subsequent sections.

(1) Process 1: Data Fetching. The purpose of this process is to enable the simplified IFFT process by avoiding redundant butterflies. To achieve this, the first task is to copy the 12 nonzero input data, i.e., data related to one PRB from external memory (containing data of all PRBs) at required places in 64-location memory attached to 64-point IFFT process. The read locations of data related to first PRB from external memory and write location in memory attached to IFFT block are arranged in Table 1.

In order to perform this data transfer, a Finite State Machine with Data path (FSMD) is designed for which (Algorithmic State Machine) ASM Chart is shown in Figure 9. The FSMD has three states, Idle, Processing, and Done. The system remains in Idle state till the time start signal is asserted and system moves to Processing State. Two registers of 4 bits and 6 bits, one to generate the read address (r_adr) for external memory and other to generate write addresses (w_adr) for memory of IFFT block, respectively, are created in this FSMD. These registers start from zero value while entering in Processing State. Since the adopted IFFT process is Radix-2 DIT bit reversed input based, the final addresses for both memories are created by reversing the bits of r_adr and w_adr registers and shall be denoted by bit_reversed_r_adr and bit_reversed_w_adr in following part of this paper. During the Processing State, the value of r_adr goes from 0 to 14 rather than going from 0 to 11. This is because when the value of r_adr is either 3, 7, or 11, the address generated for external memory through bit_reversed_r_adr signal is greater than 11 whereas the required data resides in locations from 0 to 11 for first PRB. Hence, when r_adr has these values, it is incremented by 1 and no write operation is performed; i.e., w_adr is not incremented and write_enable signal remains 0. This is shown in the first decision block of Processing State of Figure 9. The second case is when r_adr has values equal to 2, 6, 10, and 14. For these values of r_adr, 2 LSBs of r_adr will be “10” and corresponding bit_reversed_r_adr values will be 4, 6, 5, and 7. Once a data is read from these locations, it will be copied in 8 locations by, first of all, incrementing w_adr by 1 and then by generating address of IFFT process RAM through bit_reversed_w_adr, i.e., bit reversal of w_adr. For the rest of the r_adr registers values, i.e., 0, 1, 4, 5, 8, 9, 12, and 13 the bit_reversed_r_adr for external memory will be 0, 8, 2, 10, 1, 9, 3, and 11 and each read data from external memory shall be copied in 4 locations of IFFT process memory through bit_reversed_w_adr.

A total of 68 clock cycles are required to complete this process. Hence, 451 clock cycles are left from 519 clock cycles to execute remaining four processes related to one PRB. In order to copy data related to PRBs other than first PRB in IFFT process RAM, the developed mechanism is shown in Figure 10. The overall address to external memory is computed by multiplying the PRB number with 12 and adding the bit reversed address coming from the FSMD of data fetching process.

(2) Butterfly Execution. This process takes the data from IFFT process RAM on which data is arranged by data fetching process. In fact this RAM is a true dual port RAM used to take two simultaneous inputs for a butterfly computation. The other memory attached to this process contains twiddle factors. Process of butterfly execution is shown in Figure 11.

Using this process, the reduced number of butterflies, i.e., 112 for and , is computed to take 64-point IFFT of one PRB. As stated earlier, the butterfly hardware is a three-stage pipelined architecture. Hence, three clock cycles are required to perform one butterfly computation; i.e., in first clock cycle address for two inputs is generated. In second cycle, data read and execution of first stage of butterfly are performed whereas in third clock cycle final result of butterfly is computed and written at same location of dual port RAM from where data was fetched. As butterflies’ computation of first two stages is redundant (Figure 3), the computations will start from 3rd stage. Half of the butterflies are computed in 3rd stage and all butterflies in the next 3 stages. To achieve this, two counters are used for reading the data form RAM (Figure 11). Counter 1 starts from ‘000000’ and counter 2 starts from ‘000001’ binary values. During execution of 3rd stage of butterflies, the address for 2 inputs of butterfly computation will be generated in a way that four LSBs of both counters will be reversed and 2 MSBs will not be changed. Once butterfly results are generated, the output of butterflies is copied on the same location from where input data is fetched. Both counters are incremented by 4 to compute addresses for data related to next butterflies. The process will be repeated for those half butterflies which have both nonzero inputs. In the rest of three stages, all butterflies are computed. Hence, in each stage the two counters start from ‘000000’ and ‘000001’, respectively; however, after each butterfly execution the increment of 2 is performed in each counter. The final addresses are generated such that, out of 6 bits, during 4th stage butterflies’ execution only 3 LSBs are reversed, in 5th stage only 2 LSBs are reversed, and in last stage no bit is reversed. In this way required butterflies are computed. The whole process takes 336 clock cycles. Hence, out of 519 clock cycles, the first two processes, i.e., data fetching and IFFT, take clock cycles. This means that both processes can run serially without violating the timing requirements (i.e., 519 clock cycles) due to the avoidance of redundant butterfly computations (240 more clock cycles would be required). However, for other processes, a process level pipelining will be required which is discussed in section describing system level architecture.

(3) Filtering, Frequency Shifting, and Final Addition. The 64 outputs of the IFFT process are given to the filtering process. In order to fulfill the timing requirements, an architecture shown in Figure 12 is used. Here, a parallel architecture generating 4 outputs at the same time is employed. Once a sample enters into the filtering hardware architecture, each branch generates 4 outputs. Due to pipelined architecture a total of 8 cycles are required to generate 16 outputs from one input (4 outputs from each branch). Hence, in order to consume 64 inputs of filtering, 512 clock cycles are required. Keeping in view the clock cycle available, the process is completed within the given time.

Spectrum shifting is achieved by using the LUT architecture explained in Section 3.3. As soon as 4 samples of filtered data come out of the filtering process, they are multiplied with respective spectrum shifting coefficient stored in 4 dual port ROMs (EXP Memory shown in Figure 13).

Before storing the data in the memory, previous data is read first. Then spectrum shifted data is added to data read and saved in the memory. In total, data level pipelining is established in three processes of filtering, spectrum shifting, and final addition of all processed PRBs. The whole process is shown in Figure 13. The final results are saved in 4 memory blocks due to the 4 concurrent outputs from filtering and consequently from spectrum shifting hardware. Due to these pipelined processes a total of 514 clock cycles are consumed which are within the timing constraint established in terms of number of clock cycles.

4.3. System Level Architecture of UFMC Transmitter

In overall design, serial execution of group of first two processes and pipelined execution of group of last three processes individually meet timing requirements. Hence, by executing both groups of processes in a pipelined fashion, overall system level timing requirements can be met.

To materialize the pipelining of two groups of processes, the concept of ping-pong buffering/memory [24, 25] is used. Hence, two memories ‘Memory 1’ and ‘Memory 2’ are used in place of a single RAM shown in Figures 10, 11, and 13. The memory accesses on these memories by different processes is shown in Figure 14. At the start, process 1 writes the constellation symbols from external memory on Memory 1 and soon after this is completed process 2 of IFFT starts. Process 2 reads from Memory 1 and writes butterfly computation results back to same locations of Memory 1. Once IFFT process is finished, process 3 fetches data from Memory 1, performs filtering, and last two processes operate on 4 sets of output memories. While process 3 accesses Memory 1 for filtering, process 1 and process 2 use Memory 2 and it goes on until all PRBs are executed. This efficient and noncontemporary usage of two memories speeds up the whole process.

5. Implementation and Performance Results

5.1. Implementation Results

In order to obtain implementation results for FPGA, the whole project is created on Xilinx ISE design suite. Virtex 7 XC7V2000t is selected as an implementation platform. The synthesis results of different building blocks are summarized in Table 2.

It can be seen that a very few resources are required to enable simplifications in IFFT operation using processes 1 and 2. A total of 27 slice registers are required to implement address register and state register of process 1. Similarly, 28 LUTs are required to complete FSMD of process 1. Process 2 of 64-point IFFT operation requires few slice registers along with 233 slice LUTs and 4 multipliers. These FPGA resources are used to generate one complex multiplier, one complex adder, and one complex subtractor to implement one Radix-2 DIT butterfly structure along with address generator and control logic required to execute this butterfly 112 times. Two BRAMs are required to work as ping-pong memories whereas one BRAM is used to create the LUT for twiddle factors. As far as process 3 is concerned, 40 DSP48 slices are used in filtering part as we have to multiply a real valued filtering coefficient with a complex valued output of IFFT operation, i.e., two real multipliers are required in one multiplier of the filter. Hence, 40 real multipliers are used to create 20 multipliers. While looking at process 4 and 5, 20 multipliers are used for spectrum shifting. Out of these 20 multipliers, 16 multipliers are used in 4 complex multiplications, i.e., multiplication of 4 simultaneous complex valued outputs of filter with 4 complex valued spectrum shifting coefficients. Finally, the last 4 multipliers are used in the address generation mechanism of spectrum shifting coefficient LUT, i.e., multiplication of PRB number and sample number of Figure 4. Multiplication of 3 with is achieved by shift and add operation to avoid 4 multipliers. A total of 136 slice registers and 152 slice LUTs are used to create 4 dual port ROMs to generate filtering coefficients. As far as clock frequency is concerned, use of 364 MHz, which is smallest among highest achievable frequencies of all processes, fulfills the timing requirements.

5.2. Performance Results

In presented architecture of UFMC transmitter, serial execution of processes (processes 1 and 2), data sample level pipelining (3,4, and 5), and finally process group level pipelining (1 and 2 with 3,4, and 5) are used. Processes 1 and 2 jointly take 404 clock cycles and processes 3, 4, and 5 take 516 clock cycles in total. Taking 516 clock cycles as critical path to process single PRB and post place and route frequency of 364 MHz, the processing delay (latency), i.e., time from start of data input of 1 PRB till generation of UFMC waveform for 1 PRB, is (1032 clock cycles) 2.835 due to two stage pipelined process. However, in the pipelined execution, the architecture outputs one processed PRB per 516 clock cycles or 1.417 . This meets the timing requirement as established in Section 4.1. With 1096 output samples per PRB, the throughput of proposed architecture is 773.5 MSamples/sec at a post-place and route frequency of 364 MHz. Taking 50 PRBs which make one useful symbol of 10MHz channelization of LTE, the processing delay is only 1.417 or 516 clock cycles in addition to LTE symbol time of 71.3 , i.e., around 2% of LTE symbol time. In order to achieve higher throughput, multiple copies of presented hardware design can be used. In this scenario, smaller number of PRBs can be assigned to each hardware unit to achieve gain in throughput.

In order to assess out-of-band emission results achieved through FPGA a prototype system is realized. In this system, the input test vectors were generated from fixed point reference model of UFMC transmitter. These test vectors are saved in FPGA memories which are accessed by the presented UFMC transmitter. The output of proposed transmitter was then saved in FPGA internal block RAMs. The output results on block RAMs were then read through a PC (using serial link). Finally the frequency response of FPGA output was plotted along with results achieved through software based golden reference model as shown in Figure 15. The plot in blue color belongs to floating point golden reference model whereas red colored dotted plot shows the results achieved from FPGA. Due to minor difference between both plots, they are overlaid on each other. The slight difference in results of floating point and FPGA output can be observed on the edges. Figure 16 shows the zoomed portion for normalized frequency range from 0.6 to 0.7 (right edge). It can be seen that the output from FPGA has the negligible difference when comparing the out-of-band emission results attained through golden reference model.

5.3. Results Comparison

Complexity analyses based on number of computations are presented in [18, 19], but actual hardware implementation has not been discussed. In this paper, we have presented the simplest UFMC transmitter hardware implementation till date using highly reduced complexity constituent building blocks. As far as comparison with hardware solution of other waveforms is concerned, we can compare it with OFDM to assess incremental cost associated with UFMC.

While comparing OFDM and UFMC, the IFFT complexity for 10MHz channelization of OFDM has 5120 butterfly executions (single 1024-point IFFT) while in UFMC we have computed 5600 butterflies for 50 PRBs (112 actual butterflies in 64-point IFFT per PRB). Hence, clock cycles required to perform 1024 points through our hardware will take clock cycles which are less than clock cycles available for one OFDM symbol. Hence, through our proposed simplifications, the IFFT hardware for UFMC and OFDM can be considered equal. So the hardware for filtering, spectrum shifting, and final addition is additional cost for UFMC. Hence, the complexity of UFMC in case of 10 MHz channelization is 31 times higher for slice registers, 5 times higher for slice LUTs, and 16 times higher for DSP48 slices as compared to OFDM in order to have required out-of-band emissions.

Keeping in view the different parameters, i.e., different number of carriers selected for FBMC transmitter in the work presented through [21, 22], a fair comparison cannot be carried out. However, for the sake of completion of information the area overhead, throughput, and latency figures of work in [21, 22] and our work are presented in Table 3.

6. Conclusion

In this paper we have targeted real time FPGA implementation of UFMC transmitter in order to compare it with other popular waveforms. In this regard, first of all a hardware architecture of UFMC is proposed while taking LTE 10MHz channelization as a case study. By selecting efficient methods of implementing constituent blocks of a UFMC transmitter, we have achieved a solution which consumes very fewer FPGA resources. In our design, through the use of data and process level pipelining, we have achieved high operational frequency. This elevated frequency helped in increasing the overall throughput; hence, timing requirements of 10MHz channelization are met with fewer FPGA resources. To the best of our knowledge, the work presented in this paper is first ever effort in presenting dedicated hardware solution for UFMC transmitter and comparing it with OFDM and FBMC on actual FPGA implementation bases.

Data Availability

The authors have reached the conclusion by modeling the whole system in VHDL and then synthesizing and implementing the design using ISE tool from Xilinx. These results are presented in the article. The VHDL models cannot be shared as they are having commercial value for the funding agencies.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Acknowledgments

This research was partly funded by EPSRC Global Challenges Research Fund, the DARE Project EP/P028764/1. This work is funded by both Bahira University, Islamabad, Pakistan, and University of Glasgow, Scotland, UK.