Abstract
For the conventional extended minsum (EMS) algorithm, all check nodes update their checktovariable (C2V) messages in every iteration. Selected scheduling, which reduces the number of check nodes for message updating in one iteration, can effectively reduce the complexity of the decoding algorithm, but it also lead to some performance degradation. With the introduction of a metric based on node stability, we propose stabilitybased nodesubset scheduling (SNS) for the EMS algorithm, which can improve the performance of nodesubset scheduling (NS). Second, to further improve the decoding performance of SNS while maintaining low complexity, we propose the SNSEMS algorithm with a subsetreset mechanism (RSNSEMS) based on the abnormal stability found in the processing node subset, which will cause the estimated codeword to fail to converge. The RSNSEMS algorithm enhances performance through a sliding window detection and reset mechanism, and it resets the elements in the processing node subset to force all check nodes to update new messages when abnormal stability is detected. The simulation results show that the proposed algorithm can reduce complexity by approximately 25% with negligible performance degradation.
1. Introduction
Lowdensity paritycheck (LDPC) codes [1] have been shown to achieve performance close to the Shannon limit when using the iterative beliefpropagation (BP) algorithm for very long code lengths [2, 3]. LDPC codes have attracted much attention, and novel types of LDPC codes are constantly being proposed, such as rootphotograph LDPC codes [4], which were proposed for blockfading channels and were also used to design bitinterleaved coded modulation (BICM) with iterative demapping and decoding under the constraint of limited bandwidth for wireless communication [5]. Moreover, for medium and short code lengths, nonbinary LDPC (NBLDPC) codes have shown better decoding performance than their binary counterparts because they are able to avoid error floor problems [6]. NBLDPC codes were originally proposed by Davey and Mackay [7], who considered codes defined in the finite field and correspondingly proposed an implementation of the BP algorithm for these codes, usually referred to as the Qary sumproduct algorithm (QSPA). The computational complexity of the direct implementation of the QSPA is too high, which makes NBLDPC codes difficult to apply in practice. To reduce the complexity of the decoding algorithm for NBLDPC codes, the extended minsum (EMS) algorithm [8] and its variant version [9] have been proposed. In these algorithms, the calculation amount of check nodes is reduced by using the truncation rule, which can reduce the length of the message vector from field size to a specific number less than .
In addition, for an iterative algorithm like the BP algorithm, its convergence rate is crucial for its implementation. Recently, different scheduling strategies have been proposed for BP decoders of binary LDPC codes. The conventional BP algorithm has been implemented with a standard flooding scheduling strategy [1], in which all variable nodes and, subsequently, all check nodes update the message and propagate new messages to their neighbors in one iteration. Shuffled [10–12] and layered [13–15] scheduling are two commonly used sequential scheduling strategies to accelerate the convergence speed of flooding scheduling.
However, the order of message updating is fixed in these algorithms, which limits the acceleration of convergence. To further speed up convergence and improve the performance of LDPC codes, many informed dynamic strategies (IDSs) have been proposed. Residual beliefpropagation decoding [16] is the represented IDS; it prioritizes message updates dynamically based on a metric called a residual. Although the RBP algorithm can achieve faster convergence speed, it is shown to be a greedy algorithm in which some edges and nodes will not be updated throughout the entire iteration process. Such greediness will prevent the RBP algorithm from converging to a low error rate. To solve this problem, different algorithms have recently been proposed, such as variable nodebased dynamic scheduling [17], dynamic scheduling based on tabu search [18], and residualdecayingbased residual belief propagation [19].
Another subset of scheduling strategies is aimed at reducing the iteration cost of the LDPC decoder. The lazy scheduling strategy [20] was proposed to decide whether a check node should be updated at a given iteration based on its reliability and its update history. The selectiveupdate decoding algorithm [21] uses the stability of variable nodes, in which the outgoing messages are similar or not similar for the last several iterations, to determine which check nodes do not need to be updated. When all variable nodes connected to a given check node are stable and reliable, it is not useful to update them again. Check node reliabilitybased scheduling [15] led to the creation of a concept called check node reliability, which is associated with the a posteriori probability of the existence of neighboring variable nodes, to determine whether the check node should be updated in the next iteration. Check node reliabilitybased scheduling also employed a random variable to overcome the high false reliability of the check nodes. Nodesubset scheduling for the EMS (NSEMS) decoding algorithm [22] uses the largest a posteriori probability among all tentatively estimated symbols as the metric to determine whether a variable node is reliable. Then, the check node is divided into two subsets—the processing subset and nonprocessing subset—according to the number of unreliable variable nodes connected to the check nodes. This method can effectively reduce computational complexity.
In conclusion, among these scheduling strategies, the most important issue that has a significant impact on the performance and computational complexity of the decoding algorithm is to decide which check nodes should be updated. Therefore, we have carried out research on this issue, and the main contributions of this paper are twofold. First, a scheme called stabilitybased nodesubset scheduling (SNS) based on nodesubset (NS) scheduling for the EMS algorithm is proposed to improve the errorcorrecting performance of the NSEMS algorithm. In this algorithm, a metric based on node stability is introduced to determine whether a variable node is reliable, and the threshold is heuristically defined to match the iteration and communication channels. Second, to further improve performance while still maintaining low computational complexity, we studied the behavior of the processing subset of error frames in the SNSEMS algorithm and found abnormal stability in the processing subset of error frames, which would lead to the stagnation of the information transfer between check nodes and variable nodes. When the information transfer between these two nodes gradually stagnates, iterative decoding will no longer play its own role; thus, the errors in the estimated codeword will not be able to be corrected. Then, according to this property of error frames, stabilitybased nodesubset scheduling with a subsetreset mechanism (RSNS) is proposed. The kernel of the RSNSEMS algorithm is a form of sliding window detection for the processing subset. When it detects that there is such abnormal stability in the decoding process, the elements in the processing subset are reset so that all check nodes will update messages in the next iteration. Hence, abnormal stability will be alleviated, and performance will be improved. The simulation results demonstrate that the proposed algorithm can achieve better decoding performance at lower computational complexity compared to its counterpart.
The rest of this paper is organized as follows. Section 2 gives a brief introduction of the extended minsum (EMS) algorithm and nodesubset scheduling. Section 3 introduces the proposed algorithms. The performance evaluation of various algorithms is demonstrated and discussed in Section 4. Finally, Section 5 concludes the paper.
2. Preliminary Works
2.1. Normal Graph of Nonbinary LDPC Codes
Let be the finite field with elements. A nonbinary LDPC code over can be defined as the zero space of its parity check matrix with . Message vector is a legal codeword of this NBLDPC code if and only if it satisfies . We define the two following index sets:for each row of , andfor each column of .
For a given parity check matrix, we can use the normal graph shown in Figure 1 to describe the decoding progress of NBLDPC nodes. In a normal graph, edges represent variables, and vertices represent constraints. There are three kinds of vertices in a normal graph: (1) the left vertex is called the variable node, which represents a column of ; (2) the right vertex is called the check node, which represents a row of ; and (3) the middle vertex is called the intermediate node, which represents the nonzero elements in , which bridge the check node and variable node; moreover, (4) all edges connected to the variable node must take identical values, and all edges connected to the check node must add up to zero.
2.2. MEMS Algorithm
In this subsection, we will describe the specific steps of the MEMS algorithm [5] combined with a normal graph, including initialization, message truncation, and the transfer and processing of information among variable nodes, check nodes and intermediate nodes.
2.2.1. Initialization
Consider an NBLDPC code, , defined over , and is a codeword, where any of the codeword symbols can be represented by a bit binary vector. For BPSK modulation, the codeword can be mapped to a bipolar sequence , where for . Then, the sequence received over the AWGN channel is as follows:
For a given received sequence , the channel initialization vector can be calculated in the following manner. First, we calculate the real logdomain message as follows:where represents the bit of a binary vector of finite field symbol .
Let and be the two parameters to be designed. Using these two parameters, we can clip and quantize the real logdomain message to an integer message according to the following equation:where and denotes a rounding operation, that is, taking the nearest integer to it.
2.2.2. Variable Node Update
In the iterative decoding process, variable node receives message passed by check node connected thereto and updates the a posteriori message according to
The variabletocheck (V2C) message that needs to be passed to the check nodes is calculated as follows:
2.2.3. Check Node Update
We can use the forwardbackward algorithm [23] based on the trellis diagram [9] to represent message processing at check nodes.
Let and denote the forward and backward recursion vectors, respectively.
Set , and let denote the degree of the check node. Then, for and , compute the forward recursion vector and its truncated version:
Set , and then, for and , recursively compute the backward recursion vectors and its truncated version:
Next, for and , compute the extrinsic message:
Finally, for , postprocess the extrinsic message:where is a scaling factor and denotes a rounding operation.
2.2.4. NodeSubset Scheduling
Let be the set of estimated symbols of variable nodes which is connected to the check node at the iteration. The syndrome of this check node can be computed by
For a check node, if its syndrome is nonzero, then it means that the tentatively estimated symbols of variable nodes connected to the check node has not converged to the correct symbols. Therefore, these check nodes need to be updated in the next iteration. When the syndrome of a check node is zero, it does not always mean that the variable nodes connected to it have converged, it may be caused by the existence of two or more error symbols.
To distinguish which kind of situation occurs when the syndrome is zero, nodesubset (NS) scheduling [22] is based on the reliability of the variable node. If the reliability of the variable node is large enough, then the current decoding symbol is considered reliable. When the number of unreliable variable nodes connected to a check node with zero syndrome is less than two, the check node will be divided into a nonprocessing subset. These check nodes in the nonprocessing subset will not update their C2V messages in the next iteration.
NS scheduling can be described as follows. Let denote the mark of the variable node at the iteration. If the reliability of the current estimated symbol exceeds predefined threshold , then this variable node is considered reliable and marked as 0; otherwise, it is marked as 1:where .
Let be the set of all check nodes and let be the processing subset during the decoding iteration. The check nodes contained in can be determined by the following equation:where and denotes the syndrome of the check node at the iteration.
3. Proposed Schemes
3.1. Metric for the Stability of Variable Nodes
In this subsection, we propose a metric based on the stability of the variable node to determine whether this node is reliable. First, we define stability metric as follows:where denotes the secondlargest value of the logdomain message of .
In the MEMS algorithm [9], the value distribution of is affected by decoding iteration and the channel environment. The current decoding symbol of a variable node will often change between the first few finite field symbols with the highest reliability value. In other words, in the of a variable node, there will be several symbols with similar reliability values.
As shown in Figure 2, these two variable nodes have similar values of. However, there are several symbols with close values of in variable node (b), so the estimated symbol of this node is more likely to change compared to variable node (a). However, when using as the metric of reliability, both variable nodes should be considered reliable.
(a)
(b)
With the progress of decoding iteration, the variable nodes gradually converge to the correct symbol and no longer change, which means that they become stable. According to this property, when a variable node is not stable, then we consider it unreliable and in need of being updated in the next iteration.
Therefore, we use as the metric of the reliability of a variable node. For a variable node, the larger its value is, the greater the difference in between the estimated decoding symbol and the second reliable symbol is; thus, it is less likely for the estimated symbol to change, which means that the variable node will become more stable.
The method proposed in this paper for determining which variable nodes are stable is described as follows. Let represent the mark of the variable node at the decoding iteration. If the value of this variable node exceeds a designed threshold, , then it is considered sufficiently reliable and marked as 0. Otherwise, the variable node is considered unreliable and marked as 1.
For threshold , we present a method that can adaptively calculate the threshold based on decoding iteration and channel variation:where is a revision factor to be designed.
depends on the value of all variable nodes in one iteration, so it does not need to be redetermined due to the variation in the parity check matrix. Moreover, as the iterative decoding process proceeds, an increasing number of variable nodes will converge to the correct symbol and become stable. Therefore, we use the statistic mean of for all variable nodes to determine this threshold, which can reflect the average stability level of the variable nodes in this iteration. The variable nodes with values under threshold will be considered unreliable. The improved stabilitybased nodesubset schedule for the EMS algorithm (SNSEMS) can be described by pseudocode, as shown in Algorithm 1.

3.2. SNS Scheduling Combined with a SubsetReset Mechanism
In the previous section, we described the proposed SNSEMS algorithm. According to the simulation results, the SNSEMS algorithm can make a good tradeoff between decoding performance and computational complexity. However, compared with the original MEMS algorithm, its error correcting performance has a certain degree of degradation.
Since SNS scheduling mainly affects which check node will update its C2V message in iterative decoding, we take the processing subset as the point of penetration and study the variation in the processing subset in the iterative decoding process of the SNSEMS algorithm. Note that, in all experiments in this subsection, the NBLDPC code of rate 0.5 is used, and the maximum number of iterations is set to 50.
In the first experiment, we track the variation in the number of processing nodes for both converged frames and error frames at SNRs of 1.0 dB, 1.4 dB, 1.6 dB, 1.8 dB, 2.2 dB, and 2.4 dB, and their collected error frames total 50, 30, 30, 20, 10, and 10, respectively. Note that when a frame converges, we set the number of processing nodes in the subsequent iterations to 0.
Figure 3 shows the average number (over frames) of processing nodes in the iterative process of converged frames and error frames of the SNSEMS algorithm under different SNRs. We can see that there is a clear difference in the variation curves between converged frames and error frames. The number of processing nodes of converged frames decreases greatly and always shows a downward trend until dropping to a very low level, which means that this frame has converged. Moreover, as the SNR continues to increase, the convergence speed also increases. When the SNR is greater than 1.8 dB, the number of processing nodes can be reduced to a very low level after the first several iterations.
Conversely, the number of processing nodes in error frames exhibits a gentle downward trend. The number of processing nodes will no longer continue to decline after approximately 10 iterations but will remain at a high value. At this time, the number of processing nodes shows a stable state, where only a small fluctuation occurs.
Based on the above experiment, we further studied the changes in elements in the processing subset of error frames. Let and denote the processing subset of the current iteration and that of the next iteration, respectively, and let denote the intersection of and , which contains the same elements in these two sets. Consequently, the numbers of elements in and are and , respectively. We use the ratio of to as the metric, which is referred to as repeatability ratio , to intuitively reflect the changes in the elements in the processing node subset. When is close to 1, it means the elements in the processing subset are almost unchanged in two successive iterations.
In this experiment, we track the variation in the repeatability ratio at SNRs of 1.0 dB, 1.4 dB, 1.6 dB, 1.8 dB, 2.2 dB, and 2.4 dB, and their collected error frames total 50, 30, 30, 20, 10, and 10, respectively.
As shown in Figure 4, all the values of the processing subset of error frames under different SNRs are at a very high level, and as the iteration continues, the ratio gradually increases. After 20 iterations, under all SNRs is higher than 0.98, which means that the elements in the processing subset will hardly change in the subsequent iterations. These results combined with those in Figure 3 show that when the number of processing nodes of the error frames tends to be stable, almost the same check nodes update their C2V messages. This situation continues until the decoder reaches the maximum number of iterations and stops decoding, but the estimated code has not converged.
According to the above experiments, we speculate that, after dividing the node subset, some check nodes stop updating their C2V messages. Then, some of the V2C messages of variable nodes cannot be updated correspondingly, according to equations (6) and (7). As a result, will no longer be updated, which is used to tentatively decide the estimated symbols for variable nodes. Therefore, even though there are many check nodes for C2V message updating in the processing node subset, the errors in the estimated codeword still cannot be corrected.
To verify the correctness of the above speculations, we conduct research and analysis on the vector of variable nodes during the iteration process. Since the calculation of these vectors is related to both C2V and V2C messages, the changes in this message can reflect the changes in these two messages. We start by defining the vector as follows:where and are the variable node and the iteration, respectively, and is the absolute value. The elements in denote the magnitude of logdomain message for each finite field symbol between two consecutive iterations.
Then, for each variable node, a metric, , is stored to trace how many finitefield symbols in vary more than variation threshold . can approximately reflect the change in vector in the iteration process: if the value of is larger, then there are a greater number of symbols whose logdomain messages have changed in .
We take the arithmetic mean of the values of all the variable nodes in the iteration as and use it to show the overall situation of for all variable nodes as decoding iterations progress. It is worth noting that there are two situations where will not change: (1) when the estimated codeword converges and the decoder stops decoding, the calculation of will stop accordingly, and (2) when C2V messages stop being updated, the value will not change. To distinguish these two cases, we set the values of converged frames to after they have converged, and denotes the order of finite field .
The experimental results are shown in Figure 5, where . It is worth noting that the value of error frames gradually decreases as the number of iterations increases, especially after 20 iterations, and the value of remains at a very low level under all SNRs. This finding means that, during these iterations, hardly changes. Combined with the above experiments, we know that after “abnormal” stability appears in the processing subset, although there are still a large number of check nodes updating their C2V messages in the subsequent iterative decoding, it does not help with error correction. Since the information transfer between check nodes and variable nodes no longer generates new correction information, , which is used for deciding the estimated symbol of variable nodes, hardly changes. Therefore, the errors in the estimated codeword cannot be corrected, and thus, it will fail to converge to the correct codeword.
The proposed stabilitybased nodesubset scheduling with a subsetreset mechanism for EMS (RSNSEMS) algorithm has been developed based on this “abnormal” stability of the processing subset. The main idea of this method is to distinguish the frames with this property during the decoding process. Then, the elements in the processing subset are reset so that all check nodes will update their C2V messages in the next iteration. When new C2V messages are updated, it is expected that the abnormal stability of the processing subset will be alleviated and that the message transfer between the two nodes will be accelerated. Therefore, the estimated codeword can converge, thereby mitigating the degradation of the errorcorrecting performance of the SNSEMS algorithm. Note that since only a small number of error frames will be reset and the FER will gradually decrease as the number of SNRs increases, we can expect that this method will not significantly increase overall computational complexity.
To implement the previously described method, we design a sliding window detection mechanism. First, because in the first few iterations the number of processing nodes for both error frames and converged frames exhibits a downward trend, we set a threshold of iteration , and the sliding window detection mechanism will only be activated when the number of iterations is greater than . To trace the abnormal stability behavior of the decoding frame, a sliding window mechanism of width can be defined as follows: let denote the number of processing nodes in each iteration, where . We collect for the last iterations and define vector as follows:where denotes the number of processing nodes for the current iteration. The elements in represent the difference in between two consecutive iterations.
Finally, we calculate the arithmetic mean of the value in , which denotes the average value of the difference in the number of processing check nodes within successive iterations:
We use it as the measurement to determine whether the processing subset has abnormal stability. When the value of is less than the designed variation tolerance , the processing subset of this frame has abnormal stability. Then, the elements of the processing subset are reset, and all check nodes update their C2V messages in the next iteration. It should be noted that the interval within the resetting of these two subsets needs to be greater than window width . The proposed RSNSEMS algorithm can be described by pseudocode, as shown in Algorithm 2.

4. Performance Evaluation
In this section, we will evaluate the performances of the proposed SNSEMS and RSNSEMS algorithms in terms of computational complexity and errorcorrecting performance through a large number of simulations and compare them with other algorithms, including the QSPA [7], MEMS [7], SUEMS [21], and NSEMS [22] algorithms. In this evaluation, to verify the reliability of the proposed method, we select three NBLDPC codes with different finite fields and construction methods.
The first code, , is a regular NBLDPC code over with a rate of 0.5, which has been designed using optimized rows and optimized short loops according to [24, 25]. The second code, , is a regular NBLDPC code over and it has been constructed by using the progressedgegrow (PEG) algorithm [26]. The third code, , is a regular NBLDPC code over with a rate of 0.5, which has been constructed using the same method as the first code. All the simulations are performed at BPSK modulation over the AWGN channel. The maximum number of iterations is set to 30.
4.1. Complexity Analysis
In this subsection, we will analyze the complexity of the proposed algorithms. The computational complexity of the decoding algorithm of NBLDPC codes is mainly concentrated on the part of check node update. The check node update can use forwardbackward algorithms [23] on the trellis, so computational complexity is determined by the total number of live branches in the trellis. Thus, we will count the total number of live branches of different algorithms on the trellis. Conversely, different decoding algorithms may have different convergence speeds. Therefore, we will find the statistical mean (averaging over frames) of the total number of live branches involved in all the iterations for the decoding of one frame. Because other algorithms are used to reduce the computational complexity of the QSPA, we take the following complexity ratio as the measurement of complexity:which can approximately reflect the reduction of complexity compared to the QSPA.
In addition, the abovementioned complexity ratio reflects the complexity of each algorithm approximately but intuitively. We think that it is necessary to compare the computational complexity of the MEMS algorithm and the QSPA more accurately, since all selected scheduling strategies are used in the MEMS algorithm. The computation of messages in check node updates can use the method proposed in [27]. According to this paper, the complexity of check node updates in one iteration varies on the order of and with , which is a great computational reduction compared to the QSPA of .
Regarding the introduction of various mechanisms for reducing complexity in various selected scheduling strategies, some extra operations have also been introduced. Thus, we will also analyze these newly introduced operations. According to the relevant steps of these algorithms, we divide the operations into five categories: finitefield addition (FA), real addition (RA), real multiplication (RM), comparison (CM), and max operation (Max).
The extra operations introduced by different algorithms compared to the MEMS algorithm in one iteration are shown in Table 1. For the NSEMS algorithm, the newly introduced operation determines whether the variable node is reliable and divides the node subsets. These two steps require RAs, CMs, and Maxs. Compared to the NSEMS algorithm, the SNSEMS algorithm additionally needs to compute threshold and find the estimated symbol with the secondlargest logdomain message in . Thus, the extra operations required by the SNSEMS algorithm are RAs, 2 RMs, CMs, and Maxs. Due to the introduction of abnormal stability detection in the RSNSEMS algorithm, the required extra operations are RAs, 3 RMs, CMs, and Maxs. The SUEMS algorithm needs to find the difference in each finitefield symbol in of two consecutive iterations and determine whether it is greater than threshold . Therefore, the extra operations introduced by the SUEMS algorithm are RAs and CMs.
Since these scheduling strategies can reduce the number of check nodes updated in one iteration, we also list the reduction in complexity in Table 1. Because in each iteration of the QSPA and the MEMS algorithm the number of processing check nodes is a fixed value , and the calculation for the C2V messages of each iteration requires FAs, Ras, and Maxs in forwardbackward recursion. All algorithms listed in Table 1 reduce the complexity by reducing the number of processing check nodes in one iteration. Therefore, in each iteration, the reduced operations are the corresponding operations of these nonprocessing check nodes. Let denote the number of nonprocessing check nodes in one iteration for various algorithms. Then, the reduction in complexity is FAs, RAs, and Maxs.
We can notice that the additional complexity is linear and that the reduced complexity is proportional to the square of finitefield order . With regard to the case of short and medium code lengths, where NBLDPC codes have more advantages compared to binary LDPC codes, the additional operations caused by the introduction of a new mechanism are much smaller than the reduced computational complexity.
4.2. Simulation Results
Errorcorrecting performance is measured by the bit error rate (BER) and frame error rate (FER). First, we consider a regular NBLDPC code that has been designed using optimized rows and optimized short loops according to [24, 25]. The paritycheck matrix of this code has row weight and column weight . The parameters of the implemented algorithms are listed below.
(1) For the MEMS algorithm, ; for the NSEMS algorithm, ; for the SUEMS algorithm,; for the SNSEMS algorithm, ; and for the RSNSEMS algorithm, . (2) For all algorithms, the scaling factor and .
The simulation results are shown in Figures 6 and 7. Figure 6 shows the BER and FER performance of various algorithms, where the dashed lines represent the FER, and the solid lines represent the BER. Figure 7 shows the complexity ratio of various algorithms compared to the QSPA, and the subwindow shows the complexity ratio compared to the MEMS algorithm.
The optimal QSPA has the best errorcorrecting performance, and the EMS algorithm has the secondbest performance. We can see that the computational complexity of the other algorithms is much lower than that of the QSPA, almost reducing the complexity by more than 70%. Due to the introduction of the truncation rule, the MEMS algorithm reduces complexity and brings about some performance degradation. For the proposed algorithms, we can see that there is negligible (nearly no) performance degradation between the proposed RSNSEMS algorithm and the original MEMS algorithm. Moreover, it is obvious that the proposed RSNSEMS algorithm can outperform the NSEMS and SNSEMS algorithms. When the SNR is less than 1.8 dB, the complexity of the RSNSEMS algorithm is slightly higher than that of the SNSEMS algorithm, and when the SNR is greater than 1.8 dB, the complexity of the RSNSEMS algorithm is even lower than that of the SNSEMS algorithm. Compared to the SUEMS algorithm, the errorcorrecting performance of the proposed RSNSEMS algorithm is basically the same, and the complexity of the RSNSEMS algorithm is much lower than that of the SUEMS algorithm, which increases with increasing SNR and is basically the same as that of the MEMS algorithm when the SNR is greater than 2.0 dB.
Second, we consider a regular NBLDPC code, . The paritycheck matrix of this code has row weight and column weight . The parameters of the implemented algorithms are listed below.
(1) For the MEMS algorithm,; for the NSEMS algorithm, ; for the SUEMS algorithm, ; for the SNSEMS algorithm, ; and for the RSNSEMS algorithm, . (2) For all algorithms, the scaling factor and .
The simulation results are shown in Figures 8 and 9. It can be observed that the QSPA has the best errorcorrecting performance among all algorithms, but it also has the highest computational complexity. The performance of the EMS algorithm is secondbest, and it can reduce complexity by nearly 70% with approximately 0.2 dB degradation. Although the proposed SNSEMS algorithm has the overall lowest complexity, its performance has a certain degree of degradation compared to the MEMS algorithm. The proposed RSNSEMS algorithm can outperform the NSEMS and NSEMS algorithms; it has negligible (nearly no BER) degradation with an approximately 25% reduction in complexity compared to the MEMS algorithm. It is worth noting that the complexity ratio of the SUEMS algorithm starts from 0.7 and increases to be almost equal to that of the MEMS algorithm with increasing SNR. Therefore, compared to the SUEMS algorithm, the RSNS scheduling algorithm proposed in this paper has almost the same errorcorrecting performance but has lower computational complexity.
Finally, we consider a regular NBLDPC code, . The paritycheck matrix of this code has row weight and column weight . The parameters of the implemented algorithms are listed below.
(1) For the MEMS algorithm, ; for the NSEMS algorithm, ; for the SUEMS algorithm, ; for the SNSEMS algorithm, ; and for the RSNSEMS algorithm, . (2) For all algorithms, the scaling factor and .
The simulation results are shown in Figures 10 and 11. It can be observed that, in these three simulations, the complexity ratio of SUEMS algorithm increases with increasing SNR, and it is almost the same as the MEMS algorithm at high SNR. We believe the reason is that the method used by the SUEMS algorithm to decide whether a variable node is reliable is tracing the message evolution over iterations, where the variable node will be considered reliable only if its outgoing message is stable in iterations. The convergence speed of decoder is accelerated under high SNR, and it is possible that all variable nodes have already converged before being considered reliable. Therefore, the SUEMS algorithm and MEMS algorithm are basically the same in this situation, so there will be similar computational complexity. Besides, compared with Figures 8 and 9, we set in this simulation, which makes the complexity ratio of NSEMS algorithm higher than those of the proposed algorithms. In this case, the correcting performance of NSEMS algorithm is better than that of the SNSEMS algorithm and it is close to that of the RSNSEMS algorithm in lowtomoderate SNR, but it still has a certain degradation with the proposed algorithms in moderatetohigh SNR. Conversely, the proposed algorithms, especially the RSNSEMS algorithm, can reduce the complexity by nearly 25% while keeping the errorcorrecting performance almost unchanged compared to the original MEMS algorithm.
5. Conclusions
In this paper, we propose two innovative scheduling strategies for the EMS algorithm—the SNSEMS and RSNSEMS algorithms—to reduce the computational complexity of the decoder. The SNSEMS algorithm improves the errorcorrecting performance of the NSEMS algorithm by using a new metric to judge whether or not a variable node is reliable. More importantly, to further improve the performance of the SNSEMS algorithm while not significantly increasing its computational complexity, we investigate the variation in the processing node subset of both error frames and converged frames in the SNSEMS algorithm. According to the experimental results, the RSNSEMS algorithm is proposed, and it can alleviate performance degradation by employing a subsetreset mechanism. The simulation results show that the RSNSEMS algorithm can outperform the NSEMS and SNSEMS algorithms with similar computational complexity. It not only effectively alleviates the performance degradation compared to the MEMS algorithm but also reduces the computational complexity by approximately 25%. In conclusion, the proposed method can provide a good tradeoff between BER/FER and decoding complexity compared to other methods.
However, this paper only discovers that such abnormal stability in the processing subset will affect decoding performance, as well as why this characteristic occurs has not been deeply discussed. We believe that continuing to study the reasons for this abnormal stability can provide a theoretical basis for the design of this algorithm. Furthermore, when resetting the processing subset, the method in this paper chooses to put all check nodes into subsets. If we continue to study which check nodes will have a greater effect on alleviating abnormal stability and use this to reset the processing subset, then it is expected that the computational complexity of the algorithm will continue to be reduced. Finally, since there are less researches for informed dynamic schedule and selected schedule on the implementation of decoder, combining proposed scheduling with reduced (hardware) complexity decoding [28, 29] is also an interesting topic for the implementation of NBLDPC codes. We hope that this work will motivate more researches in these areas.
Data Availability
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
Acknowledgments
This research was supported by the National Key R&D Program of China (2018YFB0505103), the National Natural Science Foundation of China (61561016 and 61861008), and the funding of Science and Technology Major Project of Guangxi (AC16380014, AA17202048, and AA17202033).