Abstract

Modulation with correlated signal waveforms is considered. Such correlation arises naturally in a number of modern communications systems and channels, for example, in code-division multiple-access (CDMA) and multiple-antenna systems. Data entering the channel in parallel streams either naturally or via inverse multiplexing is transmitted redundantly by adding additional signal waveforms populating the same original time-frequency space, thus not requiring additional bandwidth or power. The transmitted data is spread over a frame of N signaling intervals by random permutations. The receiver combines symbol likelihood values, calculates estimated signals and iteratively cancels mutual interference. For a random choice of the signal waveforms, it is shown that the capacity of the expanded waveform set is nondecreasing and achieves the capacity of the Gaussian multiple access channel as its upper limit when the number of waveforms becomes large. Furthermore, it is proven that the iterative demodulator proposed here can achieve a fraction of 0.995 or better of the channel capacity irrespective of the number of transmitted data streams. It is also shown that the complexity of this iterative demodulator grows only linearly with the number of data streams.

1. Introduction

Modulation is the process of injectively mapping elements of a discrete set, called the messages, onto functions of time, called the signals, for the purpose of information transmission. The signals form a (finite-dimensional) Hilbert space, called the signal space. Geometric representations of the signals are often called signal constellations. Basic modulation methods prefer the use of orthogonal bases of the signal space as the signals themselves, since demodulation can be accomplished by projection onto these bases. For example, equidistant π‘š-ary pulse-amplitude modulation (PAM) uses discrete amplitudes on each basis [1].

Signals experience distortion during transmission which is modeled probabilistically, mainly due to the addition of noise. The received signals are therefore no longer identical with the transmitted signals. The demodulation problem is that of mapping a received signal back to a message such that the probability of the demodulated message not equalling the original transmitted message is minimized. Under the assumption of additive white Gaussian noise, picking the message whose signal is closest to the received signal using the natural Euclidean distance metric is optimal (if noise is correlated, for example, a generalized metric needs to be used) [1]. This is referred to as maximum-likelihood (ML) decoding since it minimizes the message error probability.

However, ML decoding quickly becomes practically infeasible by the β€œcurse of combinatorics,” and other methods are needed to be considered. Shannon [2] showed that every transmission channel has a maximum possible transmission rate which it can support, called the Shannon capacity, and that there exist coding and decoding methods which can operate to within πœ– of this capacity at arbitrarily low error rates. Shannon's nonconstructive proofs did not require ML decoding, opening the door to possibly low-complexity capacity-achieving signaling methods. Unfortunately, to achieve capacity requires continuous input alphabets which is highly impractical. Discrete modulations, mapped onto orthogonal bases, such as PAM, cannot achieve the Shannon capacity on the Gaussian channel. Certain high-dimensional discrete constellations, such as lattices, have been reported to achieve capacity, but in many ways their regular (discrete) structure is lost in the process [3].

In this paper we pursue another approach, abandoning the use of orthogonal bases as signals. In many practical situations the signals utilized are correlated, either by design, or by the effects that occur during transmission. An example of the former is (random) code-division multiple-access (CDMA) [4], and an example of the latter is multiple-antenna transmission (MIMO) [5]. In both cases the signals are densely correlated-which makes efficient demodulation extremely difficult. If the correlation pattern is sparse, that is, if any given signal waveform interferes with only a few other (neighboring) signal waveforms, sequence detection algorithms like the Viterbi algorithm can be used efficiently. A number of modulation methods based on superposition of individual data streams have been proposed (see [6–8]). When a number of independent signals add up in the channel, they can sometimes be decoded sequentially. Onion-peeling decoding starts from the largest power signal, decodes it treating the rest of the signals as noise and subtracts the result from the composite received signal. The decoding then continues analogously with the second strongest signal to the weakest. A number of methods based on successive decoding have been proposed and studied for various types of signals (data streams), including binary [9]. Channel capacity can be approached in the case when powers and rates of the signals follow specific precise arrangements, which is, however, challenging to accomplish in practice.

In this paper we assume a random correlation among the signals by postulating that these signals correspond to random vectors in signal space. The CDMA and MIMO channels are practical examples of such random channels [5, 10, 11]. Transmission relies on repeating the symbols of a message with random delays. Each time the symbol is modulated onto a new signal. While this increases the number of signals utilized, it allows for a very efficient iterative demodulation method to be used. This iterative demodulator forms the first stage of a two-stage receiver, where the second stage is a conventional forward error control (FEC) decoder for individual (binary) data streams, That is, the iterative first stage efficiently separates the correlated data streams. Specific adaptations of generalized modulation have recently been proposed for both CDMA [12] and MIMO channels [13].

Our contributions in this paper are two-fold. First we show that using random signals incurs no capacity loss, and furthermore, that regular-spaced PAM-type modulation on these random signals can achieve the Shannon capacity. We then discuss transmission using redundant signaling and an iterative demodulation method for which we show that it can operate close to the Shannon capacity over the entire range of operational interest. In showing this, we will only assume that we have capacity-achieving binary error control codes available, a very reasonable assumption given the current state-of-the art in error control coding [14].

2. Modulation

Generally, a discrete data stream 𝐝 is mapped onto signals from a finite set of such signals according to some mapping rule. In the ubiquitous pulse-amplitude (PAM) modulation, a discrete amplitude π‘₯π‘Ÿ for each value of π‘‘π‘Ÿ from 𝐝 is first selected then used to multiply the π‘Ÿth signal. Most commonly, one of 2𝐡 amplitude levels is selected for each 𝐡-bit data symbol. In the case of 8-PAM, for example, with 𝐡=3, the discrete equispaced amplitudes shown in Figure 1 are used on each signal. This signal constellation can be interpreted as the superposition of three simple binary constellations, where Bit 1 has 4 times the power of Bit 0, and Bit 2 has 16 times its power.

In general, any properly labeled 2𝐡-PAM modulation can be written as the superposition of 𝐡 binary antipodal amplitudes, that is,𝑣=π΅βˆ’1𝑗=02𝑗𝑏𝑗,(1) where π‘π‘—βˆˆ{βˆ’1,1}. If an entire sequence 𝐯 of 2𝐡-ary PAM symbols is considered, it may be viewed as the superposition of 𝐡 binary modulated data streams with powers 𝑃0,4𝑃0,16𝑃0,…,4π΅βˆ’1𝑃0 on binary data streams which make up the PAM symbol sequence 𝐯.

This viewpoint is quite productive in that it suggests a capacity-achieving demodulation method for large constellations based on cancellation. Consider the case where the highest-power uncanceled data stream considers all lower power data streams as noise. Its maximum rate is then given by the mutual information 𝐢𝑗𝑏=𝐼𝑗;π‘¦βˆ£π‘π‘—+1,…,π‘π΅βˆ’1ξ€Έ,(2) where 𝑦 is the output of the channel. As long as the rate on the 𝑗th binary data stream 𝑅𝑗<𝐢𝑗, it can be correctly decoded using a binary Shannon-capacity-achieving code.(While no class of binary codes with nonexponential decoding complexity exist which can provably achieve the capacity on a binary-input channel, codes which can achieve this capacity β€œpractically” with β€œimplementable” complexities have recently emerged from intense research. The most popular representatives are turbo codes and low-density parity-check codes. Both utilize iterative message passing decoding algorithms [14].) By virtue of (1), knowledge of 𝑏𝑗+1,…,π‘π΅βˆ’1 implies that these data streams can be canceled from the received signal, and 𝐢𝑗 is the capacity of the 𝑗th binary data stream. This thought model leads to a successive decoding and canceling method which can achieve the mutual information rate𝐢symmetric=𝐼(𝑣;𝑦)=π΅βˆ’1𝑗=0𝐢𝑗(3) by the chain rule of mutual information. 𝐢symmetric is of course not equal to the capacity of the additive white Gaussian noise channel 𝑦=𝑣+𝑛, since the input distribution of 𝑣 is uniform, rather than being Gaussian distributed as required to achieve the channel capacity. In fact, 𝐢symmetric loses the so-called shaping gain of 1.52 dB with respect to the capacity of the Gaussian channel [14].

3. Main Result

In this paper, we propose a generalized PAM modulation method which operates with random signals, rather than the orthogonal bases implied by the discussion in the previous section. We present a two-stage demodulator/decoder which remedies the difficulties of the onion-peeling method discussed above. Specifically, the demodulator consists of an iterative demodulator which operates in parallel and achieves a signal-to-noise ratio (SNR) improvement on each of the binary data streams. These are then decoded using external binary error control codes. The latency issue is basically confined to that of the parallel demodulator and that of the follow-up error control decoder. The iterative demodulator is based on cancelation. This means that its complexity, and that of the entire decoder scales linearly with the number of data streams.

Our main result is that we will prove that such an iterative demodulator/decoder can achieve a cumulative data rate 𝑅𝐝 per dimension such that𝑅𝐝β‰₯0.995𝐢GMACξ‚ƒβˆ’0.54bitsξ‚„dimension,(4) where 𝐢GMAC is the Shannon capacity of the Gaussian multiple-access channel. That is, our system can achieve a fraction of 0.995 of channels information theoretic capacity, irrespective of system size.

In order for the iterative demodulator to function, we require that the number of signals in the signal space is increased, but not the power or spectral resources.

4. System Model

4.1. Signaling

We are considering communication of multiple data streams ππ‘˜ using random signals π¬π‘˜ of dimension 𝑁. If 𝑁 is sufficiently large, the number of useful signals is arbitrarily large. A set of 𝐾 data symbols π‘‘π‘˜,𝑙 from the data streams ππ‘˜ is transmitted at each time interval 𝑙. There are basically two ways to do this. Conventionally each symbol π‘‘π‘˜,𝑙 is directly modulated onto an individual signal π¬π‘˜,𝑙. In this paper, however, we propose an alternative where we duplicate each symbol π‘‘π‘˜,𝑙𝑀-fold. These duplicates are then modulated onto separate signals π¬π‘˜,πœ‹π‘˜,π‘š(𝑙) at 𝑀 random time intervals within a certain signal block, where πœ‹π‘˜,π‘š(𝑙) is the random location within the block where the π‘šth copy of symbol π‘‘π‘˜,𝑙 is located. The function πœ‹π‘˜,π‘š(𝑙) is a permutation function with inverse πœ‹βˆ’1π‘˜,π‘š(𝑙). Even though we have increased the number of signals by a factor 𝑀, scaling the power with 𝑀, and requiring that the signal set {π¬π‘˜,πœ‹π‘˜,π‘š(𝑙)} occupies the original 𝑁-dimensional signal space, this will not affect total power or the total spectrum utilization. (Another form of modulation based on randomly correlated signals called β€œpartitioned transmission” has been recently proposed in [15]. Partition signalling creates redundancy and sparseness in transmitted data by partitioning 𝐾 existing 𝑁-dimensional signal waveforms and permuting the resulting partitions. Generalized modulation relies on populating the signals space with additional 𝑁-dimensional signal waveforms. The latter gives an opportunity to create the requited level 𝑀 of redundancy independently of signal dimensionality 𝑁. Further, near capacity operation with generalized modulation does not require π‘β†’βˆž.) A diagram of this modulator is given in Figure 2.

We make the convenient, but in no way necessary assumption that the channel is block-synchronous, that is, that the signal waveforms at time interval 𝑙 interfere only within that time interval, and that there is no correlation of signal waveforms between time intervals. With this we can write the channel in the linear matrix form𝐲𝑙=𝐒𝑙𝐖1/2𝐱𝑙+πœ‚π‘™,(5) where the 𝑁×𝐾𝑀 matrix 𝐒𝑙 contains the signal vectors as columns. The capacity per dimension of this channel is well known [10] and is given by𝐢𝐒=1ξ‚€12𝑁logdet𝐈+𝜎2𝐒𝐖𝐒𝑇,(6) where 𝑃𝐖=diag1𝑀,𝑃1𝑀𝑃,…,1𝑀,𝑃2𝑀𝑃,…,𝐾𝑀(7) is a 𝐾𝑀×𝐾𝑀 diagonal matrix with the powers used for transmission of the different signal vectors. We now assume that the signals π¬π‘˜,𝑙 are chosen randomly from the signal space(the individual components π‘ π‘˜,𝑙,𝑛, 𝑛=1,…,𝑁, of signals π¬π‘˜,𝑙 can, for example, be selected randomly out of the set √{βˆ’1/βˆšπ‘,1/𝑁} picking each entry with probability 1/2. However, other random selections satisfying (8) are also possible) such that the mutual pairwise expected correlation between signals isEξ‚ƒπ¬βˆ—π‘—,π‘™π¬π‘—ξ…ž,𝑙2=1𝑁;π‘—β‰ π‘—ξ…ž.(8) This model captures among others the random code-division multiple-access channel and the isotropic multiple-antenna channel model.

The capacity 𝐢𝐒 of this random vector channel is given by the expectation over 𝐒 in (6). Using Jensen's inequality𝐢𝐒≀E𝐒1ξ‚€12𝑁logdet𝐈+𝜎2𝐒𝐒𝑇(9) with equality if and only if 𝐖=𝐈 (see [16]). That is to say that an equal distribution of powers over the signals maximizes the capacity of the random vector channel (5).

We now investigate the information theoretic impact of increasing the signal population as proposed by the 𝑀-fold duplication. The following lemma addresses this issue.

Lemma 1. Keeping the transmit power tr(𝐖)=𝑃 constant, the capacity 𝐢𝐒 as a function of 𝐾 and 𝑁 approaches the capacity of the Gaussian multiple-access channel in the limit, that is, πΆπ’βŸΆπΆGMAC=12ξƒ©βˆ‘log1+πΎπ‘˜=1π‘ƒπ‘˜π‘πœŽ2ξƒͺ.(10) It approaches this limit from below as 𝑀,πΎβ†’βˆž, that is, 𝐢𝐒<𝐢GMACforall𝐾/𝑁<∞.

Proof. See Appendix A.

Lemma 1 reveals useful information in several ways. Firstly, it guarantees that the signaling strategy presented above, that is, the addition of extra random signal waveforms, incurs no capacity loss, and secondly, in the limit, arbitrary power assignments become capacity achieving, not only the equal power assignment.

4.2. Demodulation

The first stage of the demodulation process starts with matched filtering of the received signal with respect to each transmitted signal waveform in each time interval. Given the received signal embedded in Gaussian noise asπ²π‘Ÿ=πΎξ“π‘˜=1π‘€βˆ’1ξ“π‘š=0ξ‚™π‘ƒπ‘˜π‘€π‘‘π‘˜,πœ‹βˆ’1π‘˜,π‘š(π‘Ÿ)π¬π‘˜,πœ‹βˆ’1π‘˜,π‘š(π‘Ÿ)+π§π‘Ÿ,(11) these matched filter outputs are given byπ‘§π‘˜,𝑙=π¬βˆ—π‘˜,π‘™β‹…π²π‘Ÿ=πΎξ“π‘˜β€²=1π‘€βˆ’1ξ“π‘šβ€²=0ξ‚™π‘ƒπ‘˜π‘€π‘‘π‘˜β€²,π‘™β€²π¬βˆ—π‘˜,π‘™β‹…π¬π‘˜β€²,𝑙′+π‘›π‘Ÿ,𝑙=πœ‹βˆ’1π‘˜,π‘š(π‘Ÿ),π‘™ξ…ž=πœ‹π‘˜βˆ’1β€²,π‘šβ€²(π‘Ÿ),(12) where π‘›π‘Ÿ is the sampled noise of variance 𝜎2, and π¬βˆ—π‘˜,π‘™β‹…π¬π‘˜β€²,𝑙′=πœŒπ‘˜,π‘˜β€²π‘š,π‘šβ€²,𝑙,𝑙′ is the correlation value between the target signal and a given interfering signal. The matched filter outputs in (12) consist of π‘‘π‘˜,𝑙 and an interference and noise term, which is given byπΌπ‘˜,π‘š,π‘Ÿ=πΎξ“π‘˜β€²π‘˜=1β€²β‰ π‘˜π‘€βˆ’1ξ“π‘šβ€²=0ξ‚™π‘ƒπ‘˜β€²π‘€πœŒπ‘˜,π‘˜β€²π‘š,π‘šβ€²,𝑙,π‘™β€²π‘‘π‘˜β€²,𝑙′+π‘›π‘Ÿ.(13)

At this point the graphical illustration shown in Figure 3 may prove helpful, which shows how the different symbols and signals combine to generate the sequence of received signal vectors π²π‘Ÿ. Note that in the interference equation (13) above, self-interference is not included. Apart from unnecessarily complicating the notation, this self-interference is negligible as shown below. Furthermore, in many cases it is not difficult to ensure that the signal vectors used for the different signals from a user π‘˜ impinging on channel symbol π²π‘Ÿ are orthogonal, that is,π¬βˆ—π‘˜,πœ‹βˆ’1π‘˜,π‘š(π‘Ÿ)β‹…π¬π‘˜,πœ‹βˆ’1π‘˜,π‘šβ€²(π‘Ÿ)=0,βˆ€π‘˜,(14) and cause no self-interference. In [12], for example, different time intervals are used for the duplicate signals to accomplish this. The graphical representation reveals the similarity with graph-based error control codes, in particular with fountain codes [17]. Consequently, we will explore a demodulation method based on message passing.

Iterative demodulation follows the following message-passing principle. At the channel nodes, updated matched filter output signals are computed at each iteration by subtracting interference to generate 𝑧(𝑖)π‘˜,𝑙,π‘š=π¬βˆ—π‘˜,π‘™β‹…βŽ›βŽœβŽœβŽœβŽœβŽπ²π‘Ÿβˆ’πΎξ“π‘˜β€²=1(π‘˜β€²β‰ π‘˜)ξ“π‘šπ‘€βˆ’1β€²=0ξ‚™π‘ƒπ‘˜β€²π‘€ξ‚π‘‘π‘˜(π‘–βˆ’1)β€²,𝑙′,π‘šβ€²β‹…π¬π‘˜ξ…ž,π‘™ξ…žβŽžβŽŸβŽŸβŽŸβŽŸβŽ ,(15) where ξ‚π‘‘π‘˜β€²,𝑙′,π‘šβ€² is a soft symbol estimate of the π‘šξ…žth copy of π‘‘π‘˜,𝑙. Note that, following the extrinsic principle, the π‘š different estimates for the same symbol are not necessarily identical (see below). These soft symbol estimates, in turn, are computed at the symbol nodes from the 𝑀 matched filter signals for each copy of π‘‘π‘˜,𝑙. While π‘‘π‘˜,𝑙 can, in general, be any complex integer, we will concentrate on the basic binary case were π‘‘π‘˜,π‘™βˆˆ{βˆ’1,1}. We will show later how to build larger modulation alphabets from this basic binary case using the binary decomposition of PAM signals.

In the binary case, the soft symbols are calculated as 𝑑(𝑖)π‘˜,𝑙,π‘šβŽ›βŽœβŽœβŽξ“=tanhπ‘šπ‘€βˆ’1β€²=0(π‘šβ€²β‰ π‘š)ξ‚™π‘ƒπ‘˜π‘€π‘§(𝑖)π‘˜,𝑙,π‘šβ€²πœŽ2π‘˜,π‘–βŽžβŽŸβŽŸβŽ (16) which is the optimal local minimum-variance estimate of π‘‘π‘˜,𝑙 given that interference and noise combined form a Gaussian random variable with power 𝜎2π‘˜,𝑖. The variance of the symbol estimates (16) will be required in the analysis in Section 5. Defining this variance at iteration 𝑖 as 𝜎2𝑑,π‘˜,𝑖=E|π‘‘π‘˜βˆ’ξ‚π‘‘(𝑖)π‘˜,π‘š|2, and assuming that correlation between interference experienced by different replicas of the same symbol is negligible due to sufficiently large interleaving, it can be calculated adapting the development in [18] for CDMA as𝜎2𝑑,π‘˜,π‘–ξ€·ξ€·βˆš=E1βˆ’tanhπœ‡+πœ‡πœ‰ξ€Έξ€Έ2=𝑔(πœ‡),βˆ€π‘–,(17) where πœ‰βˆΌπ’©(0,1) and πœ‡=(π‘€βˆ’1)π‘ƒπ‘˜/(π‘€πœŽ2𝑖) from (16), and 𝜎2π‘˜,𝑖=𝜎2𝑖,forallπ‘˜. The function 𝑔(πœ‡) has no closed form, but the following bounds are quite tight [19]: ξ€·βˆšπ‘”(πœ‡)≀1(1+πœ‡);πœ‡<1,(18)𝑔(πœ‡)β‰€πœ‹π‘„πœ‡ξ€Έ;πœ‡β‰₯1,(19) where 𝑄(β‹…) is the complementary error function. The final output signal after 𝐼 iterations is 𝑧(𝐼)π‘˜,𝑙=βˆ‘π‘€βˆ’1π‘š=0𝑧(𝐼)π‘˜,𝑙,π‘š, which is passed to binary error control decoders for data stream π‘˜. The final signal-to-noise/interference ratio (SINR) of 𝑧(𝐼)π‘˜,𝑙 is what primarily matters for the error performance of these error control decoders.

After 𝐼 detection iterations of the first stage the data is passed to the second stage of demodulation. The second stage of the reception is the error control decoding which is executed for each of the 𝐾 data streams individually. SINR for data stream π‘˜ equals π‘ƒπ‘˜/𝜎2𝐼 and it can be argued that the residual noise and interference is Gaussian [15]. Ultimately the information rate (i.e., the rate of the error control code) of stream π‘˜ should satisfyπ‘…π‘˜β‰€πΆBIAWGNξƒ©π‘ƒπ‘˜πœŽ2𝐼ξƒͺ(20) for error-free decoding at the second stage. Here by 𝐢BIAWGN(π‘₯) we denote the capacity of the binary-input real-valued output AWGN channel with SNR π‘₯.

5. Generalized Modulation

The discussion above treated the case of binary modulation on the different signal waveforms, however, as illustrated in Section 2, we can create the regular-spaced PAM modulations with geometrically scaled binary modulations using powers𝑃04𝑏,0β‰€π‘β‰€π΅βˆ’1.(21) We assume that there are 𝐾𝑏 data streams with powers 𝑃04𝑏. Thus, the total number of streams equals βˆ‘πΎ=π΅βˆ’1𝑏=0𝐾𝑏.

Assuming large enough interleavers, the evolution of the interference in this iterative demodulator can be captured with a standard density evolution analysis. Since the average correlation between signal waveforms E[(πœŒπ‘˜,π‘˜β€²m,π‘šβ€²,𝑙,𝑙′)2]=1/𝑁 (see (8)), the interference and noise on stream π‘˜ is given by𝜎2π‘˜,𝑖=1π‘πΎξ“π‘˜β€²=1(π‘˜β€²β‰ π‘˜)π‘ƒπ‘˜ξ…žπœŽ2𝑑,π‘˜ξ…ž,π‘–βˆ’1+𝜎2≀1π‘πΎξ“π‘˜=1π‘ƒπ‘˜πœŽ2𝑑,π‘˜,π‘–βˆ’1+𝜎2=𝜎2𝑖(22) which is common to all streams. The upper bound in (22) contains the self-interference term for π‘˜ξ…ž=π‘˜, which, however, becomes negligible as 𝐾 and 𝑀 grow. Using (17) in (22) and the PAM power distribution we obtain𝜎2𝑖𝑃0=π΅βˆ’1𝑏=0𝐾𝑏4π‘βˆ’1π‘π‘”ξƒ©π‘€βˆ’1𝑀4π‘βˆ’1𝜎2π‘–βˆ’1/𝑃0ξƒͺ+𝜎2𝑃0.(23) The next theorem proves that generalized PAM modulation used with the two-stage demodulation described above can closely approach the channel capacity.

Theorem 1. Consider generalized PAM modulation (21) with 𝐡 levels and 𝐾𝑏=0.995𝑁 data streams per level for 𝑏=0,1,…,π΅βˆ’1 giving a total number of streams 𝐾=0.995𝐡𝑁. One assumes that each data stream is encoded with a binary error control code which is capacity achieving on the binary-input AWGN channel, that is, Rπ‘˜=𝐢BIAWGNξ‚΅π‘ƒπ‘˜πœŽ2βˆžξ‚Ά,(24) and let π‘€β†’βˆž. Then the resulting spectral efficiency per dimension 𝑅d=1π‘πΎξ“π‘˜=1π‘…π‘˜β‰₯0.995𝐢GMACβˆ’0.54.(25)

Proof. See Appendix B.

We note that for 𝐡1<𝐡2 the corresponding capacity approaching power profiles 𝑃0,𝑃04,…,𝑃04𝐡1βˆ’1 and 𝑃0,𝑃04,…,𝑃04𝐡2βˆ’1 coincide for 𝑏≀𝐡1βˆ’1. The importance of these results is that new data streams can always be added without affecting decodability of the existing streams.

The gap between achieved spectral efficiency and the channel capacity can also be introduced in terms of average 𝐸b/𝑁0, instead of data stream power profile. Average 𝐸b/𝑁0 for the power profile used in Theorem 1 can be upper bounded as𝐸b𝑁0=12Rdπœ‚π΅βˆ’1𝑏=0𝛾04𝑏≀42πœ‚π΅ξ€Έβˆ’13πœ‚(π΅βˆ’1+0.6706)(26) from (B.14), and therefore the corresponding capacity of AWGN channel 𝐢(𝐸b/𝑁0), using 𝐸b𝑁0=22𝐢(𝐸b/𝑁0)βˆ’1𝐸2𝐢b/𝑁0ξ€Έ(27) can be upper bounded as𝐢𝐸b/𝑁0≀𝐡+0.76.(28) As a result we obtainξ‚΅πΈπœ‚πΆb𝑁0ξ‚Άβˆ’π‘…d𝐸b𝑁0≀1.09.(29)

In Figure 4 we plot the achievable spectral efficiencies for the proposed generalized PAM modulation (21) for 𝐡=1,2,3,4 levels and assume ideal posterror control decoding with rates satisfying (24). Such performance can be closely approached with appropriate standard error control codes, which are very well developed for the binary case [20, 21]. We assume that 𝐾𝑏=𝛼𝑁/𝐡, for 𝑏=0,1,2,…,π΅βˆ’1, where parameter π›Όβˆˆ(0,∞). Each curve corresponds to fixed 𝐡 and plotted as a function of average 𝐸b/𝑁0 which is in turn the function of 𝛼. We can observe that spectral efficiency of generalized PAM modulation exceeds the capacity of the same PAM modulation using orthogonal waveforms. This is because the number of allowable correlated signal waveforms 𝐾 exceeds the number of available orthogonal dimensions 𝑁. This advantage is most noticeable for 2-PAM, where the maximum achievable capacity of 2.08 is more than twice the number of orthogonal dimensions. For higher PAM constellations, the capacity per level 𝛼𝑏=𝐾𝑏/𝑁=𝛼/𝐡→1 rapidly from above. Note that for 𝛼=πœ‚=0.995 the gap between the performance curves and the capacity curve satisfies (29). Specifically, point 𝛼=πœ‚ for 𝐡=1 gives 𝐸b/𝑁0=4.72 dB, for 𝐡=2 gives 𝐸b/𝑁0=7.74 dB, for 𝐡=3 gives 𝐸b/𝑁0=11.94 dB, and for 𝐡=4 gives 𝐸b/𝑁0=16.63 dB.

6. Conclusions

We have presented and analyzed a two-stage iterative demodulation methodology for generalized PAM constellations using correlated random signals rather than the usual orthogonal bases. The method operates by introducing redundant duplicate copies of the data symbols modulated onto extra signals. An exponential power distribution, inherently present in PAM modulations, allows this two-stage iterative demodulator to achieve 99.5% of the Shannon capacity using binary capacity-achieving error control codes for each data stream. This generalized PAM modulation format was shown to approach the channel capacity over a wide range of operating SNRs, and can exceed the capacity of traditional PAM constellations on orthogonal signals.

Appendices

A. Proof of Lemma 1

Decompose the argument of (6) as ξ‚€1det𝐈+𝜎2π’π–π’π‘‡ξ‚βˆ‘=det1+πΎπ‘˜=1π‘ƒπ‘˜π‘πœŽ2ξƒͺξƒͺ=ξƒ©βˆ‘πˆ+𝐁1+πΎπ‘˜=1π‘ƒπ‘˜π‘πœŽ2ξƒͺ𝑁det(𝐈+𝐅),(A.1) where 𝐅 is the matrix of the off-diagonal elements with 𝐹𝑖𝑗=πœ…π‘€ξ“πΎπ‘š=1ξ“π‘˜=1π‘ƒπ‘˜π‘€π‘π‘š,π‘˜,(A.2) where πœ…=(𝜎2+βˆ‘πΎπ‘˜=1π‘ƒπ‘˜)βˆ’1 and π‘π‘š,π‘˜βˆˆ{Β±1} with uniform probabilities. The entries 𝐡𝑖𝑗 have zero mean and variance 𝐹var𝑖𝑗=1𝑀2βˆ‘πΎπ‘˜=1𝑃2π‘˜ξ‚€βˆ‘πΎπ‘˜=1π‘ƒπ‘˜ξ‚2,(A.3) which vanishes as (i) π‘€β†’βˆž or (ii) πΎβ†’βˆž. Condition (ii), however, requires the Lindeberg condition to hold on the set {π‘ƒπ‘˜}.

Using (i), or (ii), the elements in 𝐅 are sufficiently small to apply Jacobi's formula, that is, 𝐹det(𝐈+𝐅)=(1βˆ’tr(𝐅))+𝑂2𝑖𝑗.(A.4) Since tr(𝐅)=0, and the second moment of 𝐹𝑖𝑗 vanishes, the limit value of the Lemma is proven.

Using Hadamard's inequality it is straightforward to show that𝐢𝐒<𝐢GMAC,(A.5) and the limit is approached from below. While det(𝐈+𝐅)β†’1 in probability, 𝐢𝐒→𝐢GMAC almost surely.

B. Proof of Theorem 1

Let us define 𝛾=𝑃0/𝜎2𝑖, 𝛾′=𝑃0/𝜎2, and π›Ύβˆž=𝑃0/𝜎2∞. Consider 𝐡=∞ here for simplicity and define πœ‚=0.995. Convergence defined by (23) (for 𝜎2𝑖, 𝑖=0,1,…) follows from1>βˆžξ“π‘=0πœ‚π›Ύ4𝑏𝑔𝛾4𝑏+π›Ύπ›Ύξ…žξ€·,forπ›Ύβˆˆ0,π›Ύβˆžξ€»,(B.1) and 𝐾𝑏=πœ‚π‘. Success of the demodulation stage happens if π›Ύβˆž is close to π›Ύξ…ž. This means that the interstream interference is canceled almost entirely. Let us choose a somewhat arbitrary lowest power 𝑃0 such that π›Ύξ…ž=4 and prove that π›Ύβˆž>1.79.

Let us define the functions 𝑑(π‘₯)=π‘₯𝑔(π‘₯),(B.2)𝑓(𝛾)=βˆžξ“π‘=0ξ€·πœ‚π‘‘π›Ύ4𝑏+π›Ύπ›Ύξ…ž=βˆžξ“π‘=0ξ€·πœ‚π‘‘π›Ύ4𝑏+𝛾4.(B.3) The function 𝑑(π‘₯) monotonically increases for 0<π‘₯<π‘₯0 and monotonically decreases for π‘₯>π‘₯0, where π‘₯0β‰ˆ1.508. To find an upper bound on 𝑓(β‹…), we consider the terms 𝑑(𝛾4𝑏) for very small and very large arguments separately, that is,𝑓(𝛾)=πœ‚βˆžξ“π‘=0𝑑𝛾4𝑏+𝛾4=πœ‚π‘s.t.𝛾4𝑏<𝐴1𝑑𝛾4𝑏+πœ‚π‘s.t.𝛾4𝑏>𝐴2𝑑𝛾4𝑏+πœ‚π‘s.t.𝐴1≀𝛾4𝑏≀𝐴2𝑑𝛾4𝑏+𝛾4.(B.4) Using the fact that 𝑔(π‘₯)≀1 for any π‘₯ we obtain for any 𝑏1β‰₯0𝑏1𝑏=0𝑑𝛾4𝑏≀𝛾𝑏1𝑏=04𝑏=𝛾4𝑏1+1ξ€Έβˆ’13<𝛾4𝑏1+13.(B.5) From (19) we obtainβˆžξ“π‘=𝑏2𝑑𝛾4π‘ξ€Έβ‰€π›Ύπœ‹βˆžξ“π‘=𝑏24𝑏𝑄2π‘βˆšπ›Ύξ‚β‰€ξ‚™π›Ύπœ‹2βˆžξ“π‘=𝑏22π‘π‘’βˆ’π›Ύ4𝑏/2β‰€ξ‚™π›Ύπœ‹22𝑏2𝑒𝑏2βˆ’π›Ύ4/2ξ€·1βˆ’π‘’βˆ’3𝛾4𝑏2/2ξ€Έ(B.6) for 𝑏2 such that 𝛾4𝑏2>1. The last inequality in (B.6) is computed by upper bounding the sum by the geometrical progression using the inequality𝑑𝛾4𝑏+1𝑑𝛾4𝑏=2π‘’βˆ’3𝛾4𝑏/2β‰€π‘’βˆ’3𝛾4𝑏2/2.(B.7) Using (B.5) we get for any 𝐴1𝑏s.t.𝛾4𝑏<𝐴1𝑑𝛾4𝑏<4𝐴13.(B.8) Analogously, (B.6) gives𝑏s.t.𝛾4𝑏>𝐴2𝑑𝛾4π‘ξ€Έβ‰€ξ‚™πœ‹π΄22π‘’βˆ’π΄2/2ξ€·1βˆ’2π‘’βˆ’3𝐴2/2ξ€Έ.(B.9) By choosing 𝐴1=0.00003 and 𝐴4=24, we compute numeric values of the bounds (B.8) and (B.9) for the tails as 𝑏s.t.𝛾4𝑏<𝐴1𝑑𝛾4𝑏<4β‹…10βˆ’5,(B.10)𝑏s.t.𝛾4𝑏>𝐴2𝑑𝛾4𝑏<4β‹…10βˆ’5.(B.11) Let us define𝑓10(𝛾)=8β‹…10βˆ’5+πœ‚9𝑏=0𝑑𝛾4𝑏+𝛾4.(B.12) It follows from (B.4), (B.10), and (B.11) that for any 𝛾𝑓(𝛾)<𝑓10(𝛾).(B.13) We also notice that it is enough to consider π›Ύβˆˆ[𝐴1,𝛾′]. Numerical calculation shows that the only root 𝛾 of 𝑓10(𝛾)βˆ’1 on the interval π›Ύβˆˆ[𝐴1,4] equals 1.79374. Thus, π›Ύβˆžβ‰₯𝛾=1.79374 due to (B.13).

We calculate the spectral efficiency (or sum-rate per dimension) as follows:𝑅d=π΅βˆ’1𝑏=0πœ‚πΆBIAWGNξ€·π›Ύβˆž4𝑏=πœ‚(0.6859+0.9835+π΅βˆ’2βˆ’πœ–)β‰₯πœ‚(π΅βˆ’1+0.6706),(B.14) where we use a bound from [15]1βˆ’πΆBIAWGN(π‘₯)≀2πœ‹3/2ξ€·πœ‹ln22ξ€Έπ‘’βˆ’8βˆ’1/2π‘₯<10π‘’βˆ’1/2π‘₯(B.15) to upper bound πœ– asπœ–=π΅βˆ’1𝑏=2πœ‚ξ€·1βˆ’πΆBIAWGNξ€·π›Ύβˆž4π‘β‰€ξ€Έξ€Έπ΅βˆ’1𝑏=210π‘’βˆ’(π›Ύβˆž4𝑏)/2≀10βˆ’5,forany𝐡.(B.16)

The capacity of the additive Gaussian channel corresponding to power profile (21) with 𝐾𝑏=πœ‚π‘ streams per level can be calculated as follows 𝐢GMAC=12log21+πœ‚π΅βˆ’1𝑏=0𝛾04𝑏ξƒͺ=12log21+4πœ‚π΅βˆ’1𝑏=04𝑏ξƒͺ=1,(B.17)2log241+4πœ‚π΅ξ€Έβˆ’13ξƒͺβ‰€π΅βˆ’1+1.21.(B.18) Combining (B.18) and (B.14), we obtain (25).