Near-Ideal M-ary LDGM Quantization with Recovery - IEEE Xplore

0 downloads 0 Views 449KB Size Report
Near-Ideal M-ary LDGM Quantization with. Recovery. Qingchuan Wang, Student Member, IEEE, Chen He, Member, IEEE, and Lingge Jiang, Member, IEEE.
1830

IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 59, NO. 7, JULY 2011

Near-Ideal ๐‘€ -ary LDGM Quantization with Recovery Qingchuan Wang, Student Member, IEEE, Chen He, Member, IEEE, and Lingge Jiang, Member, IEEE Abstractโ€”For iterative mean-square error (MSE) quantizers with alphabet size ๐‘€ = 2๐พ using low-density generator-matrix (LDGM) code constructions, an efficient recovery algorithm is proposed, which adjusts the priors used in belief propagation (BP) to limit the impact of previous non-ideal decimation steps. Based on an analysis of the BP process under ideal or nonideal decimation, the algorithm first estimates the conditional probability distributions describing the effect of non-ideal decimation, then adjusts the priors to make the distributions match the ideal situation. As shown in simulation results, the recovery algorithm can improve quantization performance greatly, reducing the shaping loss to as low as 0.012 dB, while the increase in computational complexity is modest thanks to the use of FFT techniques. Index Termsโ€”Low-density generator-matrix, quantization, decimation, recovery.

I. I NTRODUCTION

S

PARSE-GRAPH codes have recently found some use in long-block lossy source coding problems due to their potential to achieve near-ideal rate-distortion performance at a lower computational complexity than traditional methods like trellis-coded quantization (TCQ) [1]. Although structured constructions such as polar codes [2] have been shown to be fast and effective in many channel and source coding problems [3], including those involving side information and binning [4], more randomized ones based on low-density generator matrix (LDGM) codes remain attractive due to their more moderate block length requirements, efficient integration with e.g. superposition coding schemes, as well as the availability of well-established optimization methods from low-density parity-check (LDPC) literature. Indeed, being duals to LDPC codes used in channel coding [5], LDGM codes are known to be able to approach the Shannon limit under optimal encoding for the binary symmetric case [6], and with appropriate modulation mappings, for more general sources possibly requiring non-uniform reconstruction alphabets as well [7]; variants with additional parity [8] or Hamming weight [9] constraints have also been proposed to improve finite-degree performance or to allow channel coding and binning. Naturally, most practical encoding algorithms (or quantizers) for such sparse-graph codes employ some form of message passing. Although more elaborate algorithms like survey propagation exist [6], [7], ordinary belief propagation Paper approved by Z. Xiong, the Editor for Distributed Coding and Processing of the IEEE Communications Society. Manuscript received August 5, 2010; revised December 19, 2010. The authors are with the Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China (e-mail: {r6144, chenhe, lgjiang}@sjtu.edu.cn). This paper was supported by the National Natural Science Foundation of China Grants No. 60772100, 60832009, and 60872017, as well as National 863 Program Grant No. 2009AA011505. Digital Object Identifier 10.1109/TCOMM.2011.061511.100462

(BP) appears to be sufficient for quantization when a good degree distribution is used [10]. In any case, decimation steps (i.e. hard decisions) usually need to be carried out according to e.g. the BP marginals (extrinsic probabilities) in order to make BP converge to one codeword among many similarly good ones; other positive feedback mechanisms used in [11] and [12] serve essentially the same purpose. For randomized LDGM constructions, loops in the factor graph and the limited number of iterations causes the extrinsic probabilities from BP to be approximate; this leads to non-ideal decimation choices, which adversely affects future iterations and makes the performance of such BP algorithms difficult to predict theoretically. Our works mainly focus on a specific lossy source coding problem, mean-square error (MSE) quantization of Euclidean space [13, Sec. II-C], which plays an important role in highrate source coding, as well as various channel coding schemes on Gaussian channels, such as the shaping component of dirty paper coding [14]. As shown in previous works [15] and [16], an LDGM-based construction using BP and decimation for encoding can approach the Shannon limit of this problem quite well, and methods for optimizing the degree distribution, pace of decimation, etc. have been proposed. The non-ideal decimation problem above also exists here; in [16, Sec. VI-C], a recovery step run before each BP iteration to adjust the priors has been found to reduce its impact and improve quantization performance significantly. However, the recovery algorithm in [16] is somewhat ad hoc and poorly understood, and it is only applicable to binary constructions, whose alphabet limitation leaves a significant gap to the Shannon limit. In this paper, after giving an overview of the quantization algorithm in Section II, in Section III we will present some theoretical arguments in an attempt to gain a better understanding of non-ideal decimation and how recovery might be performed. Based on this analysis, a recovery algorithm will be designed in Section IV for LDGM-based MSE quantization with ๐‘€ -ary alphabet, whose computational complexity turns out to be quite modest. The simulation results in Section V demonstrate that recovery can improve quantization performance greatly and approach the Shannon limit to within 0.012 dB. Finally, Section VI concludes the paper. A. Notations and Conventions Notations are similar to those in [16]. Bold letters denote sequences, vectors or matrices whose elements are indicated by subscripts, e.g. ๐’š = (๐‘ฆ1 , . . . , ๐‘ฆ๐‘› ); conversely, a vector or matrix can also be defined element-wise, e.g. (๐‘’โˆ’j2๐œ‹๐‘˜๐‘ฃ/๐‘€ )๐‘˜๐‘ฃ is the DFT (discrete Fourier transform) matrix. (โ‹…)T and (โ‹…)H denotes matrix transposition and Hermitian transposition, โˆฅโ‹…โˆฅ is the Euclidean norm, 1[โ‹…] is 1 if the condition is true and 0 otherwise, โŠ• and โŠ– denote addition and

c 2011 IEEE 0090-6778/11$25.00 โƒ

WANG et al.: NEAR-IDEAL ๐‘€-ARY LDGM QUANTIZATION WITH RECOVERY

subtraction modulo-2 or modulo-๐‘€ (should be clear from context), (โ‹…)โ„ is the modulo-๐‘€ operation on a real number into โ„ โ‰œ [โˆ’๐‘€/2, ๐‘€/2), and (โ‹…)โ„ ๐‘› is the element-wise modulooperation into โ„ ๐‘› . In the BP algorithm, symbols like ๐œ†b๐‘– , b ๐œ‡bc ๐‘–๐‘  , ๐œˆ๐‘– are binary or ๐‘€ -ary probability tuples representing the priors, messages and extrinsic probabilities in BP; each ๐‘€ -ary probability tuple ๐œ‡ is a tuple of ๐‘€ real numbers representing a probability distribution over {0, . . . , ๐‘€ โˆ’ 1}, with each component denoted by ๐œ‡(๐‘ข), ๐‘ข = 0, . . . , ๐‘€ โˆ’1. For conciseness, all probability tuples are implicitly normalized; that is, when we define an ๐‘€ -ary probability tuple ๐œ‡ by writing ๐œ‡(๐‘ข) = ๐œ‡๐‘ข , ๐‘ข = 0, . . . , ๐‘€ โˆ’ 1, it actually means that ๐œ‡(๐‘ข) = ๐œ‡๐‘ข /(๐œ‡0 + โ‹… โ‹… โ‹… + ๐œ‡๐‘€โˆ’1 ), and later mentions of ๐œ‡(๐‘ข) refer to these normalized components. For ๐‘ข = 0, . . . , ๐‘€ โˆ’ 1, ๐‘ข denotes the โ€œsure-๐‘ขโ€ probability tuple with ๐‘ข(๐‘ข) = 1 and all 1 1 other components being zero, while โˆ— = ( ๐‘€ ,..., ๐‘€ ) denotes the โ€œunknownโ€ probability tuple. For ๐‘€ -ary probability tuples ๐œ‡โ€ฒ and ๐œ‡โ€ฒโ€ฒ , ๐œ‡โ€ฒ โŠ™ ๐œ‡โ€ฒโ€ฒ and ๐œ‡โ€ฒ โŠ• ๐œ‡โ€ฒโ€ฒ are also ๐‘€ -ary probability tuples (๐œ‡โ€ฒ โŠ™ ๐œ‡โ€ฒโ€ฒ )(๐‘ข) โ‰œ ๐œ‡โ€ฒ (๐‘ข)๐œ‡โ€ฒโ€ฒ (๐‘ข) and (๐œ‡โ€ฒ โŠ• ๐œ‡โ€ฒโ€ฒ )(๐‘ข) โ‰œ โˆ‘๐‘€โˆ’1with โ€ฒ โ€ฒ โ€ฒโ€ฒ โ€ฒ ๐‘ขโ€ฒ =0 ๐œ‡ (๐‘ข )๐œ‡ (๐‘ข โŠ– ๐‘ข ), ๐‘ข = 0, . . . , ๐‘€ โˆ’ 1, similar to the variable-node and check-node operations in LDPC literature. The informativeness of an ๐‘€ -ary probability tuple ๐œ‡ is measured in bits according to ๐ผ(๐œ‡) โ‰œ log ๐‘€ +

๐‘€โˆ’1 โˆ‘

๐œ‡(๐‘ข) log ๐œ‡(๐‘ข);

(1)

๐‘ข=0

for example, among binary probability tuples, 0 and 1 are the most informative and โˆ— is the least. Often we will have a random ๐‘€ -ary probability tuple ๐œ‡ and a random variable ๐‘ขโˆ— โˆˆ {0, . . . , ๐‘€ โˆ’ 1}, where ๐‘(๐‘ขโˆ— ) is uniform and ๐‘(๐œ‡ โˆฃ ๐‘ขโˆ— ) satisfies, for any deterministic ๐œ‡โ€ฒ , ๐‘(๐œ‡ = ๐œ‡โ€ฒ โˆฃ ๐‘ขโˆ— = 0) = ๐‘(๐œ‡ = ๐œ‡โ€ฒ โŠ• ๐‘ข โˆฃ ๐‘ขโˆ— = ๐‘ข), ๐‘ข = 0, . . . , ๐‘€ โˆ’ 1, (2) due to the dithering performed below. ๐‘(๐œ‡ โˆฃ ๐‘ขโˆ— ) can then be fully characterized by ๐‘(๐œ‡ โˆฃ ๐‘ขโˆ— = 0), and we extend similar notions in LDPC analysis [17] and call the latter the density of ๐œ‡ with respect to ๐‘ขโˆ— , and say ๐œ‡ has a symmetric density (or simply is symmetric) w.r.t. ๐‘ขโˆ— if, for any deterministic ๐œ‡โ€ฒ , ๐‘(๐‘ขโˆ— = ๐‘ข โˆฃ ๐œ‡ = ๐œ‡โ€ฒ ) = ๐œ‡โ€ฒ (๐‘ข),

๐‘ข = 0, . . . , ๐‘€ โˆ’ 1.

(3)

Eq. (3) implies that the mutual information ] [๐‘€โˆ’1 โˆ‘ โˆ— โˆ— โˆ— ๐‘(๐‘ข โˆฃ ๐œ‡) log ๐‘(๐‘ข โˆฃ ๐œ‡) ๐ผ(๐‘ข ; ๐œ‡) = log ๐‘€ + E๐œ‡

(4)

๐‘ขโˆ— =0

is equal to E [๐ผ(๐œ‡)]; if E [๐ผ(๐œ‡)] > ๐ผ(๐‘ขโˆ— ; ๐œ‡), ๐œ‡ is then said to be over-confident w.r.t. ๐‘ขโˆ— . II. OVERVIEW OF THE Q UANTIZATION P ROBLEM AND A LGORITHM We consider MSE quantization with reconstruction alphabet size ๐‘€ โ‰œ 2๐พ , which is equivalent to the follows [16]: given ๐‘€ , rate ๐‘… and block length ๐‘›, design a codebook ๐’ฐ โŠ† {0, . . . , ๐‘€ โˆ’ 1}๐‘› with โˆฃ๐’ฐโˆฃ = 2๐‘›b and ๐‘›b โ‰œ ๐‘›๐‘…, as well as a quantization algorithm that quantizes a source sequence ๐’š ห† uniformly distributed in [0, ๐‘€ )๐‘› into a ๐’– โˆˆ ๐’ฐ, such that the modulo-๐‘€ MSE ๐œŽ 2 โ‰œ E [๐‘‘(ห† ๐’š , ๐’–)] with

1831

ฮผcu sj ฮผbc is

c1

u1

c2 b1

c3

b2

u2

c4

c2nโˆ’1

bnb ฮผcb si

c2n

un

ฮผuc js Fig. 1. The factor graph of the LDGM quantizer when ๐‘€ = 4. Since the scrambling sequence ๐’‚ has already been fixed when the quantization algorithm is run, the corresponding variable nodes can be omitted. 2

๐‘‘(ห† ๐’š , ๐’–) โ‰œ ๐‘›1 โˆฅ(ห† ๐’š โˆ’ ๐’–)โ„ ๐‘› โˆฅ is minimized. As ๐‘› โ†’ โˆž, ๐œŽโˆ—2 (๐‘…) โ‰œ (2๐œ‹๐‘’(2๐‘… /๐‘€ )2 )โˆ’1 is a lower bound of ๐œŽ 2 that becomes tight as ๐‘€ increases (see [13] and [16, Sec. II]), so we define 10 log10 (๐œŽ 2 /๐œŽโˆ—2 (๐‘…)) as the shaping loss in decibels. The ๐‘€ -ary LDGM codebook is constructed in a way similar to [16, Sec. VII]. We let ๐‘ฎ be an ๐‘›b ร— ๐‘›c (๐‘›c โ‰œ ๐พ๐‘›) binary low-density generator matrix from a suitably optimized LDGM code ensemble, ๐œน โˆˆ {0, . . . , ๐‘€ โˆ’ 1}๐‘› and ๐’‚ โˆˆ {0, 1}๐พ๐‘› be predetermined i.i.d. uniform dithering and scrambling sequences known at both the encoder and the decoder. For each ๐’ƒ โˆˆ {0, 1}๐‘›b , we divide ๐’„ โ‰œ ๐’„(๐’ƒ, ๐’‚) โ‰œ ๐’ƒ๐‘ฎ โŠ• ๐’‚ into ๐‘› sub-sequences ๐’„หœ๐‘— โ‰œ (๐‘๐‘—1 , . . . , ๐‘๐‘—๐พ ) with ๐‘—๐‘˜ โ‰œ ๐พ(๐‘— โˆ’ 1) + ๐‘˜, ๐‘— = 1, . . . , ๐‘›, and map each resulting ๐’„หœ๐‘— into one ๐‘€ -ary ๐’„๐‘— ) with ๐œ™๐‘— (โ‹…) โ‰œ ๐œ™(โ‹…) โŠ• ๐›ฟ๐‘— being the symbol ๐‘ข๐‘— = ๐œ™๐‘— (หœ dithered version of Gray mapping ๐œ™(โ‹…); this yields a codeword ๐’– โ‰œ ๐’–(๐’ƒ, ๐’‚), and all 2๐‘›b codewords from the given ๐‘ฎ, ๐œน and ๐’‚ form the codebook ๐’ฐ. Fig. 1 shows the factor graph [18] describing the code; as in [16], the variable nodes for all ๐‘๐‘– โ€™s have the same right-degree, denoted by ๐‘‘b . The BP messages used in the algorithm, denoted by e.g. ๐œ‡bc ๐‘–๐‘  , are also depicted in the figure; the two subscripts are the indices of its source and destination nodes. Given parameter ๐‘ก > 0, the quantization algorithm carries out BP with a priori probabilities หœ u (๐‘ข) = ๐‘’โˆ’๐‘ก(หœ๐‘ฆ๐‘— โˆ’๐‘ข)2โ„ /๐‘„(หœ ๐œ† ๐‘ฆ๐‘— ), ๐‘ข = 0, . . . , ๐‘€ โˆ’ 1 (5) ๐‘— โˆ‘๐‘€โˆ’1 โˆ’๐‘ก(หœ๐‘ฆ๐‘— โˆ’๐‘ข)2 โ„ is the nor๐‘ฆ๐‘— ) โ‰œ for each ๐‘ข๐‘— , where ๐‘„(หœ ๐‘ข=0 ๐‘’ malization factor that will henceforth be omitted as noted in the conventions, and ๐’š หœ is an adjusted version of the source sequence ๐’š ห† using the proposed recovery algorithm. To make BP converge, a number of decimation steps are performed in each iteration, each of which fixes a certain ๐‘๐‘–โˆ— to a hard decision ๐‘โˆ— โˆˆ {0, 1} by setting its prior ๐œ†b๐‘–โˆ— to the sure message ๐‘โˆ— . The choice of ๐‘–โˆ— and ๐‘โˆ— in each decimation step is made according to the extrinsic probabilities ๐œˆ๐‘–b of the ๐‘๐‘– โ€™s from BP. Specifically, we can use either the greedy decimator (GD), which chooses ๐‘–โˆ— and ๐‘โˆ— among the undecimated positions such that ๐œˆ๐‘–bโˆ— (๐‘โˆ— ) is maximized, or the typical (probabilistic) decimator (TD), which chooses an undecimated bit index ๐‘–โˆ—

1832

IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 59, NO. 7, JULY 2011

randomly and then ๐‘โˆ— โˆˆ {0, 1} with probability proportional to ๐œˆ๐‘–bโˆ— (๐‘โˆ— ); while GD performs better in practice, TD is more amenable to analysis and thus used there. For conciseness, the quantization algorithms using TD/GD with recovery will henceforth be called TD-R and GD-R, respectively, while if recovery is not done (i.e. with ๐’šหœ made the same as ๐’šห†) they are called plain TD or GD. The number of decimation steps performed in each iteration, โŒฉ known โŒช as the pace of decimation, is controlled via ๐ผbc โ‰œ ๐ผ(๐œ‡bc ) where โŸจโ‹…โŸฉ means averaging ๐‘–๐‘  over the subscripts ๐‘– and ๐‘ . Based on the analysis in [16, Sec. VI-D], in each iteration we simply carry out decimation until ๐ผbc increases by at least ๐‘‘b โˆ’2 2(๐‘‘b โˆ’ 1) ๐‘‘๐ผbc = (1 โˆ’ ๐ผbc ) 2(๐‘‘b โˆ’1) ๐‘‘๐‘™ ๐ฟ0 ๐‘‘b

(6)

since the last iteration, which ensures that the algorithm finishes within ๐ฟ0 iterations. When the block length ๐‘› is small, the actual iteration count ๐ฟ may be significantly smaller than ๐ฟ0 ; a โ€œthrottlingโ€ mechanism has been used in [16] to reduce the shaping loss in this case at the cost of higher ๐ฟ, but since it does not improve the performance-speed tradeoff significantly, it is not adopted here for simplicity. The entire quantization algorithm used in this paper can now be outlined in Fig. 2, which is largely the same as that in [16, Sec. VII]. The main difference is the introduction of the recovery algorithm at the beginning of each iteration, which is the focus of this paper. As the recovery algorithm หœ u โ€™s must be updated recomputes ๐’šหœ in every iteration, the ๐œ† ๐‘— accordingly, rather than being computed once before BP; to reduce the computational cost, (5) is pre-evaluated for different ๐‘ฆหœ๐‘— at step size 1/32 and then approximated via linear interpolation, which turns out to have negligible performance impact. Another difference is that the ๐œ‡cu ๐‘ ๐‘— โ€™s are now updated after the decimation steps, unlike in [16] where they were cu updated at the same time as the ๐œ‡cb ๐‘ ๐‘– โ€™s. This allows the ๐œ‡๐‘ ๐‘— โ€™s to include the information in the ๐œ‡bc ๐‘–๐‘  โ€™s from the current iteration, thus making the ๐œˆ๐‘—u โ€™s used in recovery in the next iteration more up-to-date; the importance of this will be explained in Section III-D. Finally, the introduction of ๐œน and ๐’‚, which is necessary in analysis, also leads to some minor changes. III. T HEORETICAL A RGUMENTS FOR R ECOVERY A. Analysis Framework for the Decimation Process Decimation in the LDGM quantization algorithm is both a significant obstacle in the algorithmโ€™s analysis and the main reason for the use of recovery. The analysis in [16] was for TD and under ideal decimation in the sense defined below, which was sufficient for the purpose of degree distribution optimization. On the other hand, when designing a recovery algorithm, a good understanding of the behavior under nonideal decimation is necessary, so we will attempt below to extend the ideas in [16] to model this. To avoid complications, we assume without loss of generality that the bits in ๐’ƒ are decimated by TD sequentially rather than in a random order, and the number of decimation steps carried out in each iteration is likewise assumed to be deterministic over the ensemble defined below, unaffected by statistical fluctuations in ๐ผbc .

Input: Quantizer parameters ๐‘ฎ, ๐œน, ๐’‚, ๐‘ก, source sequence ๐’š ห† Output: Quantized codeword ๐’– = ๐’–(๐’ƒ, ๐’‚) labeled by ๐’ƒ cu ๐œ‡bc ๐‘–๐‘—๐‘˜ โ‡ โˆ—, ๐œ‡๐‘—๐‘˜ ๐‘— โ‡ โˆ—, ๐‘– = 1, . . . , ๐‘›b , ๐‘— = 1, . . . , ๐‘›, ๐‘˜ = 1, . . . , ๐พ ๐œ†b๐‘– โ‡ โˆ—, ๐‘– = 1, . . . , ๐‘›b โ„ฐ โ‡ {1, 2, . . . , ๐‘›b } {the set of bits not yet decimated} repeat {beliefโˆpropagation iteration} cu ๐‘ ), ๐‘— = 1, . . . , ๐‘›, ๐‘ข = ๐œ™ (หœ ๐œˆ๐‘—u (๐‘ข) โ‡ ๐พ ๐‘— ๐’„) = 0, . . . , ๐‘€ โˆ’ 1 ๐‘˜ ๐‘˜=1 ๐œ‡๐‘—๐‘˜ ๐‘— (หœ Compute ๐’š หœ from ๐’š ห† and the ๐œˆ๐‘—u โ€™s using the recovery algorithm หœ u (๐‘ข) โ‡ ๐‘’โˆ’๐‘ก(๐‘ฆหœ๐‘— โˆ’๐‘ข)2โ„ , ๐‘— = 1, . . . , ๐‘›, ๐‘ข = 0, . . . , ๐‘€ โˆ’ 1 ๐œ† ๐‘— for ๐‘— = 1 to ๐‘› โˆ‘ and ๐‘˜ = 1 to ๐พ do โˆ หœu ๐œ‡uc ๐’„)) ๐‘˜โ€ฒ โˆ•=๐‘˜ ๐œ‡cu ๐‘๐‘˜โ€ฒ ), ๐‘ = 0, 1 ๐’„ หœ:หœ ๐‘๐‘˜ =๐‘ ๐œ†๐‘— (๐œ™๐‘— (หœ ๐‘—๐‘—๐‘˜ (๐‘) โ‡ ๐‘—๐‘˜โ€ฒ ๐‘— (หœ end for for ๐‘  = ๐‘—๐‘˜ = 1 to ๐‘›c do( ) uc bc cb ๐œ‡cb bc โˆ–{๐‘–} ๐œ‡๐‘–โ€ฒ ๐‘  , ๐‘– โˆˆ ๐’ฉ๐‘ โ‹… ๐‘ ๐‘– โ‡ (๐œ‡๐‘—๐‘  โŠ• ๐‘Ž๐‘  ) โŠ• โŠ•๐‘–โ€ฒ โˆˆ๐’ฉโ‹…๐‘  end for for ๐‘– = 1 to ๐‘›b (do ) b cb bc ๐œ‡bc ๐‘–๐‘  โ‡ ๐œ†๐‘– โŠ™ โŠ™๐‘ โ€ฒ โˆˆ๐’ฉ cb โˆ–{๐‘ } ๐œ‡๐‘ โ€ฒ ๐‘– , ๐‘  โˆˆ ๐’ฉ๐‘–โ‹… โ‹…๐‘–

๐œˆ๐‘–b โ‡ โŠ™๐‘ โ€ฒ โˆˆ๐’ฉ cb ๐œ‡cb โ€ฒ โ‹…๐‘– ๐‘  ๐‘– end for while โ„ฐ โˆ•= โˆ… and more decimation is necessary in this iteration do Choose the bit index ๐‘–โˆ— to decimate and its value ๐‘โˆ— bc โˆ— โˆ— ๐œ†b๐‘–โˆ— โ‡ ๐‘โˆ— , ๐œ‡bc ๐‘–โˆ— ๐‘  โ‡ ๐‘ , ๐‘  โˆˆ ๐’ฉ๐‘–โˆ— โ‹… {decimate ๐‘๐‘– to ๐‘ } โ„ฐ โ‡ โ„ฐโˆ–{๐‘–โˆ— } end while ( ) bc ๐œ‡cu bc ๐œ‡๐‘–โ€ฒ ๐‘  , ๐‘  = ๐‘—๐‘˜ = 1, . . . , ๐‘›c ๐‘ ๐‘— โ‡ ๐‘Ž๐‘  โŠ• โŠ•๐‘–โ€ฒ โˆˆ๐’ฉโ‹…๐‘  until โ„ฐ = โˆ… ๐‘๐‘– โ‡ 0 (resp. 1) if ๐œ†b๐‘– = 0 (or 1), ๐‘– = 1, . . . , ๐‘›b

Fig. 2. A brief outline of the quantization algorithm. ๐’ฉ๐‘–โ‹…bc = ๐’ฉโ‹…๐‘–cb is the cb = ๐’ฉ bc is the positions of 1โ€™s in the ๐‘–-th row of ๐‘ฎ, and similarly ๐’ฉ๐‘ โ‹… โ‹…๐‘  positions of 1โ€™s in the ๐‘ -th column of ๐‘ฎ.

The analysis uses a fully-adjusted sequence ๐’š, obtained from an idealized recovery algorithm described in Secหœ tion III-C, in place of ๐’š หœ in (5) to define the priors ๐œ†u๐‘— โ€™s, while ๐’š u หœ and the ๐œ†๐‘— โ€™s used in practice are regarded as approximations. The ๐œˆ๐‘–b used in each TD decimation step then approximates the corresponding โ€œtrueโ€ extrinsic probabilities โˆ‘ โˆ ๐œ†u๐‘— (๐‘ข๐‘— (๐’ƒโ€ฒ , ๐’‚โ€ฒ )) ๐œˆ๐‘–bโˆ— (๐‘) โ‰œ (๐’ƒโ€ฒ ,๐’‚โ€ฒ )โˆˆ๐’ž ๐‘— ๐‘โ€ฒ๐‘– =๐‘

=

โˆ‘

โ€ฒ

โ€ฒ

๐‘’โˆ’๐‘›๐‘ก๐‘‘(๐’š,๐’–(๐’ƒ ,๐’‚ )) ,

(7)

(๐’ƒโ€ฒ ,๐’‚โ€ฒ )โˆˆ๐’ž ๐‘โ€ฒ๐‘– =๐‘

where ๐’ž is defined as the set of (๐’ƒโ€ฒ , ๐’‚โ€ฒ ) consistent with previous decimation choices, meaning here that ๐’‚โ€ฒ is equal to the scrambling sequence ๐’‚, while each ๐‘โ€ฒ๐‘–โ€ฒ is equal to ๐‘โˆ— โˆˆ {0, 1} if ๐‘๐‘–โ€ฒ has previously been decimated to ๐‘โˆ— (i.e. ๐œ†b๐‘–โ€ฒ = ๐‘โˆ— ), but can be either 0 or 1 if ๐‘๐‘–โ€ฒ has not been decimated yet (i.e. ๐œ†b๐‘–โ€ฒ = โˆ—). The i.i.d. choice of each ๐‘Ž๐‘  in the scrambling sequence ๐’‚, which determines the codebook ๐’ฐ, can likewise be viewed as a TD decimation step using โˆ— as the extrinsic probabilities. Assuming for convenience that ๐‘Ž1 , ๐‘Ž2 , . . . , ๐‘Ž๐‘›c are chosen sequentially, the corresponding true extrinsic probabilities for the choice of ๐‘Ž๐‘  can similarly be given by โˆ‘ โˆ ๐œ†u๐‘— (๐‘ข๐‘— (๐’ƒโ€ฒ , ๐’‚โ€ฒ )), (8) ๐œˆ๐‘ aโˆ— (๐‘Ž) โ‰œ (๐’ƒโ€ฒ ,๐’‚โ€ฒ )โˆˆ๐’ž ๐‘— ๐‘Žโ€ฒ๐‘  =๐‘Ž

where ๐’ž is still the set of (๐’ƒโ€ฒ , ๐’‚โ€ฒ ) consistent with previous decimation choices, i.e. ๐‘Žโ€ฒ๐‘ โ€ฒ = ๐‘Ž๐‘ โ€ฒ for ๐‘ โ€ฒ < ๐‘ , while the other bits in ๐’‚โ€ฒ as well as the entire ๐’ƒโ€ฒ are arbitrary. We can thus define the true typical decimator (TTD) as an idealized version

WANG et al.: NEAR-IDEAL ๐‘€-ARY LDGM QUANTIZATION WITH RECOVERY

1833

of TD that uses respectively ๐œˆ๐‘ aโˆ— and ๐œˆ๐‘–bโˆ— in the choice of each ๐‘Ž๐‘  and ๐‘๐‘– , and say ideal decimation has been performed if all the bits in ๐’‚ and ๐’ƒ decimated so far have used the TTD.1 Using TTD in all decimation steps yields each ๐’– โˆˆ {0, . . . , ๐‘€ โˆ’ 1}๐‘› with probability proportional to ๐‘’โˆ’๐‘›๐‘ก๐‘‘(๐’š,๐’–) , giving

with recovery, and say a certain ๐‘(โ‹…) is ideal if it matches the corresponding ๐‘ยฏ(โ‹…). Now the reference codeword is simply the result of TTD after all decimation steps, so

โˆซ

๐‘€โˆ’1 โˆ‘

โˆ’๐‘ก(๐‘ฆโˆ’๐‘ข)2โ„

๐‘€ 1 ๐‘’ E [๐‘‘(๐’š, ๐’–)] = ๐œŽ๐‘ก2 โ‰œ ๐‘‘๐‘ฆ (๐‘ฆ โˆ’ ๐‘ข)2โ„ โ‹… ๐‘€ 0 ๐‘„(๐‘ฆ) ๐‘ข=0 โˆซ ๐‘€/2 2 ๐‘’โˆ’๐‘ก๐‘ง ๐‘‘๐‘ง = ๐‘ง2 โ‹… ๐‘„(๐‘ง) โˆ’๐‘€/2

โˆ— 2

๐‘ยฏ(๐‘ขโˆ—๐‘— โˆฃ ๐‘ฆ๐‘— ) = ๐‘ยฏ(๐‘ขโˆ—๐‘— โˆฃ ๐’š) = ๐‘’โˆ’๐‘ก(๐‘ฆ๐‘— โˆ’๐‘ข๐‘— )โ„ /๐‘„(๐‘ฆ๐‘— ),

which implies ๐‘ยฏ(๐‘ขโˆ—๐‘— โˆฃ ๐œ†u๐‘— ) = ๐œ†u๐‘— (๐‘ขโˆ—๐‘— ), i.e. ๐œ†u๐‘— has a symmetric density w.r.t. ๐‘ขโˆ—๐‘— . As ๐‘ยฏ(๐‘ฆ๐‘— ) is uniform, this density can be computed from โˆ— 2

๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— ) = ๐‘’โˆ’๐‘ก(๐‘ฆ๐‘— โˆ’๐‘ข๐‘— )โ„ /๐‘„(๐‘ฆ๐‘— ).

(9)

(denoted by ๐‘ƒ๐‘ก in [16]); for example, if ๐’š = ๐’šห†, then the achieved MSE is simply ๐œŽ๐‘ก2 . TD can yield the same result if each ๐œˆ๐‘ aโˆ— can be made equal to โˆ— and each ๐œˆ๐‘–bโˆ— equal to the ๐œˆ๐‘–b from BP. Similar approaches have been used in [19] and [3] to analyze the decimation process in satisfiability problems and quantization with polar codes. We now consider an ensemble of quantizer instances with a probability measure over it, so that probabilistic analytical approaches including density evolution (DE) can be carried out. Recall that in the DE analysis of the LDPC decoder [17], the probabilities are defined over the possible channel realizations and the LDPC code ensemble with a specific degree distribution, and DE is performed in reference to the transmitted codeword (usually assumed to be all-zero). In contrast, when analyzing the LDGM quantizer with TD or TTD, each quantizer instance in the ensemble has not only a specific ๐’šห†, a ๐‘ฎ in the LDGM code ensemble, a dither ๐œน but also a specific sequence of values used as the random source for decimation of both ๐’‚ and ๐’ƒ; for example, for each ๐‘๐‘– the random source can be denoted by an i.i.d. uniform variable ๐œ”๐‘–b over [0, 1), such that ๐‘๐‘– is decimated to 1[๐œ”๐‘–b โ‰ฅ ๐œˆหœ๐‘–b (0)], ๐œˆหœ๐‘–b being ๐œˆ๐‘–b for TD and ๐œˆ๐‘–bโˆ— for TTD. Over this ensemble, ๐’š as well as all the priors, BP messages, etc. are now random variables whose distributions depend on whether TD or TTD (ideal decimation) is used, and whether recovery is performed. The reference codeword used by DE, denoted by (๐’ƒโˆ— , ๐’‚โˆ— ) or the corresponding ๐’–โˆ— โ‰œ ๐’–(๐’ƒโˆ— , ๐’‚โˆ— ), is defined as the quantization result that would be obtained if previous decimation choices were kept and the remaining decimation steps used TTD. Therefore, (๐’ƒโˆ— , ๐’‚โˆ— ) is random (specific to each quantizer instance); given ๐’š and ๐’ž, it is any (๐’ƒโ€ฒ , ๐’‚โ€ฒ ) โˆˆ ๐’ž โ€ฒ โ€ฒ with probability proportional to ๐‘’โˆ’๐‘›๐‘ก๐‘‘(๐’š,๐’–(๐’ƒ ,๐’‚ )) . Over the entire ensemble, by definition it remains unchanged after a decimation step using TTD, but is otherwise also specific to each decimation step. B. Analysis of Ideal Decimation The analysis of a decimation step under ideal decimation (in previous steps), as was done in the DE analysis for degree distribution optimization [16, Sec. V], can be expressed in the above framework as follows. As recovery is not necessary in this case, ๐’š is identical to ๐’š ห†. For clarity, we use ๐‘ยฏ(โ‹…) to denote a pdf under ideal decimation, ๐‘(โ‹…) when TD is used 1 Note that the definition of TTD here also encompasses the concept of true typical source generator (TTSG) introduced in [16, Sec. V-A], by viewing the choice of bits in ๐’‚ as decimation steps as well.

(10)

(11)

For asymptotically large ๐‘›, DE can then be carried out to obtain the densities of the ๐œˆ๐‘ a โ€™s and ๐œˆ๐‘–b โ€™s after any given number of BP iterations with different initial symmetric densities of BP messages. The corresponding ๐œˆ๐‘ aโˆ— and ๐œˆ๐‘–bโˆ— are bound by these densities in terms of physical degradation, making it possible to evaluate how close each ๐œˆ๐‘ aโˆ— matches โˆ— and each ๐œˆ๐‘–bโˆ— matches ๐œˆ๐‘–b . For given degree distributions, this DE analysis yields a monotonicity threshold ๐‘กthr such that, as long as ๐‘ก โ‰ค ๐‘กthr , when the iteration count ๐ฟ goes to infinity along with ๐‘› (i.e. decimation is performed slowly), the mean-square difference between each ๐œˆ๐‘ aโˆ— , ๐œˆ๐‘–bโˆ— and respectively โˆ—, ๐œˆ๐‘–b goes to zero. In other words, assuming ideal decimation in previous steps, when the current decimation step is carried out using TD after a sufficiently large number of BP iterations, it will yield the same result as the TTD with high probability and thus be ideal as well. This MSE ๐œŽ๐‘ก2thr potentially achievable by TD, as defined in (9), can be very close to the theoretical limit ๐œŽโˆ—2 (๐‘…) when using degree distributions optimized for a high ๐‘กthr . The extrinsic probabilities ๐œˆ๐‘—u of each ๐‘ข๐‘— , given by ๐œˆ๐‘—u (๐‘ข) โ‰œ

๐พ โˆ

๐œ‡cu ๐‘๐‘˜ ), ๐‘—๐‘˜ ๐‘— (หœ

๐‘ข = ๐œ™๐‘— (หœ ๐’„) = 0, . . . , ๐‘€ โˆ’ 1, (12)

๐‘˜=1

are used by the proposed recovery algorithm as a summarization of the BP messages. The corresponding true extrinsic probabilities can be defined as โˆ‘ โˆ ๐œ†u๐‘— โ€ฒ (๐‘ข๐‘— โ€ฒ (๐’ƒโ€ฒ , ๐’‚โ€ฒ )). (13) ๐œˆ๐‘—uโˆ— (๐‘ข) โ‰œ (๐’ƒโ€ฒ ,๐’‚โ€ฒ )โˆˆ๐’ž ๐‘— โ€ฒ โˆ•=๐‘— ๐‘ข๐‘— (๐’ƒโ€ฒ ,๐’‚โ€ฒ )=๐‘ข

As the ๐œ†u๐‘— โ€ฒ โ€™s have symmetric densities, it is easy to prove that each ๐œˆ๐‘—uโˆ— is symmetric w.r.t. ๐‘ขโˆ—๐‘— as well, i.e. ๐‘ยฏ(๐‘ขโˆ—๐‘— โˆฃ ๐œˆ๐‘—uโˆ— ) = ๐œˆ๐‘—uโˆ— (๐‘ขโˆ—๐‘— ); ๐‘ฆ๐‘— (or ๐œ†u๐‘— ) โ€” ๐‘ขโˆ—๐‘— โ€” ๐œˆ๐‘—uโˆ— can also be shown to form a Markov chain, as ๐œˆ๐‘—uโˆ— only depends on the ๐‘ฆ๐‘— โ€ฒ โ€™s with ๐‘— โ€ฒ โˆ•= ๐‘—. Using (11), we thus have ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—uโˆ— ) = =

๐‘€โˆ’1 โˆ‘ ๐‘ข=0 ๐‘€โˆ’1 โˆ‘

๐‘ยฏ(๐‘ขโˆ—๐‘— = ๐‘ข โˆฃ ๐œˆ๐‘—uโˆ— ) โ‹… ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— = ๐‘ข) (14) ๐œˆ๐‘—uโˆ— (๐‘ข)

โ‹…

๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘—

= ๐‘ข).

๐‘ข=0

Since the neighborhood in the factor graph involved in computing ๐œˆ๐‘—u is loop-free with high probability as ๐‘› โ†’ โˆž, for asymptotically large ๐‘›, ๐œˆ๐‘—u can be shown to be symmetric w.r.t. ๐‘ขโˆ—๐‘— and ๐‘ฆ๐‘— โ€” ๐‘ขโˆ—๐‘— โ€” ๐œˆ๐‘—u is a Markov chain as well, so (14) remains true with ๐œˆ๐‘—uโˆ— replaced by ๐œˆ๐‘—u .

1834

C. Decimation Errors and Idealized Recovery Plain TD gives poor quantization performance in practice, and its reason can be understood as follows. The finite ๐‘› and ๐ฟ used in practice leaves a small but finite mean-square difference between ๐œˆ๐‘–bโˆ— and ๐œˆ๐‘ aโˆ— and respectively ๐œˆ๐‘–b and โˆ—, so each TD decimation step has a finite probability to give a result different from that of TTD, which we call decimation errors for convenience. Without recovery (๐’š = ๐’š ห†), such erroneous decimations will usually not favor codewords close to ๐’šห† as much as the TTD, and after ๐’ž shrinks due to decimation, the subsequent reference codeword will, on average, have a larger distance to ๐’š ห†. The quantization error (ห† ๐’š โˆ’ ๐’–โˆ— )โ„ ๐‘› is analogous to the noise in the LDPC decoder; its increase in magnitude will change the density of the ๐œ†u๐‘— โ€™s compared to ideal decimation, usually making them over-confident, which will slow down BP convergence and reduce the informativeness of the ๐œˆ๐‘–b โ€™s in future BP iterations, thus worsening the quality of future decimation choices. The recovery algorithm intends to allow BP to recover from decimation errors. It is inspired by the analysis of the similar issue in binary erasure quantization (BEQ) [16, Sec. VI-B], where erroneous decimations are evident as contradictions among the BP messages and the source sequence ๐’šห†, and can similarly result in less informative ๐œˆ๐‘–b โ€™s in the future. In that case, a solution is to flip the contradictory bits in the source sequence ๐’šห† and use the resulting flipped sequence ๐’š หœ to generate the ๐œ†u๐‘— โ€™s for subsequent iterations (also reminiscent of the approach in [12]). These flipped bits intuitively represent errors already made which we do not attempt to fix, and flipping them can be shown to make future ๐œˆ๐‘–b โ€™s as informative as if no decimation errors occurred, so fewer decimation errors will be made in the future. The proposed recovery algorithm for ๐‘€ -ary MSE quantization uses a similar idea, which can be better explained by first considering an idealized recovery algorithm that defines the ๐’š above and is approximated by the actual one. Whenever ๐’ž and thus the reference codeword ๐’–โˆ— changes due to decimation errors, the algorithm adjusts ๐’š accordingly to make each ๐‘(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— ) over the ensemble identical to its ideal version (11). This keeps the density of each ๐œ†u๐‘— generated from this ๐‘ฆ๐‘— , and thus that of each ๐œˆ๐‘–b computed from these ๐œ†u๐‘— โ€™s in subsequent iterations, identical to their counterparts under ideal decimation (which are computable via DE for asymptotically large ๐‘›), so decimation errors no longer degrade the quality of future decimations. When the quantization algorithm finishes after decimating all bits in ๐’ƒ, ๐’ž will contain only the resulting (๐’ƒ, ๐’‚), which will also be the reference codeword then; since ๐‘(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— ) is ideal, E [๐‘‘(๐’š, ๐’–)] = E [๐‘‘(๐’š, ๐’–โˆ— )] is also equal to the ideal ๐œŽ๐‘ก2 , while the difference between ๐’šห† and ๐’š correspond to decimation errors that cause the actual MSE ๐’š , ๐’–)] to be higher. ๐œŽ 2 = E [๐‘‘(ห† While adjustment of each ๐‘ฆ๐‘— causes ๐‘ขโˆ—๐‘— to change as well, the desired ๐‘(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— ) can still be achieved in idealized recovery using ๐œˆ๐‘—uโˆ— from (13), which may be approximated by ๐œˆ๐‘—u in practice. To see this, recall that given ๐’š and ๐’ž, the reference codeword is any (๐’ƒโ€ฒ , ๐’‚โ€ฒ ) โˆˆ ๐’ž with probability โ€ฒ โ€ฒ proportional to ๐‘’โˆ’๐‘›๐‘ก๐‘‘(๐’š,๐’–(๐’ƒ ,๐’‚ )) , so ignoring normalization,

IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 59, NO. 7, JULY 2011

๐‘(๐‘ขโˆ—๐‘— = ๐‘ข โˆฃ ๐’š, ๐’ž) = =

โˆ‘ โ€ฒ

โ€ฒ

(๐’ƒ ,๐’‚ )โˆˆ๐’ž ๐‘ข๐‘— (๐’ƒโ€ฒ ,๐’‚โ€ฒ )=๐‘ข

โˆ

โˆ‘

โ€ฒ

โ€ฒ

๐‘’โˆ’๐‘›๐‘ก๐‘‘(๐’š,๐’–(๐’ƒ ,๐’‚ ))

(๐’ƒโ€ฒ ,๐’‚โ€ฒ )โˆˆ๐’ž ๐‘ข๐‘— (๐’ƒโ€ฒ ,๐’‚โ€ฒ )=๐‘ข

๐œ†u๐‘— โ€ฒ (๐‘ข๐‘— โ€ฒ (๐’ƒโ€ฒ , ๐’‚โ€ฒ )) = (๐œ†u๐‘— โŠ™ ๐œˆ๐‘—uโˆ— )(๐‘ข);

(15)

๐‘—โ€ฒ

that is, the dependence on ๐’š and ๐’ž can be subsumed into ๐œ†u๐‘— (a function of ๐‘ฆ๐‘— ) and ๐œˆ๐‘—uโˆ— (a function of ๐’ž and the other elements of ๐’š). ๐‘(๐‘ขโˆ—๐‘— โˆฃ ๐‘ฆ๐‘— , ๐œˆ๐‘—uโˆ— ) is thus also given by (15) and identical to the ideal ๐‘ยฏ(๐‘ขโˆ—๐‘— โˆฃ ๐‘ฆ๐‘— , ๐œˆ๐‘—uโˆ— ). Using this fact, we let ๐นยฏ๐œˆ๐‘—uโˆ— (โ‹…) and ๐น๐œˆ๐‘—uโˆ— (โ‹…) be the cdfs corresponding to respectively ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—uโˆ— ) and ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—uโˆ— ), and define ๐‘ฆ๐‘— โ‰œ ๐นยฏ๐œˆโˆ’1 ๐‘ฆ๐‘— )) uโˆ— (๐น๐œˆ uโˆ— (ห† ๐‘— ๐‘—

(16)

as the fully-adjusted version of ๐‘ฆห†๐‘— . Conditioned on ๐œˆ๐‘—uโˆ— , ๐‘ฆ๐‘— ) is uniformly distributed over [0, 1], so the appli๐น๐œˆ๐‘—uโˆ— (ห† uโˆ— cation of ๐นยฏ๐œˆโˆ’1 ยฏ(๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—uโˆ— ). We then uโˆ— (โ‹…) makes ๐‘(๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘— ) = ๐‘ ๐‘— have ๐‘(๐‘ฆ๐‘— , ๐‘ขโˆ—๐‘— โˆฃ ๐œˆ๐‘—uโˆ— ) = ๐‘ยฏ(๐‘ฆ๐‘— , ๐‘ขโˆ—๐‘— โˆฃ ๐œˆ๐‘—uโˆ— ) and ๐‘(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— , ๐œˆ๐‘—uโˆ— ) = ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— , ๐œˆ๐‘—uโˆ— ) = ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— ), the final equality implied by the Markov chain property in Section III-B, so ๐‘(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— ) now matches ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— ), as desired. It should be noted that updating a certain ๐‘ฆ๐‘— in this manner changes the corresponding ๐œ†u๐‘— and thus the ๐œˆ๐‘—uโˆ—โ€ฒ โ€™s for ๐‘— โ€ฒ โˆ•= ๐‘—, so ๐‘ฆ๐‘— โ€ฒ has to be updated again according to the new ๐‘(ห† ๐‘ฆ๐‘— โ€ฒ โˆฃ ๐œˆ๐‘—uโˆ—โ€ฒ ) in โˆ— order to maintain the desired ๐‘(๐‘ฆ๐‘— โ€ฒ โˆฃ ๐‘ข๐‘— โ€ฒ ). A rigorous definition of ๐’š will thus require an iterative sequential update process, with each step updating one ๐‘ฆ๐‘— ; examination of the BEQ case leads us to conjecture that this process will always converge to a fixed point. In any case, such details are no longer relevant when ๐œˆ๐‘—uโˆ— is approximated by ๐œˆ๐‘—u below. D. Principles of the Proposed Recovery Algorithm The proposed recovery algorithm approximates the idealized version by using ๐œˆ๐‘—u instead of ๐œˆ๐‘—uโˆ— : let ๐นยฏ๐œˆ๐‘—u (โ‹…) and ๐น๐œˆ๐‘—u (โ‹…) be the cdfs corresponding to respectively ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) and ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ), the algorithm yields, for each ๐‘—, ๐‘ฆหœ๐‘— โ‰œ ๐นยฏ๐œˆโˆ’1 ๐‘ฆ๐‘— )). u (๐น๐œˆ u (ห† ๐‘— ๐‘—

(17)

While ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) is well approximated by the asymptotic result (14) (with ๐œˆ๐‘—u in place of ๐œˆ๐‘—uโˆ— ), ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) must, in practice, be u estimated from the (ห† ๐‘ฆ๐‘— , ๐œˆ๐‘— ) samples in the actual quantizer instance. Approximating ๐œˆ๐‘—uโˆ— with ๐œˆ๐‘—u causes the following inaccuracies: u โˆ™ ๐œˆ๐‘— is computed with a finite number of BP iterations and หœ uโ€ฒ โ€™s from thus fails to include the information in the ๐œ† ๐‘— faraway nodes in the factor graph. หœuโ€ฒ โ€™s from the ๐’šหœ given by โˆ™ The priors used are the ๐œ† ๐‘— the actual recovery algorithm, which can be inaccurate compared to the ๐œ†u๐‘— โ€ฒ โ€™s from idealized recovery due to both estimation errors in ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) and the incomplete recovery issue discussed below. หœ uโ€ฒ โ€™s involved in the computation of the current ๐œˆ u โˆ™ The ๐œ† ๐‘— ๐‘— come from recovery steps in earlier iterations, which do not take more recent decimation errors into account. As the result, while (17) makes ๐‘(หœ ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) identical to ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) in the absence of estimation errors, this does not imply that

WANG et al.: NEAR-IDEAL ๐‘€-ARY LDGM QUANTIZATION WITH RECOVERY

๐‘(หœ ๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— ) will match ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— ), as is the case for idealized recovery. In general, the effect is that ๐’š หœ is insufficiently adjusted compared to ๐’š; for example, when ๐œˆ๐‘—u = โˆ—, ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) and ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) will likely both be uniform distributions, so ๐‘ฆหœ๐‘— will be identical to ๐‘ฆห†๐‘— even if ๐œˆ๐‘—uโˆ— indicates that ๐‘ฆ๐‘— should be adjusted to some other value. However, such incomplete recovery does not cause much performance degradation in practice (actually none in case of BEQ). Roughly speaking, if ๐œˆ๐‘—u is uninformative at some iteration in an quantizer instance, most of the corresponding cb outgoing BP messages ๐œ‡uc ๐‘—๐‘  and ๐œ‡๐‘ ๐‘– will be uninformative as well and not much affected by the difference between ๐‘ฆหœ๐‘— and ๐‘ฆ๐‘— due to incomplete recovery, while the affected messages are also prevented from propagating far with the use of recovery in future iterations. In effect, recovery is being carried out incrementally, adjusting ๐‘ฆหœ๐‘— toward ๐‘ฆ๐‘— as the corresponding ๐œˆ๐‘—u becomes informative. Due to the reliance on future recovery to stop the propagation of BP messages affected by incomplete recovery, it is important that recovery be carried out in every iteration, and u the ๐œ‡cu ๐‘ ๐‘— โ€™s be updated late to make the ๐œˆ๐‘— โ€™s more up-to-date (see Section II). Indeed, the algorithm often exhibits oscillatory behavior otherwise, which can affect ๐ผbc and thus disrupt the pace of decimation so much as to increase the shaping loss above 1dB. IV. T HE R ECOVERY A LGORITHM Based on the above principles, we will now explain in detail how to estimate ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) and perform the recovery using (17) with reasonable computational complexity. A. Modeling ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) Since the LDGM code ensemble is symmetric with respect to symbol permutation, ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) is the same for any ๐‘—. Its estimation must still rely on the ๐‘› (ห† ๐‘ฆ๐‘— , ๐œˆ๐‘—u ) samples from the current quantizer instance, so some assumptions must be made on it in order to reduce the number of parameters to estimate and minimize random estimation errors. Recall that ๐œˆ๐‘—u is asymptotically symmetric w.r.t. ๐‘ขโˆ—๐‘— as ๐‘› โ†’ โˆž under ideal decimation; this asymptotic symmetry should remain true under non-ideal decimation but idealized recovery, so it is reasonable to assume approximately that it holds under the actual recovery algorithm as well. Analogous to (14), we can then express ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) as ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) = =

๐‘€โˆ’1 โˆ‘ ๐‘ข=0 ๐‘€โˆ’1 โˆ‘

๐‘(๐‘ขโˆ—๐‘— = ๐‘ข โˆฃ ๐œˆ๐‘—u )๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— = ๐‘ข, ๐œˆ๐‘—u ) (18) ๐œˆ๐‘—u (๐‘ข)๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— = ๐‘ข, ๐œˆ๐‘—u ).

๐‘ข=0

In general, the deviation of ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— , ๐œˆ๐‘—u ) from the ideal โˆ— ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐‘ข๐‘— ) is related to the pace of decimation and can vary with ๐œˆ๐‘—u ; analysis of BEQ suggests that there is usually more deviation when ๐œˆ๐‘—u is informative. However, this effect is difficult to estimate because only the (ห† ๐‘ฆ๐‘— , ๐œˆ๐‘—u ) samples with u informative ๐œˆ๐‘— โ€™s are useful in the estimation. In practice, attempts to model this dependence do not improve the results, so we ignore it and model ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— , ๐œˆ๐‘—u ) as ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— , ๐œˆ๐‘—u ) = ๐‘z ((ห† ๐‘ฆ๐‘— โˆ’ ๐‘ขโˆ—๐‘— )โ„ ),

(19)

1835

such that ๐‘(ห† ๐‘ฆ๐‘— โˆฃ ๐œˆ๐‘—u ) =

๐‘€โˆ’1 โˆ‘

๐œˆ๐‘—u (๐‘ข)๐‘z ((ห† ๐‘ฆ๐‘— โˆ’ ๐‘ข)โ„ ).

(20)

๐‘ข=0

๐‘z (โ‹…) is further made to satisfy the constraints ๐‘z (๐‘ง) = ๐‘z (โˆ’๐‘ง), 0 โ‰ค ๐‘ง < ๐‘€/2, ๐‘€โˆ’1 โˆ‘

๐‘z ((๐‘ฆ โˆ’ ๐‘ข)โ„ ) = 1, ๐‘ฆ โˆˆ [0, ๐‘€ ),

(21) (22)

๐‘ข=0

where (22) corresponds to theโˆซuniformity of ๐‘(ห† ๐‘ฆ๐‘— ) and implies the normalization condition โ„ ๐‘z (๐‘ง) ๐‘‘๐‘ง = 1. Note that the ideal ๐‘ยฏ(๐‘ฆ๐‘— โˆฃ ๐‘ขโˆ—๐‘— ) in (11) also has the form of (19) with the 2 corresponding ๐‘ยฏz (๐‘ง) โ‰œ ๐‘’โˆ’๐‘ก๐‘ง /๐‘„(๐‘ง) (๐‘ง โˆˆ โ„) satisfying (21) and (22). The interval [0, ๐‘€ ) of possible ๐‘ฆห†๐‘— values is now equally divided into ๐‘€ ๐‘† subintervals ๐’ด๐‘ ๐‘ฃ , ๐‘  = 0, . . . , ๐‘† โˆ’ 1, ๐‘ฃ = 0, . . . , ๐‘€ โˆ’ 1, with each ๐’ด๐‘ ๐‘ฃ โ‰œ โˆซ [๐‘ฃ + ๐‘ /๐‘†, ๐‘ฃ + (๐‘  + 1)/๐‘†). This discretizes ๐‘z (โ‹…) into ๐‘๐‘ ๐‘ฃ โ‰œ ๐’ด๐‘ ๐‘ฃ ๐‘z ((๐‘ฆ)โ„ ) ๐‘‘๐‘ฆ, which can be further grouped into column vectors ๐’‘๐‘  โ‰œ (๐‘๐‘ ๐‘ฃ )๐‘€โˆ’1 ๐‘ฃ=0 , and the constraints above become, for all ๐‘  and ๐‘ฃ, ๐‘๐‘ ๐‘ฃ = ๐‘๐‘†โˆ’1โˆ’๐‘ ,๐‘€โˆ’1โˆ’๐‘ฃ , or ๐’‘๐‘  = R๐’‘๐‘†โˆ’1โˆ’๐‘  , ๐‘€โˆ’1 โˆ‘

๐‘๐‘ ๐‘ฃ = 1/๐‘†, or 1T ๐’‘๐‘  = 1/๐‘†,

(23) (24)

๐‘ฃ=0

where R = RT = Rโˆ’1 is such a matrix that R๐’‘ is ๐’‘ with its elements in reverse order, and 1 is the all-one vector. By (23), it is sufficient to estimate those ๐’‘๐‘  with ๐‘  = 0, . . . , ๐‘†/2 โˆ’ 1, each satisfying (24), from the ๐‘› (ห† ๐‘ฆ๐‘— , ๐œˆ๐‘—u ) samples according to the discretized version of (20), โˆ‘ ] ๐‘€โˆ’1 [ ๐œˆ๐‘—u (๐‘ข)๐‘๐‘ ,๐‘ฃโŠ–๐‘ข , Pr ๐‘ฆห†๐‘— โˆˆ ๐’ด๐‘ ๐‘ฃ โˆฃ ๐œˆ๐‘—u =

โˆ€๐‘—, ๐‘ , ๐‘ฃ.

(25)

๐‘ข=0

B. Computing the Raw Estimate of ๐‘๐‘ ๐‘ฃ Maximum-likelihood estimation using (25) turns out to be computationally intractable, so we adopt a regularized leastsquares method instead. For each[๐‘—, ๐‘  and ๐‘ฃ, the left-hand side ] of (25) is just the expectation E 1[ห† ๐‘ฆ๐‘— โˆˆ ๐’ด๐‘ ๐‘ฃ โˆฃ ๐œˆ๐‘—u ] , so we can approximate it with its sample 1[ห† ๐‘ฆ๐‘— โˆˆ ๐’ด๐‘ ๐‘ฃ ] obtained from the actual ๐‘ฆห†๐‘— , yielding 1[ห† ๐‘ฆ๐‘— โˆˆ ๐’ด๐‘ ๐‘ฃ ] =

๐‘€โˆ’1 โˆ‘

๐œˆ๐‘—u (๐‘ข)๐‘๐‘ ,๐‘ฃโŠ–๐‘ข + ๐‘ค๐‘ ๐‘—๐‘ฃ ,

โˆ€๐‘—, ๐‘ , ๐‘ฃ,

(26)

๐‘ข=0

๐‘ค๐‘ ๐‘—๐‘ฃ being the sampling error. These equations can be grouped by ๐‘  and re-expressed as ๐‘‘๐‘ ๐‘—๐‘ฃ =

๐‘€โˆ’1 โˆ‘

โ€ฒ

๐ด๐‘ฃ๐‘—๐‘ฃ ๐‘๐‘ ๐‘ฃโ€ฒ + ๐‘ค๐‘ ๐‘—๐‘ฃ , or ๐’…๐‘  = ๐‘จ๐’‘๐‘  + ๐’˜๐‘  ,

(27)

๐‘ฃ โ€ฒ =0 โ€ฒ

where ๐ด๐‘ฃ๐‘—๐‘ฃ โ‰œ ๐œˆ๐‘—u (๐‘ฃ โŠ– ๐‘ฃ โ€ฒ ), ๐‘‘๐‘ ๐‘—๐‘ฃ โ‰œ 1[ห† ๐‘ฆ๐‘— โˆˆ ๐’ด๐‘ ๐‘ฃ ], and they form respectively an ๐‘€ ๐‘›ร—๐‘€ matrix ๐‘จ and ๐‘€ ๐‘›ร—1 vectors ๐’…๐‘  . For each ๐‘—, the corresponding ๐‘€ rows of ๐‘จ and ๐’…๐‘  are denoted ๐‘จ๐‘— and ๐’…๐‘ ๐‘— , respectively; note that each ๐‘จ๐‘— is a circulant matrix that does not vary with ๐‘ .

1836

IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 59, NO. 7, JULY 2011

Since ๐’‘๐‘  = R๐’‘๐‘†โˆ’1โˆ’๐‘  , for each ๐‘  โˆˆ {0, . . . , ๐‘†/2 โˆ’ 1}, (27) for ๐‘  and ๐‘† โˆ’ 1 โˆ’ ๐‘  can be combined to yield [ ] [ ] ] [ ๐’…๐‘  ๐‘จ ๐’˜๐‘  = ๐’‘ + , (28) ๐‘จR ๐‘  ๐’…๐‘†โˆ’1โˆ’๐‘  ๐’˜๐‘†โˆ’1โˆ’๐‘  หœ ๐‘ +๐’˜ or simply ๐’…หœ๐‘  = ๐‘จ๐’‘ หœ๐‘  . The estimation of ๐’‘๐‘  is then formulated as the following regularized least-squares problem 2 1  1 หœ 2 หœ ๐‘  ๐’… minimize โˆ’ ๐‘จ๐’‘   + 2 โˆฅ๐’‘๐‘  โˆ’ ๐’‘ยฏ๐‘  โˆฅ ๐‘  2 ๐œŽ๐‘ค ๐œŽ๐‘ (29) T subject to 1 ๐’‘๐‘  = 1/๐‘†, โˆซ where we have let ๐‘ยฏ๐‘ ๐‘ฃ โ‰œ ๐’ด๐‘ ๐‘ฃ ๐‘ยฏz (๐‘ง) ๐‘‘๐‘ง, or ๐’‘ยฏ๐‘  in vector form, be the โ€œidealโ€ ๐’‘๐‘  (which also satisfies 1T ๐’‘ยฏ๐‘  = 1/๐‘†), and 2 and ๐œŽ๐‘2 can respectively be the regularization parameters ๐œŽ๐‘ค understood as the variance of each ๐‘ค๐‘ ๐‘—๐‘ฃ and the mean-square deviation of each ๐’‘๐‘  from ๐’‘ยฏ๐‘  . Solving (29), we obtain the estimate หœ + ๐œŒI)โˆ’1 (๐‘จ หœT ๐‘จ หœT ๐’…หœ๐‘  + ๐œŒ๐’‘ยฏ๐‘  โˆ’ ๐œ‡1) ๐’‘๐‘  = (๐‘จ = (๐‘จT ๐‘จ + R๐‘จT ๐‘จR + ๐œŒI)โˆ’1 T

(30)

T

โ‹… (๐‘จ ๐’…๐‘  + R๐‘จ ๐’…๐‘†โˆ’1โˆ’๐‘  + ๐œŒ๐’‘ยฏ๐‘  โˆ’ ๐œ‡1), 2 where ๐œŒ โ‰œ ๐œŽ๐‘ค /๐œŽ๐‘2 and ๐œ‡ is the Lagrange dual variable determined by the constraint 1T ๐’‘๐‘  = 1/๐‘†. T There remains the computation of the ๐‘€ ร— ๐‘€ matrix ๐‘จ โˆ‘ ๐‘จ T T T and the ๐‘€ ร— 1 vectors ๐‘จ ๐’…๐‘  . Note that ๐‘จ ๐‘จ = ๐‘— ๐‘จ๐‘— ๐‘จ๐‘— ; T since each ๐‘จ๐‘— is circulant, so is ๐‘จT ๐‘— ๐‘จ๐‘— and thus ๐‘จ ๐‘จ, and the matrix inversion in (30) can be computed efficiently using FFT. Specifically, we first use the circulant property of ๐‘จT ๐‘จ to see that (๐‘จT ๐‘จ + R๐‘จT ๐‘จR + ๐œŒI)1 = (2๐›ผ + ๐œŒ)1, where ๐›ผ is the sum of any row in ๐‘จT ๐‘จ; this allows the term with ๐œ‡ in (30) to be separated out, yielding

๐’‘๐‘  = ๐’‘หœ๐‘  โˆ’ ๐œ‡โ€ฒ 1,

(31)

where ๐’‘หœ๐‘  is (30) sans the ๐œ‡-term, and ๐œ‡โ€ฒ โ‰œ ๐œ‡/(2๐›ผ + ๐œŒ) can be computed from the 1T ๐’‘๐‘  = 1/๐‘† constraint as ๐œ‡โ€ฒ = (1T ๐’‘หœ๐‘  โˆ’ 1/๐‘†)/๐‘€ . Let F โ‰œ (๐‘’โˆ’j2๐œ‹๐‘˜๐‘ฃ/๐‘€ )๐‘˜๐‘ฃ be the DFT matrix with Fโˆ’1 = FH /๐‘€ (both the row index ๐‘˜ and the column index ๐‘ฃ start at 0 for convenience), then (30) can be transformed into F๐’‘หœ๐‘  = (F๐‘จT ๐‘จFโˆ’1 + FR๐‘จT ๐‘จRFโˆ’1 + ๐œŒI)โˆ’1 โ‹… F(๐‘จT ๐’…๐‘  + R๐‘จT ๐’…๐‘†โˆ’1โˆ’๐‘  + ๐œŒ๐’‘ยฏ๐‘  ).

(32)

โˆ’1 Now F๐‘จ๐‘— Fโˆ’1 = diag(A๐‘— ) and F๐‘จT = diag(A๐‘— )โˆ— , ๐‘— F โˆ— where (โ‹…) denotes complex conjugation and A๐‘— (not to be confused with ๐‘จ๐‘— ) is an ๐‘€ ร— 1 vector representing the DFT of ๐œˆ๐‘—u , i.e. ๐‘€โˆ’1 โˆ‘ ๐‘’โˆ’j2๐œ‹๐‘˜๐‘ข/๐‘€ ๐œˆ๐‘—u (๐‘ข). (33) (A๐‘— )๐‘˜ โ‰œ ๐‘ข=0 โˆ—

Moreover, FR = ฮฆF where ฮฆ โ‰œ diag((๐‘’j2๐œ‹๐‘˜/๐‘€ )๐‘€โˆ’1 ๐‘˜=0 ). Therefore, โˆ‘ โˆ’1 F๐‘จT ๐‘จFโˆ’1 = F๐‘จT = diag(Q), (34) ๐‘— ๐‘จ๐‘— F ๐‘— T

โˆ’1

FR๐‘จ ๐‘จRF = ฮฆ โ‹… diag(Q)โˆ— ฮฆโˆ’1 = diag(Q), (35) โˆ‘ 2 where Q โ‰œ ๐‘— โˆฃA๐‘— โˆฃ is an ๐‘€ ร— 1 vector and โˆฃโ‹…โˆฃ takes the complex magnitude element-wise. On the other hand, ๐‘จT ๐’…๐‘  =

โˆ‘

T ๐‘จT ๐‘— ๐’…๐‘ ๐‘— , with each ๐‘จ๐‘— ๐’…๐‘ ๐‘— given by { โˆ‘ โ€ฒ ๐œˆ๐‘—u (๐‘ฃ โŠ– ๐‘ฃ โ€ฒ ), if ๐‘ฆห†๐‘— โˆˆ ๐’ด๐‘ ๐‘ฃ , (๐‘จT ๐ด๐‘ฃ๐‘—๐‘ฃ ๐‘‘๐‘ ๐‘—๐‘ฃ = ๐‘— ๐’…๐‘ ๐‘— )๐‘ฃ โ€ฒ = / โˆช๐‘ฃ ๐’ด๐‘ ๐‘ฃ , 0, if ๐‘ฆห†๐‘— โˆˆ ๐‘ฃ (36) and ๐‘จT ๐’…๐‘†โˆ’1โˆ’๐‘  can be computed analogously. ๐’‘หœ๐‘  can thus be obtained from (32), with the left-multiplication by F and Fโˆ’1 carried out using FFT and inverse FFT respectively, and ๐’‘๐‘  follows from (31). ๐‘—

C. Further Regularization of the Estimate and Computation of the Adjusted ๐’šหœ As the final step of the recovery algorithm, we regularize the estimated ๐‘๐‘ ๐‘ฃ โ€™s, compute the corresponding ๐น๐œˆ๐‘—u (โ‹…)โ€™s, and use (17) to obtain the adjusted ๐’š หœ. By (20), each ๐น๐œˆ๐‘—u (โ‹…) is a linear combination ๐น๐œˆ๐‘—u (๐‘ฆ) =

๐‘€โˆ’1 โˆ‘

๐œˆ๐‘—u (๐‘ข)๐น๐‘ข (๐‘ฆ)

(37)

๐‘ข=0

of ๐น๐‘ข (โ‹…), the cdfs corresponding to ๐‘z ((โ‹… โˆ’ ๐‘ข)โ„ ). For convenience, we define the โ€œperiodically extendedโ€ cdf ๐น (๐‘ฆ) โ‰œ โˆซ๐‘ฆ โ€ฒ ๐‘ ((๐‘ฆ )โ„ ) ๐‘‘๐‘ฆ โ€ฒ for ๐‘z (โ‹…), so that ๐น๐‘ข (๐‘ฆ) becomes simply z 0 ๐น (๐‘ฆ โˆ’ ๐‘ข) โˆ’ ๐น (โˆ’๐‘ข). After discretization, for ๐‘ฆ โˆˆ [0, ๐‘€ ] that are integer multiples of 1/๐‘†, โˆ‘ ๐น (๐‘ฆ) = ๐‘๐‘ ๐‘ฃ , (38) (๐‘ ,๐‘ฃ):๐‘ฃ+๐‘ /๐‘†