The present invention relates to the field of information coding/decoding, and more particularly to a system and method for Slepian-Wolf coded nested quantization for Wyner-Ziv coding, e.g., for use in distributed source encoding/decoding.
Issues related to distributed compression of correlated sources are relevant for a wide variety of applications, such as distributed sensor networks and multi-source video distribution, both wired and wireless, coding for relay channels, and digital communications, among others. Distributed source coding (DSC), whose theoretical foundation was laid by Slepian and Wolf as early as 1973 (see D. Slepian and J. K. Wolf, “Noiseless coding of correlated information sources,” IEEE Trans. On Information Theory, vol. IT-19, pp. 471-480, July 1973.), refers to the compression of the outputs of two or more physically separated sources that do not communicate with each other (hence distributed coding). These sources send their compressed outputs to a central point (e.g., the base station) for joint decoding. DSC is related to the well-known “CEO problem” (in which a source is observed by several agents, who send independent messages to another agent (the chief executive officer (CEO)), who attempts to recover the source to meet a fidelity constraint, where it is usually assumed that the agents observe noisy versions of the source, with the observation noise being independent from agent to agent), and is part of network information theory.
The Wyner-Ziv coding problem deals with source coding with side information under a fidelity criterion. Although the theoretical limits for this problem have been known for some time (see, e.g., A. Wyner and J. Ziv, “The rate-distortion function for source coding with side information at the decoder,” IEEE Trans. Inform. Theory, vol. 22, pp. 1-10, January 1976; and A. Wyner, “The rate-distortion function for source coding with side information at the decoder-II: general sources”, Inform. Contr., vol. 38, pp. 60-80, 1978.), practical approaches to Wyner-Ziv coding have only recently been determined (see, e.g., S. Servetto, “Lattice quantization with side information,” Proc. DCC'00, Snowbird, Utah, March 2000; X. Wang and M. Orchard, “Design of trellis codes for source coding with side information at the decoder,” Proc. DC'01, Snowbird, Utah, March 2001; P. Mitran, and J. Bajcsy, “Coding for the Wyner-Ziv problem with turbo-like codes,” Proc. ISIT'2, Lausanne, Switzerland, June/July 2002; A. Aaron, R. Zhang and B. Girod, “Wyner-Ziv coding of motion video,” Proc. 36th Asilomar Conf., Pacific Grove, Calif., November 2002; S. Pradhan and K. Ramchandran, “Distributed source coding using syndromes (DISCUS): Design and construction,” IEEE Trans. Inform. Theory, vol. 49, pp. 626-643, March 2003; D. Rebollo-Monedero, R. Zhang, and B. Girod, “Design of optimal quantizers for distributed source coding,” Proc. IEEE Data Compression Conference, Snowbird, Utah, April 2003; J. Chou, S. Pradhan, and K. Ramchandran, “Turbo and trellis-based constructions for source coding with side information,” Proc. DCC'03, Snowbird, Utah, March 2003; A. Liveris, Z. Xiong and C. Georghiades, “Nested convolutional/turbo codes for the binary Wyner-Ziv problem,” Proc. ICIP'03, Barcelona, Spain, September 2003; Z. Xiong, A. Liveris, S. Cheng, and Z. Liu, “Nested quantization and Slepian-Wolf coding: A Wyner-Ziv coding paradigm for i.i.d. sources,” Proc. IEEE Workshop on Statistical Signal Processing, St.Louis, Mo., September 2003; and Y. Yang, S. Cheng, Z. Xiong, and W. Zhao “Wyner-Ziv coding based on TCQ and LDPC codes,” Proc. 37th Asilomar Conf., Pacific Grove, Calif., November 2003.). Zamir et al. (see R. Zamir, S. Shamai, and U. Erez, “Nested linear/lattice codes for structured multiterminal binning,” IEEE Trans. Inform. Theory, vol. 48, pp. 1250-1276, June 2002; and R. Zamir and S. Shamai, “Nested linear/lattice codes for Wyner-Ziv encoding,” Proc. IEEE Information Theory Workshop, pp. 92-93, Killarney, Ireland, June 1998) first outlined some constructive mechanisms using a pair of nested linear/lattice codes for binary/Gaussian sources, where the fine code in the nested pair plays the role of source coding while the coarse code does channel coding. They also proved that, for the quadratic Gaussian case, the Wyner-Ziv rate-distortion (R-D) function is asymptotically achievable using nested lattice codes, with the assumption that the lattice is ideally sphere-packed as the lattice dimension goes to infinity. Servetto (see S. Servetto, “Lattice quantization with side information,” Proc. DCC'00, Snowbird, Utah, March 2000) proposed explicit nested lattice constructions with the same dimensional lattice source codes and lattice channel codes. At low dimensions, the lattice channel coding component is not as strong as the lattice source coding component in the setup of Servetto. This is because, as the dimensionality increases, lattice source codes reach the R-D function much faster than lattice channel codes approach the capacity (see, e.g., V. Tarokh, A. Vardy, and K. Zeger, “Universal Bound on the performance of lattice codes,” IEEE Trans. Inform. Theory, vol. 45, pp. 670-681, March 1999; and J. H. Conway and N. J. A. Sloane, Sphere Packings. Lattices and Groups, New York: Springer, 1998.), when measuring the gaps to the corresponding theoretical limits in dBs for the MSE and channel SNR, respectively.
Consequently, one needs channel codes of much higher dimensions than source codes to achieve the same loss in terms of the gap to the corresponding limit, and the Wyner-Ziv limit may be approached with nesting codes of different dimensions in practice (see R. Zamir, S. Shamai, and U. Erez, “Nested linear/lattice codes for structured multiterminal binning,” IEEE Trans. Inform. Theory, vol. 48, pp. 1250-1276, June 2002).
In one set of embodiments, a system and method for realizing a Wyner-Ziv encoder may involve the following steps: (a) applying nested quantization to input data from an information source in order to generate intermediate data; and (b) encoding the intermediate data using an asymmetric Slepian-Wolf encoder in order to generate compressed output data representing the input data.
In another set of embodiments, a system and method for realizing a Wyner-Ziv decoder involves the following steps:
A better understanding of the present invention can be obtained when the following detailed description of the preferred embodiment is considered in conjunction with the following drawings, in which:
While the invention is susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and are herein described in detail. It should be understood, however, that the drawings and detailed description thereto are not intended to limit the invention to the particular form disclosed, but on the contrary, the intention is to cover all modifications, equivalents and alternatives falling within the spirit and scope of the present invention as defined by the appended claims.
Incorporation by Reference
The following references are hereby incorporated by reference in their entirety as though fully and completely set forth herein:
This application includes eight appendices labeled A —H.
Appendix A comprises a paper titled: “Design of Slepian-Wolf Codes by Channel Code Partitioning” by Vladimir M. Stankovic, Angelos D. Liveris, Zixiang Xiong, and Costas N. Georghiades.
Appendix B comprises a paper titled: “On Code Design for the Slepian-Wolf Problem and Lossless Multiterminal Networks” by Vladimir M. Stankovic, Angelos D. Liveris, Zixiang Xiong, and Costas N. Georghiades.
Appendix C comprises a paper titled: “Slepian-Wolf Coded Nest Quantization (SWC-NQ) for Wyner-Ziv Coding: Performance Analysis and Code Design” by Zhixin Liu, Samuel S. Cheng, Angelos D. Liveris & Zixiang Xiong.
Appendix D comprises a paper titled: “Slepian-Wolf Coded Nested Lattice Quantization for Wyner-Ziv Coding: High Rate Performance Analysis and Code Design” by Zhixin Liu, Samuel S. Cheng, Angelos D. Liveris & Zixiang Xiong.
Appendix E comprises a paper titled: “Layered Wyner-Ziv Video Coding” by Qian Xu and Zixiang Xiong.
Appendix F comprises a paper titled: “A Turbo Code Tutorial” by William E. Ryan.
Appendix G comprises a paper titled: “Generalized Coset Codes for Symmetric Distributed Source Coding” by S. Sandeep Pradhan and Kannan Ramchandran.
Appendix H comprises a paper titled: “Compression of Binary Sources with Side Information at the Decoder Using LDPC Codes” by Angelos D. Liveris, Zixiang Xiong and Costas N. Georghiades.
Terms
The following is a glossary of terms used in the present application:
Memory Medium—Any of various types of memory devices or storage devices. The term “memory medium” is intended to include an installation medium, e.g., a CD-ROM, floppy disks 104, or tape device; a computer system memory or random access memory such as DRAM, DDR RAM, SRAM, EDO RAM, Rambus RAM, etc.; or a non-volatile memory such as a magnetic media, e.g., a hard drive, or optical storage. The memory medium may comprise other types of memory as well, or combinations thereof. In addition, the memory medium may be located in a first computer in which the programs are executed, or may be located in a second different computer which connects to the first computer over a network, such as the Internet. In the latter instance, the second computer may provide program instructions to the first computer for execution. The term “memory medium” may include two or more memory mediums which may reside in different locations, e.g., in different computers that are connected over a network.
Carrier Medium—a memory medium as described above, as well as signals such as electrical, electromagnetic, or digital signals, conveyed via a communication medium such as a bus, network and/or a wireless link.
Programmable Hardware Element—includes various types of programmable hardware, reconfigurable hardware, programmable logic, or field-programmable devices (FPDs), such as one or more FPGAs (Field Programmable Gate Arrays), or one or more PLDs (Programmable Logic Devices), such as one or more Simple PLDs (SPLDs) or one or more Complex PLDs (CPLDs), or other types of programmable hardware. A programmable hardware element may also be referred to as “reconfigurable logic”.
Medium—includes one or more of a memory medium, carrier medium, and/or programmable hardware element; encompasses various types of mediums that can either store program instructions/data structures or can be configured with a hardware configuration program. For example, a medium that is “configured to perform a function or implement a software object” may be 1) a memory medium or carrier medium that stores program instructions, such that the program instructions are executable by a processor to perform the function or implement the software object; 2) a medium carrying signals that are involved with performing the function or implementing the software object; and/or 3) a programmable hardware element configured with a hardware configuration program to perform the function or implement the software object.
Program—the term “program” is intended to have the full breadth of its ordinary meaning. The term “program” includes 1) a software program which may be stored in a memory and is executable by a processor or 2) a hardware configuration program useable for configuring a programmable hardware element.
Software Program—the term “software program” is intended to have the full breadth of its ordinary meaning, and includes any type of program instructions, code, script and/or data, or combinations thereof, that may be stored in a memory medium and executed by a processor. Exemplary software programs include programs written in text-based programming languages, such as C, C++, Pascal, Fortran, Cobol, Java, assembly language, etc.; graphical programs (programs written in graphical programming languages); assembly language programs; programs that have been compiled to machine language; scripts; and other types of executable software. A software program may comprise two or more software programs that interoperate in some manner.
Hardware Configuration Program—a program, e.g., a netlist or bit file, that can be used to program or configure a programmable hardware element.
Graphical User Interface—this term is intended to have the full breadth of its ordinary meaning. The term “Graphical User Interface” is often abbreviated to “GUI”. A GUI may comprise only one or more input GUI elements, only one or more output GUI elements, or both input and output GUI elements.
The following provides examples of various aspects of GUIs. The following examples and discussion are not intended to limit the ordinary meaning of GUI, but rather provide examples of what the term “graphical user interface” encompasses:
A GUI may comprise a single window having one or more GUI Elements, or may comprise a plurality of individual GUI Elements (or individual windows each having one or more GUI Elements), wherein the individual GUI Elements or windows may optionally be tiled together.
A GUI may be associated with a graphical program. In this instance, various mechanisms may be used to connect GUI Elements in the GUI with nodes in the graphical program. For example, when Input Controls and Output Indicators are created in the GUI, corresponding nodes (e.g., terminals) may be automatically created in the graphical program or block diagram. Alternatively, the user can place terminal nodes in the block diagram which may cause the display of corresponding GUI Elements front panel objects in the GUI, either at edit time or later at run time. As another example, the GUI may comprise GUI Elements embedded in the block diagram portion of the graphical program.
Computer System—any of various types of computing or processing systems, including a personal computer system (PC), mainframe computer system, workstation, network appliance, Internet appliance, personal digital assistant (PDA), television system, grid computing system, or other device or combinations of devices. In general, the term “computer system” can be broadly defined to encompass any device (or combination of devices) having at least one processor that executes instructions from a memory medium.
The computer system 82 may include a memory medium(s) on which one or more computer programs or software components according to any subset of the method embodiments of the present invention may be stored. For example, the memory medium may store one or more programs which are executable to perform the methods described herein. The memory medium may also store operating system software, as well as other software for operation of the computer system. Various embodiments further include receiving or storing instructions and/or data implemented in accordance with the foregoing description upon a carrier medium.
The computer may include at least one central processing unit or CPU (processor) 160 which is coupled to a processor or host bus 162. The CPU 160 may be any of various types, including an x86 processor, e.g., a Pentium class, a PowerPC processor, a CPU from the SPARC family of RISC processors, as well as others. A memory medium, typically comprising RAM and referred to as main memory, 166 is coupled to the host bus 162 by means of memory controller 164. The main memory 166 may store programs operable to implement Wyner-Ziv encoding and/or decoder according to any of various embodiments described herein. The main memory may also store operating system software, as well as other software for operation of the computer system.
The host bus 162 may be coupled to an expansion or input/output bus 170 by means of a bus controller 168 or bus bridge logic. The expansion bus 170 may be the PCI (Peripheral Component Interconnect) expansion bus, although other bus types can be used. The expansion bus 170 includes slots for various devices such as described above. The computer 82 further comprises a video display subsystem 180 and hard drive 182 coupled to the expansion bus 170.
As shown, a device 190 may also be connected to the computer. The device 190 may include a processor and memory which may execute a real time operating system. The device 190 may also or instead comprise a programmable hardware element (such as an FPGA). The computer system may be operable to deploy a program to the device 190 for execution of the program on the device 190.
FIGS. 3A and 3B—Exemplary Systems
Various embodiments of the present invention may be directed to sensor systems, wireless or wired video transmission systems, or any other type of information processing or distribution system utilizing information coding such as Wyner-Ziv coding.
For example, as
The step of applying nested quantization to the input data may include: (a) quantizing values of the input data with respect to a fine lattice to determine corresponding points of the fine lattice; and (b) computing values of the intermediate data based on the fine lattice points, respectively, wherein the values of the intermediate data designate cosets of the fine lattice modulo a coarse lattice, wherein the coarse lattice Λ2 is a sublattice of the fine lattice Λ1.
The information source may be a source of audio information, video information, image information, text information, information derived from physical measurements, etc.
In step 510, compressed input data is received. The compressed input data represents a block of samples of a first source X. In step 512, a block of samples of a second source Y is received. Steps 510 and 512 may be performed in parallel.
In step 514, an asymmetric Slepian-Wolf decoder is applied to the compressed input data in order to generate a block of intermediate values.
In step 516, joint decoding is performed on each intermediate value and a corresponding sample of the second source block to obtain a corresponding decompressed output value. The joint decoding includes computing a centroid of a conditional probability density of the first source X given said corresponding sample of the second source block restricted to a region of space corresponding to the intermediate value, wherein the centroid determines the decompressed output value. The intermediate values specify cosets of a fine lattice with respect to a coarse lattice, wherein the coarse lattice is a sublattice of the fine lattice.
Wyner-Ziv Coding
Various embodiments of the present invention provide a new framework for Wyner-Ziv coding of independent identically distributed (i.i.d.) sources based on Slepian-Wolf coded nested quantization (SWC-NQ). The role of SWC in SWCNQ is to exploit the correlation between the quantized source and the side information for further compression and to make the overall channel code stronger. SWC-NQ generalizes the classic source coding approach of quantization (O) and entropy coding (EC) in the sense that the quantizer performs quite well alone and can exhibit further rate savings by employing a powerful Slepian-Wolf code.
For the quadratic Gaussian case, a high-rate performance of SWC-NQ with low-dimensional nested quantization and ideal SWC has been established, and it is shown herein that SWC-NQ achieves the same performance of classic entropy-constrained lattice quantization. For example, 1-D/2-D SWC-NQ performs 1.53/1.36 dB away from the Wyner-Ziv R-D function of the quadratic Gaussian source at high rate assuming ideal SWC.
A recent work (D. Rebollo-Monedero, A. Aaron, and B. Girod “Transforms for high-rate distributed source coding,” Proc. 37th Asilomar Conf., Pacific Grove, Calif., November 2003) starts with non-uniform quantization with index reuse and Slepian-Wolf coding and shows the same high-rate theoretical performance as embodiments of the present invention when the quantizer becomes an almost uniform one without index reuse. This agrees with the assertion that at high rate, the nested quantizer asymptotically becomes a non-nested regular one so that strong channel coding is guaranteed.
Embodiments implementing 1-D and 2-D nested lattice quantizers in the rate range of 1-7 bits per sample are presented. Although analysis has shown that in some embodiments nesting does not benefit at high rate, experiments using nested lattice quantizers together with irregular LDPC codes for SWC obtain performances close to the corresponding limits at low rates. Some of the embodiments presented herein thus show that SWC-NQ provides an efficient scheme for practical Wyner-Ziv coding with low-dimensional lattice quantizers at low rates.
It is noted that in some embodiments estimation in the decoder may yield significant gains only for low rates and for high rates it may not help noticeably. This is confirmed by the agreement of the high rate analysis results described herein, which assume that no estimation is used, with the high rate simulation results, for which an estimation method is always used.
Lattices and Nested Lattices
Lattices
For a set of n basis vectors {g1, . . . , gn} an unbounded n-D lattice Λ is defined by Λ={ι=Gi: i ε n} and its generator matrix G=[g1|g2| . . . |gnquantizer QΛ(•) associated with Λ is given by QΛ(χ)=arg minιεΛ|χ−ι|. The basic Voronoi cell of Λ, which specifies the shape of the nearest-neighbor decoding region, is ν={χ: QΛ(χ)=0}. Associated with the Voronoi cell V are several important quantities: the cell volume V, the second moment σ2 and the normalized second moment G(Λ), defined by
respectively. The minimum of G(Λ) over all lattices in n is denoted as Gn. By (J. H. Conway and N. J. A. Sloane, Sphere Packings. Lattices and Groups, New York: Springer, 1998), Gn≧½πe, ∀n; and limn→∞ Gn=½πe.
Nested Lattices
A pair of n-D lattices (Λ1, Λ2) with corresponding generator matrices G1 and G2 is nested, if there exists an n×n integer matrix P such that G2=G1×P and det{P}>1. In this case V2/V1 is called the nesting ratio, and Λ1 and Λ2 are called the fine and coarse lattice, respectively.
For a pair of nested lattices (Λ1, Λ2), the points in the set Λ1/Λ2{Λ1∩V2}are called the coset leaders of Λ2 relative to Λ1, where ν2 is the basic Voronoi cell of Λ2. For each υ ε Λ1/Λ2 the set of shifted lattice points C(v){u+ι,∀ι ε Λ2}is called a coset of Λ2 relative to Λ1. The j-th point of C(ν) is denoted as cj(ν). Then C(0)={cj(0),∀j ε =Λ2, ∪υεΛ
Nested Lattice Quantization
Throughout the present description, the correlation model of X=Y+Z is used, where X is the source to be coded and the side information Y˜N(0,σY2) and the noise Z˜N(0,σZ2) are independent Gaussian. Zamir et al.'s nested lattice quantization scheme (see R. Zamir, S. Shamai, and U. Erez, “Nested linear/lattice codes for structured multiterminal binning,” IEEE Trans. Inform. Theory, vol. 48, pp. 1250-1276, June 2002; and R. Zamir and S. Shamai, “Nested linear/lattice codes for Wyner-Ziv encoding,” Proc. IEEE Information Theory Workshop, pp. 92-93, Killarney, Ireland, June 1998) works as follows: Let the pseudo random vector U (the dither), known to both the quantizer encoder and the decoder, be uniformly distributed over the basic Voronoi cell ν1 of the fine lattice Λ1. For a given target average distortion D, denote
as the estimation coeficient.
Given the realizations of the source, the side information and the dither as x, y and u, respectively, then according to R. Zamir, S. Shamai, and U. Erez, “Nested linear/lattice codes for structured multiterminal binning,” IEEE Trans. Inform. Theory, vol. 48, pp. 1250-1276, June 2002, the nested quantizer encoder quantizes αx+u to the nearest point
computes
which is the coset shift of
with respect to Λ2, and transmits the index corresponding to this coset shift.
The nested quantizer decoder receives
forms
and reconstructs χ as {circumflex over (χ)}=y+α(w−QΛ
It is shown in R. Zamir, S. Shamai, and U. Erez, “Nested linear/lattice codes for structured multiterminal binning,” IEEE Trans. Inform. Theory, vol. 48, pp. 1250-1276, June 2002, that the Wyner-Ziv R-D function DWZ(R)=σX|Y22−2R is achievable with infinite dimensional nested lattice quantization. Presented herein, the high rate performance of low-dimensional nested lattice quantization is analyzed, which is of more practical interest as high-dimensional nested lattice quantization may currently be too complex to implement.
The analysis is based on the high-resolution assumption, which means 1) V1 is small enough so that the pdf of X, ƒ(x), is approximately constant over each Voronoi cell of Λ1 and 2) dithering can be ignored. With the high-rate assumption, α≈1 and the encoder/decoder described above simplifies to
1) The encoder quantizes χ to
computes
and transmits an index corresponding to the coset leader υ.
Upon receiving υ, the decoder forms w=υ−y and reconstructs x as {circumflex over (χ)}υ=y+w−QΛ
For the remainder of this description, the discuss will be limited to this simplified nested lattice quantization scheme for high rate, which was also used in S. Servetto, “Lattice quantization with side information,” Proc. DCC'00, Snowbird, Utah, March 2000.
High Rate Performance
Before discussing the main results, the following Lemma is provided.
Lemma 1: If a pair of n-D nested lattices (Λ1,Λ2) with nesting ratio N=V2=V1 is used for nested lattice quantization, the distortion per dimension in Wyner-Ziv coding of X with side information Y at high rate is
Proof: The average distortion for a given realization of the side information Y=y is
where (a) comes from the high resolution assumption and ∫χεR
Therefore, the average distortion per dimension over all realizations of Y is
Note that for a fixed pair of nested lattices (Λ1,Λ2), Dn, only depends on Z, i.e, the correlation between X and Y. It is independent of the marginal distribution of X (and Y). The first term,
in the Dn is due to lattice quantization in source coding. It is determined by the geometric structure and the Voronoi cell's volume of Λ1. The second term,
is the loss due to nesting (or the channel coding component of the nested lattice code). It depends on V2 and the distribution of Z. From Lemma 1, the lower bound of the high-rate R-D performance of low dimensional nested lattice quantizers for Wyner-Ziv coding may be obtained, when Z is Gaussian.
Theorem 1: For X=Y+Z, Y˜N(0,σY2) and Z˜N(0,σZ2), the R-D performance of Wyner-Ziv coding for X with side information Y using n-D nested lattice quantizers is lower-bounded at high rate by
where
γn is the n-D Hermite's constant (see J. H. Conway and N. J. A. Sloane, Sphere Packings. Lattices and Groups, New York: Springer, 1998; and V. Tarokh, A. Vardy, and K. Zeger, “Universal Bound on the performance of lattice codes,” IEEE Trans. Inform. Theory, vol. 45, pp. 670-681, March 1999), and u(t) is defined in V. Tarokh, A. Vardy, and K. Zeger, “Universal Bound on the performance of lattice codes,” IEEE Trans. Inform. Theory, vol. 45, pp. 670-681, March 1999, as
Specifically, when n=1, the best possible high rate performance is
where
Proof:
1) Rate computation: Note that the nested lattice quantizer is a fixed rate quantizer with rate
2) Distortion computation: Define
and PZ(L)=Pr(∥Z∥>L).
In the 1-D (scalar) case, PZ can be expressed in term of the Q function and EZ[∥QΛ
In the n-D (n>1) case, note that
(see G. D. Forney Jr., “Coset codes-Part II: Binary lattices and related codes,” IEEE Trans. Inform. Theory, vol. 34, pp. 1152-1187, 1988). From V. Tarokh, A. Vardy, and K. Zeger, “Universal Bound on the performance of lattice codes,” IEEE Trans. Inform. Theory, vol. 45, pp. 670-681, March 1999, one gets
where Λ(t)=∫0∞ut−1e−udu is Euler's gamma function, γ(Λ2) is the Hermite's constant of lattice Λ2 (see J. H. Conway and N. J. A. Sloane, Sphere Packings. Lattices and Groups, New York: Springer, 1998; and V. Tarokh, A. Vardy, and K. Zeger, “Universal Bound on the performance of lattice codes,” IEEE Trans. Inform. Theory, vol. 45, pp. 670-681, March 1999), and Pe(•) is defined in V. Tarokh, A. Vardy, and K. Zeger, “Universal Bound on the performance of lattice codes,” IEEE Trans. Inform. Theory, vol. 45, pp. 670-681, March 1999, as the symbol error probability under maximum-likelihood decoding while transmitting the lattice points υ ε Λ over an AWGN channel. A lower bound of Pe(•) was also given as Pe(t)≧u(t).
Then Lemma 1 and (9) give
Using
may be eliminated in Dn and a lower bound obtained of
Regarding
Slepian-Wolf Coded Nested Lattice Quantization (SWC-NQ)
Motivation of SWC-NQ
The performance of lattice source codes and lattice channel codes are analyzed in V. Tarokh, A. Vardy, and K. Zeger, “Universal Bound on the performance of lattice codes,” IEEE Trans. Inform. Theory, vol. 45, pp. 670-681, March 1999, and J. H. Conway and N. J. A. Sloane, Sphere Packings. Lattices and Groups, New York: Springer, 1998, respectively. In the 1-D nested scalar quantizer, the source code Λ1 performs 1.53 dB (see J. H. Conway and N. J. A. Sloane, Sphere Packings. Lattices and Groups, New York: Springer, 1998; and A. Gersho and R. Gray, Vector Quantization and Signal Compression, Kluwer Academic Publishers, Boston, Mass., 1992) worse than the R-D function at high rate, whereas the channel code Λ2 suffers more than 6.5 dB loss with respect to the channel capacity in terms of channel SNR (with Pe=106) (see V. Tarokh, A. Vardy, and K. Zeger, “Universal Bound on the performance of lattice codes,” IEEE Trans. Inform. Theory, vol. 45, pp. 670-681, March 1999). As the dimensionality increases, lattice source codes reach the R-D function much faster than the lattice channel codes approach the capacity. This explains the increasing gap between {overscore (D)}n(R) and DWZ(R) in
Consequently one needs channel codes of much higher dimension than source codes to achieve the same loss, and the Wyner-Ziv limit should be approached with nesting codes of different dimensionality in practice.
Thus a framework is proposed for Wyner-Ziv coding of i.i.d. sources based on SWC-NQ, which is NQ followed by SWC. Despite the fact that there is almost no correlation among the nested quantization indices that identify the coset leaders υ ε Λ1/Λ2 of the pair of nested lattices (Λ1, Λ2), there still remains correlation between υ and the side information Y. Ideal SWC can be used to compress V to the rate of R=H(υ|Y). In practice, it has already been made clear that SWC is a channel coding problem (see R. Zamir, S. Shamai, and U. Erez, “Nested linear/lattice codes for structured multiterminal binning,” IEEE Trans. Inform. Theory, vol. 48, pp. 1250-1276, June 2002). State-of-the-art channel codes, such as LDPC codes, can be used to approach the Slepian-Wolf limit H(υ|Y) (see A. Liveris, Z. Xiong and C. Georghiades, “Compression of binary sources with side information at the decoder using LDPC codes,” IEEE Communications Letters, vol. 6, pp. 440-442, October 2002). The role of SWC in SWC-NQ is to exploit the correlation between υ and Y for further compression, thus making the overall channel code stronger.
High Rate Performance
For the quadratic Gaussian case, the high-rate performance of SWC-NQ using low-dimensional nested lattices is given in this section. Start with the following Lemma.
Lemma 2: The R-D performance of ideally Slepian-Wolf coded nested lattice quantization at high rate using a low-dimensional nested lattice pair (1;2) is
where h′(X, Λ)−∫xεR
Proof:
1) Rate Computation: The rate for SWC-NQ is: R=1/nH(υ|Y). Since at high rate,
where g(x)Σj=∞∞ƒX|Y(x+cj(0)), and X|Y˜N(0,σX|Y2). Then the achievable rate of SWC-NQ is
where (a) comes from the periodic property of g(•), i.e. g(χ−ι|y)=g(χ|y), ∀ι ε Λ2.
Thus the achievable rate of SWC-NQ is
nR=H(υ|Y)=h′(X,Λ2)+log2 σX|Yn−log2 V1 (14)
2) Distortion Computation: From Lemma 1, the average distortion of nested lattice quantization over all realizations of
Because SWC is lossless, the distortion of SWC-NQ is also Dn. Combining Dn and R through V1, the R-D performance may be obtained of SWC-NQ with a pair of n-D nested lattice (Λ1, Λ2) as
Based on Lemma 2, a search may be made for the optimal V2 to minimize Dn(R), and the following theorem presented. Due to space limitations, only a high-level discussion of the proof is presented herein.
Theorem 2: For the quadratic Gaussian case, the R-D performance of SWC-NQ using low-dimensional nested lattices for Wyner-Ziv coding at high rate is
Dn(R)=2πeGnσX|Y22−2R (16)
Sketch of the proof: If (15) is rewritten as
and compare (17) with 2πeG(Λ1)σX|Y22−2R, it is found that Dn(R) for SWC-NQ has an extra distortion
with a rate saving
due to nesting.
It may thus be concluded for the quadratic Gaussian source and at high rate, SWC-NQ performs the same as traditional entropy-constrained lattice quantization with side information available at both the encoder and decoder. Specifically, the R-D functions with 1-D (scalar) lattice and 2-D (hexagonal) lattice are 1.53 dB and 1.36 dB away from the Wyner-Ziv bound, respectively.
Regarding
and the rate saving
due to nesting as a function of V2, for n=2 and σZ2=0.01.
Remarks: It has been found that for finite rate R and small n (e.g., n=1 and 2), the optimal V2, denoted as V2, that minimizes the distortion Dn(R) in (15) is also finite.
Regarding
Code Design and Simulation Results
Slepian-Wolf coding can be implemented using syndrome coding based on the most powerful channel codes available, such as LDPC codes and/or turbo codes. Let J(0≦J≦N−1) denote the index of the coset leader υ, one needs to code J using Slepian-Wolf codes with Y as the side information. Instead of coding J as a whole, J may be coded bit-by-bit using the multi-layer Slepian-Wolf coding. Assume J=(BmBm-1 . . . B2B1)2, where Bm is the most significant bit (MSB) of J, and B1 is the least significant bit (LSB). At first B1 is coded using a Slepian-Wolf code with Y as side information at rate R1=H(B1|Y); then B2 is coded with side information R2=H(B2|Y,B1); . . . ; Bm is coded with side information {Y, B1, B2, . . . , Bm-1} at rate Rm=H(Bm|Y, B1, . . . Bm-1). Hence the total rate is H(J|Y)=H(υ|Y).
Estimation at the decoder plays an important role for low-rate implementation. Thus, an optimal non-linear estimator may be applied at the decoder at low rates in simulations. This estimator degenerates to a linear one at high rates. 1-D nested lattice quantizer design was performed for different sources with 106 samples of X in each case. For σY2=1 and σZ2=0.01,
For 2-D nested lattice quantization, the A2 hexagonal lattices may be used again with σY2=1, σZ2=0.01.
At high rate, the former case exhibits a 4.06-8.48 dB gap from DWZ(R) for R=1.40-5.00 b/s, again in agreement with the high rate lower bound of Theorem 1. It may be observed that the gap between results with ideal SWC (measured in the simulation) and DWZ(R) is 1.36 dB. With practical SWC based on irregular LDPC codes (of length 106 bits), this gap is 1.67-1.72 dB for R=0.95-2.45 b/s.
It may thus be seen that using optimal estimation, simulation results with either 1-D or 2-D nested quantization (and practical Slepian-Wolf coding) are almost a constant gap away from the Wyner-Ziv limit for a wide range of rates.
Although the embodiments above have been described in considerable detail, numerous variations and modifications will become apparent to those skilled in the art once the above disclosure is fully appreciated. It is intended that the following claims be interpreted to embrace all such variations and modifications.