Lecture Notes in Information Theory Volume II - Website of Professor ...
Recommend Documents
1 Jun 2010 ... iii. Lecture Notes in Information Technology. 2012 International Conference on
Affective Computing and. Intelligent Interaction (ICACII 2012).
provided by UNASUS/UFMA using the. Monitoring System (MonSyns). Márcia
M P Rendeiro. Universidade do Estado Rio de Janeiro/Departamento de ...
Email: [email protected]. Rab Nawaz. Department ... Email: [email protected]. AbstractâPeer to .... and promote C2C along with B2B and B2C.
Jun 8, 2010 - It is about a philosopy, or a belief, which, in a nutshell, is the ... every information bit in the universe there is a certain amount of energy. ...... Jij's are i.i.d. RV's, which are normally modeled to have a zeroâmean Gaussian pd
microeconomics course for PhD or Master's degree economics students. The lecture notes were developed over a period of almost 15 years during.
a microeconomics course for PhD (or MA) economics students. The ... I published
the book for the first time in 2007 with some hesitation since several superb ...
Sep 15, 2009 ... Theory of Computation Lecture. Notes. Prof. Yuh-Dauh Lyuu. Dept. Computer
Science & Information Engineering and. Department of Finance.
published by Springer Lecture Notes in Electrical Engineering, which include
about ... Biswanath Vokkarane Society on Social Implications of Technology and.
area called by Carl Friedrich Gauss âthe queen of mathematicsâ. One must give .... perspective, that is the ring of Gaussian integers: Z[i] := {a + bi|a, b â Z}. The.
It took 200 years before the first book on graph theory was written. This was “The
... D.B. WEST, “Introduction to Graph Theory”, Prentice Hall, 1996. R.J. WILSON ...
6.2 Harmonic oscillator correlation functions; path integrals . . . . . . . . . . . . 44 ...
However, to give a quantum-mechanical treatment of field theory is much more ...
Matematik ... These are lecture notes on integration theory for a eight$week
course at the ... Chapter 2 deals with integration and is very similar to [F] and
most.
Introduction These brief notes include major definitions and theorems of the graph theory lecture held by Prof. Maria Axenovich at KIT in the winter term 2013/14.
The interested reader may consult the Dudley book [D]. Exercises. 1. (X,
Dec 10, 2013 ... cryptography as a tool for communication is to secure data transmitted in ... the
Kasiski method supported by the Friedman test. To supplement ...
ISSN: 2070-1918. Distributed worldwide by. Information Engineering Research Institute. 100 Continental Dr, Newark, DELAWARE 19713, Unite State, USA.
German Research Center for Artificial Intelligence (DFKI) .... Nigel Shadbolt presents problems that arose in designing an integrated knowl- edge acquisition ...
1. NGSSC/OOSP/OOAD. OOAD part III. On OO methodologies,. UML,. Design
patterns. NGSSC/OOSP/OOAD. On OO methodologies. Historical perspectives.
W.L. Winston's "Operations Research: Applications and Algorithms“. (slides
submitted by Brooks/Cole, a division of Thomson Learning, Inc.) and Prof.
CHROMOSOME MUTATION II: CHANGES IN NUMBERS. I. Aberrant ...
Polyploids (>2x) are by definition abnormal although they exist in plants in nature
. 5.
CALCULUS MATH 166 FALL 2013 (COHEN) LECTURE NOTES .... “limit process
,” especially the development of infinite sequences and series. Before we turn ...
viii) Illustrate the basic mechanisms of some catalytic reactions incorporating
alkenes. ix) ... R. B. Jordan, Reaction Mechanisms of Inorganic and
Organometallic ...
The major goal of Principles of Communication Engineering (I,II) is to teach
students ... Robert G. Gallager: Principles of Digital Communication, Cambridge ...
23 Sep 2009 ... As shown from the list, the lecture notes are divided into two volumes. ... In
addition, you may obtain the latest version of the lecture notes from.
Lecture Notes in Information Theory Volume II by Po-Ning Chen† and Fady Alajaji‡
†
Department of Electrical Engineering Institute of Communication Engineering National Chiao Tung University 1001, Ta Hsueh Road Hsin Chu, Taiwan 30056 Republic of China Email: [email protected] ‡
Department of Mathematics & Statistics, Queen’s University, Kingston, ON K7L 3N6, Canada Email: [email protected] December 7, 2014
c Copyright by
Po-Ning Chen† and Fady Alajaji‡ December 7, 2014
Preface The reliable transmission of information bearing signals over a noisy communication channel is at the heart of what we call communication. Information theory—founded by Claude E. Shannon in 1948—provides a mathematical framework for the theory of communication; it describes the fundamental limits to how efficiently one can encode information and still be able to recover it with negligible loss. This course will examine the basic concepts of this theory. What follows is a tentative list of topics to be covered. 1. Volume I: (a) Fundamentals of source coding (data compression): Discrete memoryless sources, entropy, redundancy, block encoding, variable-length encoding, Kraft inequality, Shannon code, Huffman code. (b) Fundamentals of channel coding: Discrete memoryless channels, mutual information, channel capacity, coding theorem for discrete memoryless channels, weak converse, channel capacity with output feedback, the Shannon joint source-channel coding theorem. (c) Source coding with distortion (rate distortion theory): Discrete memoryless sources, rate-distortion function and its properties, rate-distortion theorem. (d) Other topics: Information measures for continuous random variables, capacity of discrete-time and band-limited continuous-time Gaussian channels, rate-distortion function of the memoryless Gaussian source, encoding of discrete sources with memory, capacity of discrete channels with memory. (e) Fundamental backgrounds on real analysis and probability (Appendix): The concept of set, supremum and maximum, infimum and minimum, boundedness, sequences and their limits, equivalence, probability space, random variable and random process, relation between a source and a random process, convergence of sequences of random variables, ergodicity and laws of large numbers, central limit theorem, concavity and convexity, Jensen’s inequality.
ii
2. Volumn II: (a) General information measure: Information spectrum and Quantile and their properties, R´enyi’s informatino measures. (b) Advanced topics of losslesss data compression: Fixed-length lossless data compression theorem for arbitrary channels, Variable-length lossless data compression theorem for arbitrary channels, entropy of English, Lempel-Ziv code. (c) Measure of randomness and resolvability: Resolvability and source coding, approximation of output statistics for arbitrary channels. (d) Advanced topics of channel coding: Channel capacity for arbitrary single-user channel, optimistic Shannon coding theorem, strong capacity, ε-capacity. (e) Advanced topics of lossy data compressing (f) Hypothesis testing: Error exponent and divergence, large deviations theory, Berry-Esseen theorem. (g) Channel reliability: Random coding exponent, expurgated exponent, partitioning exponent, sphere-packing exponent, the asymptotic largest minimum distance of block codes, Elias bound, Varshamov-Gilbert bound, Bhattacharyya distance. (h) Information theory of networks: Distributed detection, data compression over distributed source, capacity of multiple access channels, degraded broadcast channel, Gaussian multiple terminal channels. As shown from the list, the lecture notes are divided into two volumes. The first volume is suitable for a 12-week introductory course as that given at the Department of Mathematics and Statistics, Queen’s University at Kingston, Canada. It also meets the need of a fundamental course for senior undergraduates as that given at the Department of Computer Science and Information Engineering, National Chi Nan University, Taiwan. For an 18-week graduate course as given in Department of Communications Engineering, National Chiao-Tung University, Taiwan, the lecturer can selectively add advanced topics covered in the second volume to enrich the lecture content, and provide a more complete and advanced view on information theory to students. The authors are very much indebted to all people who provided insightful comments on these lecture notes. Special thanks are devoted to Prof. Yunghsiang S. Han with the Department of Computer Science and Information Engineering in National Chi Nan University, Taiwan, for his enthusiasm in testing these lecture notes at his school, and providing the authors valuable feedback.
iii
Notes to readers. In these notes, all the assumptions, claims, conjectures, corollaries, definitions, examples, exercises, lemmas, observations, properties, and theorems are numbered under the same counter for ease of their searching. For example, the lemma that immediately follows Theorem 2.1 will be numbered as Lemma 2.2, instead of Lemma 2.1. In addition, you may obtain the latest version of the lecture notes from http://shannon.cm.nctu.edu.tw. Interested readers are welcome to return comments to [email protected].
iv
Acknowledgements Thanks are given to our families for their full support during the period of writing these lecture notes.
The asymptotic CDFs of a sequence of random variables, {An }∞ n=1 . u¯(·) = sup-spectrum of An ; u(·) = inf-spectrum of An ; U1− = ¯ ¯ limξ↑1 Uξ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ¯ The spectrum h(θ) for Example 2.7. . . . . . . . . . . . . . . . . The spectrum ¯i(θ) for Example 2.7. . . . . . . . . . . . . . . . . The limiting spectrums of (1/n)hZ n (Z n ) for Example 2.8 . . . . The possible limiting spectrums of (1/n)iX n ,Y n (X n ; Y n ) for Example 2.8. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
. 8 . 22 . 22 . 23 . 24
Behavior of the probability of block decoding error as block length n goes to infinity for an arbitrary source X. . . . . . . . . . . . . 35 ¯ ε (X) + Illustration of generalized AEP Theorem. Fn (δ; ε) , Tn [H ¯ ε (X) − δ] is the dashed region. . . . . . . . . . . . . . . . 37 δ] \ Tn [H Notations used in Lempel-Ziv coder. . . . . . . . . . . . . . . . . 48 Source generator: {Xt }t∈I (I = (0, 1)) is an independent random process with PXt (0) = 1 − PXt (1) = t, and is also independent of the selector Z, where Xt is outputted if Z = t. Source generator of each time instance is independent temporally. . . . . . . . . . . 72 The ultimate CDF of −(1/n) log PX n (X n ): P r{hb (Z) ≤ t}. . . . . 74 The ultimate CDFs of (1/n) log PN n (N n ). . . . . . . . . . . . . The ultimate CDF of the normalized information density for Example 5.14-Case B). . . . . . . . . . . . . . . . . . . . . . . . . The communication system. . . . . . . . . . . . . . . . . . . . . The simulated communication system. . . . . . . . . . . . . . .
. 89 . 91 . 94 . 94
7.1 7.2
λZ,f (Z) (D + γ) > ε ⇒ sup[θ : λZ,f (Z) (θ) ≤ ε] ≤ D + γ. . . . . . . . 115 The CDF of (1/n)ρn (Z n , fn (Z n )) for the probability-of-error distortion measure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117
8.1
The geometric meaning for Sanov’s theorem. . . . . . . . . . . . . 126
xi
8.2 8.3 8.4
The divergence view on hypothesis testing. . . . . . . . . . . . . . 134 Function of (π/6)u − h(u). . . . . . . . . . . . . . . . . . . . . . . 168 The Berry-Esseen constant as a function of the sample size n. The sample size n is plotted in log-scale. . . . . . . . . . . . . . . . . . 170
9.1
BSC channel with crossover probability ε and input distribution (p, 1 − p). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.2 Random coding exponent for BSC with crossover probability 0.2. Also plotted is s∗ = arg sup0≤s≤1 [−sR − E0 (s)]. Rcr = 0.056633. 9.3 Expurgated exponent (solid line) and random coding exponent (dashed line) for BSC with crossover probability 0.2 (over the range of (0, 0.192745)). . . . . . . . . . . . . . . . . . . . . . . . 9.4 Expurgated exponent (solid line) and random coding exponent (dashed line) for BSC with crossover probability 0.2 (over the range of (0, 0.006)). . . . . . . . . . . . . . . . . . . . . . . . . . 9.5 Partitioning exponent (thick line), random coding exponent (thin line) and expurgated exponent (thin line) for BSC with crossover probability 0.2. . . . . . . . . . . . . . . . . . . . . . . . . . . . c ; (b) The shaded area is Acm,m0 . . . . . 9.6 (a) The shaded area is Um ¯ X (R) asymptotically lies between ess inf(1/n)µ(X ˆ n , X n ) and 9.7 Λ ˆ n , X n )|µn (X ˆ n , X n )] for R ¯ p (X) < R < R ¯ 0 (X). . . . (1/n)E[µn (X ¯ X (R). . . . . . . . . . . . . . . . . . . . . . . 9.8 General curve ofΛ 9.9 Function of sup −sR − s log 2s 1 − e−1/2s . . . . . . . . . . s>0 9.10 Function of sup −sR − s log 2 + e−1/s /4 . . . . . . . . . .
. 187 . 188
. 195
. 196
. 201 . 247 . 248 . 248 . 249 . 249
s>0
10.1 The multi-access channel. . . . . . . . . . . . . . . . . . . . . . . 10.2 Distributed detection with n senders. Each observations Yi may come from one of two categories. The final decision D ∈ {H0 , H1 }. 10.3 Distributed detection in Sn . . . . . . . . . . . . . . . . . . . . . . 10.4 Bayes error probabilities associated with gˆ and g¯. . . . . . . . . . 10.5 Upper and lower bounds on e∗NP (α) in Case B. . . . . . . . . . . .
xii
253 254 260 266 275
Chapter 1 Introduction This volume will talk about some advanced topics in information theory. The mathematical background on which these topics are based can be found in Appendices A and B of Volume I.
1.1
Notations
Here, we clarify some of the easily confused notations used in Volume II of the lecture notes. For a random variable X, we use PX to denote its distribution. For convenience, we will use interchangeably the two expressions for the probability of X = x, i.e., Pr{X = x} and PX (x). Similarly, for the probability of a set characterizing through an inequality, such as f (x) < a, its probability mass will be expressed by either PX {x ∈ X : f (x) < a} or Pr {f (X) < a} . In the second expression, we view f (X) as a new random variable defined through X and a function f (·). Obviously, the above expressions can be applied to any legitimate function f (·) defined over X , including any probability function PXˆ (·) (or log PXˆ (x)) of a ˆ Therefore, the next two expressions denote the probability random variable X. of f (x) = PXˆ (x) < a evaluated under distribution PX : PX {x ∈ X : f (x) < a} = PX {x ∈ X : PXˆ (x) < a} and Pr {f (X) < a} = Pr {PXˆ (X) < a} .