On Distortion Bounds for Dependent Sources Over Wireless Networks Shirin Jalali

Michelle Effros

Center for Mathematics of Information California Institute of Technology Pasadena, California, 91125 Email: [email protected]

Department of Electrical Engineering California Institute of Technology Pasadena, California, 91125 Email: [email protected]

Abstract—Characterizing the set of achievable distortions in lossy transmission of dependent sources over multi-user channels is an open problem. Even for a channel such as the multiple access channel, where the capacity is well-understood, the complete characterization of achievable distortions is still unknown. Since separation of source coding and channel coding is sub-optimal for this case, bounds on achievable distortions require tools beyond prior channel capacities and rate-distortion bounds. In this paper we propose a systematic approach for finding lower bounds on the set of distortions achievable through joint source-channel coding.

I. P ROBLEM S TATEMENT Shannon’s famous separation theorem states that in a network described by a single point-to-point channel, separating source coding and channel coding operations is optimal [1]. Unfortunately, this result does not generalize to multi-user channels with dependent sources [2], [3]. Characterizing the set of distortion vectors achievable by joint source-channel (JSC) coding over such channels is an open problem. Even for channels with few inputs and outputs, such as a 2-user multiple access channel (MAC) or a 2-user broadcast channel (BC), this characterization is still unknown. Inner and outer bounds on the set of achievable distortions are useful until a complete characterization becomes available. While evaluating the performance of separate source and channel codes provides a systematic method for finding upper bounds on minimum achievable distortions, it seems that lower bounds are more difficult to obtain since they require that we develop an understanding of JSC codes. The only known methods for deriving lower bounds are based on cut-set bounds [4], [5]. As a result, very little is known even for the smallest networks. For example, several different achievability schemes for JSC coding over a BC are proposed in the literature, but no non-trivial lower bound on distortion was known until recently (see [6], [7] and refenences therein). Combining the idea of bounding networks from [8] with the idea of coordination capacity described in [9], we here propose a systematic method for finding lower bounds on the set of achievable distortion vectors in multi-user memoryless channels with dependent sources. The idea is to replace the multi-user discrete memoryless channel (DMC) with a noiseless network of point-to-point bit-pipes capable of simulating the memoryless joint distribution that would otherwise be

established by the DMC. We show that after this replacement, the distortion achievable across the new network serves as a lower bound on the distortion achievable by JSC coding across the original network. Since the new network is a noiseless network, characterizing its achievable distortion region is a pure source coding problem. The organization of this paper is as follows. Section II summarizes our notation and definitions. Section III describes the problem setup. Section IV briefly reviews the ideas of coordination capacity and coordination-rate region. In Section V, we present our approach for finding lower bounds on the the set of achievable distortions for dependent sources over multiuser channels. Section VI concludes the paper.

II. N OTATION

AND

D EFINITIONS

Finite sets are denoted by script letters such as X and Y. The size of a finite set A is denoted by |A|. Random variables are denoted by upper case letters such as X and Y . The alphabet of a random variable X is represented by X . Bold face letters represent vectors. A random vector is represented by upper case bold letters like X or Y. The dimension of a vector is implied in the context. The ℓth element of vector X is denoted by Xℓ . A vector x = (x1 , . . . , xn ) (X = (X1 , . . . , Xn )) is sometimes represented as xn (X n ). For 1 ≤ i ≤ j ≤ n, xji = (xi , xi+1 , . . . , xj ). For a set A ⊆ {1, 2, . . . , n}, xA = (xi )i∈A , where the elements are sorted in ascending order of their indices. For integers n1 ≤ n2 , let [n1 : n2 ] , {n1 , n1 + 1, . . . , n2 }. For two vectors x, y ∈ Rr , x ≤ y iff ri ≤ yi for all 1 ≤ i ≤ r The ℓ1 distance between vectorsPx and y of the same r dimension r is denotedP by kx−yk1P= i=1 |xi −yi |. If x and r r y represent pmfs, i.e., i=1 xi = i=1 yi = 1 and xi , yi ≥ 0 for all i ∈ {1, . . . , r}, then the total variation distance between x and y is defined as kx − ykTV = 0.5kx − yk1 . Definition 1: For a sequence xn ∈ X n , the empirical distribution of xn is defined as |{i : xi = x}| , (1) π(x|xn ) , n for all x ∈ X . Similarly, for (xn , y n ) ∈ X n × Y n , the joint

ˆ i→1 }i∈β(1) {U

Y1 U1

D

X1

E

ˆ i→2 }i∈β(2) {U

Y2 X2

U2

D

DMC

E

p(y|x)

where d(i→j) : Ui × Uˆi→j → R+ is a per-letter distortion measure. Let dmax , maxj,i∈β(j),a,b d(i→j) (a, b) < ∞. The distortion vector D , (Di→j )j∈J ,i∈β(j) is said to be achievable at a rate κ, if for any ǫ > 0, there exists a pair (L, n) with L/n ≤ κ, and a block code of source blocklength L and channel blocklength n such that (i→j)

E[dL Uℓ

Xℓ

E

Fig. 1.

Ym

D

ˆ i→m }i∈β(m) {U

A multi-user channel with dependent sources.

empirical distribution of (xn , y n ) is defined as π(x, y|xn , y n ) ,

|{i : (xi , yi ) = (x, y)}| , n

(2)

for all (x, y) ∈ X × Y. III. T HE S ET-U P Consider an ℓ-input, Qm by p(y|x), Qℓ m-output DMC described where x = xℓ ∈ i=1 Xi and y = y m ∈ j=1 Yj . (See Fig. 1.) Let I , [1 : ℓ] and J , [1 : m]. Transmitter i, i ∈ I, observes the random process {Ui,t }∞ t=1 . The sources are jointly distributed and memoryless. That is, for any T ∈ N, P ((U1 , U2 , . . . , Uℓ ) = (u1 , u2 , . . . , uℓ )) =

T Y

p(u1,t , u2,t , . . . , uℓ,t ),

(3)

t=1

where for i ∈ I, Ui = (Ui,1 , Ui,2 , . . . , Ui,T ) and ui = (ui,1 , ui,2 , . . . , ui,T ). Receiver j ∈ J is interested in lossy reconstruction of a subset β(j) ⊆ I of sources observed by the transmitters. A block code of source blocklength L and channel blocklength n is described as follows. Transmitter i ∈ I observes a block Ui = (Ui,1 , Ui,2 , . . . , Ui,L ) of length L. Encoder i maps the observed block Ui to the encoded block Xi of length n, i.e., Xi : UiL → Xin . Receiver j ∈ J receives Yj = (Yj,1 , Yj,2 , . . . , Yj,n ), and Decoder j maps the received block into reconstruction blocks ˆ i→j }i∈β(j) , i.e., for j ∈ J and i ∈ β(j), {U L ˆ i→j : Yjn → Uˆi→j U .

The rate κ , L n is a parameter of the code. The performance of the described coding scheme is measured by the induced expected average distortions between source and reconstruction blocks, i.e., for j ∈ J and i ∈ β(j), (i→j)

Di→j , E[dL

L

= E[

ˆ i→j )] (Ui , U

1 X (i→j) ˆi→j,t )], d (Ui,t , U L t=1

ˆ i→j )] ≤ Di→j + ǫ, (Ui , U

(4)

for any j ∈ J and i ∈ β(j). Let D(κ, p(y|x)) denote the set of distortions achievable at rate κ on the ℓ-input, m-output DMC described by p(y|x). Given a bit-pipe network Nb with links of capacity R, let D(κ, R) denote the set of distortions achievable at rate κ on Nb . Our aim is to replace the JSC region D(κ, p(y|x)) by the distortion-rate region D(κ, R) for a suitable network Nb . When ℓ = m = 1, D(κ, p(y|x)) is fully characterized for all source distributions p(u) and channel distributions p(y|x). For ℓ > 1 or m > 1, the problem remains unsolved. IV. C OORDINATION C APACITY The idea of coordination capacity was introduced in [9]. Instead of considering a network of bit-pipes as a platform for communicating information or reconstructing random processes, this work uses the network to generate desired dependencies among the users’ processes. In this section, we briefly review this idea. In the next section we show how it can be combined with an idea from [8] for use in deriving lower bounds on the distortions of JSC codes. Consider the ℓ-input, m-output acyclic graph Nb shown in Fig. 2. Network Nb is represented by a directed graph G = (V, E) with E ⊂ V×V. Let In(v) = {(v1 , v) : (v1 , v) ∈ E} and Out(v) = {(v, v1 ) : (v, v1 ) ∈ E} denote the sets of incoming and outgoing edges of Node v. Given a rate vector R = (Re : e ∈ E) we use (Nb , R) to represent a bit-pipe network in which each directed edge e = (v1 , v2 ) ∈ E represents a noiseless link of capacity Re . An i.i.d. jointly distributed ranN dom process {(X1,t , . . . , Xℓ,t )}t≥1 with P((xN 1 , . . . , xℓ ) = QN N N (x1 , . . . , xℓ )) = t=1 po (x1,t , . . . , xℓ,t ) is available to the network input nodes 1, . . . , ℓ. A coordination code of block length N for network Nb is a code for sending messages across network Nb in order to establish a joint distribution p(xℓ , y m ) = po (xℓ )p(y m |xℓ ) between the channel inputs and outputs. The message We carried over link e has alphabet We = {1, 2, . . . , 2⌊N Re ⌋ }. Input node vi , i ∈ I, observes a block XiN of length N and maps it to the messages (We : e ∈ Out(vi )) sent over its outgoing links, giving We : XiN → We for all i ∈ I, and e ∈ Out(vi ). Any node v ∈ V\I ∪ J inside the network maps its incoming messages (We : e ∈ In(v))Qto its outgoing messages (We : e ∈ Out(v)), i.e., W(v,v′ ) : e∈In(v) We → W(v,v′ ) for all (v, v ′ ) ∈ Out(v). Output node vj′ , j ∈ J , maps its incoming messages (We : e ∈ In(vj′ )) to Y˜jN , i.e., Q Y˜jN : e∈In(j) We → YjN . The performance of the coordination code is measured by the total variation distance between po (x)p(y|x) and ˜ N ). π(x, y|XN , Y

Y˜1N

Y1

X1N

U

X1

E

noiseless network

Y2

D

U

R1

E

N Y˜m

R1

D

R1

D

R2

Fig. 2.

ˆ2 U

(a) JSC coding over a BC

Nb XℓN

ˆ1 U

p(y|x)

Y˜2N X2N

D

R1 κ−1

U

For a given po (x) conditional distribution p(y|x) is said to be achievable on Nb at rate R, if there exists a sequence of codes such that

= {R : p(y|x) achievable on Nb at rate R}, where for a set A, A denotes the closure of the set A. V. O UTER B OUNDS In [8], Koetter et al. introduced systematic strategy for bounding the capacity of a network of stochastic multiterminal channels by the capacity of a network of noiseless bit-pipes. The key to deriving the outer bounds presented in that work is to show that if the bit-pipe network can simulate the distribution of the stochastic network then any code for the stochastic network can also be run across the bit-pipe network. Combining the new proof technique and the channel simulation results of [9], we next derive a corresponding strategy for finding outer bounds on the distortions achievable by JSC codes. We begin with an example. Example 5.1: Broadcast channel Let ℓ = 1 and m = 2. Then p(y|x) describes a BC as shown in Fig. 3(a). Fig. 3(b) shows the bit-pipe model Nb proposed in [8], which has one bit pipe of capacity R2 and 3 bit pipes of capacity R1 . Given a JSC code operating at rate κ, and distortion D on the BC, we build a code for Nb as follows. (See Fig. 3(c)) The encoder first maps U = U L to a channel input X1 using the JSC encoder for the BC. The encoder then applies the mappings from a coordination code to map the channel input into a pair of messages of rates R1 and R2 for transmission across Nb . The coordination code simulates distribution po (x)p(y|x) across Nb , where po (x) is 1 Note that this definition is slightly different than the definition presented in [9]. Here we require almost sure convergences, but the convergences in [9] are in probability.

D

E D

ˆ1 U

ˆ2 U

R2 κ−1

(c) Translation of JSC coding rate κ to the network of bit pipes.

˜ N )kTV → 0, kpo (x)p(y|x) − π(x, y|XN , Y

Rpo (Nb , p(y|x))

ˆ2 U

(b) The BC in Fig. 3(a) is replaced by 4 point-topoint bit pipes

Simulating the channel p(y|x) using network Nb

almost surely as N grows without bound1. The coordinationrate region of Network Nb under input distribution po (x) is defined as [9]

ˆ1 U

Fig. 3.

Example of replacing a DMC with a noiseless network of bit pipes

the channel input distribution employed by the JSC and p(y|x) is the defining distribution for the BC. Each receiver applies the coordination decoder for the given node and then applies ˆ L . Theorem the decoder for the JSC to build reconstruction U 1 shows that the code achieves distortion approaching D as the block length of the coordination code grows without bound, provided that (R1 , R2 ) ∈ Rpo (Nb , p(y|x)), which shows D(κ, p(y|x)) ⊆ D(κ, R). T Theorem 1: If R ∈ p(x) Rp (Nb , p(y|x)), then D(κ, p(y|x)) ⊆ D(κ, R).

(5)

Proof outline: Given any D ∈ D(κ, p(y|x)), consider a JSC code for p(y|x) with source blocklength L, channel ˆ ≤ D + ǫ · 1. For blocklength n, L = ⌊nκ⌋, and distortion D ˆ ˆ i→j , for some the rest of the proof, let U = Ui and U = U j ∈ J and i ∈ β(j). By the law of iterated expectations, we can express the distortion of the JSC code on channel p(y|x) as ˆ Di→j = E[dL (U, U)] h i ℓ ˆ = E E[dL (U, U)|(X , Ym )] X ℓ ˆ = E[dL (U, U)|(X , Ym ) = (xℓ , ym )] (xℓ ,ym )

× P((Xℓ , Ym ) = (xℓ , ym )) X ℓ ˆ = E[dL (U, U)|(X , Ym ) = (xℓ , ym )] (xℓ ,ym )

× P(Xℓ = xℓ ) P(Ym = ym |Xℓ = xℓ ) X ℓ ˆ = E[dL (U, U)|(X , Ym ) = (xℓ , ym )] (xℓ ,ym )

(6)

× ×

n Y

t=1 n Y

P(XI,t = xI,t |XI,1:t−1 = xI,1:t−1 ) p(y1,t , y2,t , . . . , ym,t |x1,t , x2,t , . . . , xℓ,t ),

(7)

t=1

where XI,1:t = (XI,1 , . . . Q , XI,t ) and YJQ = ,1:t (YJ ,1 , . . . , YJ ,t ), where XI,t ∈ i∈I Xi , YJ ,t ∈ j∈J Yj denote the ℓ inputs and m outputs of the channel at time t. Since the JSC code may employ a different distribution on the channel inputs at each time t, we cannot combine the JSC code with a coordination code applied across time. Instead, following the proof from [8], we send a sequence of independent messages and apply the coordination code to a block of channel inputs resulting from different messages at the same time step t in our JSC code. We begin showing that coding a sequence of N independent messages does not change the performance of the code on channel p(y|x). Assume that each source i ∈ I observes a block UiN L of length N L, and breaks it into N non-overlapping sub-blocks of length L, UiN L = (Ui (1), Ui (2), . . . , Ui (N )). During N independent sessions, these sub-blocks are described to the decoders applying the same code that achieves distortion vector D. In session k, k ∈ [1 : N ], each encoder transmits the k th sub-block Ui (k). Each decoder reconstructs the k th ˆ i→j (k)}j∈J ,i∈β(j) . Since sub-block giving reconstructions {U the sessions are independent, the code’s expected average ˆ N L )] = distortion at receivers is given by E[dN L (UiN L , U i→j PN 1 ˆ ˆ k=1 E[dL (Ui (k), Ui→j (k))] = Di→j . N To show that D ∈ D(κ, R), we next combine the given JSC code with a coordination code and run that code across network Nb with edge capacity vector R. Since R ∈ T p(x) Rp(x) (Nb , p(y|x)), Nb can simulate the distribution p(x)p(y|x) at the outputs using a coordination code. We combine the JSC code with the coordination code as follows. Each input node vi , i ∈ I first encodes blocks (Ui (1), . . . , Ui (N )) using the node-i encoder for the JSC independently on each block. At each time t ∈ [1 : n], node vi blocks together the tth symbols of the N sub-blocks of the JSC output, giving Xi,t = (Xi,t (1), Xi,t (2), . . . , Xi,t (N )). Since the source sub-vectors Ui (1), . . . , Ui (N ) are independent and the sub-blocks are coded independently, Xi,t (1), Xi,t (2), . . . , Xi,t (N ) is distributed i.i.d. pt (xi,t ). XI,t (1), XI,t (2), . . . , XI,t (N ) is distributed i.i.d. pt (xI,t ), where pt (xi,t ) is the distribution on the channel input from node vi and pt (xI,t ) is the distribution on the channel input from all nodes induced by Q the JSC code at time t in channel p(y|x). For each xI,t ∈ i∈I Xi , X X pt (xI,t ) = p(xI,1:n |uL )p(uL ), (8) xI,1:t−1 ,xI,t+1:n uL L

L

where p(xI,1:n |u ) equals 1 if the JSC encoders encode U to xI,1:n , and 0 otherwise. At each time t ∈ [1 : n], we employ a coordination code of blocklength N on Network Nb that simulates conditional distribution p(y|x) for the input distribution pt (x) defined

T in (8). Since R ∈ p(x) Rp(x) (Nb , p(y|x)) by assumption, for each pt (x) and all N sufficiently large there exists a ˜ tN for which kpt (x)p(y|x)− coordination code with outputs Y N ˜N π(x, y|Xt , Yt )kTV ≤ ǫ. Fix such a coordination code for each t ∈ [1 : n]. Applying these coordination codes, at time t, the output node vj′ , j ∈ J , maps the messages it receives from ˜ j,t , (Y˜j,t (1), . . . , Y˜j,t (N )). After its incoming bit pipes to Y n time steps, decoder j reassemples N sub-blocks of length ˜ jN = (Y ˜ j (1), . . . , Y ˜ j (N )), where Y ˜ j (k) = Y ˜ j,1:n (k). n as Y Node vj then applies the JSC decoder independently on each ˜ j (k), k ∈ [1 : N ]. The resulting reconstruction sub-block Y ˜ N }i∈β(j) . blocks are {U i→j To bound the expected distortion of this code, define ˜ m (k)) = the event Ak,xℓ ,ym as Ak,xℓ ,ym , {( Xℓ (k), Y ℓ m (x , y )}. Note that for any k ∈ [1 : N ], X 1Ak,xℓ ,ym = 1. (9) xℓ ,ym :xi ∈Xin ,yj ∈Yjn

˜ i→j )j∈J ,i∈β(j) denote the expected average ˜ = (D Let D distortion performance of the new code. As before, let UN = ˜N = U ˜ N , for some j ∈ J and i ∈ β(j). Since UN i , and U i→j the sources and the channel are memoryless, h i ℓ ˜ ˜ m (k)) = (xℓ , ym ) E dL (U(k), U(k))|(X (k), Y h i ℓ ˆ = E dL (U, U)|(X , Ym ) = (xℓ , ym ) . (10) Hence,

˜ i→j = E[dN L (UN , U ˜ N )] D =

=

N 1 X ˜ E[dL (U(k), U(k))] N k=1 

N 1 X   ˜ E dL (U(k), U(k)) N  k=1

=

(xℓ ,ym ): xi ∈Xin ,yj ∈Yjn



1Ak,xℓ ,ym  

N 1 X ˜ E 1Ak,xℓ ,ym dL (U(k), U(k)) N

X

h i ℓ ˜m ˆ E dL (U, U)|(X , Y ) = (xℓ , ym )

(xℓ ,ym ): xi ∈Xin ,yj ∈Yjn

k=1



#

X

(xℓ ,ym ): xi ∈Xin ,yj ∈Yjn

=

"

X



  × E π xℓ , ym |Xℓ (1 : N ), Ym (1 : N ) .

On the other hand,   ˜ m (1 : N ) π xℓ , ym |Xℓ (1 : N ), Y =

˜ m (k)) = (xℓ , ym )}| |{k : (Xℓ (k), Y N (11)

=

n Y

0.4

t=1

D2

˜ J ,1:t (1 : N )) π(xI,1:t , yJ ,1:t |XI,1:t (1 : N ), Y , ˜ J ,1:t−1 (1 : N )) π(xI,1:t−1 , yJ ,1:t−1 |XI,1:t−1 (1 : N ), Y (12)

Di n Do u t

0.3

0.2

0.1

˜ J ,1:t−1 (1 : where π(xI,1:t−1 , yJ ,1:t−1 |XI,1:t−1 (1 : N ), Y N )) , 1, for t = 0. Comparing (7) and (12) reveals that to finish the proof, it is enough to show that, for any t ∈ [1 : n], ˜ J ,1:t (1 : N )) π(xI,1:t , yJ ,1:t |XI,1:t (1 : N ), Y , ˜ J ,1:t−1 (1 : N )) π(xI,1:t−1 , yJ ,1:t−1 |XI,1:t−1 (1 : N ), Y (13)

0.3

D1

0.4

0.5

Fig. 4. Inner and outer bounds described in Example 5.2 for sending a binary source over a binary BC. (q = 0.5, p1 = 0.25, p2 = 0.05, κ = 1)

VI. C ONCLUSION

converges to P(XI,t = xI,t |XI,1:t−1 = xI,1:t−1 )p(yJ ,t |xI,t ),

0 0.2

(14)

almost surely, as N → ∞. Then, since π(·|·) is a positive and bounded function, by the bounded convergence theorem, ˜ i→j → D ˆ i→j , as N → ∞. The almost sure convergence D ˜ J ,t (1 : N )) to p(xI,t , yJ ,t ) of π(xI,t , yJ ,t |XI,t (1 : N ), Y follows from the (modified) definition of coordination capacity. Since the cooperation codes are chosen independently to simulate the channel p(y|x) at each time t ∈ [1 : n], and since there is no feedback from the outputs to the inputs of the channel, the almost sure convergence for each time t implies the convergence of (13) to (14), which gives the desired result. Example 5.2 (Binary source over a binary symmetric BC): Consider sending an i.i.d. binary Bern(q) source over a 2-user binary symmetric BC with input X and outputs Y1 = X + Z1 and Y2 = X + Z2 , where for i ∈ {1, 2}, Zi ∼ Bern(pi ), p2 < p1 , and Z1 and Z2 are independent. The point-to-point rate-distortion function of the source is R(D) , max(0, h(q) − h(D)), where h(β) = −β log β − (1 − β) log(1 − β) [10]. Define δ(r) , R−1 (r), i.e., δ(r) = d iff R(d) = r. Since the source is successively refinable [11], the set of achievable distortions by separating source and channel coding over the BC can be described as [   (D1 , D2 ) : D1 ≥ δ κ−1 (1 − h(p1 ∗ α)) , Din (κ) = α∈[0,1]

D2 ≥ δ κ−1 (1 − h(p1 ∗ α) + h(α ∗ p2 ) − h(p2 ))



.

On the other hand, R1 ≥ 1−h(p1 ) and R2 ≥ h(p1 )−h(p2 ) is sufficient for simulating the described binary BC [8]. Hence, by Theorem 1,   Dout (κ) = (D1 , D2 ) : D1 ≥ δ κ−1 (1 − h(p1 )) ,  D2 ≥ δ κ−1 (1 − h(p2 )) ,

defines an outer bound on the set achievable distortions, i.e., Din (κ) ⊂ D(κ, p(y1 , y2 |x)) ⊂ Dout (κ). Fig. 4 demonstrates an example of the inner and outer bounds for the case of q = 0.5, p1 = 0.25, p2 = 0.05, κ = 1. In this case the derived lower bound coincides with the cut-set lower bound.

We considered the problem of sending dependent sources over multi-user memoryless channels and proposed a systematic approach for finding lower bounds on the set of achievable distortions. In such networks, separation of source coding and channel coding is sub-optimal. Upper bounds on the performance can be derived by evaluating the performance of separation-based coding schemes. Combining the ideas from coordination capacity and network equivalence, we here proposed a systematic approach for finding lower bounds on the achievable distortion. We gave a small example where outer bounds were previously unknown. ACKNOWLEDGMENTS This work is partly supported by the Center for Mathematics of Information at Caltech and the NSF grant number CCF1018741. R EFERENCES [1] C. E. Shannon. A mathematical theory of communication: Parts I and II. Bell Syst. Tech. J., 27:379–423 and 623–656, 1948. [2] T. Cover, A. El Gamal, and M. Salehi. Multiple access channels with arbitrarily correlated sources. IEEE Transactions on Information Theory, 26(6):648–657, November 1980. [3] U. Mittal and N. Phamdo. Hybrid digital-analog (HDA) joint sourcechannel codes for broadcasting and robust communications. IEEE Trans. Inform. Theory, 48(5):1082–1102, May 2002. [4] M. Gastpar. Cut-set arguments for source-channel networks. In Proc. IEEE Int. Symp. Inform. Theory, Chicago, IL, June 2004. [5] A.A. Gohari and V. Anantharam. A generalized cut-set bound. In Proc. IEEE Int. Symp. Inform. Theory, pages 99–103, Seoul, Korea, July 2009. [6] Z. Reznic, M. Feder, and R. Zamir. Distortion bounds for broadcasting with bandwidth expansion. IEEE Transactions on Information Theory, 52(8):3778–3788, Aug. 2006. [7] C. Tian, S. Diggavi, and S. Shamai. Approximate characterization of the gaussian source broadcast distortion region. arXiv:0906.3183v1. [8] R. Koetter, M. Effros, and M. M´edard. A theory of network equivalence, parts I and II. arXiv:1007.1033v2. [9] P.W. Cuff, H.H. Permuter, and T.M. Cover. Coordination capacity. IEEE Transactions on Information Theory, 56(9):4181–4206, Sept. 2010. [10] T. Cover and J. Thomas. Elements of Information Theory. Wiley, New York, 2nd edition, 2006. [11] W.H.R. Equitz and T.M. Cover. Successive refinement of information. IEEE Transactions on Information Theory, 37(2):269–275, Mar. 1991.

On Distortion Bounds for Dependent Sources Over ...

multiple access channel (MAC) or a 2-user broadcast channel .... Receiver j ∈ J is interested in lossy ... Receiver j ∈ J receives Yj = (Yj,1,Yj,2,...,Yj,n), and.

122KB Sizes 1 Downloads 235 Views

Recommend Documents

New bounds on the rate-distortion function of a binary ...
and Hb(x) = −x log 2−(1−x log(1−x)). Beyond Dc, even for this simple case, currently only lower and upper bounds bounds on R(D) are known. In 1977, Berger found explicit lower and upper bounds on. R(D), Rl(D) and Ru(D) respectively, which do

Deterministic Extractors for Affine Sources over Large ...
May 16, 2007 - We denote by Fq the finite field of q elements. We denote by Fq the algebraic closure of Fq and by Fq[t] the ring of formal polynomials over Fq. We denote by F ...... Tools from higher algebra. In R. L. Graham & M. Grotschel & L. Lovas

Extractors for Polynomials Sources over Constant-Size ...
Sep 22, 2011 - In this work, we construct polynomial source extractors over much smaller fields, assuming the characteristic of the field is significantly smaller than the field size. Theorem 1 (Main — Extractor). Fix a field Fq of characteristic p

On upper bounds for high order Neumann eigenvalues of convex ...
Let Ω be a bounded convex domain in Euclidean space Rn. By considering a manifold ... where f1/(n−1) is a positive and concave function on the interval (0,1).

Deterministic Performance Bounds on the Mean Square Error for Near ...
the most popular tool [11]. However ... Date of publication November 27, 2012; ... of this manuscript and approving it for publication was Dr. Benoit Champagne.

Lower Bounds on Deterministic Schemes for the ...
of space and 2 probes is presented in a paper by Radhakrishnan, Raman and Rao[2]. ... BigTable. Google uses nondeterministic space efficient data structures ...

Deterministic Performance Bounds on the Mean Square Error for Near ...
mean square error applied to the passive near field source localization. More precisely, we focus on the ... Index Terms—Deterministic lower bounds, mean square error, near field source localization, performance analysis ..... contained in the samp

On upper bounds for high order Neumann eigenvalues ...
Let Ω be a bounded convex domain in Euclidean space Rn. By considering a .... It is easy to check that G is constant on every interval where g = f1/(n−1) is ...

Bounds on the Automata Size for Presburger Arithmetic
notice is given that copying is by permission of the ACM, Inc. To copy otherwise, .... However, it is possible to express k · x by a ...... Let Ad|t+c := (Q, Σr, δ, qI,F) be.

Structural Risk Minimization over Data-Dependent ...
Nov 28, 1997 - [email protected]. Peter L. Bartlett. Department of Systems Engineering. Australian National University. Canberra 0200 Australia. Peter.

Setting Lower Bounds on Truthfulness
prove our lower bounds for this problem in the inter-domain routing setting ..... 100 otherwise. Let M be a deterministic, weakly-monotone, mechanism that ...... We assume that for every i vi(∅) = 0 (free disposal) and for every two bundles of ...

SPECTRAL DISTORTION MODEL FOR ... - Research at Google
[27] T. Sainath, O. Vinyals, A. Senior, and H. Sak, “Convolutional,. Long Short-Term Memory, Fully Connected Deep Neural Net- works,” in IEEE Int. Conf. Acoust., Speech, Signal Processing,. Apr. 2015, pp. 4580–4584. [28] E. Breitenberger, “An

Growth and distortion theorems on a complex Banach ...
[1] R.W. Barnard, C.H. FitzGerald, S. Gong, A distortion theorem for biholo- morphic mappings in C2, Trans. Amer. Math. Soc. 344 (1994) 907–924. [2] C.-H. Chu, H. Hamada, T. Honda, G. Kohr, Distortion theorems for convex mappings on homogeneous bal

On some upper bounds on the fractional chromatic ...
This work was carried out while the author was at the University of Wisconsin at. Madison, USA. Thanks are due to professor Parmesh Ramanathan for suggesting this direction. References. [1] B. Bollobás. Modern Graph Theory. Springer, Graduate Texts i

Device and method for compensating for phase distortion in base ...
May 13, 2010 - Networks, Part 16: Air Interface for Fixed Broadband Wireless .... It is an advantage of the present invention to provide a device and method for ...

On explicit bounds for the spectral gap on compact ...
We have shown in [2] that the smallest positive eigenvalue of the Laplace- ... note is to give explicit lower bounds for the eigenvalues of the corresponding Sturm ...

RESONANCES AND DENSITY BOUNDS FOR CONVEX CO ...
Abstract. Let Γ be a convex co-compact subgroup of SL2(Z), and let Γ(q) be the sequence of ”congruence” subgroups of Γ. Let. Rq ⊂ C be the resonances of the ...

Learning Bounds for Domain Adaptation - Alex Kulesza
data to different target domain with very little training data. .... the triangle inequality in which the sides of the triangle represent errors between different decision.

Improved Competitive Performance Bounds for ... - Semantic Scholar
Email: [email protected]. 3 Communication Systems ... Email: [email protected]. Abstract. .... the packet to be sent on the output link. Since Internet traffic is ...

EFFICIENCY BOUNDS FOR SEMIPARAMETRIC ...
Nov 1, 2016 - real-valued functions on Rk. Assume that we are given a function ψ which maps Rp ×Rk into Rq with ..... rt = Λft + ut,. (6) with Λ = (λ1,λ2) ∈ R2, ft a common factor that is a R-valued process such that E(ft|Ft−1) = 0 and. Var

Rademacher Complexity Bounds for Non-I.I.D. Processes
Department of Computer Science. Courant Institute of Mathematical Sciences. 251 Mercer Street. New York, NY 10012 [email protected]. Abstract.

BOUNDS FOR TAIL PROBABILITIES OF ...
E Xk = 0 and EX2 k = σ2 k for all k. Hoeffding 1963, Theorem 3, proved that. P{Mn ≥ nt} ≤ Hn(t, p), H(t, p) = `1 + qt/p´ p+qt`1 − t´q−qt with q = 1. 1 + σ2 , p = 1 − q, ...

On Generalized Weights for Codes over Zk
Jun 22, 2011 - introduced the GHWR of a linear code C over a finite chain ring and studied some properties of the GHWR. For any g, 1 ≤ g ≤ rank(C), we define the higher weight spectrum as. A g i = |{D : D is a Zk-submodule of C with rank(D) = g a