You are on page 1of 11

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. IT-26, NO.

2, MARCH 1980 155

Communicafionr. Urbana, IL: Univ. of Illinois, 1949. 1’71 J. B. Anderson and S. Mohan, “A systematic analysis of cost for
I121 T. Berger, Rate Disfortion Theoty. Englewood Cliffs, NJ: Pren- sequential coding alaorithms,” Commun. Res. Lab.. Facultv of
tice-Hall, 1971. En&, McMaster-Univ., Hamilton, ON, Canada, Tech. Rep.-No. .,
I131 R. E. Blahut, “Computation of channel capacity and rate-distor- CRL-56, July 1978.
tion functions,” IEEE Trans. Inform. Theory, vol. IT-18, pp. I181 J. B. Anderson, “Algorithms for tree coding with a fidelity criter-
460-473, July 1972. ion,” Ph.D. dissertation, School of Elec. Eng., Cornell Univ.,
[I41 W . A. Finamore, “Optimal encoding of continuous-amplitude Ithaca, NY, May 1972.
sources with constrained size output alphabet,” Ph.D. dissertation, 1191 T. Berger, F. Jelinek, and J. K. Wolf, “Permutation codes for
Dep. Elec. Comput. Eng., Univ. Wisconsin-Madison, Madison, sources,” IEEE Trans. Inform. Theoty, vol. IT-18, pp. 160-169,
W I, May 1978. Jan. 1972.
iI51 A. J. Viterbi and J. K. Omura, “Trellis encoding of memoryless 1201Y. Linde and R. M. Gray, “A fake process approach to data
discrete-time sources with a fidelity criterion,” IEEE Trans. In- compression,” IEEE Trans. Comm., vol. COM-26, pp. 840-847,
form. Theory, vol. IT-20, pp. 325-332, May 1974. June, 1978.
1161 C. R. Davis and M. E. Hellman, “On tree coding with a fidelity Pll F. Jelinek and J. B. Anderson, “Instrumentable tree encoding of
criterion,” IEEE Trans. Inform. Theory, vol. IT-21, pp. 373-378, information sources,” IEEE Trans. Inform. Theory, vol. IT-17, pp.
July 1975. 118-l 19. Jan. 1971.

Towards a General Theory of Source Networks


IMRE CSISZAR and JANOS KGRNER

Absttzzct-A unified approach to muhitermlnal source codll problem fied sources from the codewords it has received. A prob-
not lnvolvlng rate4Jistortlon theory is presented. It is shown that, for
determlnlng the achievable rate region, attention may be restricted to
lem of primary interest is to determine the achievable rate
source networks of a relatively simple structure. A product space char- region, i.e., those rates of the individual encoders for
acterization of the achievable rate region pinpoints the mathematical which the reproduction requirements can be met. These
problem to be solved for getting a single letter characterization. The m u ltiterminal source coding problems are substantially
complex&y of this problem depends on a structural condltiou, viz., the more difficult than the standard textbook material ‘con-
number of encoders of a certain kind in the source network. Tbls approach
cerning the encoding of a single source. An essential
yields all the known slugle-letter cbaracterizatious of achievable rate
regions and a number of new ones for more complex networks. As a feature is that the answer cannot be given in terms of the
digression, for a class of source networks including that of Slepiau and individual sources only; rather, it strongly depends on the
Wolf, exponential error bounds are derived which are attainable by unlver- joint statistics of the sources, which are assumed to be
sal codes. These huunds are tight lu a neighborhood of the boundary of the known.
achievable rate region. The first m a jor contribution to m u ltiterminal source
coding was the paper of Slepian and W o lf [l]. They
I. INTRODUCTION
considered the case when each source is connected to a
separate encoder, and these encoders are connected to a
ANY COMMUNICATION situations can be single destination which must be able to reproduce all the
M m o d e led by a graph (or network) having informa-
tion sources, encoders, and destinations located at its
sources with a small probability of error. They derived the
surprising result that the sum of the rates of the individual
vertices. Each encoder operates on messages from the encoders can be m a d e as small as the rate of a common
sources connected to it, and the resulting codeword is encoder connected with all sources, i.e., as small as the
m a d e available (without noise) to all destinations to which joint entropy of these sources. M u ltiterminal source
the encoder is connected. Each destination must be able coding theory has been further developed by Wyner and
to reproduce accurately the messagesfrom certain speci- Z iv [2], Ahlswede and Kiirner [3], Wyner [4], Cover [5],.
and quite a few others, cf. the surveys by Korner [6],
Manuscript received January 19, 1979. This paper was presented at the Berger [7], and G e lfand and Prelov 181.
Czech-Hungarian-Soviet Seminar on Information Theory, Tsahkadsor, Most of the presently known m u ltiterminal source
Armenian SSR, USSR, September 1978, and at the IEEE International coding theorems have been obtained by rather similar
Symposium on Information Theory, Grignano, Italy, June 25-29, 1979.
The authors are with the Mathematical Institute of the Hungarian methods. This suggeststhat a general theory should exist
Academy of Sciences, H-1053 Budapest, Realtanoda u. 13-15, Hungary. containing them as special cases.One aim of this paper is

0018-9448/80/0300-0155$00.75 01980 IEEE


156 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. IT-26, NO. 2, MARCH 1980

to attempt such a theory, i.e., a unified approach to a II. DEFINITIONS


large class of multiterminal source coding problems. This
class will be specified by the definition of a source network We shall consider random variables (RV’s) X, taking
given in Section II. Informally we restrict ourselves to 1) values in finite sets %a where a ranges over some finite
discrete memoryless sources, 2) graphs in which no edge index set &?.For any subset C of @ we write
joins two sources or two encoders or two destinations, and f& A & %, xe k vc7Le*
3) graphs in which the destinations are required to repro-
duce the messages of the specified sources with small X,” resp. Xi will be a shorthand notation for a k-length
probability of error. sequence of RV’s {Xa,i}$C1 resp. {XL,i}$=i.
It should be noted that a few interesting results (cf. Definition 1: A discrete memoryless multiple source
Wyner-Ziv [9], Berger [7]) do not fit into this restricted (DMMS) with generic variables X,, a E &?, is a sequence
framework. While our definitions could be easily extended lx@,i>E 1
of independent replicas of X, The sequence
to cover multiterminal rate-distortion problems, no pre- ixa,i>E-l of independent replicas of X, is called the
sently known general methods yield single-letter char- component source of index a of this DMMS.
acterizations of achievable rate regions for the type of Let disjoint finite sets Ci?,% , t?, and a directed graph
problems we have excluded. with vertex set @ u ‘$J u C?be given such that no edge ends
It will be shown in Section IV that from the point of in @, no edge starts in 6?, and there are no edges between
view of determining the achievable rate region attention vertices in 3. The elements of Ci?,Ci3,and 6? will be called
may be restricted to source networks of a relatively simple input, intermediate, and output vertices, respectively. For
structure called normal source networks. A subclass of any vertex o we denote by S, the set of starting points of
normal source networks will be studied in more detail in edges leading to v. Further, for each output vertex c, let
Section III. For this subclass, exponential error bounds 5 c 6? denote the set of all input vertices from which
will be obtained which are tight in a neighborhood of the thcereis a path to c.
boundary of the achievable rate region. Further, these We shall regard the vertices in &?, %I, and 6? as loca-
bounds are universally attainable, i.e., they are attainable tions of sources, encoders, and destinations, respectively,
by encoders and decoders not depending on the source the edges of the graph describing their connections. Then
statistics. For the particular case of the network of Slepian YC gives the locations of those sources from which infor-
and Wolf [l], the results of Section III were obtained mation can be conveyed to the destination at c. We shall
jointly with K. Marton. A characterization of the achiev- assign to each output vertex c E C?a set qC c TC of input
able rate region of arbitrary normal source networks is vertices interpreted as the locations of the sources to be
given in Section V. As one would expect, this is a non- reproduced for the destination at c.
computable “product space characterization.” Its main Definition 2: A source network is defined by a graph as
relevance is to show which mathematical problem ought above, a DMMS with generic variables X,, a E 8, and sets
to be solved in order to get a single-letter characterization. ‘QCc TC,,c E C?.The generic variables of the component
Thus one is led to the entropy characterization problem of sources will also be called the generic variables of the
Section VI. The approach of Sections V and VI gener- source network.
alizes the one introduced by Ahlswede and Korner [3] for Source networks will often be given graphically. The
solving the “side information” problem. Of course, a vertices of the graph will be represented by boxes. We
complete solution of the entropy characterization problem write the generic variables of the component sources into
is a long way off. Still the special case already settled the boxes representing the input vertices and put the
yields the single-letter characterization of the achievable symbol X0 into the box of an output c whenever a E ‘%I=;
rate region of quite a few source networks, including all e.g., the source network of Slepian and Wolf [l] is given
those for which such a characterization has been pub- by Fig. 1, and the zigzag source network of Kiirner and
lished previously. This will be illustrated by the examples Marton [lo] is given by Fig. 2.
of Section VII. Definition 3: A k-length block code for a source
Much of the ideas of this paper were influenced by our network is a family of mappings
previous joint research with R. Ahlswede, P. Gacs, and K.
Marton. Also, a referee called our attention to a paper of fb: %&+Gs(tb, bEa; vcpe:&, ‘&+J%i,, cEC
Han and Kobayashi [18] submitted to this TRANSACTIONS
e
about the same time as our paper. In [ 181source networks where “srt,, b E C?Iare arbitrary finite sets and %b k 5%:
similar to our normal source networks are considered and if b E 6?. The mappings fb are called encoders and the ‘p,
an inner bound to the achievable rate region is derived. are called decoders.
The results of the two papers do not overlap in any Performing the encoding and decoding operations pre-
significant manner. In particular, while [18] also contains scribed by a k-length block code on the k-length random
a product space characterization of the achievable rate messageXi of the DMMS, we obtain at each,output c E C?
region, our characterization in Section V is more explicit an RV
and, unlike the one in [18], yields single-letter characteri-
zations in quite a few cases as shown in Section VII.
CSIS& AND KhUiER: GENERAL THEORY OF SOURCE NETWORKS 157

iv) for each pair of output vertices c’ and c”, the


inclusion $5, c SC,,implies OC,c gi),,,.
Intuitively SC,c SC,,means that all information availa-
ble at the output c’ is also available at c”. Thus the
m e a n ing of condition iv) is that a more informed destina-
tion has more demanding reproduction requirements.
Fig. 1. SW sourcenetwork. Because of the one-to-one correspondence of ii) be-
tween &? and 9, for NSN’s we shall assume that the
component sources are also assigned to the intermediate
vertices b E ‘33, and the generic variables X, resp. alpha-
bets x0, a E @ , will be denoted also by X, resp. !?&,
bE93.
For NSN’s it will be significant whether or not there
exist vertices b E ?i3 such that some destination having
Fig. 2. Zigzag sourcenetwork. available information from the component source
assigned to b does not require the reproduction of this
source. For such a destination, the only role of informa-
For the destination c, an error occurs if and only if tion stemming from the encoder at b is to help in repro-
J?kC#XGC. Thus the probability of error of this code at c ducing the messagesof some other component sources.
is Definition 6: A vertex b E 93 of an arbitrary source
network is a helper if there exists a c E C?with b ES, such
that 5,~ qC.
In addition to the error probabilities, the other m a in For example the zigzag source network of F ig. 2 has
indices of the performance of a code are the rates two helpers. However, by replacing ? in the second out-
put box by z?, the resulting NSN (which has the same
achievable rate region as the original source network) has
but one helper.
Here, as usual, l]&l] stands for the cardinality of the range
off,, i.e., II&II p P&I. III. NORMAL SOURCENETWORKS W ITH No HELPER
4: A vector R= { Rb}bE3 with nonnegative
Definition
components is an c-achievable rate vector for a source NSN’s without helpers are characterized by the prop-
network if for every 6 > 0 and all sufficiently large k there erty that at each output c E E! all component sources
exist k-length block codes for this source network with assigned to the vertices b E S, have to be reproduced.
error probabilities These source networks are uniquely determined by the
e, <E for every cEC?, DMMS {X~,i}~~ and the sets S,c?2i3, CE&?. They are
the most direct generalizations of the network of F ig. 1
and rates considered by Slepian and W o lf [l].
~wll <%+s foreverybE%. Theorem 1: The achievable rate region of an NSN
without helpers equals the set of those vectors R=
Further, g is called an achievable rate vector if it is which satisfy the inequalities
{Rb)bEB
e-achievable for every E> 0. The set of all these vectors is
the achievable rate region of the source network. 2
bEf2
Rb > H(XdXS,,!2) (1)
Notice that the achievable rate region of every source
network is a closed set by definition. Further, by a stan- for every output c E L? and set C c S, which is disjoint
dard time-sharing argument, it is also a convex set. A from every S, c S, (c’E (3).
primary goal of m u ltiterminal source coding theory is to Remark: This theorem contains the statement that the
determine (in a computable manner) the achievable rate above inequalities imply the validity of (1) for every
region of source networks. W e shall show that it is CCS,.
sufficient to solve this problem for the following simpler
Proof: The necessity of the inequalities (1) for every
class of source networks.
c E C?and C c S, is obvious. If these inequalities hold for
Definition 5: A source network is called a normal
every c and C as in the Theorem then they also hold for
source network (NSN) if
every c E &? and C c $5,. This can be shown recursively,
9 there are no edges from inputs to outputs, using the fact that for every e = C, u f$ with c, c SC,c s,,
ii) )& ( = 1 %] and the edges from &? to 3 define a k c 5, \S, we have
one-to-one correspondencebetween input and inter-
m e d iate vertices, wwkc\~) ~~(X,*IXScYG)
+ wf~lX*c%)~
iii) all the sets S,, c E c? are different, and Hence it remains to show that the vectors R satisfying the
158 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. IT-26, NO. 2, MARCH 1980

inequalities (1) for every c E 6?, lZ c 5, are achievable rate where zs={??b}bES runs over all IS I-tuples of RV’s with
vectors. This can be done easily by the random partition values in xb, b E 5, and Itl+ e max(t,O).
method of Cover [5]. We omit the details as we shall prove
a stronger result in Theorem 2 (cf. the remark following Remark: One immediately sees that the exponents
that theorem). E,(R,XS ) are all positive whenever X, satisfies the condi-
tions (1)’with strict inequalities for every c E e, l? c 5,.
We now derive universally attainable exponential error Thus the codes of Theorem 2, which are independent of
bounds for NSN’s without helpers. These results will not the source statistics, yield exponentially small error proba-
be needed in further sections. For the case I@I = 2 the bilities whenever R is an inner point of the achievable rate
results below were obtained jointly with K. Marton. Par- region of the source network with the particular X,.
tial results for that case (with nonuniversal decoding) had
been derived earlier by Gallager [ 1l] and KoSelev [12]. Consider the family ‘% of of all 1%I-tuples of
Proof
Universally attainable error exponents in channel coding mappingsf=ifb)bEB where fb: %~+-Gx, for some fixed
have been given by Csiszar, Korner, and Marton [13], cf. sets 9Rb satisfying
also Csiszar and Korner [14].
We need some notation. For any pair of Rys ? and Y
with values in the same finite set 9, let D( Y II Y) be the
informational divergence of their distributions, i.e., ForeveryfE%andsets0#CcSc% let G(S,C,f)be
Pr{ ?=r} the set of those 1%(-tuples U= { ?i?b} bEJ of k-length
D( YII P) A x Pr{ f=y}log sequences XbE 3Eg for which the ME decoder correspond-
Pr{ Y=y} ’
YEY
ing to the Cartesian product of the mappings f,, bES,
For any k-length sequence YE qk let q,-(y) denote the yields a correct reproduction of the sequences Xb if b E
relative frequency of the symbol y in this sequence. The S\C but not if bEC.
distribution P7 A { PY_(y)}YEq is the type of J. The entropy Let $? be the set of those distributions on 96%which are
of the sequencejj is defined as the entropy of its type, i.e., possible types of sequences U E %$. For every P E 9, let
Tp denote the set of U of type P. Since clearly
H(j) A H(Pj).
A minimum entropy (ME) decoder corresponding to a
19 I< (k+ l)‘%‘, (5)
mapping f: qk --+a is a mapping ‘p: %-+qk such that it follows from Lemma A of the Appendix that for
sufficiently large k
=yiff(y)=mandH(y)<H(.?‘)
o(m) whenever f( j7) = m for f #j,
1 is arbitrary if there is no such YE ‘?Jk.
For NSN’s without helpers, we shall consider codes with
.exp( k( b~~R,-H(P,)+H(P,“))) <l,
ME decoders. These codes are determined by the en-
coders fb: %~--+9?& b E %I, letting ‘p, be an ME decoder
corresponding to the Cartesian product fsc: g, + where P, is the marginal distribution of P on !?&
XbE5, ?&, of the mappings fb, b E S,, for every output = x xb. This implies the existence of an f E 9 such
CEt?. thabt?& every P E 9 and every 0# C c 5 c 91 we have
Theorem 2: Given the set xb, b E ‘% , and any S > 0,
for sufficiently large k (depending only on the sizes I’%,] IqwLf)r--lET,I
Rb-H(-?,.#s,I?) ,
and S) to every vector R = { Rb} bEb with nonnegative ITPI

components there exist mappings f, with domain %t and


where x,={X,,}b~% denotes a dummy RV with distribu-
$logllfbil <Rb+ 6, bE%, (2) tion P. We claim that this f = { fb} bEa satisfies the asser-
tion of the theorem. In fact, for a DMMS with generic
such that for every NSN without helpers with generic variables X,, b E %!I,all sequences in ?Jp are equiprobable.
variables X, taking values in the sets xb, b E ‘81, the code Since
, with ME decoders determined by the encoders fb has
probability of error Pr{ X&E Tp} < exp{ - kD(j?,@&}
e,<exp( -kE,(E,XsC)+k8), c~C?. (3) (cf., e.g., Dueck and Korner [15, Lemma 11)it follows that
Here for every 5 c $8
E,(E,X,) k mm Pr{X$E&(S,C,f)n9p} <exp -k D(g:ellX,)
i (
xs
+ min
e:Ozecs b~cRb-H(‘&%,,)
I1
+ ) + b;e % - H(XddS\C>
CSIS& AND KhiF3R: GENERAL THEORY OF SOURCE NETWORKS 159

whence by (5) { R;}G~G with


4, if be%,
R$ K, if bECB3,,
1 0, if b E Q2,
is achievable for $; and
iii) 5 has no more helpers than S .
As the ME decoder at c commits an error if and only if
GE u0#lszsc & (S,, C,f), the bound (3) for large k Lemma I: To-every source network S there exists a
follows from the last inequality and the well-known fact source network S without edges from inputs to outputs
that O(xsllX,) <d(x,llX,). and having the same set of helpers such that the achiev-
able rate region of S is a projection of that of $.
Rather surprisingly, the universally attainable error
bounds of Theorem 2 are asymptotically optimal in a Proof 5 is obtained from S as follows. Introduce a
neighborhood of the boundary of the achievable rate new intermediate vertex 6 and a new output vertex E for
region of any NSN without helpers, i.e., no essentially every input a E @ connected with some output vertex. Set
smaller error probabilities are attainable even if codes D,- k {a}, introduce a path a&?, and connect 6 with all
depending on the source statistics are allowed. This is seen those output vertices CEC? which are connected by an
from the following result. edge with a. F inally, delete all edges from a to output
vertices.
Theorem 3: For every 6 > 0 and sufficiently large k,
every k-length block code for an NSN with no helper Lemma 2: To every source network 5 without edges
satisfying from inputs to outputs there exists a source network S of
the same kind such that
i) the edges between the input and intermediate vert-
has error probabilities ices of 5 establish a one-to-one correspondence;
ii) the set 6 % of intermediate vertic_esof S may be
e,>exp( -kE,(E,X,C)-kS), cEC?. identified with a subset of the set ‘$Bof intermediate
Here for every s c ‘??I vertices of $ so that a vector R= { R,,}bE4 belongs
to the achievable rate region of S if and only if
E,(R,XS) A m-in~(~s/lX,) setting
*.s
where X,= {X,},,, runs over those IS I-tuples of RV’s for
which
+
m in x R, - H(zf&,) =O. the vector $= {&,},,G is an element of the achiev-
e: 0zecs bEe able rate region of s; and
This theorem can be proved by routine arguments. W e iii) with the identification of ii), S and $I have the same
o m it the proof. set of helpers.
Proof: From the point of view of an encoder resp.
IV. REDUCTIONOF SOURCENETWORKSTO encoder-decoder pair of a given source network, it is
NORMALONES natural to consider as a single source the collection of
those component sources which are encoded resp. both
In this section we reduce the problem of determining encoded and reproduced together. The idea of the proof is
the achievable rate region of an arbitrary source network to introduce a new DMMS with these sources as compo-
to the corresponding problem for an NSN. More exactly, nents.
we establish through a series of three lemmas the follow- Formally, for every fixed b E ‘% consider the nonvoid
ing reduction theorem. subsets of @ which are of the form S, n 9, for some edge
bc of 5. The collection consisting of these sets together
Theyem 4: To every source network S there exists an with S, will be denoted by A(b). Let the intermediate
NSN S with the following properties: vertices of 5 be the pairs (a’, b) fof every b E ??I and
i) for the set ‘% I and ‘% Iof intermediate vertices of S @ ’E A(b). Let the input vertices of S be in one-to-one
and S one has correspondence with its intermediate vertices; connect
each of them with the corresponding intermediate one.
Further, identifying the output vertices of 5 with those of
S , let 6= (a’, b) E ‘% be connected with c E C?if and only
ii) R= { Rb}bEs is an achievable rate vector for Ss if if b is connected with c in S and either 8’ = S, or
and only if for some constant K the vector R= @ ‘=S,n9c.
160 IEEE TRANSACTIONS ON INFORhfATION THEORY, VOL. IT-26, NO. 2, MARCH 1980

Finally, consider a DMMS with generic variables Y,-,


~E%I, where
:ry
Y,-= %,’ if b”=(@‘,b).
The definition of $ is completed by assigning the compo-
nent source with generic variable Y,- to the input vertex + L-l
connected with 6 and requiring at each output c the Fig. 3. “Two-help-one”sourcenetwork.
reproduction of this component source if and only if the
vertex b = (@‘,b) is connected with c and @’c 9=.
The source network S obtained in this way clearly with an a E qc. Then G, c S, and Definition 6 can be
satisfies the assertions of the lemma if we identify the rewritten as follows.
intermediate vertices (S,, b), b E ‘Zi3,with the original b. Definition 6* : A vertex b E 9 of an NSN is a helper if
Lemma 3: Given any source network, one can modify bES,\&, for some cE(?.
the sets Qc, c E (2, without changing the achievable rate Theorem 5: The achievable rate region of an NSN with
region and without increasing the set of helpers so that for set of helpers X equals the closure of the set of those
every pair of output vertices c’ and C” the inclusion vectors R= {Rb}bES to which for some n > 1 there exist
S, c S,. implies 6iJc,c 9cPPfor the modified qc. After this functions fb, b E X with domain ?$ (and arbitrary range)
modification, if there are several outputs with the same such that
S,, one can delete all but one of them without changing
the achievable rate region. %a ~~(f,(x,")) (6)
Proof: For a fixed pair of outputs with S, c SC,,the
addition of the component sources to be reproduced at c’
to those to be reproduced at c” does not change the
achievable rate region. Then proceed iteratively. The last
assertion is obvious.
+ ,,Fn, fwxx,“>) (7)
Lemmas l-3 prove that to every source network one
can easily construct an NSN (with, at most, the same for every c E e and C c G, disjoint from every &,, such
number of helpers) such that the achievable rate region of that S, c S, with c’#c.
the former can be immediately obtained from that of the Proof: To prove the converse part, let R’ = { RL}bE3
latter. In particular, in order to get a computable char- be an achievable rate vector. Then, by Definition 4, to
acterization of the achievable rate region of an arbitrary arbitrary small 6 > 0 and E> 0 there exists an n-length
source network it suffices to derive a computable char- block code (for sufficiently large n) such that the encoders
acterization of that of the corresponding NSN. It should fb, b E 9~ satisfy
be noted, however, that for determining the achievable
rate region of a particular source network, a complete +ogI~f,ll <R;+ 6, bE% (8)
solution of the corresponding NSN problem may not be
needed, as is obvious from the reduction steps above. and the error probabilities satisfy
ec <c, CEC?. (9
V. NORMALSOURCENETWORKS WITH ANY It suffices to show that setting Rb p R,‘+26 and properly
NIJMBER OF HELPERS choosing E, the encoders fb for the helpers b E X satisfy all
While for NSN’s without helpers the achievable rate the stated inequalities.
region is easy to characterize, for arbitrary NSN’s no Clearly (8) implies the inequality (6) for every b E x ;
single-letter characterization is known. Already the simple thus it is enough to show that
NSN of Fig. 3 represents a very hard problem, cf. Kiirner
and Marton [16]. We give next a (noncomputable) prod- x & 2 ~H(X~Ix~c,, {fb(Xbn))bE&“%) (lo)
bet?\%
uct space characterization of the achievable rate region of for every c E (? and l? c &, . On account of (8) and the fact
arbitrary NSN’s. The main significance of this result con- that $5,= G, u (S, n X) (cf. Definition 6*), we have
sists in reducing the problem of single-letter characteriza-
tions to a common mathematical problem of a rather
simple structure. As another application of independent
interest we shall show at the end of this section that the
so-called “optimistic” and “pessimistic” viewpoints in def-
ining the achievable rate region of a source network lead
to the same result.
For an arbitrary NSN, let us denote by G, the set of
those intermediate vertices b E %I which are connected
CSISZh AND K&NER: GENERAL THEORY OF SOURCE NETWORKS 161

Using (9) and Fano’s inequality the last term is lower i-length block code for S with rates
bounded by

and error probabilities satisfying (9).


(where h(e) stands for the binary entropy function), so Now we show that tl+ implies the same result for every
that (10) follows. large k, and therefore R is an achievable rate vector for S
Turning to the direct part, it is enough to show that if R as we have claimed. To this end, consider any k with
satisfies the inequalities (6) and (7) of the theorem for (I - 1) n <k <In and consider a code of block length i= In
some n and functions fb, b E X, then E is an achievable for S with encoders fb, b E%, and decoders $$,, CE e,
rate vector. To this end we construct, for the given NSF satisfying (12) and (9). One can apply this code to k-
S and the functions fb, bE X, an auxiliary NSN S length sequencesby adding to the latter arbitrary suffixes
without helpers. of-length i-k. Formally, for every L={Zb}bE9, ?$,E
Write & P u .Ee&,. To every b E X n & introduce a %z- k, define k-length block encoders fb,i by
new intermediate vertex b+ and a new input vertex con-
nected with b +. Connect b+ with those c for which
b E G ,. Further, to every b E X introduce a new output where Z,,Zbstands for the jUXtapO$tiOnO f .?band ?& Let 9$
vertex and connect it with b. Define a new DMMS with be defined by deleting the last k- k symbols from every
generic variables assigned to the intermediate vertices of component sequenceof the values of $. Denoting by e,,,
the m o d ified graph, setting the error probabilities of the code with encoders fb,L,
b E $i3 and decoders ‘pc,c E e, (9) implies

Y, p X,“, if bBX,
if bEX,
fb(Xt37
Y,+ 42 x;, if bEXn&. It follows that for some ZE!?&” we have e.., ;<elel08 for
Let $ be the NSN without helpers uniquely defined by
lx
every c E e. This completes the proof of our ilaim con-
thi assignment.
this cerning (11).
W e claim that if a vector 2,
R, with components ib,
Now the direct part of the theorem will follow if we
show that whenever E satisfies the inequality of the state-
b E 3, and %+, b E X n & isanachieva
bE%,andi?b+,bEXn& is an achievable rate vector for ment of the theorem then
$ then ,

R, g
$%, if b@Xn&,
(11)
kb+ = nRb - H( fb(X:)),
if bEYC,
if b@X,
_ if b E X n & ,
(13)

;( ib + ib+), if bEXn&,
defines an achievable rate vector d for 5. Notice that $ is
defines an achievable rate vector i k { Rb} bEb for S . an NSN without helpers for which the set of intermediate
TO, see this, observe first that every I-length block code vertices connected with an output vertex c E e is S, u &:,
for S gives rise to an In-length block code with at most where
the same error probabilities for S ; further, the rates of
this code for S are related to the rates of the code for S G: b {b+; bE&,nX},
asin(ll).Infact,theencodersf?,,b_E%,and~~+,bEXf Further, the outputs of 5 not belonging to e are in
G , of any I-length block code for S define m a p p ings Fb one-to-one correspondence with the elements of X, each
resp. Fb+ with d o m a in Xk such that being connecte$only with the corresponding b E x. Thus
I;,+(X’“) =fb+( Y;+). by Theorem 1 R is an achievable rate vector for $, if
&ix”) =fb( ybl)p
Consider the m a p p ings Fb, b 6 % n G , resp. the pairs &, 2 H( y,,), bEX, (14)
(Fb,Fb+), b EX n G , as encoders of an In-length block
and
code for S. Every decoder I$~ of the code for S which
corresponds to an output vertex c E e has a range defined x ‘b+ x kb+ >H(Y,Y,+IYS,\e,Y,:,,+) (15)
as a Cartesian product containing all the sets % F , b E F,, bEf? b+EfZ.+
as components. Thus one can define the corresponding for every C c S, \X, C+ c G : such that C u C+ is disjoint
decoders of the code for S by projecting the value of @c from every S, u G : properly contained in S, u &: . Since
on % tc. Clearly the code obtained in this manner for the S,\X=&,\X and S,u&~cS,u&,+ if and only if S,
source network S d_oeshave the announced properties. c S,, the inequalities (14) and (15) follow from (6), (7),
This proves that if R” is an achievable rate vector for 5 and (13) by the definition of the RV’s Y, and Y,+.
and E is defined by (1 l), then to every e > 0, 6 > 0, and Notice that Definition 4 of the achievable rate region
sufficiently large integral m u ltiple k^= In of n there exists a represents a “pessimistic” viewpoint. Instead of postulat-
162 IEEE TRANSACTIONS
ON INFORMATIONTHEORY,VOL. 1~26, NO. 2, MARCH 1980

ing the existence of good codes for every sufficiently large Clearly, if this problem could be solved for every
k, an optimist would already call an F an achievable rate DMMS, then Theorem 5 would yield a single-letter char-
vector if to every E> 0 and 6 > 0 there exists a k-length acterization of the achievable rate region of every NSN
block code (for just one k) satisfying the inequalities of with one helper. More generally, the same is true for
Definition 4. This alternative definition is quite common NSN’s with any number of helpers provided that no
in the literature. output is connected with two helpers and no helper is
An equivalent way of formulating the “optimistic” defi- connected with two outputs.
nition is to say that E is an achievable rate vector if for Similarly, one could formulate an r-parameter entropy
some sequence of positive integers k, there exist k,-length characterization problem, the solution of which would yield
block codes of rates converging to the components of R a single-letter characterization of the achievable rate re-
and error probabilities converging to zero. The point is gion of every NSN with r helpers. Unfortunately, even the
that here k,, may be any sequence, for example, even two-parameter entropy characterization problem is be-
k,+,/k,+w is possible. yond the reach of our present methods. In particular, it
Corollary: For every source network the optimistic and seems already very hard to give a computable characteri-
pessimistic definitions lead to the same achievable rate zation of the set of all three-dimensional vectors
region.
f ~(X”lf(X”)), ; NY7 g( Y”)),
Proof: As in the reduction of the problem of de- (
termining achievable rate regions to the special case of
NSN’s, one sees that it is enough to prove the theorem for ; H(Z”lf(X”M Y ”)))
normal source networks. Thus it suffices to prove that
Theorem 5 holds also with the optimistic definition of the for a D M M S with generic variables X, Y, Z. Such a
achievable rate region. As the optimist’s rate region could characterization would lead to that of the achievable rate
only be larger, the statement follows since the proof of the region of the “two-help-one” source network of Fig. 3; the
converse part of Theorem 5 applies literally to the “opti- difficulties of the latter problem are discussed in Korner
m istic” case. and Marton [16], cf. also Csiszar and KSrner [14].
While we believe that entropy characterization prob-
VI. THE ENTROPY CHARACTERIZATION PROBLEM lems are the mathematical core of source network theory,
it is not the aim of this paper to enter into them in detail.
Because of Theorem 5, to give a single-letter characteri- We lim it ourselves to illustrating the power of our ap-
zation of the achievable rate region of an NSN it is proach by showing (in the next section) that even the
sufficient to characterize the range of certain vectors of presently available partial results on the one-parameter
information quantities involving as many arbitrary func- entropy characterization problem yield the achievable rate
tions as there are helpers in this NSN. region of quite a few source networks. For this purpose we
In particular, in the case of one helper, it suffices to now state without proof two results from Csiszar and
determine the set of all possible vectors of form Kiirner [14] which were obtained jointly with K. Marton.
The first one is the solution of the one-parameter entropy
characterization problem for D M M S ’s with three compo-
nents. The second result, a slight generalization of the
first, is a partial result on the same problem for D M M S ’s
Here n ranges over the positive integers and f ranges over with four components.
functions with domain Fb,, where b, is the helper. In fact
the information quantities figuring in (6) and (7) for Theorem 6: Given a D M M S with generic variables X,
X = {b,} can be immediately expressed in terms of the Y, 2, the closure of the set of all vectors of the form
components of such a vector and the known joint entro-
pies of the generic variables X,, b E ‘33. Since the RV’s X,, (+wlf(X”))>; w~nlf(x"))?~ H(z"lfW)),
J c $i3 can be considered as the generic variables of a new
DMMS, it will be convenient to formulate the following. (with arbitrary positive integer n and function f) consists
One-Parameter Entropy Characterization Problem: For of the vectors (a, b, c) satisfying
an arbitrary D M M S with generic variables Y, taking
values in finite sets % ,, a E @ , give a single-letter char- max[I(X/\YlU),I(X/\ZIU)]+t<a<H(XIU)+t
acterization of the set of all J&]-dimensional vectors of the b=H(Y(U)+t, c=H(Z(U)+t
form
for some RV U and nonnegative number t where
i) the range % of U satisfies 1% I < 1% I + 2,
ii) the RV’s U, X, YZ form a Markov chain in this
where n=1,2;-. and f ranges over functions with order, and
domain 9:,, a, being some fixed element of 8. iii) 0 < t < m in[1( U/\ Y), I( U/\ Z)].
CSISZuiR AND KiiRNER: GENERAL THE?ORY OF SOURCE NETWORKS 163

Theorem 7: G iven a DMMS with generic variables X,


S, Y, Z such that X, S, YZ form a Markov chain in this
order, the closure of the set of all vectors of the form

(with arbitrary positive integer n and function f) consists


Fig. 4. Source network with two helpers.
of the vectors (a, b, c) satisfying
max[I(X/\YlU),I(Xr\ZIU)]+H(SIX)+t
<a<H(SIU)+t,
b=H(YlU)+t, c=H(ZIU)+t,
for some RV U and nonnegative number t where
i) the range % of U satisfies IG21 I < I% I + 3,
ii) the RV’s U, X, S, YZ form a Markov chain in this Fig. 5. NSN corresponding to zigzag source network.
order, and
iii) 0 < t < min[ I( U/\ Y), I( Ur\Z)].
Remark: Setting X= Z in Theorem 6 we see that the
closure of the set of all vectors of the form

consists of the vectors (a, b) satisfying


Fig. 6. Source network of Example 3.
b-H(YIX)<a<H(XjU)+t, b=H(Y(U)+t,
for some U and t as in Theorem 6. It is easy to show that and ii) of Theorem 6. The result of this example is due to
the set of these (a, b) is the same as the set of (a, b) with Sgarro [ 171.
b-H(YIX)<a<H(XJU), b=H(YlU). Example 2: Consider the source network of Fig. 2,
This special case of Theorem 6 has been obtained by interchanging for notational convenience the RV’s X and
Ahlswede and Korner [3]. Y. By Lemma 3, this source network has the same achiev-
able rate region as the NSN with one helper shown in Fig.
5. By Theorem 5, the achievable rate region is the closure
VII. EXAMPLES of the set of those triples (Rx, R,, R,) which for some n
and mappings f with domain % ” satisfy
Example 1: The NSN of Fig. 4 has two helpers. Still,
as no two helpers are connected with the same output and R, 2 fJ( Y),
each helper is connected but with one output, the achiev- Rx > ; H(X”I Y”,f(X”)) + + H(f(X”)),
able rate region of this source network can be determined
via a one-parameter entropy characterization problem. By
Theorem 5, the achievable rate region is the closure of the R, > $ H(Z”I f(X”)).
set of those triples (R,, R,, R,) which for some positive Noticing that
integer n and mappings f and h with domain %?’resp. %*
satisfy Hf(X”))+ NX”I Y”,f(X”))
= nH(X, Y) - H( Y”( f(X”)),
Rx> +(f(X")), R, 2 ; ff( Y”lf(x”)), it follows from Theorem 6 that (Rx, R,, R,) is achievable
if and only if
R, > +f( Y”lh(Z”)), R, > ; H(h(Z”)). R,>H(X,Y)-H(YIU)-t,
R,=(Y),
Writing H(f(X”)) = nH(X) - H(X”] f(X”)), by the remark
at the end of Section VI the first two conditions have the R,>H(ZIU)+t,
single letter form where U and t are the same as in Theorem 6. The
single-letter characterization of the achievable rate region
Rx > I(XA U), R,>H(YIU); of this source network has been given by Korner and
similarly, the last two conditions become Marton [lo].
Example 3: By Theorem 5, the achievable rate region
R, 2 ff( YI v>, R, >I(ZA V), of the NSN with one helper of Fig. 6 is the closure of the
where U and V are arbitrary RV’s satisfying conditions i) set of those triples (R,,R,,R,) which for some n and
164 IEEE TRANSACTIONS ON INFORMATIONTHEORY,VOL. IT-26,N0. 2, MARCH 1980

mapping f satisfy
R,=(YIX),

Rx 2 ; H(X”( Yn,f(X”)) + ; ff(f(X”)),

Rx+ R,> ;H(X’Y’lf(X”))+ $H(f(X”)),

R, > ; H(Z”l f(X”)).

These conditions can be equivalently written as


R, > H( YIX),
R, >H(XY) - $ H( Y”J f(T)),

Rx + R, > H(XY),
Fig. 7. Source network and corresponding NSN.

R, 2 $ H(Z”(f(X”)).

By Theorem 6, we get the following single-letter char-


acterization of the achievable rate region:
R,)H(YlX),
R,>H(X,Y)-H(YJU)-t,
Rx + R, > H(XY),
Fig. 8. Source network of Example 5.
R,>H(Z)U)+t,
where U and t are as in Theorem 6. Notice that Example see that (R,, R,, R,) is an achievable triple for the NSN if
3 contains Example 2 as a special case. and only if for some positive integer n and mapping f with
Example 4: Consider the source network of Fig. 7(a), domain %? it satisfies the inequalities
denoting the three intermediate vertices by 1, 2, and 3. By
the reduction steps of Lemmas 1-3, (R,,R,,R,) is an R, ) ; H( Y”lf(X”)),
achievable rate vector for this source network if and only
if (H( Y),R,,R,, R3) is an achievable rate vector for the
R, ) ; H(Z”lf(X”)),
NSN of Fig. 7(b). This NSN has one helper, viz., the
vertex 1. By Theorem 5 the achievable rate region is the
closure of the set of those vectors (R,, R,, R,, R3) which for Rx > ; H(X” I Y”Xn,f(X’)) + ; H( f(X”)).
some n and f satisfy Noticing that
R, ) ff( Y), fqX”I YnZn,f(X9) + Wf(X”))
R, ) ; H(X”I Y”,f(X”)) + i H(f(X”)), = nH(X, Y, Z) - H( YT” If(X”)),
we see that Theorem 7 applies. In fact, considering a
R,) $ W(X”If(X”)), DMMS with generic variables X, S, Y, Z such that S
equals the pair YZ, the Markov chain condition of that
R, > ; H(Z”(f(X”)).
theorem is fulfilled. Thus our previous inequalities yield
that (R,, R,, R,) is an achievable rate triple for our NSN
By the same observation as in the previous example, it if and only if
follows from Theorem 6 that (R,, R,, R,) is an achievable
rate vector of the source network of Fig. 7(a) if and only if R,>H(Y(U)+t,
R, >H(ZIU)+ t,
R,>H(X,Y)-H(YIU)-t,
R,>H(X,Y,Z)--H(YZIU)-t
R, > max[ 1(X/\ YI U),I(XAZI U)] + t,
for some U and t as in Theorem 7.
R,>H(ZIU)+t,

for some U and t as in Theorem 6. VIII. CONCLUSION


Example 5: Consider the source network of Fig. 8. By
Lemma 3, this source network has the same achieyable We have attempted to develop a general theory of
rate
-IA region as the NSN obtained upon replacing X by source networks. We have shown that in order to
XYZ at the middle output vertex. Applying Theorem 5 we determine the achievable rate region of an arbitrary
CSISzliR AND KihER: GENERAL THEORY OF SOURCE NETWORKS 165

source network it is sufficient to solve this problem for a Now


corresponding normal source network. A product space
characterization of the achievable rate region of NSN’s
led us to the entropy characterization problem as the
mathematical core of source network theory. The achiev- . some CE X(U) satisfies (i)} ]
y&Ji?T !{f.
able rate region of an NSN is determined by a boundary
(from some direction) of the set involved in the corre-
sponding entropy characterization problem. At present, < &J i~Fp I~~(Y)~{f:fb(~b)=fb(Xb)foTeve~b~e)l
P
only the one-parameter entropy characterization problem
can be tackled. Using recent results in this direction
jointly obtained with K. Marton, single-letter characteri-
zations of the achievable rate region of many source Substituting (Al) and (4), the assertion follows.
networks can be derived. This is illustrated by five exam-
ples, three of which are original. Our approach offers the
REFERENCES
possibility of a systematic exploration of the power of
presently known methods. It appears that the difficulty of
111 D. Slepian and J. K. Wolf, “Noiseless coding of correlated infor-
a source network problem depends on the number of mation sources,”IEEE Trans. Inform Theory,, vol. IT-19, pp.
helpers and their connections to the decoders in the 471-480. Reprinted in Key Papers in the Development of Informa-
tion Theory, D. Slepian, Ed. New York: IEEE Press, 1974.
corresponding NSN. In particular, for the case of more
PI A. D. Wyner and J. Ziv, “A theorem on the entropy of certain
than one helper, essentially new ideas will be needed, binary sequences and applications I-II.” IEEE Trans. Inform.
unless all the edges between helpers and outputs are Theory, vol. IT-19, pp. 769-778, Nov. 1973.
[31 R. Ahlswede and J. Khmer, “Source coding with side information
disjoint. at the decoder and a converse for degraded broadcast channels,”
One aim of this paper was to convince the reader that IEEE Trans. Inform. Theory, vol. IT-21, pp. 629-637, Nov. 1975.
future research in multiterminal source coding theory [41 A. D. Wyner, “On source coding with side information at the
decoder,” IEEE Trans. Inform. Theory, vol. IT-21, pp. 294-300,
should be concentrated on the entropy characterization May 1975.
problem, rather than on individual source networks. [51 T. M. Cover, “A proof of the data compression theorem of Slepian
and Wolf for ergodic sources,” IEEE Trans. Inform. Theory, vol.
IT-21, pp. 226-228, Mar. 1975.
APPENDIX 161 J. Khmer, “Some methods in multi-user communication-a tu-
torial survey,” Information Theory-New Trends and Open Prob-
Lemma A: With the notation of the proof of Theorem 2, we lems (CISM Courses and Lectures No. 219), G. Longo, Ed.
Wien: Springer-Verlag, 1975, pp. 173-224.
haveforeveryPE??,0#CcScQ source coding,” in The Information The-
171 T. Bergen “Multiterminal
ory Approach to Communications (CISM Courses and Lecture
Notes No. 229), G. Longo, Ed. Wien and New York: Springer-
Verlag, 1977, pp. 171-231.
PI S. I. Gelfand and V. V. Prelov, “Multi-user communication” (in
Russian), in Probability Theory, Mathematical Statistics and The-
. exp b$ % - H(P,) + ff(f’s,e) + ; . oretical Cybernetics, Vol. IS, P. V. Gamkrelidze, Ed., 1978, pp.
123-162.
191 A. D. Wyner and J. Ziv, “The rate-distortion function for source
Proof: For a fixed fE%, a 1%I-tuple tiC(%b of coding with side information at the decoder,” IEEE Trans. Inform.
sequences XbE %$ belongs to J(S) lZ,jJ if and only if there exists Theory, vol. IT-22, pp. l-l 1, Jan. 1976.
an 1s I-tuple i?= { yb}bES withjbE%t, bES such that WI J. Kiimer and K. Marton, “Images of a set via two channels and
their role in multi-user communication,” IEEE Trans. Inform
i) fb(jb)‘fb(fb) for every bE s , Theory, vol. IT-23, pp. 751-761, Nov. 1977.
ii) jb#Eb for bee andjb=Zb for bES\e, 1111R. G. Gallager, “Source coding with side information and univer-
sal coding,” (preprint), presented at the IEEE Int. Symp. on In-
iii) H(5) < H(i&) where iis p { -iFb}bo$,. form. Theoty, Ronneby, Sweden, June 21-24, 1976.
In the last condition, the types of 3 and ti, are defined interpret- WI V. N. Koielev, “On a problem of separate coding of two depen-
dent sources” (in Russian), Probl. Peredach. Inform., vol. 13, no. 1,
ing them as k-length sequences of elements of !?‘&. pp. 26-32, 1977.
For every UE~~ denote by X(B) the set of those 3 for which 1131 I. Csisz&, J. Khmer, and K. Marton, “A new look at the error
ii) and iii) hold. Then the type Q of a 6~3C(17) (which is a exponent of a discrete memoryless channel” (preprint), presented
at the IEEE Int. Symp. on Inform. Theory., Ithaca, NY, Oct. 10-14,
distribution on CX& must satisfy H(Q) < H(P,), and the margi- 1977.
nal of Q on &,e must equal PS,e For every such Q we have by 1141 I. Csistir and J. Khmer, Information Theoty: Coding Theoremsfor
symmetry Discrete Memovless Systems. New York: Academic, to appear.
[I51 G. Dueck and J. Khmer, “Reliability function of a discrete mem-
IX(u)nEibl=l~~lll~~p,,,l. oryless channel at rates above capacity,” IEEE Trans. Inform.
Theory, vol. IT-25, pp. 82-85, Jan. 1979.
Hence (cf., e.g., Dueck and Korner [15, Lemma 11)
WI J. Khmer and K. Marton, “How to encode the modulo 2 sum of
two binary sources,” IEEE Trans. Inform. Theory, vol. IT-25, pp.
jfJC(a)nTQ,I<(k+l) ‘s’exp{k(H(Q)- ff(J’~d))- 219-221, Mar. 1979.
v71 A. Sgarro, “Source coding with side information at several de-
Since X(Z) is the union of its intersections with less than coders,” IEEE Trans. Inform. Theory, vol. IT-23, pp. 179-182,
Mar. 1977.
(k + 1)1%1different TQ for each of which H(Q) < H(P,), we have
WI T. S. Han and K. Kobayashi, “A unified achievable rate region for
a general class of multiterminal source coding systems,” IEEE
lW4l G (k+ 1)2’3i”‘exp(k(H(P,)-H(P,,,))). (Al) Trans. Inform. Theory, vol. IT-26, no. 3, May 1980, to appear.

You might also like