scispace - formally typeset
Open AccessJournal ArticleDOI

Capacity of the Gaussian arbitrarily varying channel

TLDR
The Gaussian arbitrarily varying channel with input constraint Gamma and state constraint Lambda admits input sequences x=(x/sub 1/,---,X/sub n/) of real numbers with Sigma x/sub i//sup 2/ > as discussed by the authors.
Abstract
The Gaussian arbitrarily varying channel with input constraint Gamma and state constraint Lambda admits input sequences x=(x/sub 1/,---,X/sub n/) of real numbers with Sigma x/sub i//sup 2/ >

read more

Content maybe subject to copyright    Report

IELE
TKANSACTIONS
ON
INFORMATION THEORY,
VOL.
37.
NO.
I,
JANUARY
1991
Capacity
of
the Gaussian Arbitrarily
Varying Channel
Imre
Csiszhr and
Prakash
Narayan,
Member,
ZEEE
Abstract
-The Gaussian arbitrarily varying channel with input con-
straint
r
and state constraint
2
admits input sequences
x
=
(xl,
,
x,~)
of real numbers with
Cxf
5
nT
and state sequences
s
=
(5
,.
,
T,~)
of
real numbers with
LY~
5
HA;
the output sequence
x
+
s
+
V,
where
V
=
(
VI,
,
y,)
is a sequence of independent and identically distributed
Gaussian random variables with mean
0
and variance
c’.
It is proved
that the capacity of this arbitrarily varying channel for deterministic
codes and the average probability of error criterion equals
log(l+
r/(
2
t
c2))
if
Z
<
r
and is
0
otherwise.
Index
Terms
-Arbitrarily varying channel, Gaussian, capacity.
I.
INTRODUCTION
RBITRARILY varying channels (AVCs) were intro-
A
duced by Blackwell
et
al. [5]
to
model communication
channels with unknown parameters that may vary with time
in an arbitrary and unknown manner during the transmission
of a codeword. In this paper, attention is restricted to AVC’s
without memory; further, it is assumed that the sequence of
channel states is selected arbitrarily subject to a constraint
specified later, and possibly depending on the codebook
but
independently of the codeword actually sent.
AVC’s exhibit various mathematical complexities even in
the case of discrete alphabets (cf. Csiszir-Korner [6, Section
2.61). In particular, their capacity may depend on whether or
not random codes are permitted, and whether the average or
maximum probability of error criterion is used. The random
coding capacity admits a simple characterization as a min-max
of mutual information,
a result dating back to Blackwell
et
al.
[5].
In contrast, the problem of capacity
for
determinis-
tic codes is much harder. In particular, for the maximum
probability of error criterion, a single-letter capacity formula
is known only under certain conditions on the structure of
the AVC (cf. Ahlswede [2] and Csiszir-Korner
[7]).
Unless stated otherwise, the term capacity will hereafter
always refer to capacity for deterministic codes and the ui‘er-
age probability
of
error
criterion. In the absence of state
constraints, Ahlswede
[l]
proved that this capacity was either
equal to the random coding capacity or otherwise
to
zero.
The necessary and sufficient condition for positive capacity,
Manuscript received April
25.
1989;
revised February
13,
1990.
I.
Csiszir was supported by the Hungarian National Foundation Scientific
Research Grant
No.
1806.
P.
Narayan was supported by the Systems
Research Center at the University
of
Maryland under NSF Grant
OIR-85-00108. This work was presented at the
IEEE
International
Symposium on Information Theory, San Diego, CA, January
14-19,
1990.
I.
Csiszir
is
with the Mathematical Institute
of
the Hungarian Academy
of
Sciences. H-1364 Budapest.
POB
127,
Hungary.
P.
Narayan is with the Electrical Engineering Department and the
Systems Research Center, University
of
Maryland, College Park.
MD
20742.
IEEE
Log Number
9038858.
as well as capacity under a state constraint, have been
determined by Csiszhr-Narayan
[8];
it was further shown
that Ahlswede’s alternatives do not necessarily obtain under
a state constraint.
Less attention has been bestowed in the literature on the
capacity of AVC’s with continuous alphabets. Presumably
motivated by random coding capacity, there have been
game-theoretic considerations concentrating on the min-max
of mutual information (cf. McEliece
[l
11). Hughes-Narayan
[lo] have used a geometric approach to determine the ran-
dom coding capacity of the Gaussian AVC defined formally
in the following paragraph. Blachman [4] has provided lower
and upper bounds on capacity in a communication situation
differing from ours in that the interference (i.e., state se-
quence) could depend on the actual codeword transmitted.
Our incomplete understanding of his paper seems to indicate
that he,
too,
considered random coding capacity.
To
our
knowledge, Ahlswede’s [3] is the only paper treating the
capacity of a continuous alphabet AVC for deterministic
codes. His AVC (a Gaussian channel with the noise variance
arbitrarily varying but not exceeding a given bound) allowed
a very simple approach, which may not be extendable
to
other cases of interest.
In this paper, we determine the capacity of the Gaussian
AVC formally defined
as
follows. Let the input and output
alphabets, and the set of states, be the real line. For any
input sequence
x
=
(x,;
.
.,
x,i)
and state sequence
s
=
(si;..,
s,?),
let the
output
be
x
+
s
+
V,
where
V=
(V,;
.
.,
V,,)
is a sequence of independent and identically
distributed (i.i.d.1 Gaussian random variables with mean
0
and variance
(T’.
We adopt an input constraint
r
and state
constraint
A,
namely the permissible input sequences
of
length
n
are those satisfying
n
11x11~~
C~,21nr,
(no)
(1.1)
r=l
and the permissible state sequences are those satisfying
n
Ilsll’=
zs,?~nA,
(A>O).
(1.2)
i=l
A code of block-length
n
comprises a set of codewords
xi;.
’,x,,,,
each in
R“,
and a decoder
cp:
R”
+{O;..,
N).
The average probability of error of this code, used on the
Gaussian AVC as above when the state sequence is
s,
equals
1.
F(s)=-
Pr{cp(x,+s+V)#i}. (1.3)
N,=I
0018-9448/01/0100-0018$01.00
U
1991
IEEE

CSISZAR AND NARAYAN: CAPACITY
OF
THE GAUSSIAN ARBITRARILY VARYING CHANNEL
I
‘j
The capacity C of the Gaussian AVC with input constraint
r
and state constraint
A
is the largest number with the
property that for every
6
>
0
and sufficiently large n, there
exist codes with
N
2
exp{n(C
-
a)}
codewords, each satisfy-
ing
(l.l),
such that the supremum
of
F(s)
subject
to
(1.2)
converges to
0
as n
+m.
Our main result is the following.
Theorem
1:
The capacity of the Gaussian AVC with in-
put constraint
r
and state constraint
A
is
if
riil
According to Hughes-Narayan
[lo],
the random coding
capacity of the Gaussian AVC equals
f
log(l+
r/(A
+
u2)).
Thus, in this case Ahlswede’s alternatives do obtain. Yet a
proof of the theorem above by the elegant elimination
technique” of Ahlswede
[1]
is not apparent. Rather, we shall
use the straightforward but more computational method of
CsiszAr-Narayan [SI. Suitable approximation arguments
would enable a derivation of our theorem directly from the
results of
[SI.
Instead, we prefer to present a more transpar-
ent and direct proof, which will also serve to keep this paper
self-contained.
We also determine the capacity of the noiseless additive
AVC whose output is
r+s
rather than
r+s
+V.
The
capacity of this AVC is defined similarly to that of the
Gaussian AVC with the exception that
(1.3)
is now replaced
by
1
N
F(s)
=
-I(i:
cp(x,
+
s)
#
i}I.
Theorem
2:
The capacity of the noiseless additive AVC
with input constraint
r
and state constraint
A
is
if
TsA.
Whereas this result is not a formal special case of Theo-
rem 1, both theorems can be proved by the same method.
We shall prove the simpler Theorem
2
first
so
that the
reader may better understand the key ideas. Observe that
Theorem
1
requires a separate proof only in the case
A
+
u2
2
r.
In fact, since
(1.2)
implies for an arbitrary
E
>
0
that
11s
+
VJ12
I
n(A
+
u2
+
E)
with probability arbitrarily close to
1
if n is sufficiently large, in the case
A
+
u2
<
the
assertion of Theorem
1
follows immediately from that of
Theorem
2.
Actually, we shall show that the capacity as claimed in
Theorems
1
and
2
can be achieved using the minimum-dis-
tance decoder, namely
(1
5)
if
II~
-
.rill2
<
~ly
-
xjl12,
if
no such
1
I
i
I
N
exists.
for
j
+
i
d(Y)
=
{
i,
0,
It is worth pointing out that the result of Theorem
2
with this
decoder provides a solution to a weakened version of the
unsolved sphere-packing problem. This problem seeks the
exponential rate of the maximum number of nonintersecting
spheres of radius
JT?
in
R”
with centers in a sphere of radius
n.
In our case, the spheres may intersect but for any given
s
in
R”
of norm
I&,
only for a vanishingly small fraction
of sphere centers
xi
can
xi
+
s
be closer to another sphere
center than to
xi.
The number
C
in Theorem
2
then gives
the exponential rate of the maximum number of spheres
satisfying this condition. A similar weakened version of the
sphere-packing problem in Hamming space was solved in
[8]
as a special case of the coding theorem for the binary adder
AVC.
11.
PROOF
OF
THE
MAIN
RESULT
The proof
of
the converse parts of Theorems 1 and
2
being standard, is relegated to the Appendix. The essential
contribution
of
this paper consists in the direct part of
coding Theorems
1
and
2.
Our goal is to show that, when
r
>
A,
for all sufficiently
large n there exist
N
=
exp(nR) codewords
xI;
.
.,
x,,,
in
R“
satisfying
Ilx,l12
I
nT,
i
=
1;
.
.,
N,
with
R
arbitrarily close to
the asserted capacity value, such that for a suitable decoder
cp
the average probability
of
error
F(s)
is arbitrarily small
uniformly subject to
llsl12
5
nA.
Using the minimum distance decoder
4
of
(1.5)
for the
noiseless AVC,
(1.4)
becomes
and for the Gaussian case,
(1.3)
gives
iN
Xjl12
I
11s
+
VI12,
for some
j
z
i}.
We can assume without any loss of generality that
(2.2)
=
1,
0
<
A
<
1.
Further,
(2.1)
and
(2.2)
remain unchanged
if
all
vectors are multiplied by
l/&.
Hence it suffices to prove
that for every sufficiently small
6
>
0
and sufficiently large n
there exist
N
=
exp(nR) unit vectors
xI;.
.,xN
in
R“
with
C
-26
<
R
<
C
-
6
where
C
=
f
log(1
+
l/A)
for the noise-
less AVC and C
=
+
log(1
+
l/(A
+
u2))
for the Gaussian
AVC, such that
F’(s)
is arbitrarily small, uniformly subject to
lls1I2
I
A,
where
in the noiseless case, and
for some
j
#
i}
(2.4)
in the Gaussian case where
V=
(VI;
..,V,,)
is now a se-
quence of i.i.d. Gaussian random variables with mean
0
and
variance u2/n.
We claim that the unit vectors
xI;
.
.,
xN
of the following
Lemma
1
do possess the property above
if
7
and
E
are
sufficiently small.
Lemma
1
(Codeword Properties): For every
E
>
0,
86
<
q
<
1,
K
>
2~
and
N
=
exp(nR) with
2~
I
RI
K,
for n
2
n,,(E,q,K) there exist unit vectors
x,;
“,xN
in
R”
such that
for every unit vector
U
in
R”
and constants
CY,
p
in [0,1], we

20
IEEE
TRANSACTIONS ON INFORMATION THEORY.
VOL.
37.
NO.
I,
JANUARY
I991
Here
(.;)
denotes inner product and
1.1’
denotes “the
positive part
of.”
This lemma is an analog of the key Lemma
3
of
[SI,
and can be proved similarly. The proof is in the
Appendix.
Commencing with the noiseless case, in order to bound
(2.3)
for
lls112
I
A,
note that
Ilx;
+
s
-
x,l12
=
11x,1I2
+
1lS1l2 +
llXjl12
Hence
1
N
t?’(s)=-({i:
(x,,s)+(x,,x,)2
I+(x,,s),
1
-N
<
-
1
{
i
:
(
x;
,
s
)
I
-
77)
I
for some
j
f
i}
To
complete the proof of the direct part
of
Theorem
2,
it
suffices to check for every
(a,
p)
=
(1 -277,O)
and
(ak,
1
-
27
-
ak)/fi),
k
=
1;.
.,
K,
the condition
a’
+
p2
>
1
+
7
-
exp(-2R)
of
2)
of
Lemma
1.
(The condition
a
2
7
is clearly
satisfied provided
7
<
min{+,(l-
6)/2}.)
Differentiation
shows that
a2
+(1-27
-
a)’/A
is minimized by
a
=
(1
-
277)/1+
A,
and the minimum equals
(1
-2~)~/
1
+
A.
Thus,
the condition to be satisfied is
1
N
Obviously, if C
-26
<
R
<
C
-
6
for any fixed
6
>
0,
where
+
-
({i:
(x,,s)
+(xi,
xi)
>
1
-
7,
for some
j
z
i)
1.
1
1
c=-log
1+-
=--log
1--
(2.6) 2
(
i)
2
(
1;Aj’
The first term of the sum in
(2.6)
can be bounded by Lemma
I(i). In fact, letting
U
be the unit vector such that
(x,.
s)
I
-
77
implies by the assumption
A
<
1
that
(x,,
U)
I
-
7.
Thus
if
R
>
-
log(1-
v2),
we get that
s
exp
{n(e
-
z))
-+
0,
as
n
-+W.
(2.8)
The second term
of
the sum in
(2.6)
can be bounded using
2)
of Lemma
1
by suitably partitioning the set
of
possible values
of
the inner product
(x,,x,).
To
this end, let
al
=
1
-
77
-fi<a2<
...
<a,=1-2~,
with
ar+I-ak~q,
k=
1;
.
.,
K
-
1.
Then
(x,,
s)+
(x,, x,)
>
1
-
77
implies that
(x,, x,)
1
aI,
and
if
aL
I
(xI,
x,)
I
+
I
then necessarily
(x,,
s)
>
1-27
-
ah.
The latter, in turn, implies by
(2.7)
that
the inequality
(2.9)
will be satisfied if
77
is sufficiently small.
The proof for the Gaussian case (Theorem
1)
is similar but
bounding
(2.4)
is not as easy. We first present two simple
technical lemmas.
Lemma
2:
Let the r.v.
U
be uniformly distributed on the
unit n-sphere. Then for every vector
U
on this sphere and
any
0
<
a
<
1,
we have
1
01
-
1)/2
~r{l(~,u)l>
a)
I
2(1-
a2)
,
if
a
2
-
fi’
Proof:
Denote the angle between the unit vectors
U
and
U
by O(U,u). Then by Shannon
[12, (28)1,
With
a
=
cos$, it follows that
1
ifa2-
+zz
The proof is completed by observing that Pr{(U,
U)
I
-
a}
=
Lemma
3:
Let
U
and
U
be unit vectors with I(u,u)~~
7.
of
x
orthogo-
Pr{(U,
U)
2
a).
U
Then for any unit vector
x,
the component
x
nal to span
{U,
v)
has norm
IIX~II*I
1-(u,x)2-(U,x)2+47.
(2.10)


ILtE
TRANSACTIONS ON INFORM,\I ION THFOKY.
VOL 17
NO
I.
JANUARY
I‘NI
Further, by Lemma
3,
if
xi
=
x,’(i,u)
represents the unit
vector in the direction
of
xy
‘1,
for
j
E
F,h/
we have
IIxy
,zII
I
J1-(x,,x,)’-(x,,u)’+477
IJl-a:-pf+477
if
IKx,,
u)l
I
7.
Hence
Pr(llVII’
5
u2
+
77,
(xy’,,,~)
>
1-57
-
ah
-
p,fi>
(2.19)
where
U
is a r.v. distributed uniformly on the unit n-sphere
and
U‘
is any fixed unit vector in
R”.
Together with (2.17),
this implies that (2.16) is overbounded by
C,,,,,,,A‘,;‘,
where
Hence it suffices to show that
A$;)+
0 as
n
+CO
for every
(k,l)
E
G.
Since
(k,
I)
E
G,
there are
two
cases to consider:
a)
a;
+
P:
>
1
+
17
-
exp
(
-
2~),
and
b)
ax.
I
7,
a:
+
p,?
I
1
+
77
-exp(
-2~).
We first observe that in both cases
1-
ah
-p/fi-jV
>
0
(2.21)
provided that
7
is chosen sufficiently small. Indeed, in case
a), the expression in
(2.21)
is
2
1
-
fi
-677. In case b), the
assumption
R
<
C
-
6
=
+
log(1
+
1/(A
+
u2))-
6
implies
that
1
a:
+p:5
1
+
77-
1
exp(26)
l+y
A+u
11
<
1
+
77
-
-
exp(26).
1+A
Since
ah
+
P/fi
5
d(
a;
+
pf)(
1
+
11)
(as can be directly verified by squaring both sides), this yields
1
-a,
-p/fi-577
>
1-57
-
41
-
A(exp(26)
-
1)
+
v(
1
+
A)
>
0
if
77
is sufficiently small.
Now, in case
a)
we obtain, using Lemma
2,
that
if
E
and
77
are chosen small enough.
using Lemma
2
we obtain from (2.20) that
In case
b),
we have
R
+
+
log(1-
ai
-
P:
+
77)
>
0.
Then,
a;
-p;
+
q)+
e)
11
i
(l-a,<-P/fi-577)*
(
v2
+
q)(l-
ai
-
P,?
+47)
1
JI
/
Evaluating the maximum
of
(1
-
a
-pa
-5q)*
2
-
p2
+477
-
y(a,P)=l-a
>
U2fV
we obtain by differentiation that the maximum is attained at
and the value
of
the maximum is
(1
-w2
l+477-
1+
A+u2+v

Citations
More filters
Journal ArticleDOI

Capacity of fading channels with channel side information

TL;DR: The Shannon capacity of a fading channel with channel side information at the transmitter and receiver, and at the receiver alone is obtained, analogous to water-pouring in frequency for time-invariant frequency-selective fading channels.
Journal ArticleDOI

The effect upon channel capacity in wireless communications of perfect and imperfect knowledge of the channel

TL;DR: A model for time-varying communication single-access and multiple-access channels without feedback is presented and the rate of time variation of the channel is related to the loss in mutual information due to imperfect knowledge of the measured channel.
Journal ArticleDOI

Reliable communication under channel uncertainty

TL;DR: In many communication situations, the transmitter and the receiver must be designed without a complete knowledge of the probability law governing the channel over which transmission takes place.
Journal ArticleDOI

The method of types [information theory]

TL;DR: While the method of types is suitable primarily for discrete memoryless models, its extensions to certain models with memory are also discussed, and a wide selection of further applications are surveyed.
Journal ArticleDOI

The worst additive noise under a covariance constraint

TL;DR: The problem becomes one of extremizing the mutual information over all noise processes with covariances satisfying the correlation constraints R/sub 0/,..., R/ sub p/ for high signal powers, the worst additive noise is Gauss-Markov of order p as expected.
References
More filters
Book

Information Theory: Coding Theorems for Discrete Memoryless Systems

TL;DR: This new edition presents unique discussions of information theoretic secrecy and of zero-error information theory, including the deep connections of the latter with extremal combinatorics.
Journal ArticleDOI

Probability of error for optimal codes in a Gaussian channel

TL;DR: Upper and lower bounds are found for the error probability in decoding with optimal codes and decoding systems for a continuous channel with an additive gaussian noise and subject to an average power limitation at the transmitter.
Journal ArticleDOI

Elimination of correlation in random codes for arbitrarily varying channels

TL;DR: Under a mild regularity condition this formula for the average error capacity in case of randomized encoding turns out to be valid and follows as consequence from either a or b.
Journal ArticleDOI

The capacity of the arbitrarily varying channel revisited: positivity, constraints

TL;DR: The capacity of the AVC is determined with constraints on the transmitted codewords as well as on the channel state sequences, and it is demonstrated that it may be positive but less than the corresponding random code capacity.
Frequently Asked Questions (1)
Q1. What have the authors contributed in "Capacity of the gaussian arbitrarily varying channel" ?

In this paper, the capacity of the arbitrarily varying channel with input constraint r and state constraint 2 admits input sequences x = ( x l,,, x, ~ ) of real numbers with Cxf 5 nT and state sequences s = ( 5,., T, ~ ) of real number with L Y ~ 5 H A ; the output sequence x + s + V, where V = ( VI,,, y, ) is a sequence of independent and identically distributed Gaussian random variables with mean 0 and variance c '.