scispace - formally typeset
Search or ask a question
Journal ArticleDOI

MIMO Detection Methods: How They Work [Lecture Notes]

17 Apr 2009-IEEE Signal Processing Magazine (IEEE)-Vol. 26, Iss: 3, pp 91-95
TL;DR: An overview of approaches for detection for MIMO, in the communications receiver context, finds that notions that are important in slow fading are less important in fast fading, where diversity is provided anyway by time variations.
Abstract: The goal of this lecture has been to provide an overview of approaches, in the communications receiver context. Which method is the best in practice? This depends much on the purpose of solving : what error rate can be tolerated, what is the ultimate measure of performance (e.g., frame-error-rate, worst-case complexity, or average complexity), and what computational platform is used. Additionally, the bits in s may be part of a larger code word and different s vectors in that code word may either see the same H (slow fading) or many different realizations of H (fast fading). This complicates the picture, because notions that are important in slow fading (such as spatial diversity) are less important in fast fading, where diversity is provided anyway by time variations. Detection for MIMO has been an active field for more than ten years, and this research will probably continue for some time.

Summary (2 min read)

RELEVANCE

  • The most important motivating application for the discussion here is receivers for multiple-antenna systems such as multiple-input, multiple-output (MIMO), where several transmit antennas simultaneously send different data streams.
  • Essentially the same problem occurs in systems where the channel itself introduces time- or frequency-dispersion, in multiuser detection, and in cancellation of crosstalk.

PROBLEM STATEMENT

  • For simplicity of their discussion, the authors assume that all quantities are real-valued.
  • If H has structure, for example, if it is a Toeplitz matrix, then one should use algorithms that can exploit this structure.
  • (2) Problem (2) is a finite-alphabet-constrained least-squares (LS) problem, which is known to be nondeterministic polynomial-time (NP)-hard.

SOLUTIONS

  • (4) Problem (4) can be visualized as a decision tree with n1 1 layers, |S| branches emanating from each nonleaf node, and |S|n leaf nodes.
  • Finally, to each node, the authors associate the symbols 5s1, c, sk6 it takes to reach there from the root.
  • Clearly, a naive but valid way of solving (4) would be to traverse the entire tree to find the leaf node with the smallest cumulative metric.
  • Such a brute-force search is extremely inefficient, since there are |S|n leaf nodes to examine.
  • The authors will now review some efficient, popular, but approximate solutions to (4).

ZF DETECTOR WITH DECISION FEEDBACK (ZF-DF)

  • Consider again ZF, and suppose the authors use Gaussian elimination to compute s| in (5).
  • In the decision-tree perspective, ZF-DF can be understood as just examining one single path down from the root.
  • Clearly, after n steps the authors end up at one of the leaf nodes, but not necessarily in the one with the smallest cumulative metric.
  • In its simplest form (as explained above), ZF-DF detects sk in the natural order, but this is not optimal.
  • Even with the optimal ordering, error propagation severely limits the performance.

SPHERE DECODING (SD)

  • The SD [2], [9] first selects a user parameter R, called the sphere radius.
  • Effectively, the authors will adaptively prune the decision tree, and visit much fewer nodes than those in the original sphere.
  • In particular, the algorithm does not examine any branches stemming from the node “5” in the right subtree.
  • The SD algorithm can be improved in many other ways, too.
  • The symbols can be sorted in an arbitrary order, and this order can be optimized.

FIXED-COMPLEXITY SPHERE DECODER (FCSD)

  • FCSD [3] is, strictly speaking, not really sphere decoding, but rather a clever combination of brute-force enumeration and a low-complexity, approximate detector.
  • Authorized licensed use limited to: Linkoping Universitetsbibliotek.
  • To form its symbol decisions, FCSD selects the leaf, among the leaves it has visited, which has the smallest cumulative metric f1 1s1 2 1c1 fn 1s1, c, sn 2 .
  • FCSD solves (2) with high probability even for small r, it runs in constant time, and it has a natural parallel structure.

SEMIDEFINITE-RELAXATION (SDR) DETECTOR

  • The idea behind SDR [5], [6] is to relax the finite-alphabet constraint on s into a matrix inequality and then use semidefinite programming to solve the resulting problem.
  • The authors explain how it works, for binary phase-shift keying (BPSK) symbols (sk [ 5616).
  • It then proceeds by minimizing Trace {CS} with respect to S, but relaxes the rank constraint and instead requires that S be positive semidefinite.
  • This relaxed problem is convex, and can be efficiently solved using so-called interior point methods.
  • The error incurred by the relaxation is generally small.

LATTICE REDUCTION (LR) AIDED DETECTION

  • The idea behind LR [8], [9] is to transform the problem into a domain where the effective channel matrix is better conditioned than the original one.
  • Naturally, there are many such matrices (T5 6 I is one trivial example).
  • Namely, some of its elements may be beyond the borders of the original constellation.
  • Hence a clipping-type operation is necessary and this will introduce some loss.

SOFT DECISIONS

  • It is then of interest to take decisions on the individual bits bk,i, and often, also to quantify how reliable these decisions are.
  • Fortunately, (9) can be relatively well approximated by replacing the two sums in (9) with their largest terms.
  • This is naturally accomplished by many of the methods the authors discussed, by simply including the terms corresponding to all leaf nodes in the decision tree that the algorithm has visited.

CONCLUSIONS

  • This depends much on the purpose of solving (2): what error rate can be tolerated, what is the ultimate measure of performance (e.g., frame-error-rate, worst-case complexity, or average complexity), and what computational platform is used.
  • This complicates the picture, because notions that are important in slow fading (such as spatial diversity) are less important in fast fading, where diversity is provided anyway by time variations.
  • Detection for MIMO has been an active field for more than ten years, and this research will probably continue for some time.

Did you find this useful? Give us your feedback

Content maybe subject to copyright    Report

Linköping University Post Print
MIMO Detection Methods: How They Work
Erik G. Larsson
N.B.: When citing this work, cite the original article.
©2009 IEEE. Personal use of this material is permitted. However, permission to
reprint/republish this material for advertising or promotional purposes or for creating new
collective works for resale or redistribution to servers or lists, or to reuse any copyrighted
component of this work in other works must be obtained from the IEEE.
Erik G. Larsson, MIMO Detection Methods: How They Work, 2009, IEEE signal processing
magazine, (26), 3, 91-95.
http://dx.doi.org/10.1109/MSP.2009.932126
Postprint available at: Linköping University Electronic Press
http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-21997

IEEE SIGNAL PROCESSING MAGAZINE [91] MAY 2009
[
lecture
NOTES
]
Digital Object Identifier 10.1109/MSP.2009.932126
I
n communications, the receiver
often observes a linear superposition
of separately transmitted informa-
tion symbols. From the receiver’s
perspective, the problem is then to
separate the transmitted symbols. This is
basically an inverse problem with a
finite-alphabet constraint. This lecture
will present an accessible overview of
state-of-the-art solutions to this problem.
RELEVANCE
The most important motivating applica-
tion for the discussion here is receivers
for multiple-antenna systems such as
multiple-input, multiple-output (MIMO),
where several transmit antennas simul-
taneously send different data streams.
However, essentially the same problem
occurs in systems where the channel
itself introduces time- or frequency-dis-
persion, in multiuser detection, and in
cancellation of crosstalk.
PREREQUISITES
General mathematical maturity is
required along with knowledge of basic
linear algebra and probability.
PROBLEM STATEMENT
Concisely, the problem is to recover the
vector
s
[ R
n
from an observation of
the form
y
5 Hs 1 e,
y
[ R
m
, (1)
where H [
R
m3n
is a known (typically,
estimated beforehand) channel matrix
and e [
R
m
represents noise. We
assume that e , N
1
0, sI
2
. The ele-
ments of
s
, say s
k
, belong to a finite
alphabet
S
of size
S
. Hence there are
|
S
|
n
possible vectors
s
. For simplicity of
our discussion, we assume that all
quantities are real-valued. This is most-
ly a matter of notation, since
C
n
is iso-
morphic to
R
2
n
. We also assume that
m
$
n, that is, (1) is not underdeter-
mined, and that H has full column
rank. This is so with probability one in
most applications. We also assume that
H has no specific structure. If H has
structure, for example, if it is a Toeplitz
matrix, then one should use algorithms
that can exploit this structure.
We want to detect
s
in the maxi mum-
likelihood (ML) sense. This is equivalent to
The problem:
min
s
[ S
n
7y 2 Hs7. (2)
Problem (2) is a finite-alphabet-con-
strained least-squares (LS) problem,
which is known to be nondeterministic
polynomial-time (NP)-hard. The compli-
cating factor is of course the constraint
s
[
S
n
, otherwise (2) would be just clas-
sical LS regression.
SOLUTIONS
As a preparation, we introduce the
QL-decomposition of H : H 5 QL, where
Q [
R
m3n
is orthonormal (Q
T
Q 5 I),
and
L
[ R
n3n
is lower triangular. Then
7y 2 Hs7
2
5 7QQ
T
1
y2Hs
2
7
2
1 7
1
I2QQ
T
21
y 2Hs
2
7
2
57Q
T
y 2 Ls7
2
1 7
1
I 2 QQ
T
2
y7
2
,
where the last term does not depend on
s
.
It follows that we can reformulate (2) as
Equivalent problem: min
s
[ S
n
7y
|
2 Ls 7,
where
y
|
! Q
T
y
(3)
or, in yet another equivalent form, as
min
5
s
1
,c, s
n
6
s
k
PS
5
f
1
1
s
1
2
1 f
2
1
s
1
, s
2
2
1
c
1 f
n
1
s
1
, c, s
n
26
,
where
f
k
1
s
1
, c, s
k
2
!
a
y
|
k
2
a
k
l
51
L
k, l
s
l
b
2
. (4)
Problem (4) can be visualized as a
decision tree with n 1 1 layers,
|
S
|
branches emanating from each nonleaf
node, and
|
S
|
n
leaf nodes. See Figure 1.
To any branch, we associate a hypotheti-
cal decision on s
k
, and the branch metric
f
k
1
s
1
, c, s
k
2
. Also, to any node (except
the root), we associate the cumulative
metric f
1
1
s
1
2
1
c
1 f
k
1
s
1
, c, s
k
2
,
which is just the sum of all branch met-
rics accumulated when traveling to that
node from the root. Finally, to each node,
we associate the symbols
5
s
1
, c, s
k
6
it
takes to reach there from the root.
Clearly, a naive but valid way of solving
(4) would be to traverse the entire tree to
find the leaf node with the smallest cumu-
lative metric. However, such a brute-force
search is extremely inefficient, since there
are
|
S
|
n
leaf nodes to examine. We will
now review some efficient, popular, but
approximate solutions to (4).
ZERO-FORCING (ZF) DETECTOR
The ZF detector first solves (2), neglect-
ing the constraint
s
[
S
n
s
|
! arg min
s
[ R
n
7y 2 Hs7
5 arg min
s
[ R
n
7y
,
2 Ls 7 5 L
21
y
,
. (5)
Of course,
L
21
does not need to be explic-
itly computed. For example, one can do
Gaussian elimination: take s
|
1
5 y
|
1
/L
1,1
,
then s
|
2
5
1
y
|
2
2 s
|
1
L
2,1
2
/L
2,2
, and so forth.
ZF then approximates (2) by projecting
each s
|
k
onto the constellation
S
Erik G. Larsson
MIMO Detection Methods: How They Work
1053-5888/09/$25.00©2009IEEE
Authorized licensed use limited to: Linkoping Universitetsbibliotek. Downloaded on April 21, 2009 at 05:22 from IEEE Xplore. Restrictions apply.

IEEE SIGNAL PROCESSING MAGAZINE [92] MAY 2009
[
lecture
NOTES
]
continued
s
^
k
5
3
s
|
k
4
! arg min
s
k
[ S
|
s
k
2 s
|
k
|
. (6)
We see that s
|
5 s 1 L
21
Q
T
e, so s
|
in (5) is
free of intersymbol interference. This is
how ZF got its name. However,
unfortunately ZF works poorly unless H is
well conditioned. The reason is that the
correlation between the noises in s
|
k
is
neglected in the projection operation (6).
This correlation can be very strong, espe-
cially if H is ill conditioned (the covariance
matrix of s
|
is S ! s
#
1
L
T
L
2
21
2
. There are
some variants of the ZF approach. For
example, instead of computing s
|
as in (5),
one can use the MMSE estimate (take
s
|
5 E
3
s|
y
4
). This can improve perfor-
mance somewhat, but it does not overcome
the fundamental problem of the approach.
ZF DETECTOR WITH
DECISION FEEDBACK (ZF-DF)
Consider again ZF, and suppose we use
Gaussian elimination to compute s
|
in
(5). ZF-DF [1] does exactly this, with the
modification that it projects the symbols
onto the constellation
S
in each step of
the Gaussian elimination, rather than
afterwards. More precisely,
1) Detect s
1
via s
^
1
5 arg min
s
1
[ S
f
1
1
s
1
2
5
c
y
|
1
L
1,1
d
.
2) Consider s
1
known (s
1
5 s
^
1
) and
detect s
2
via s
^
2
5 arg min
s
2
[ S
f
2
1
s
^
1
, s
2
2
5
c
y
|
2
2 s
^
1
L
2,1
L
2,2
d
.
3) Continue for k 5 3, . . . , n
s
^
k
5 arg min
s
k
[ S
f
k
1
s
^
1
, c, s
^
k21
, s
k
2
5
c
y
|
k
2S
k
21
l5 1
L
k, l
s
^
l
L
k, k
d
.
In the decision-tree perspective,
ZF-DF can be understood as just examin-
ing one single path down from the root.
When deciding on s
k
, it considers
s
1
, c, s
k
21
known and takes the s
k
that
corresponds to the smallest branch met-
ric. Clearly, after n steps we end up at one
of the leaf nodes, but not necessarily in the
one with the smallest cumulative metric.
In Figure 2(a), ZF-DF first chooses
the left branch originating from the root
(since 1 , 5), then the right branch
(since 2 . 1) and at last the left branch
(because 3 , 4), reaching the leaf node
with cumulative metric 1 1 1 1
3
5
5
.
The problem with ZF-DF is error
propagation. If, due to noise, an incor-
rect symbol decision is taken in any of
the n steps, then this error will propa-
gate and many of the subsequent deci-
sions are likely to be wrong as well. In
its simplest form (as explained above),
ZF-DF detects s
k
in the natural order,
but this is not optimal. The detection
order can be optimized to minimize the
effects of error propagation. Not sur-
prisingly, it is best to start with the sym-
bol for which ZF produces the most
reliable result: that is, the symbol s
k
for
which
S
k, k
is the smallest, and then
proceed to less and less reliable sym-
bols. However, even with the optimal
ordering, error propagation severely
limits the performance.
SPHERE DECODING (SD)
The SD [2], [9] first selects a user
parameter R, called the sphere radius. It
then traverses the entire tree (from left
to right, say). However, once it encoun-
ters a node with cumulative metric
larger than R, then it does not follow
down any branch from this node. Hence,
in effect, SD enumerates all leaf nodes
which lie inside the sphere
7y
|
2 Ls 7
2
# R. This also explains the
algorithm’s name.
In Figure 2(b), we set the sphere radi-
us to R 5 6. The SD algorithm then tra-
verses the tree from left to right. When it
encounters the node “7” in the right sub-
tree, for which 7 . 6 5 R, SD does not
follow any branches emanating from it.
Similarly, since 8 . 6, SD does not visit
Root Node
s
1
= 1
s
2
= 1
s
3
= 1 s
3
= 1 s
3
= 1s
3
= +1 s
3
= +1 s
3
= 1s
3
= +1 s
3
= +1
s
2
= 1
s
2
= +1
s
2
= +1
f
1
(1) = 1
f
2
(1, 1) = 2
f
3
(. . .) = 4 f
3
(. . .) = 1
343119
Leaves
{1, 1, 1}
{1, 1, 1}{1, 1, 1}{1, 1, 1}{1, 1, 1}{1, 1, 1}{1, 1, 1}{1, 1, 1}
f
2
(1, 1) = 1 f
2
(1, 1) = 2 f
2
(1, 1) = 3
f
1
(1) = 5
s
1
= +1
15
72
756108917
3 8
4
[FIG1] Problem (4) as a decision tree, exemplified for binary modulation (S = {–1, +1}, |S| = 2) and n = 3. The branch metrics f
k
(s
1
, . . ., s
k
)
are in blue written next to each branch. The cumulative metrics f
1
(s
1
)+ . . . + f
k
(s
1
, . . . , s
k
) are written in red in the circles representing
each node. The double circle represents the optimal (ML) decision.
Authorized licensed use limited to: Linkoping Universitetsbibliotek. Downloaded on April 21, 2009 at 05:22 from IEEE Xplore. Restrictions apply.

IEEE SIGNAL PROCESSING MAGAZINE [93] MAY 2009
any branches below the node “8” in the
rightmost subtree.
SD in this basic form can be much
improved by a mechanism called prun-
ing. The idea is this: Every time we reach
a leaf node with cumulative metric M, we
know that the solution to (4) must be
contained in the sphere 7y
|
2 Ls 7
2
# M.
So if M , R, we can set R J M, and con-
tinue the algorithm with a smaller sphere
radius. Effectively, we will adaptively
prune the decision tree, and visit much
fewer nodes than those in the original
sphere. Figure 2(c) exemplifies the prun-
ing. Here the radius is initialized to
R 5`, and then updated any time a leaf
node is visited. For instance, when visit-
ing the leaf node “4,” R will be set to
R 5 4. This means that the algorithm
will not follow branches from nodes that
have a branch metric larger than four. In
particular, the algorithm does not exam-
ine any branches stemming from the
node “5” in the right subtree.
The SD algorithm can be improved in
many other ways, too. The symbols can be
sorted in an arbitrary order, and this order
can be optimized. Also, when traveling
down along the branches from a given
node, one may enumerate the branches
either in the natural order or in a zigzag
fashion (e.g., s
k
5
5
25, 23, 21, 21, 3, 5
6
versus s
k
5
5
21, 1, 23, 3, 25, 5
6
). The
SD algorithm is easy to implement,
although the procedure cannot be directly
parallelized. Given large enough initial radi-
us R, SD will solve (2). However, depending
on H, the time the algorithm takes to finish
will fluctuate, and may occasionally be
very long.
FIXED-COMPLEXITY
SPHERE DECODER (FCSD)
FCSD [3] is, strictly speaking, not really
sphere decoding, but rather a clever
combination of brute-force enumeration
and a low-complexity, approximate detec-
tor. In view of the decision tree, FCSD
visits all
|
S
|
r
nodes on layer
r
, where
r
,
0
# r # n is a user parameter. For each
node on layer
r
, the algorithm considers
5
s
1
, ..., s
r
6
fixed and formulates and
solves the subproblem
min
5
s
r11
, c, s
n
6
s
k
[
S
5
f
r11
1
s
1
, c, s
r11
2
1
. . .
1 f
n
1
s
1
, c, s
n
26
. (7)
1
5
21 2
3
4
1343119
1
5
3
2
78
745610
8
9
17
ZF-DF
1
5
21 23
4
1343119
1
5
3
2
78
745610
8
9
17
SD, No Pruning
(here: R = 6)
15
21
23
41 34 31 19
1
5
327
8
7456
10
8
9
17
SD, Pruning
(here: R = )
15
2
1
23
41 3
4
31 19
1
5
327
8
7456
10
8
9
17
FCSD
(here: r = 1)
(a)
(b)
(c)
(d)
[FIG2] Illustration of detection algorithms as a tree search. Solid-line nodes and branches are visited. Dashed nodes and branches are
not visited. The double circles represent the ultimate decisions on
s
. (a) ZF-DF: At each node, the symbol decision is based on choosing
the branch with the smallest branch metric. (b) SD, no pruning: Only nodes with S
n
k
51
f
k
1
s
1
, . . . , s
k
2
#
R are visited. (c) SD, pruning:
Like SD, but after encountering a leaf node with cumulative metric M, the algorithm will set R :
5
M. (d) FCSD: Visits all nodes on the
r th layer, and proceeds with ZF-DF from these.
Authorized licensed use limited to: Linkoping Universitetsbibliotek. Downloaded on April 21, 2009 at 05:22 from IEEE Xplore. Restrictions apply.

IEEE SIGNAL PROCESSING MAGAZINE [94] MAY 2009
[
lecture
NOTES
]
continued
In effect, by doing so, FCSD will reach
down to
|
S
|
r
of the
|
S
|
n
leaves. To form its
symbol decisions, FCSD selects the leaf,
among the leaves it has visited, which
has the smallest cumulative metric
f
1
1
s
1
2
1
c
1 f
n
1
s
1
, c, s
n
2
.
The subproblem (7) must be solved
once for each combination
5
s
1
, ..., s
r
6
,
that is
|
S
|
r
times. FCSD does this approx-
imately, using a low-complexity method
(ZF or ZF-DF are good choices). This
works well because (7) is overdetermined:
there are n observations ( y
|
1
, c, y
|
n
),
but only n 2 r unknowns (s
r11
, c, s
n
).
More precisely, the equivalent channel
matrix when solving (7) will be a tall sub-
matrix of H, which is generally much
better conditioned than H.
Figure 2(d) illustrates the algorithm.
Here
r
5 1. Thus, both nodes “1” and “5”
in the layer closest to the root node are
visited. Starting from each of these two
nodes, a ZF-DF search is performed.
Naturally, the symbol ordering can be
optimized. The optimal ordering is the
one which renders the problem (7) most
well-conditioned. This is achieved by
sorting the symbols so that the most
“difficult” symbols end up near the tree
root. Note that “difficult symbol” is non-
trivial to define precisely here, but intui-
tively think of it as a symbol s
k
for which
S
k,k
is large.
The choice of
r
offers a tradeoff
between complexity and performance.
FCSD solves (2) with high probability
even for small
r
, it runs in constant time,
and it has a natural parallel structure.
Relatives of FCSD that produce soft out-
put also exist [4].
SEMIDEFINITE-RELAXATION
(SDR) DETECTOR
The idea behind SDR [5], [6] is to relax the
finite-alphabet constraint on
s
into a
matrix inequality and then use semidefi-
nite programming to solve the resulting
problem. We explain how it works, for
binary phase-shift keying (BPSK) symbols
(s
k
[
5
6 1
6
). Define
s
2
!
c
s
1
d
, S ! s
2
s
2
T
5
c
s
1
d
3
s
T
1
4
,
C !
c
L
T
L 2 L
T
y
,
2 y
, T
L 0
d
.
Then
7y
|
2 Ls 7
2
5 s
T
cs 1 7y
|
7
2
5 Trace
5
CS
6
1 7y
|
7
2
so solving (3) is the same as finding the
vector s [
S
n
that minimizes Trace {CS}.
SDR exploits that the constraint
s [
S
n
is equivalent to requiring that
rank {S} 5 1, s
n
11
5 1 and diag {S} 5
{1, . . . , 1}. It then proceeds by minimizing
Trace {CS} with respect to
S
, but relaxes
the rank constraint and instead requires
that
S
be positive semidefinite. This
relaxed problem is convex, and can be effi-
ciently solved using so-called interior
point methods. Once the matrix
S
is
found, there are a variety of ways to deter-
mine
s
, for example to take the dominant
eigenvector of
S
(forcing the last element
to unity) and then project each element
onto
S
like in (6). The error incurred by
the relaxation is generally small.
LATTICE REDUCTION (LR)
AIDED DETECTION
The idea behind LR [8], [9] is to trans-
form the problem into a domain where
the effective channel matrix is better
conditioned than the original one. How
does it work? If the constellation
S
is
uniform, then
S
may be extended to a
scaled enumeration of all integers, and
S
n
may be extended to a lattice
S
n
. For
illustration, if S 5
5
23, 2 1, 1, 3
6
, then
S
n
5
5
c, 23, 21, 1, 3, c
6
3
c
3
5
c, 23, 21, 1, 3, c
6
. LR decides
first on an n 3 n matrix
T
that has inte-
ger elements (
T
k,l
[ Z ) and which maps
the lattice
S
n
onto itself:
Ts [
S
n
4s [
S
n
. That is,
T
should be
invertible, and its inverse should have
integer elements. This happens precisely
if its determinant has unit magnitude:
|
T
|
561. Naturally, there are many
such matrices (
T
56I is one trivial
example). LR chooses such a matrix
T
for which, additionally, H
T
is well condi-
tioned. It then computes
s
^
r
! arg min
s9P
S
n
|| y 2
1
HT
2
s
r
||. (8)
Problem (8) is comparatively easy, since
H
T
is well conditioned, and simple
methods like ZF or ZF-DF generally
work well. Once s
^
r
is found, it is trans-
formed back to the original coordinate
system by taking s
^
5 T
21
s
^
r
.
LR contains two critical steps. First, a
suitable matrix
T
must be found. There
are good methods for this (e.g., see refer-
ences in [8], [9]). This is computationally
expensive, but if the channel H stays
constant for a long time then the cost of
finding
T
may be shared between many
instances of (2) and complexity is less of
an issue. The other problem is that while
the solution s
^
always belongs to
S
2
n
, it
may not belong to
S
n
. Namely, some of
its elements may be beyond the borders
of the original constellation. Hence a
clipping-type operation is necessary and
this will introduce some loss.
SOFT DECISIONS
In practice, each symbol s
k
typically is
composed of information-carrying bits,
say
5
b
k, 1
, c, b
k, p
6
. It is then of interest
to take decisions on the individual bits
b
k,
i
, and often, also to quantify how reli-
able these decisions are. Such reliability
information about a bit is called a “soft
decision,” and is typically expressed via
the probability ratio
P
1
b
k,i
5 1| y
2
P
1
b
k,i
5 0| y
2
5
g
s:b
k,i
1
s
2
51
P
1
s| y
2
g
s:b
k,i
1
s
2
50
P
1
s| y
2
5
g
s:b
k,i
1
s
2
51
exp
a
2
1
s
||
y 2 Hs||
2
b
P
1
s
2
g
s:b
k,i
1
s
2
50
exp
a
2
1
s
||
y 2 Hs||
2
b
P
1
s
2
.
(9)
Here “
s
:b
k,i
1
s
2
5b” means all s for which
the ith bit of s
k
is equal to
b
, and P
1
s
2
is
the probability that the transmitter sent s.
To derive (9), use Bayes rule and the
Gaussian assumption made on e [4].
Fortunately, (9) can be relatively well
approximated by replacing the two sums
in (9) with their largest terms. To find
these maximum terms is a slightly
modified version of (2), at least if all s
are equally likely so that P
1
s
2
5 1/|S|
n
.
Hence, if (2) can be solved, good approx-
imations to (9) are available too. An
even better approximation to (9) is
obtained if more terms are retained, i.e.,
Authorized licensed use limited to: Linkoping Universitetsbibliotek. Downloaded on April 21, 2009 at 05:22 from IEEE Xplore. Restrictions apply.

Citations
More filters
Journal ArticleDOI
TL;DR: The gains in multiuser systems are even more impressive, because such systems offer the possibility to transmit simultaneously to several users and the flexibility to select what users to schedule for reception at any given point in time.
Abstract: Multiple-input multiple-output (MIMO) technology is maturing and is being incorporated into emerging wireless broadband standards like long-term evolution (LTE) [1]. For example, the LTE standard allows for up to eight antenna ports at the base station. Basically, the more antennas the transmitter/receiver is equipped with, and the more degrees of freedom that the propagation channel can provide, the better the performance in terms of data rate or link reliability. More precisely, on a quasi static channel where a code word spans across only one time and frequency coherence interval, the reliability of a point-to-point MIMO link scales according to Prob(link outage) ` SNR-ntnr where nt and nr are the numbers of transmit and receive antennas, respectively, and signal-to-noise ratio is denoted by SNR. On a channel that varies rapidly as a function of time and frequency, and where circumstances permit coding across many channel coherence intervals, the achievable rate scales as min(nt, nr) log(1 + SNR). The gains in multiuser systems are even more impressive, because such systems offer the possibility to transmit simultaneously to several users and the flexibility to select what users to schedule for reception at any given point in time [2].

5,158 citations

Journal ArticleDOI
TL;DR: Very large MIMO as mentioned in this paper is a new research field both in communication theory, propagation, and electronics and represents a paradigm shift in the way of thinking both with regards to theory, systems and implementation.
Abstract: This paper surveys recent advances in the area of very large MIMO systems. With very large MIMO, we think of systems that use antenna arrays with an order of magnitude more elements than in systems being built today, say a hundred antennas or more. Very large MIMO entails an unprecedented number of antennas simultaneously serving a much smaller number of terminals. The disparity in number emerges as a desirable operating condition and a practical one as well. The number of terminals that can be simultaneously served is limited, not by the number of antennas, but rather by our inability to acquire channel-state information for an unlimited number of terminals. Larger numbers of terminals can always be accommodated by combining very large MIMO technology with conventional time- and frequency-division multiplexing via OFDM. Very large MIMO arrays is a new research field both in communication theory, propagation, and electronics and represents a paradigm shift in the way of thinking both with regards to theory, systems and implementation. The ultimate vision of very large MIMO systems is that the antenna array would consist of small active antenna units, plugged into an (optical) fieldbus.

2,717 citations

Journal ArticleDOI
TL;DR: In this article, the authors provide a recital on the historic heritages and novel challenges facing massive/large-scale multiple-input multiple-output (LS-MIMO) systems from a detection perspective.
Abstract: The emerging massive/large-scale multiple-input multiple-output (LS-MIMO) systems that rely on very large antenna arrays have become a hot topic of wireless communications. Compared to multi-antenna aided systems being built at the time of this writing, such as the long-term evolution (LTE) based fourth generation (4G) mobile communication system which allows for up to eight antenna elements at the base station (BS), the LS-MIMO system entails an unprecedented number of antennas, say 100 or more, at the BS. The huge leap in the number of BS antennas opens the door to a new research field in communication theory, propagation and electronics, where random matrix theory begins to play a dominant role. Interestingly, LS-MIMOs also constitute a perfect example of one of the key philosophical principles of the Hegelian Dialectics, namely, that “quantitative change leads to qualitative change.” In this treatise, we provide a recital on the historic heritages and novel challenges facing LS-MIMOs from a detection perspective. First, we highlight the fundamentals of MIMO detection, including the nature of co-channel interference (CCI), the generality of the MIMO detection problem, the received signal models of both linear memoryless MIMO channels and dispersive MIMO channels exhibiting memory, as well as the complex-valued versus real-valued MIMO system models. Then, an extensive review of the representative MIMO detection methods conceived during the past 50 years (1965–2015) is presented, and relevant insights as well as lessons are inferred for the sake of designing complexity-scalable MIMO detection algorithms that are potentially applicable to LS-MIMO systems. Furthermore, we divide the LS-MIMO systems into two types, and elaborate on the distinct detection strategies suitable for each of them. The type-I LS-MIMO corresponds to the case where the number of active users is much smaller than the number of BS antennas, which is currently the mainstream definition of LS-MIMO. The type-II LS-MIMO corresponds to the case where the number of active users is comparable to the number of BS antennas. Finally, we discuss the applicability of existing MIMO detection algorithms in LS-MIMO systems, and review some of the recent advances in LS-MIMO detection.

626 citations

Journal ArticleDOI
TL;DR: It is found that regardless of the Ricean K-factor, in the case of perfect CSI, the approximations converge to the same constant value as the exact results, as the number of base station antennas grows large, while the transmit power of each user can be scaled down proportionally to 1/M.
Abstract: This paper investigates the uplink achievable rates of massive multiple-input multiple-output (MIMO) antenna systems in Ricean fading channels, using maximal-ratio combining (MRC) and zero-forcing (ZF) receivers, assuming perfect and imperfect channel state information (CSI). In contrast to previous relevant works, the fast fading MIMO channel matrix is assumed to have an arbitrary-rank deterministic component as well as a Rayleigh-distributed random component. We derive tractable expressions for the achievable uplink rate in the large-antenna limit, along with approximating results that hold for any finite number of antennas. Based on these analytical results, we obtain the scaling law that the users' transmit power should satisfy, while maintaining a desirable quality of service. In particular, it is found that regardless of the Ricean K-factor, in the case of perfect CSI, the approximations converge to the same constant value as the exact results, as the number of base station antennas,, grows large, while the transmit power of each user can be scaled down proportionally to. If CSI is estimated with uncertainty, the same result holds true but only when the Ricean K-factor is non-zero. Otherwise, if the channel experiences Rayleigh fading, we can only cut the transmit power of each user proportionally to 1 root M. In addition, we show that with an increasing Ricean K-factor, the uplink rates will converge to fixed values for both MRC and ZF receivers.

526 citations

Journal ArticleDOI
TL;DR: In this paper, a joint precoding/decoding design that maximizes the end-to-end (e2e) performance of full-duplex relaying in amplify-and-forward (AF) cooperative networks is investigated.
Abstract: In this paper, we deal with the deployment of full-duplex relaying in amplify-and-forward (AF) cooperative networks with multiple-antenna terminals. In contrast to previous studies, which focus on the spatial mitigation of the loopback interference (LI) at the relay node, a joint precoding/decoding design that maximizes the end-to-end (e2e) performance is investigated. The proposed precoding incorporates rank-1 zero-forcing (ZF) LI suppression at the relay node and is derived in closed-form by solving appropriate optimization problems. In order to further reduce system complexity, the antenna selection (AS) problem for full-duplex AF cooperative systems is discussed. We investigate different AS schemes to select a single transmit antenna at both the source and the relay, as well as a single receive antenna at both the relay and the destination. To facilitate comparison, exact outage probability expressions and asymptotic approximations of the proposed AS schemes are provided. In order to overcome zero-diversity effects associated with the AS operation, a simple power allocation scheme at the relay node is also investigated and its optimal value is analytically derived. Numerical and simulation results show that the joint ZF-based precoding significantly improves e2e performance, while AS schemes are efficient solutions for scenarios with strict computational constraints.

244 citations

References
More filters
Journal ArticleDOI
TL;DR: This work provides a simple method to iteratively detect and decode any linear space-time mapping combined with any channel code that can be decoded using so-called "soft" inputs and outputs and shows that excellent performance at very high data rates can be attained with either.
Abstract: Recent advancements in iterative processing of channel codes and the development of turbo codes have allowed the communications industry to achieve near-capacity on a single-antenna Gaussian or fading channel with low complexity. We show how these iterative techniques can also be used to achieve near-capacity on a multiple-antenna system where the receiver knows the channel. Combining iterative processing with multiple-antenna channels is particularly challenging because the channel capacities can be a factor of ten or more higher than their single-antenna counterparts. Using a "list" version of the sphere decoder, we provide a simple method to iteratively detect and decode any linear space-time mapping combined with any channel code that can be decoded using so-called "soft" inputs and outputs. We exemplify our technique by directly transmitting symbols that are coded with a channel code; we show that iterative processing with even this simple scheme can achieve near-capacity. We consider both simple convolutional and powerful turbo channel codes and show that excellent performance at very high data rates can be attained with either. We compare our simulation results with Shannon capacity limits for ergodic multiple-antenna channel.

2,291 citations

Journal ArticleDOI
TL;DR: Using this joint space-time approach, spectral efficiencies ranging from 20-40 bit/s/Hz have been demonstrated in the laboratory under flat fading conditions at indoor fading rates.
Abstract: The signal detection algorithm of the vertical BLAST (Bell Laboratories Layered Space-Time) wireless communications architecture is briefly described. Using this joint space-time approach, spectral efficiencies ranging from 20-40 bit/s/Hz have been demonstrated in the laboratory under flat fading conditions at indoor fading rates. Early results are presented.

1,791 citations


"MIMO Detection Methods: How They Wo..." refers background in this paper

  • ...For each node on layer r, the algorithm considers 5s 1 , ..., s r 6 fixed and formulates and solves the subproblem min ) 1 5 2 1 2 3 4 1 3 4 3 1 1 9 1 5 3 2 7 8 7 4 5 6 1 0 8 9 17 ZF-DF 1 5 2 1 2 3 4 1 3 4 3 1 1 9 1 5 3 2 7 8 7 4 5 6 1 0 8 9 17 SD, No Pruning (here: R = 6) 1 5 2 1 2 3 4 1 3 4 3 1 1 9 1 5 3 2 7 8 7 4 5 6 10 8 9 17 SD, Pruning (here: R = ∞) 1 5 2 1 2 3 4 1 3 4 3 1 1 9 1 5 3 2 7 8 7 4 5 6 10 8 9...

    [...]

  • ...the ultimate decisions on s. (a) ZF-DF: At each node, the symbol decision is based on choosing the branch with the smallest branch metric....

    [...]

  • ...(8) Problem (8) is comparatively easy, since HT is well conditioned, and simple methods like ZF or ZF-DF generally work well....

    [...]

  • ...FCSD does this approximately , using a low-complexity method (ZF or ZF-DF are good choices)....

    [...]

  • ...In its simplest form (as explained above), ZF-DF detects s k in the natural order, but this is not optimal....

    [...]

Journal ArticleDOI
TL;DR: By judicious choice of the decoding radius, it is shown that this maximum-likelihood decoding algorithm can be practically used to decode lattice codes of dimension up to 32 in a fading environment.
Abstract: We present a maximum-likelihood decoding algorithm for an arbitrary lattice code when used over an independent fading channel with perfect channel state information at the receiver. The decoder is based on a bounded distance search among the lattice points falling inside a sphere centered at the received point. By judicious choice of the decoding radius we show that this decoder can be practically used to decode lattice codes of dimension up to 32 in a fading environment.

1,760 citations


Additional excerpts

  • ...The SD [2], [9] first selects a user parameter R, called the sphere radius....

    [...]

Journal ArticleDOI
TL;DR: An efficient closest point search algorithm, based on the Schnorr-Euchner (1995) variation of the Pohst (1981) method, is implemented and is shown to be substantially faster than other known methods.
Abstract: In this semitutorial paper, a comprehensive survey of closest point search methods for lattices without a regular structure is presented. The existing search strategies are described in a unified framework, and differences between them are elucidated. An efficient closest point search algorithm, based on the Schnorr-Euchner (1995) variation of the Pohst (1981) method, is implemented. Given an arbitrary point x /spl isin/ /spl Ropf//sup m/ and a generator matrix for a lattice /spl Lambda/, the algorithm computes the point of /spl Lambda/ that is closest to x. The algorithm is shown to be substantially faster than other known methods, by means of a theoretical comparison with the Kannan (1983, 1987) algorithm and an experimental comparison with the Pohst (1981) algorithm and its variants, such as the Viterbo-Boutros (see ibid. vol.45, p.1639-42, 1999) decoder. Modifications of the algorithm are developed to solve a number of related search problems for lattices, such as finding a shortest vector, determining the kissing number, computing the Voronoi (1908)-relevant vectors, and finding a Korkine-Zolotareff (1873) reduced basis.

1,616 citations


Additional excerpts

  • ...The SD [2], [9] first selects a user parameter R, called the sphere radius....

    [...]

Journal ArticleDOI
TL;DR: A least-mean-square adaptive filter with a variable step size, allowing the adaptive filter to track changes in the system as well as produce a small steady state error, is introduced.
Abstract: A least-mean-square (LMS) adaptive filter with a variable step size is introduced. The step size increases or decreases as the mean-square error increases or decreases, allowing the adaptive filter to track changes in the system as well as produce a small steady state error. The convergence and steady-state behavior of the algorithm are analyzed. The results reduce to well-known results when specialized to the constant-step-size case. Simulation results are presented to support the analysis and to compare the performance of the algorithm with the usual LMS algorithm and another variable-step-size algorithm. They show that its performance compares favorably with these existing algorithms. >

966 citations

Frequently Asked Questions (9)
Q1. What are the contributions in "Mimo detection methods: how they work" ?

In this paper, the authors provide an overview of approaches to ( 2 ) in the communications receiver context, and which method is the best in practice. 

If the constellation S is uniform, then S may be extended to a scaled enumeration of all integers, and S n may be extended to a lattice S n. 

The most important motivating application for the discussion here is receivers for multiple-antenna systems such as multiple-input, multiple-output (MIMO), where several transmit antennas simultaneously send different data streams. 

This depends much on the purpose of solving (2): what error rate can be tolerated, what is the ultimate measure of performance (e.g., frame-error-rate, worst-case complexity, or average complexity), and what computational platform is used. 

Once ŝ r is found, it is transformed back to the original coordinate system by taking ŝ5 T21ŝ r.LR contains two critical steps. 

It then computesŝr ! arg min s9PSn || y2 1HT 2s r||. (8) Problem (8) is comparatively easy, since HT is well conditioned, and simple methods like ZF or ZF-DF generallywork well. 

The ZF detector first solves (2), neglecting the constraint s [ S ns| ! arg min s[Rn 7y2Hs 7 5 arg mins[Rn 7 y, 2 Ls 7 5 L21y,. (5)Of course, L21 does not need to be explicitly computed. 

Such reliability information about a bit is called a “soft decision,” and is typically expressed via the probability ratioP 1bk,i5 1| y 2 P 1bk,i5 0| y 2 5 g s:bk,i 1s251 P 1s| y 2 g s:bk,i 1s250 P 1s| y 25 g s:bk,i 1s251expa2 1s || y2Hs||2bP 1s 2 g s:bk,i 1s250expa2 1s || y2Hs||2bP 1s 2 .(9)Here “s:bk,i 1s 2 5 b” means all s for which the ith bit of sk is equal to b, and P 1s 2 is the probability that the transmitter sent s. 

This is computationally expensive, but if the channel H stays constant for a long time then the cost of finding T may be shared between many instances of (2) and complexity is less of an issue.