scispace - formally typeset
Search or ask a question
Posted Content

Sphere packings I

TL;DR: A program to prove the Kepler conjecture on sphere packings is described and it is shown that every Delaunay star that satisfies a certain regularity condition satisfies the conjecture.
Abstract: We describe a program to prove the Kepler conjecture on sphere packings. We then carry out the first step of this program. Each packing determines a decomposition of space into Delaunay simplices, which are grouped together into finite configurations called Delaunay stars. A score, which is related to the density of packings, is assigned to each Delaunay star. We conjecture that the score of every Delaunay star is at most the score of the stars in the face-centered cubic and hexagonal close packings. This conjecture implies the Kepler conjecture. To complete the first step of the program, we show that every Delaunay star that satisfies a certain regularity condition satisfies the conjecture.
Citations
More filters
Journal ArticleDOI
TL;DR: This work provides a simple method to iteratively detect and decode any linear space-time mapping combined with any channel code that can be decoded using so-called "soft" inputs and outputs and shows that excellent performance at very high data rates can be attained with either.
Abstract: Recent advancements in iterative processing of channel codes and the development of turbo codes have allowed the communications industry to achieve near-capacity on a single-antenna Gaussian or fading channel with low complexity. We show how these iterative techniques can also be used to achieve near-capacity on a multiple-antenna system where the receiver knows the channel. Combining iterative processing with multiple-antenna channels is particularly challenging because the channel capacities can be a factor of ten or more higher than their single-antenna counterparts. Using a "list" version of the sphere decoder, we provide a simple method to iteratively detect and decode any linear space-time mapping combined with any channel code that can be decoded using so-called "soft" inputs and outputs. We exemplify our technique by directly transmitting symbols that are coded with a channel code; we show that iterative processing with even this simple scheme can achieve near-capacity. We consider both simple convolutional and powerful turbo channel codes and show that excellent performance at very high data rates can be attained with either. We compare our simulation results with Shannon capacity limits for ergodic multiple-antenna channel.

2,291 citations


Additional excerpts

  • ...(or its approximation) of the lattice [35]....

    [...]

Journal ArticleDOI
TL;DR: By judicious choice of the decoding radius, it is shown that this maximum-likelihood decoding algorithm can be practically used to decode lattice codes of dimension up to 32 in a fading environment.
Abstract: We present a maximum-likelihood decoding algorithm for an arbitrary lattice code when used over an independent fading channel with perfect channel state information at the receiver. The decoder is based on a bounded distance search among the lattice points falling inside a sphere centered at the received point. By judicious choice of the decoding radius we show that this decoder can be practically used to decode lattice codes of dimension up to 32 in a fading environment.

1,760 citations


Cites background or methods from "Sphere packings I"

  • ...For the basic notations in lattice theory the reader can refer to [1]....

    [...]

  • ...Many very efficient algorithms are now available for ML decoding some well-known root lattices [1]....

    [...]

Journal ArticleDOI
TL;DR: An efficient closest point search algorithm, based on the Schnorr-Euchner (1995) variation of the Pohst (1981) method, is implemented and is shown to be substantially faster than other known methods.
Abstract: In this semitutorial paper, a comprehensive survey of closest point search methods for lattices without a regular structure is presented. The existing search strategies are described in a unified framework, and differences between them are elucidated. An efficient closest point search algorithm, based on the Schnorr-Euchner (1995) variation of the Pohst (1981) method, is implemented. Given an arbitrary point x /spl isin/ /spl Ropf//sup m/ and a generator matrix for a lattice /spl Lambda/, the algorithm computes the point of /spl Lambda/ that is closest to x. The algorithm is shown to be substantially faster than other known methods, by means of a theoretical comparison with the Kannan (1983, 1987) algorithm and an experimental comparison with the Pohst (1981) algorithm and its variants, such as the Viterbo-Boutros (see ibid. vol.45, p.1639-42, 1999) decoder. Modifications of the algorithm are developed to solve a number of related search problems for lattices, such as finding a shortest vector, determining the kissing number, computing the Voronoi (1908)-relevant vectors, and finding a Korkine-Zolotareff (1873) reduced basis.

1,616 citations


Cites background from "Sphere packings I"

  • ...It is known [16] that Voronoi regions Manuscript submitted October 26, 2000....

    [...]

Journal ArticleDOI
TL;DR: A quantized maximum signal-to-noise ratio (SNR) beamforming technique is proposed where the receiver only sends the label of the best beamforming vector in a predetermined codebook to the transmitter.
Abstract: Transmit beamforming and receive combining are simple methods for exploiting the significant diversity that is available in multiple-input multiple-output (MIMO) wireless systems. Unfortunately, optimal performance requires either complete channel knowledge or knowledge of the optimal beamforming vector; both are hard to realize. In this article, a quantized maximum signal-to-noise ratio (SNR) beamforming technique is proposed where the receiver only sends the label of the best beamforming vector in a predetermined codebook to the transmitter. By using the distribution of the optimal beamforming vector in independent and identically distributed Rayleigh fading matrix channels, the codebook design problem is solved and related to the problem of Grassmannian line packing. The proposed design criterion is flexible enough to allow for side constraints on the codebook vectors. Bounds on the codebook size are derived to guarantee full diversity order. Results on the density of Grassmannian line packings are derived and used to develop bounds on the codebook size given a capacity or SNR loss. Monte Carlo simulations are presented that compare the probability of error for different quantization strategies.

1,542 citations


Cites background from "Sphere packings I"

  • ...This problem is quite mathematically challenging from an analytical point of view [35]....

    [...]

Journal ArticleDOI
29 Jun 1997
TL;DR: In this article, the problem of finding quantum error-correcting codes is transformed into one of finding additive codes over the field GF(4) which are self-orthogonal with respect to a trace inner product.
Abstract: The unreasonable effectiveness of quantum computing is founded on coherent quantum superposition or entanglement which allows a large number of calculations to be performed simultaneously. This coherence is lost as a quantum system interacts with its environment. In the present paper the problem of finding quantum-error-correcting codes is transformed into one of finding additive codes over the field GF(4) which are self-orthogonal with respect to a certain trace inner product. Many new codes and new bounds are presented, as well as a table of upper and lower bounds on such codes of length up to 30 qubits.

1,525 citations

References
More filters
Book
02 Jan 2016
TL;DR: In this paper, the authors present an ALGOLGOL-based approach for the inclusion of complex Zeros of polynomials of a function of one real variable in a system of linear systems of equations.
Abstract: Preface to the English Edition. Preface to the German Edition. Real Interval Arithmetic. Further Concepts and Properties. Interval Evaluation and Range of Real Functions. Machine Interval Arithmetic. Complex Interval Arithmetic. Metric, Absolute, Value, and Width in. Inclusion of Zeros of a Function of One Real Variable. Methods for the Simultaneous Inclusion of Real Zeros of Polynomials. Methods for the Simultaneous Inclusion of Complex Zeros of Polynomials. Interval Matrix Operations. Fixed Point Iteration for Nonlinear Systems of Equations. Systems of Linear Equations Amenable to Interation. Optimality of the Symmetric Single Step Method with Taking Intersection after Every Component. On the Feasibility of the Gaussian Algorithm for Systems of Equations with Intervals as Coefficients. Hansen's Method. The Procedure of Kupermann and Hansen. Ireation Methods for the Inclusion of the Inverse Matrix and for Triangular Decompositions. Newton-like Methods for Nonlinear Systems of Equations. Newton-like Methods without Matrix Inversions. Newton-like Methods for Particular Systems of Nonlinear Equations. Newton-like Total step and Single Step Methods. Appendix A. The Order of Convergence of Iteration Methods in vn(Ic) and Mmn(iC) ). Appendix B. Realizations of Machine Interval Arithmetics in ALGOL 60. Appendix C. ALGOL Procedures. Bibliography. Index of Notation. Subject Index.

2,054 citations


"Sphere packings I" refers methods in this paper

  • ...ilable [H6]. Floating-point operations on computers are subject to round-off errors, making many machine computations unreliable. Methods of interval arithmetic give users control over round-off errors [Int]. These methods may be reliably implemented on machines that allow arithmetic with directed rounding, for example those conforming to the IEEE/ANSI standard 754 [W],[IEEE], [NR]. Interval arithmetic p...

    [...]

Journal ArticleDOI
David E. Goldberg1
TL;DR: This paper presents a tutorial on the aspects of floating-point that have a direct impact on designers of computer systems, and concludes with examples of how computer system builders can better support floating point.
Abstract: Floating-point arithmetic is considered as esoteric subject by many people. This is rather surprising, because floating-point is ubiquitous in computer systems: Almost every language has a floating-point datatype; computers from PCs to supercomputers have floating-point accelerators; most compilers will be called upon to compile floating-point algorithms from time to time; and virtually every operating system must respond to floating-point exceptions such as overflow. This paper presents a tutorial on the aspects of floating-point that have a direct impact on designers of computer systems. It begins with background on floating-point representation and rounding error, continues with a discussion of the IEEE floating point standard, and concludes with examples of how computer system builders can better support floating point.

1,372 citations

BookDOI
01 Jan 1953

640 citations

Journal ArticleDOI
TL;DR: The sphere packing problem as mentioned in this paper asks whether any packing of spheres of equal radius in three dimensions has density exceeding that of the face-centered-cubic lattice packing (of density φ 18 ).

100 citations