scispace - formally typeset
Search or ask a question
Institution

Bell Labs

Company
About: Bell Labs is a based out in . It is known for research contribution in the topics: Laser & Optical fiber. The organization has 36499 authors who have published 59862 publications receiving 3190823 citations. The organization is also known as: Bell Laboratories & AT&T Bell Laboratories.


Papers
More filters
Journal ArticleDOI
A.D. Wyner1
TL;DR: Shannon-theoretic limits for a very simple cellular multiple-access system, and a scheme which does not require joint decoding of all the users, and is, in many cases, close to optimal.
Abstract: We obtain Shannon-theoretic limits for a very simple cellular multiple-access system. In our model the received signal at a given cell site is the sum of the signals transmitted from within that cell plus a factor /spl alpha/ (0/spl les//spl alpha//spl les/1) times the sum of the signals transmitted from the adjacent cells plus ambient Gaussian noise. Although this simple model is scarcely realistic, it nevertheless has enough meat so that the results yield considerable insight into the workings of real systems. We consider both a one dimensional linear cellular array and the familiar two-dimensional hexagonal cellular pattern. The discrete-time channel is memoryless. We assume that N contiguous cells have active transmitters in the one-dimensional case, and that N/sup 2/ contiguous cells have active transmitters in the two-dimensional case. There are K transmitters per cell. Most of our results are obtained for the limiting case as N/spl rarr//spl infin/. The results include the following. (1) We define C/sub N/,C/spl circ//sub N/ as the largest achievable rate per transmitter in the usual Shannon-theoretic sense in the one- and two-dimensional cases, respectively (assuming that all signals are jointly decoded). We find expressions for limN/spl rarr//spl infin/C/sub N/ and limN/spl rarr//spl infin/C/spl circ//sub N/. (2) As the interference parameter /spl alpha/ increases from 0, C/sub N/ and C/spl circ//sub N/ increase or decrease according to whether the signal-to-noise ratio is less than or greater than unity. (3) Optimal performance is attainable using TDMA within the cell, but using TDMA for adjacent cells is distinctly suboptimal. (4) We suggest a scheme which does not require joint decoding of all the users, and is, in many cases, close to optimal. >

787 citations

Proceedings ArticleDOI
01 Jun 1993
TL;DR: A counting algorithm that tracks the number of alternative derivations (counts) for each derived tuple in a view, and shows that the count for a tuple can be computed at little or no cost above the cost of deriving the tuple.
Abstract: We present incremental evaluation algorithms to compute changes to materialized views in relational and deductive database systems, in response to changes (insertions, deletions, and updates) to the relations. The view definitions can be in SQL or Datalog, and may use UNION, negation, aggregation (e.g. SUM, MIN), linear recursion, and general recursion.We first present a counting algorithm that tracks the number of alternative derivations (counts) for each derived tuple in a view. The algorithm works with both set and duplicate semantics. We present the algorithm for nonrecursive views (with negation and aggregation), and show that the count for a tuple can be computed at little or no cost above the cost of deriving the tuple. The algorithm is optimal in that it computes exactly those view tuples that are inserted or deleted. Note that we store only the number of derivations, not the derivations themselves.We then present the Delete and Rederive algorithm, DRed, for incremental maintenance of recursive views (negation and aggregation are permitted). The algorithm works by first deleting a superset of the tuples that need to be deleted, and then rederiving some of them. The algorithm can also be used when the view definition is itself altered.

787 citations

Journal ArticleDOI
George Sperling1
TL;DR: A model for visual recall tasks was presented in terms of visual information storage, scanning, rehearsal, and auditory information storage and the main implication of the model for human factors is the importance of the auditory coding in visual tasks.
Abstract: A model for visual recall tasks was presented in terms of visual information storage (VIS), scanning, rehearsal, and auditory information storage (AIS). It was shown first that brief visual stimuli...

784 citations

Journal ArticleDOI
TL;DR: This paper analyzes a model of a multiplexer for packetized voice and data using the index of dispersion for intervals (IDI), which describes the cumulative covariance among successive interarrival times.
Abstract: This paper analyzes a model of a multiplexer for packetized voice and data. A major part of the analysis is devoted to characterizing the aggregate packet arrival process resulting from the superposition of separate voice streams. This is done via the index of dispersion for intervals (IDI), which describes the cumulative covariance among successive interarrival times. The IDI seems very promising as a measurement tool to characterize complex arrival processes. This paper also describes the delays experienced by voice and data packets in the multiplexer using relatively simple two-parameter approximations.

783 citations


Authors

Showing all 36526 results

NameH-indexPapersCitations
Yoshua Bengio2021033420313
David R. Williams1782034138789
John A. Rogers1771341127390
Zhenan Bao169865106571
Stephen R. Forrest1481041111816
Bernhard Schölkopf1481092149492
Thomas S. Huang1461299101564
Kurt Wüthrich143739103253
John D. Joannopoulos137956100831
Steven G. Louie13777788794
Joss Bland-Hawthorn136111477593
Marvin L. Cohen13497987767
Federico Capasso134118976957
Christos Faloutsos12778977746
Robert J. Cava125104271819
Network Information
Related Institutions (5)
IBM
253.9K papers, 7.4M citations

90% related

Georgia Institute of Technology
119K papers, 4.6M citations

89% related

University of California, Santa Barbara
80.8K papers, 4.6M citations

89% related

Massachusetts Institute of Technology
268K papers, 18.2M citations

88% related

Princeton University
146.7K papers, 9.1M citations

87% related

Performance
Metrics
No. of papers from the Institution in previous years
YearPapers
20233
202245
2021479
2020712
2019750
2018862