Institution

# Alcatel-Lucent

Stuttgart, Germany•

About: Alcatel-Lucent is a based out in Stuttgart, Germany. It is known for research contribution in the topics: Signal & Network packet. The organization has 37003 authors who have published 53332 publications receiving 1430547 citations. The organization is also known as: Alcatel-Lucent S.A. & Alcatel.

Topics: Signal, Network packet, Base station, Optical fiber, Node (networking)

##### Papers published on a yearly basis

##### Papers

More filters

••

Bell Labs

^{1}, École Normale Supérieure^{2}, AT&T^{3}, École Polytechnique de Montréal^{4}, Alcatel-Lucent^{5}TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.

Abstract: Multilayer neural networks trained with the back-propagation algorithm constitute the best example of a successful gradient based learning technique. Given an appropriate network architecture, gradient-based learning algorithms can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters, with minimal preprocessing. This paper reviews various methods applied to handwritten character recognition and compares them on a standard handwritten digit recognition task. Convolutional neural networks, which are specifically designed to deal with the variability of 2D shapes, are shown to outperform all other techniques. Real-life document recognition systems are composed of multiple modules including field extraction, segmentation recognition, and language modeling. A new learning paradigm, called graph transformer networks (GTN), allows such multimodule systems to be trained globally using gradient-based methods so as to minimize an overall performance measure. Two systems for online handwriting recognition are described. Experiments demonstrate the advantage of global training, and the flexibility of graph transformer networks. A graph transformer network for reading a bank cheque is also described. It uses convolutional neural network character recognizers combined with global training techniques to provide record accuracy on business and personal cheques. It is deployed commercially and reads several million cheques per day.

42,067 citations

••

Aalto University

^{1}, Université de Montréal^{2}, École Polytechnique de Montréal^{3}, AT&T^{4}, Alcatel-Lucent^{5}TL;DR: In this paper, the encoder and decoder of the RNN Encoder-Decoder model are jointly trained to maximize the conditional probability of a target sequence given a source sequence.

Abstract: In this paper, we propose a novel neural network model called RNN Encoder‐ Decoder that consists of two recurrent neural networks (RNN). One RNN encodes a sequence of symbols into a fixedlength vector representation, and the other decodes the representation into another sequence of symbols. The encoder and decoder of the proposed model are jointly trained to maximize the conditional probability of a target sequence given a source sequence. The performance of a statistical machine translation system is empirically found to improve by using the conditional probabilities of phrase pairs computed by the RNN Encoder‐Decoder as an additional feature in the existing log-linear model. Qualitatively, we show that the proposed model learns a semantically and syntactically meaningful representation of linguistic phrases.

19,998 citations

••

Alcatel-Lucent

^{1}TL;DR: There are several arguments which support the observed high accuracy of SVMs, which are reviewed and numerous examples and proofs of most of the key theorems are given.

Abstract: The tutorial starts with an overview of the concepts of VC dimension and structural risk minimization. We then describe linear Support Vector Machines (SVMs) for separable and non-separable data, working through a non-trivial example in detail. We describe a mechanical analogy, and discuss when SVM solutions are unique and when they are global. We describe how support vector training can be practically implemented, and discuss in detail the kernel mapping technique which is used to construct SVM solutions which are nonlinear in the data. We show how Support Vector machines can have very large (even infinite) VC dimension by computing the VC dimension for homogeneous polynomial and Gaussian radial basis function kernels. While very high VC dimension would normally bode ill for generalization performance, and while at present there exists no theory which shows that good generalization performance is guaranteed for SVMs, there are several arguments which support the observed high accuracy of SVMs, which we review. Results of some experiments which were inspired by these arguments are also presented. We give numerous examples and proofs of most of the key theorems. There is new material, and I hope that the reader will find that even old material is cast in a fresh light.

15,696 citations

••

Alcatel-Lucent

^{1}TL;DR: In this paper, the authors investigate the use of multiple transmitting and/or receiving antennas for single user communications over the additive Gaussian channel with and without fading, and derive formulas for the capacities and error exponents of such channels, and describe computational procedures to evaluate such formulas.

Abstract: We investigate the use of multiple transmitting and/or receiving antennas for single user communications over the additive Gaussian channel with and without fading. We derive formulas for the capacities and error exponents of such channels, and describe computational procedures to evaluate such formulas. We show that the potential gains of such multi-antenna systems over single-antenna systems is rather large under independenceassumptions for the fades and noises at different receiving antennas.

12,542 citations

••

TL;DR: An algorithm for non-negative matrix factorization is demonstrated that is able to learn parts of faces and semantic features of text and is in contrast to other methods that learn holistic, not parts-based, representations.

Abstract: Is perception of the whole based on perception of its parts? There is psychological and physiological evidence for parts-based representations in the brain, and certain computational theories of object recognition rely on such representations. But little is known about how brains or computers might learn the parts of objects. Here we demonstrate an algorithm for non-negative matrix factorization that is able to learn parts of faces and semantic features of text. This is in contrast to other methods, such as principal components analysis and vector quantization, that learn holistic, not parts-based, representations. Non-negative matrix factorization is distinguished from the other methods by its use of non-negativity constraints. These constraints lead to a parts-based representation because they allow only additive, not subtractive, combinations. When non-negative matrix factorization is implemented as a neural network, parts-based representations emerge by virtue of two properties: the firing rates of neurons are never negative and synaptic strengths do not change sign.

11,500 citations

##### Authors

Showing all 37011 results

Name | H-index | Papers | Citations |
---|---|---|---|

George M. Whitesides | 240 | 1739 | 269833 |

Yoshua Bengio | 202 | 1033 | 420313 |

John A. Rogers | 177 | 1341 | 127390 |

Zhenan Bao | 169 | 865 | 106571 |

Thomas S. Huang | 146 | 1299 | 101564 |

Federico Capasso | 134 | 1189 | 76957 |

Robert S. Brown | 130 | 1243 | 65822 |

Christos Faloutsos | 127 | 789 | 77746 |

Robert J. Cava | 125 | 1042 | 71819 |

Ramamoorthy Ramesh | 122 | 649 | 67418 |

Yann LeCun | 121 | 369 | 171211 |

Kamil Ugurbil | 120 | 536 | 59053 |

Don Towsley | 119 | 883 | 56671 |

Steven P. DenBaars | 118 | 1366 | 60343 |

Robert E. Tarjan | 114 | 400 | 67305 |