scispace - formally typeset
Search or ask a question

Showing papers by "Dumitru Erhan published in 2009"


Proceedings Article
15 Apr 2009
TL;DR: The experiments confirm and clarify the advantage of unsupervised pre- training, and empirically show the influence of pre-training with respect to architecture depth, model capacity, and number of training examples.
Abstract: Whereas theoretical work suggests that deep architectures might be more e cient at representing highly-varying functions, training deep architectures was unsuccessful until the recent advent of algorithms based on unsupervised pretraining. Even though these new algorithms have enabled training deep models, many questions remain as to the nature of this di cult learning problem. Answering these questions is important if learning in deep architectures is to be further improved. We attempt to shed some light on these questions through extensive simulations. The experiments confirm and clarify the advantage of unsupervised pre-training. They demonstrate the robustness of the training procedure with respect to the random initialization, the positive e ect of pre-training in terms of optimization and its role as a regularizer. We empirically show the influence of pre-training with respect to architecture depth, model capacity, and number of training examples.

408 citations


Proceedings Article
15 Apr 2009
TL;DR: A simple yet effective method to introduce inhibitory and excitatory interactions between units in the layers of a deep neural network classifier is investigated, and it is presented for the first time that lateral connections can significantly improve the classification performance of deep networks.
Abstract: We investigate a simple yet effective method to introduce inhibitory and excitatory interactions between units in the layers of a deep neural network classifier The method is based on the greedy layer-wise procedure of deep learning algorithms and extends the denoising autoencoder (Vincent et al, 2008) by adding asymmetric lateral connections between its hidden coding units, in a manner that is much simpler and computationally more efficient than previously proposed approaches We present experiments on two character recognition problems which show for the first time that lateral connections can significantly improve the classification performance of deep networks

32 citations