Open AccessProceedings Article
The Cascade-Correlation Learning Architecture
Scott E. Fahlman,Christian Lebiere +1 more
- Vol. 2, pp 524-532
Reads0
Chats0
TLDR
The Cascade-Correlation architecture has several advantages over existing algorithms: it learns very quickly, the network determines its own size and topology, it retains the structures it has built even if the training set changes, and it requires no back-propagation of error signals through the connections of the network.Abstract:
Cascade-Correlation is a new architecture and supervised learning algorithm for artificial neural networks. Instead of just adjusting the weights in a network of fixed topology. Cascade-Correlation begins with a minimal network, then automatically trains and adds new hidden units one by one, creating a multi-layer structure. Once a new hidden unit has been added to the network, its input-side weights are frozen. This unit then becomes a permanent feature-detector in the network, available for producing outputs or for creating other, more complex feature detectors. The Cascade-Correlation architecture has several advantages over existing algorithms: it learns very quickly, the network determines its own size and topology, it retains the structures it has built even if the training set changes, and it requires no back-propagation of error signals through the connections of the network.read more
Citations
More filters
How Not to Be Frustrated with Neural Networks
TL;DR: The purpose of this article is to evaluate the reasons for these frustrations and show how to make these neural networks successful.
Book Chapter
Connectionist modelling of lexical segmentation and vocabulary acquisition
TL;DR: This chapter describes some computational simulations proposing ways in which cues and strategies for the acquisition of lexical segmentation can be integrated with the infants acquisition of the meanings of words.
Journal ArticleDOI
Extrapolation of Mackey-Glass data using Cascade Correlation
D. David Ensley,Dale E. Nelson +1 more
TL;DR: Findings of a ten- week research period on using the Cascade Correlation ontogenic neural network to extrapolate (predict) a chaotic time series generated from the Mackey-Glass equation are presented.
Investigation into the robustness of artificial neural networks for a case study in civil engineering
TL;DR: The results indicate that good performance of ANN models on the data used for model calibration and validation does not guarantee that the models will perform in a robust fashion over a range of data similar to that used in the model calibration phase.
References
More filters
Book ChapterDOI
Learning internal representations by error propagation
TL;DR: This chapter contains sections titled: The Problem, The Generalized Delta Rule, Simulation Results, Some Further Generalizations, Conclusion.
MonographDOI
Parallel Distributed Processing: Explorations in the Microstructure of Cognition: Foundations
Book
Learning internal representations by error propagation
TL;DR: In this paper, the problem of the generalized delta rule is discussed and the Generalized Delta Rule is applied to the simulation results of simulation results in terms of the generalized delta rule.
Journal ArticleDOI
Increased Rates of Convergence Through Learning Rate Adaptation
TL;DR: A study of Steepest Descent and an analysis of why it can be slow to converge and four heuristics for achieving faster rates of convergence are proposed.