scispace - formally typeset
Search or ask a question
Journal ArticleDOI

A dynamic architecture for artificial neural networks

01 Jan 2005-Neurocomputing (Elsevier)-Vol. 63, pp 397-413
TL;DR: This paper introduces a model that uses a different architecture compared to the traditional neural network, to capture and forecast nonlinear processes, and shows that this approach performs well when compared with traditional models and established research.
Abstract: Artificial neural networks (ANN), have shown to be an effective, general-purpose approach for pattern recognition, classification, clustering, and prediction. Traditional research in this area uses a network with a sequential iterative learning process based on the feed-forward, back-propagation algorithm. In this paper, we introduce a model that uses a different architecture compared to the traditional neural network, to capture and forecast nonlinear processes. This approach utilizes the entire observed data set simultaneously and collectively to estimate the parameters of the model. To assess the effectiveness of this method, we have applied it to a marketing data set and a standard benchmark from ANN literature (Wolf's sunspot activity data set). The results show that this approach performs well when compared with traditional models and established research.
Citations
More filters
Journal ArticleDOI
TL;DR: The empirical results with three well-known real data sets indicate that the proposed model can be an effective way to improve forecasting accuracy achieved by artificial neural networks, and can be used as an appropriate alternative model for forecasting task, especially when higher forecasting accuracy is needed.
Abstract: Artificial neural networks (ANNs) are flexible computing frameworks and universal approximators that can be applied to a wide range of time series forecasting problems with a high degree of accuracy. However, despite all advantages cited for artificial neural networks, their performance for some real time series is not satisfactory. Improving forecasting especially time series forecasting accuracy is an important yet often difficult task facing forecasters. Both theoretical and empirical findings have indicated that integration of different models can be an effective way of improving upon their predictive performance, especially when the models in the ensemble are quite different. In this paper, a novel hybrid model of artificial neural networks is proposed using auto-regressive integrated moving average (ARIMA) models in order to yield a more accurate forecasting model than artificial neural networks. The empirical results with three well-known real data sets indicate that the proposed model can be an effective way to improve forecasting accuracy achieved by artificial neural networks. Therefore, it can be used as an appropriate alternative model for forecasting task, especially when higher forecasting accuracy is needed.

663 citations


Cites methods from "A dynamic architecture for artifici..."

  • ...The data series is regarded as nonlinear and non-Gaussian and is often used to evaluate the effectiveness of nonlinear models (Ghiassi & Saidane, 2005)....

    [...]

Journal ArticleDOI
TL;DR: The effectiveness of neural network models which are known to be dynamic and effective in stock-market predictions are evaluated, including multi-layer perceptron (MLP), dynamic artificial neural network (DAN2) and the hybrid neural networks which use generalized autoregressive conditional heteroscedasticity (GARCH) to extract new input variables.
Abstract: Forecasting stock exchange rates is an important financial problem that is receiving increasing attention. During the last few years, a number of neural network models and hybrid models have been proposed for obtaining accurate prediction results, in an attempt to outperform the traditional linear and nonlinear approaches. This paper evaluates the effectiveness of neural network models which are known to be dynamic and effective in stock-market predictions. The models analysed are multi-layer perceptron (MLP), dynamic artificial neural network (DAN2) and the hybrid neural networks which use generalized autoregressive conditional heteroscedasticity (GARCH) to extract new input variables. The comparison for each model is done in two view points: Mean Square Error (MSE) and Mean Absolute Deviate (MAD) using real exchange daily rate values of NASDAQ Stock Exchange index.

641 citations


Cites background from "A dynamic architecture for artifici..."

  • ...DAN2, is a new NN architecture first developed by Ghiassi and Saidane (2005), clearly outperforms the other methods....

    [...]

  • ...Ghiassi and Saidane (2005) will be analyzed in comparison to classical Multi-Layer Perceptron (MLP) model....

    [...]

Journal ArticleDOI
01 Mar 2011
TL;DR: Empirical results with three well-known real data sets indicate that the proposed model can be an effective way to improve forecasting accuracy achieved by traditional hybrid models and also either of the components models used separately.
Abstract: Improving forecasting especially time series forecasting accuracy is an important yet often difficult task facing decision makers in many areas. Both theoretical and empirical findings have indicated that integration of different models can be an effective way of improving upon their predictive performance, especially when the models in combination are quite different. Artificial neural networks (ANNs) are flexible computing frameworks and universal approximators that can be applied to a wide range of forecasting problems with a high degree of accuracy. However, using ANNs to model linear problems have yielded mixed results, and hence; it is not wise to apply ANNs blindly to any type of data. Autoregressive integrated moving average (ARIMA) models are one of the most popular linear models in time series forecasting, which have been widely applied in order to construct more accurate hybrid models during the past decade. Although, hybrid techniques, which decompose a time series into its linear and nonlinear components, have recently been shown to be successful for single models, these models have some disadvantages. In this paper, a novel hybridization of artificial neural networks and ARIMA model is proposed in order to overcome mentioned limitation of ANNs and yield more general and more accurate forecasting model than traditional hybrid ARIMA-ANNs models. In our proposed model, the unique advantages of ARIMA models in linear modeling are used in order to identify and magnify the existing linear structure in data, and then a neural network is used in order to determine a model to capture the underlying data generating process and predict, using preprocessed data. Empirical results with three well-known real data sets indicate that the proposed model can be an effective way to improve forecasting accuracy achieved by traditional hybrid models and also either of the components models used separately.

631 citations

Journal ArticleDOI
TL;DR: This research introduces an approach to supervised feature reduction using n-grams and statistical analysis to develop a Twitter-specific lexicon for sentiment analysis, and develops sentiment classification models using this reduced lexicon and the DAN2 machine learning approach, which has demonstrated success in other text classification problems.
Abstract: Twitter messages are increasingly used to determine consumer sentiment towards a brand. The existing literature on Twitter sentiment analysis uses various feature sets and methods, many of which are adapted from more traditional text classification problems. In this research, we introduce an approach to supervised feature reduction using n-grams and statistical analysis to develop a Twitter-specific lexicon for sentiment analysis. We augment this reduced Twitter-specific lexicon with brand-specific terms for brand-related tweets. We show that the reduced lexicon set, while significantly smaller (only 187 features), reduces modeling complexity, maintains a high degree of coverage over our Twitter corpus, and yields improved sentiment classification accuracy. To demonstrate the effectiveness of the devised Twitter-specific lexicon compared to a traditional sentiment lexicon, we develop comparable sentiment classification models using SVM. We show that the Twitter-specific lexicon is significantly more effective in terms of classification recall and accuracy metrics. We then develop sentiment classification models using the Twitter-specific lexicon and the DAN2 machine learning approach, which has demonstrated success in other text classification problems. We show that DAN2 produces more accurate sentiment classification results than SVM while using the same Twitter-specific lexicon.

434 citations


Cites background or methods from "A dynamic architecture for artifici..."

  • ...In Ghiassi and Saidane (Ghiassi & Saidane, 2005), they present several nonlinear optimization strategies to estimate the nonlinear parameter lk....

    [...]

  • ...In Ghiassi and Saidane (2005), the authors offer a number of alternatives to solve the intermediate nonlinear optimization problem....

    [...]

  • ...In Ghiassi and Saidane (Ghiassi & Saidane, 2005), the authors show that this normalization can be represented by the trigonometric function Cosine (lkai + hk)....

    [...]

  • ...A detailed description of the architecture and its properties are fully presented in (Ghiassi & Saidane, 2005)....

    [...]

  • ...The transformation process defines a reference vector R = {rj; j = 1, 2, . . . m}, where m represents the number of attributes of the observation records, and projects each observation record onto this vector to normalize the data as discussed in (Ghiassi & Saidane, 2005)....

    [...]

Journal ArticleDOI
TL;DR: A dynamic neural network model for forecasting time series events that uses a different architecture than traditional models is presented and shows that this approach is more accurate and performs significantly better than the traditional neural network and autoregressive integrated moving average (ARIMA) models.
Abstract: Neural networks have shown to be an effective method for forecasting time series events. Traditional research in this area uses a network with a sequential iterative learning process based on the feed-forward, back-propagation approach. In this paper we present a dynamic neural network model for forecasting time series events that uses a different architecture than traditional models. To assess the effectiveness of this method, we forecasted a number of standard benchmarks in time series research from forecasting literature. Results show that this approach is more accurate and performs significantly better than the traditional neural network and autoregressive integrated moving average (ARIMA) models.

251 citations

References
More filters
Journal ArticleDOI
01 Jan 1988-Nature
TL;DR: Back-propagation repeatedly adjusts the weights of the connections in the network so as to minimize a measure of the difference between the actual output vector of the net and the desired output vector, which helps to represent important features of the task domain.
Abstract: We describe a new learning procedure, back-propagation, for networks of neurone-like units. The procedure repeatedly adjusts the weights of the connections in the network so as to minimize a measure of the difference between the actual output vector of the net and the desired output vector. As a result of the weight adjustments, internal ‘hidden’ units which are not part of the input or output come to represent important features of the task domain, and the regularities in the task are captured by the interactions of these units. The ability to create useful new features distinguishes back-propagation from earlier, simpler methods such as the perceptron-convergence procedure1.

23,814 citations

Journal ArticleDOI
TL;DR: A model of a system having a large number of simple equivalent components, based on aspects of neurobiology but readily adapted to integrated circuits, produces a content-addressable memory which correctly yields an entire memory from any subpart of sufficient size.
Abstract: Computational properties of use of biological organisms or to the construction of computers can emerge as collective properties of systems having a large number of simple equivalent components (or neurons). The physical meaning of content-addressable memory is described by an appropriate phase space flow of the state of a system. A model of such a system is given, based on aspects of neurobiology but readily adapted to integrated circuits. The collective properties of this model produce a content-addressable memory which correctly yields an entire memory from any subpart of sufficient size. The algorithm for the time evolution of the state of the system is based on asynchronous parallel processing. Additional emergent collective properties include some capacity for generalization, familiarity recognition, categorization, error correction, and time sequence retention. The collective properties are only weakly sensitive to details of the modeling or the failure of individual devices.

16,652 citations

Journal ArticleDOI
TL;DR: In this paper, the authors describe a self-organizing system in which the signal representations are automatically mapped onto a set of output responses in such a way that the responses acquire the same topological order as that of the primary events.
Abstract: This work contains a theoretical study and computer simulations of a new self-organizing process. The principal discovery is that in a simple network of adaptive physical elements which receives signals from a primary event space, the signal representations are automatically mapped onto a set of output responses in such a way that the responses acquire the same topological order as that of the primary events. In other words, a principle has been discovered which facilitates the automatic formation of topologically correct maps of features of observable events. The basic self-organizing system is a one- or two-dimensional array of processing units resembling a network of threshold-logic units, and characterized by short-range lateral feedback between neighbouring units. Several types of computer simulations are used to demonstrate the ordering process as well as the conditions under which it fails.

8,247 citations

Journal ArticleDOI
TL;DR: It is shown that standard multilayer feedforward networks with as few as a single hidden layer and arbitrary bounded and nonconstant activation function are universal approximators with respect to L p (μ) performance criteria, for arbitrary finite input environment measures μ.
Abstract: We show that standard multilayer feedforward networks with as few as a single hidden layer and arbitrary bounded and nonconstant activation function are universal approximators with respect to L p (μ) performance criteria, for arbitrary finite input environment measures μ, provided only that sufficiently many hidden units are available. If the activation function is continuous, bounded and nonconstant, then continuous mappings can be learned uniformly over compact input sets. We also give very general conditions ensuring that networks with sufficiently smooth activation functions are capable of arbitrarily accurate approximation to a function and its derivatives.

5,593 citations

Journal ArticleDOI
TL;DR: In this paper, the authors present a state-of-the-art survey of ANN applications in forecasting and provide a synthesis of published research in this area, insights on ANN modeling issues, and future research directions.
Abstract: Interest in using artificial neural networks (ANNs) for forecasting has led to a tremendous surge in research activities in the past decade. While ANNs provide a great deal of promise, they also embody much uncertainty. Researchers to date are still not certain about the effect of key factors on forecasting performance of ANNs. This paper presents a state-of-the-art survey of ANN applications in forecasting. Our purpose is to provide (1) a synthesis of published research in this area, (2) insights on ANN modeling issues, and (3) the future research directions.

3,680 citations

Trending Questions (1)
How do you train artificial neural networks?

Artificial neural networks (ANN), have shown to be an effective, general-purpose approach for pattern recognition, classification, clustering, and prediction.