scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Artificial intelligence for fault diagnosis of rotating machinery: A review

01 Aug 2018-Mechanical Systems and Signal Processing (Academic Press)-Vol. 108, pp 33-47
TL;DR: This paper attempts to present a comprehensive review of AI algorithms in rotating machinery fault diagnosis, from both the views of theory background and industrial applications.
About: This article is published in Mechanical Systems and Signal Processing.The article was published on 2018-08-01. It has received 1287 citations till now. The article focuses on the topics: Literature survey.
Citations
More filters
Journal ArticleDOI
TL;DR: A review and roadmap to systematically cover the development of IFD following the progress of machine learning theories and offer a future perspective is presented.

1,173 citations

Journal ArticleDOI
TL;DR: New deep learning methods, namely, deep residual shrinkage networks, are developed to improve the feature learning ability from highly noised vibration signals and achieve a high fault diagnosing accuracy.
Abstract: This article develops new deep learning methods, namely, deep residual shrinkage networks, to improve the feature learning ability from highly noised vibration signals and achieve a high fault diagnosing accuracy. Soft thresholding is inserted as nonlinear transformation layers into the deep architectures to eliminate unimportant features. Moreover, considering that it is generally challenging to set proper values for the thresholds, the developed deep residual shrinkage networks integrate a few specialized neural networks as trainable modules to automatically determine the thresholds, so that professional expertise on signal processing is not required. The efficacy of the developed methods is validated through experiments with various types of noise.

520 citations

Journal ArticleDOI
TL;DR: This paper aims to fulfill the gap by presenting the highlights of the traditional methods and provide a comprehensive review of the most recent applications of ML and DL algorithms utilized for vibration-based structural damage detection in civil structures.

440 citations


Cites background from "Artificial intelligence for fault d..."

  • ...The high efficiency of vibration-based SDD systems via SHM of rotating machines [35] is based on the fact that the vibration response of such machines is minimally affected by the operational and environmental conditions [36–38]....

    [...]

Journal ArticleDOI
TL;DR: Wang et al. as mentioned in this paper proposed a new intelligent fault diagnosis framework, i.e., deep transfer network (DTN), which generalizes deep learning model to domain adaptation scenario, by extending the marginal distribution adaptation to joint distribution adaptation (JDA).
Abstract: In recent years, an increasing popularity of deep learning model for intelligent condition monitoring and diagnosis as well as prognostics used for mechanical systems and structures has been observed. In the previous studies, however, a major assumption accepted by default, is that the training and testing data are taking from same feature distribution. Unfortunately, this assumption is mostly invalid in real application, resulting in a certain lack of applicability for the traditional diagnosis approaches. Inspired by the idea of transfer learning that leverages the knowledge learnt from rich labeled data in source domain to facilitate diagnosing a new but similar target task, a new intelligent fault diagnosis framework, i.e., deep transfer network (DTN), which generalizes deep learning model to domain adaptation scenario, is proposed in this paper. By extending the marginal distribution adaptation (MDA) to joint distribution adaptation (JDA), the proposed framework can exploit the discrimination structures associated with the labeled data in source domain to adapt the conditional distribution of unlabeled target data, and thus guarantee a more accurate distribution matching. Extensive empirical evaluations on three fault datasets validate the applicability and practicability of DTN, while achieving many state-of-the-art transfer results in terms of diverse operating conditions, fault severities and fault types.

321 citations

Journal ArticleDOI
TL;DR: A new TCNN with the depth of 51 convolutional layers is proposed for fault diagnosis of ResNet-50 and achieves state-of-the-art results, which demonstrates that TCNN(ResNet- 50) outperforms other DL models and traditional methods.
Abstract: With the rapid development of smart manufacturing, data-driven fault diagnosis has attracted increasing attentions. As one of the most popular methods applied in fault diagnosis, deep learning (DL) has achieved remarkable results. However, due to the fact that the volume of labeled samples is small in fault diagnosis, the depths of DL models for fault diagnosis are shallow compared with convolutional neural network in other areas (including ImageNet), which limits their final prediction accuracies. In this research, a new TCNN(ResNet-50) with the depth of 51 convolutional layers is proposed for fault diagnosis. By combining with transfer learning, TCNN(ResNet-50) applies ResNet-50 trained on ImageNet as feature extractor for fault diagnosis. Firstly, a signal-to-image method is developed to convert time-domain fault signals to RGB images format as the input datatype of ResNet-50. Then, a new structure of TCNN(ResNet-50) is proposed. Finally, the proposed TCNN(ResNet-50) has been tested on three datasets, including bearing damage dataset provided by KAT datacenter, motor bearing dataset provided by Case Western Reserve University (CWRU) and self-priming centrifugal pump dataset. It achieved state-of-the-art results. The prediction accuracies of TCNN(ResNet-50) are as high as 98.95% ± 0.0074, 99.99% ± 0 and 99.20% ± 0, which demonstrates that TCNN(ResNet-50) outperforms other DL models and traditional methods.

319 citations


Cites background or methods from "Artificial intelligence for fault d..."

  • ...In generally, these machine learning methods are combined with the feature extraction processes [11]....

    [...]

  • ...With the rapid development of smart manufacturing, datadriven fault diagnosis has become a hot research topic [11, 18]....

    [...]

References
More filters
Journal ArticleDOI
28 May 2015-Nature
TL;DR: Deep learning is making major advances in solving problems that have resisted the best attempts of the artificial intelligence community for many years, and will have many more successes in the near future because it requires very little engineering by hand and can easily take advantage of increases in the amount of available computation and data.
Abstract: Deep learning allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. These methods have dramatically improved the state-of-the-art in speech recognition, visual object recognition, object detection and many other domains such as drug discovery and genomics. Deep learning discovers intricate structure in large data sets by using the backpropagation algorithm to indicate how a machine should change its internal parameters that are used to compute the representation in each layer from the representation in the previous layer. Deep convolutional nets have brought about breakthroughs in processing images, video, speech and audio, whereas recurrent nets have shone light on sequential data such as text and speech.

46,982 citations

Journal ArticleDOI
01 Jan 1998
TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Abstract: Multilayer neural networks trained with the back-propagation algorithm constitute the best example of a successful gradient based learning technique. Given an appropriate network architecture, gradient-based learning algorithms can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters, with minimal preprocessing. This paper reviews various methods applied to handwritten character recognition and compares them on a standard handwritten digit recognition task. Convolutional neural networks, which are specifically designed to deal with the variability of 2D shapes, are shown to outperform all other techniques. Real-life document recognition systems are composed of multiple modules including field extraction, segmentation recognition, and language modeling. A new learning paradigm, called graph transformer networks (GTN), allows such multimodule systems to be trained globally using gradient-based methods so as to minimize an overall performance measure. Two systems for online handwriting recognition are described. Experiments demonstrate the advantage of global training, and the flexibility of graph transformer networks. A graph transformer network for reading a bank cheque is also described. It uses convolutional neural network character recognizers combined with global training techniques to provide record accuracy on business and personal cheques. It is deployed commercially and reads several million cheques per day.

42,067 citations

Book
Vladimir Vapnik1
01 Jan 1995
TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?
Abstract: Setting of the learning problem consistency of learning processes bounds on the rate of convergence of learning processes controlling the generalization ability of learning processes constructing learning algorithms what is important in learning theory?.

40,147 citations

Book
18 Nov 2016
TL;DR: Deep learning as mentioned in this paper is a form of machine learning that enables computers to learn from experience and understand the world in terms of a hierarchy of concepts, and it is used in many applications such as natural language processing, speech recognition, computer vision, online recommendation systems, bioinformatics, and videogames.
Abstract: Deep learning is a form of machine learning that enables computers to learn from experience and understand the world in terms of a hierarchy of concepts. Because the computer gathers knowledge from experience, there is no need for a human computer operator to formally specify all the knowledge that the computer needs. The hierarchy of concepts allows the computer to learn complicated concepts by building them out of simpler ones; a graph of these hierarchies would be many layers deep. This book introduces a broad range of topics in deep learning. The text offers mathematical and conceptual background, covering relevant concepts in linear algebra, probability theory and information theory, numerical computation, and machine learning. It describes deep learning techniques used by practitioners in industry, including deep feedforward networks, regularization, optimization algorithms, convolutional networks, sequence modeling, and practical methodology; and it surveys such applications as natural language processing, speech recognition, computer vision, online recommendation systems, bioinformatics, and videogames. Finally, the book offers research perspectives, covering such theoretical topics as linear factor models, autoencoders, representation learning, structured probabilistic models, Monte Carlo methods, the partition function, approximate inference, and deep generative models. Deep Learning can be used by undergraduate or graduate students planning careers in either industry or research, and by software engineers who want to begin using deep learning in their products or platforms. A website offers supplementary material for both readers and instructors.

38,208 citations

Journal ArticleDOI
28 Jul 2006-Science
TL;DR: In this article, an effective way of initializing the weights that allows deep autoencoder networks to learn low-dimensional codes that work much better than principal components analysis as a tool to reduce the dimensionality of data is described.
Abstract: High-dimensional data can be converted to low-dimensional codes by training a multilayer neural network with a small central layer to reconstruct high-dimensional input vectors. Gradient descent can be used for fine-tuning the weights in such "autoencoder" networks, but this works well only if the initial weights are close to a good solution. We describe an effective way of initializing the weights that allows deep autoencoder networks to learn low-dimensional codes that work much better than principal components analysis as a tool to reduce the dimensionality of data.

16,717 citations