scispace - formally typeset
Journal ArticleDOI

KDnet-RUL: A Knowledge Distillation Framework to Compress Deep Neural Networks for Machine Remaining Useful Life Prediction

TLDR
A knowledge distillation framework, entitled KDnet-RUL, to compress a complex LSTM-based method for RUL prediction and demonstrates that the proposed method significantly outperforms state-of-the-art KD methods.
Abstract
Machine remaining useful life (RUL) prediction is vital in improving the reliability of industrial systems and reducing maintenance cost Recently, long short-term memory (LSTM) based algorithms have achieved state-of-the-art performance for RUL prediction due to their strong capability of modeling sequential sensory data In many cases, the RUL prediction algorithms are required to be deployed on edge devices to support real-time decision making, reduce the data communication cost, and preserve the data privacy However, the powerful LSTM-based methods which have high complexity cannot be deployed to edge devices with limited computational power and memory To solve this problem, we propose a knowledge distillation framework, entitled KDnet-RUL, to compress a complex LSTM-based method for RUL prediction Specifically, it includes a generative adversarial network based knowledge distillation (GAN-KD) for disparate architecture knowledge transfer, a learning-during-teaching based knowledge distillation (LDT-KD) for identical architecture knowledge transfer, and a sequential distillation upon LDT-KD for complicated datasets We leverage simple and complicated datasets to verify the effectiveness of the proposed KDnet-RUL The results demonstrate that the proposed method significantly outperforms state-of-the-art KD methods The compressed model with 128 times less weights and 462 times less total float point operations even achieves a comparable performance with the complex LSTM model for RUL prediction

read more

Citations
More filters
Journal ArticleDOI

Prediction of remaining useful life of multi-stage aero-engine based on clustering and LSTM fusion

TL;DR: In this article, an improved multi-stage Long Short Term Memory network with Clustering (ILSTMC) is proposed to predict the remaining useful life of an aero-engine.
Journal ArticleDOI

Bi-LSTM-Based Two-Stream Network for Machine Remaining Useful Life Prediction

TL;DR: A series of new handcrafted feature flows (HFFs) are proposed, which can suppress the raw signal noise and thus improve the encoded sequential information for the RUL prediction, and a novel bidirectional LSTM (Bi-LSTM)-based two-stream network is proposed.
Journal ArticleDOI

Position Encoding Based Convolutional Neural Networks for Machine Remaining Useful Life Prediction

TL;DR: Zhang et al. as discussed by the authors proposed a position encoding based CNN to enhance the sequential information encoded by a CNN, which showed competitive results to RNN-based methods and achieved state-of-the-art performance.
Journal ArticleDOI

An industrial-grade solution for agricultural image classification tasks

TL;DR: This work uses a fine-tuning strategy to train models to better performance in classification tasks and utilizes neural network pruning techniques to reduce neural network size and computational cost, and retrain models by knowledge distillation to minimize pruned model performance loss.
Journal ArticleDOI

Remaining Useful Life Prediction for AC Contactor Based on MMPE and LSTM With Dual Attention Mechanism

TL;DR: In this paper , a RUL prediction method based on modified multiscale permutation entropy (MMPE) and long short-term memory (LSTM) with dual attention (DA) mechanism is proposed.
References
More filters
Posted Content

Distilling the Knowledge in a Neural Network

TL;DR: This work shows that it can significantly improve the acoustic model of a heavily used commercial system by distilling the knowledge in an ensemble of models into a single model and introduces a new type of ensemble composed of one or more full models and many specialist models which learn to distinguish fine-grained classes that the full models confuse.
Proceedings ArticleDOI

Convolutional Neural Networks for Sentence Classification

TL;DR: The CNN models discussed herein improve upon the state of the art on 4 out of 7 tasks, which include sentiment analysis and question classification, and are proposed to allow for the use of both task-specific and static vectors.
Posted Content

Convolutional Neural Networks for Sentence Classification

TL;DR: In this article, CNNs are trained on top of pre-trained word vectors for sentence-level classification tasks and a simple CNN with little hyperparameter tuning and static vectors achieves excellent results on multiple benchmarks.
Proceedings Article

Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding

TL;DR: Deep Compression as mentioned in this paper proposes a three-stage pipeline: pruning, quantization, and Huffman coding to reduce the storage requirement of neural networks by 35x to 49x without affecting their accuracy.
Journal ArticleDOI

Face recognition: a convolutional neural-network approach

TL;DR: A hybrid neural-network for human face recognition which compares favourably with other methods and analyzes the computational complexity and discusses how new classes could be added to the trained recognizer.
Related Papers (5)