scispace - formally typeset
Search or ask a question
Journal Article•DOI•

Review of Recent Research on Data-Based Process Monitoring

01 Mar 2013-Industrial & Engineering Chemistry Research (American Chemical Society)-Vol. 52, Iss: 10, pp 3543-3562
TL;DR: The natures of different industrial processes are revealed with their data characteristics analyzed and a corresponding problem is defined and illustrated, with review conducted with detailed discussions on connection and comparison of different monitoring methods.
Abstract: Data-based process monitoring has become a key technology in process industries for safety, quality, and operation efficiency enhancement. This paper provides a timely update review on this topic. First, the natures of different industrial processes are revealed with their data characteristics analyzed. Second, detailed terminologies of the data-based process monitoring method are illustrated. Third, based on each of the main data characteristics that exhibits in the process, a corresponding problem is defined and illustrated, with review conducted with detailed discussions on connection and comparison of different monitoring methods. Finally, the relevant research perspectives and several promising issues are highlighted for future work.
Citations
More filters
Journal Article•DOI•
TL;DR: The state-of-the-art of data mining and analytics are reviewed through eight unsupervisedLearning and ten supervised learning algorithms, as well as the application status of semi-supervised learning algorithms.
Abstract: Data mining and analytics have played an important role in knowledge discovery and decision making/supports in the process industry over the past several decades. As a computational engine to data mining and analytics, machine learning serves as basic tools for information extraction, data pattern recognition and predictions. From the perspective of machine learning, this paper provides a review on existing data mining and analytics applications in the process industry over the past several decades. The state-of-the-art of data mining and analytics are reviewed through eight unsupervised learning and ten supervised learning algorithms, as well as the application status of semi-supervised learning algorithms. Several perspectives are highlighted and discussed for future researches on data mining and analytics in the process industry.

657 citations

Journal Article•DOI•
Zhiqiang Ge1•
TL;DR: A systematic review on data-driven modeling and monitoring for plant-wide processes is presented in this paper, where the authors provide an overview of the state-of-the-art data processing and modeling procedures for the plantwide process monitoring.

462 citations

Journal Article•DOI•
TL;DR: An FD technique combining the generalized CCA with the threshold-setting based on the randomized algorithm is proposed and applied to the simulated traction drive control system of high-speed trains and shows that the proposed method is able to improve the detection performance significantly in comparison with the standard generalized C CA-based FD method.
Abstract: In this paper, we first study a generalized canonical correlation analysis (CCA)-based fault detection (FD) method aiming at maximizing the fault detectability under an acceptable false alarm rate. More specifically, two residual signals are generated for detecting of faults in input and output subspaces, respectively. The minimum covariances of the two residual signals are achieved by taking the correlation between input and output into account. Considering the limited application scope of the generalized CCA due to the Gaussian assumption on the process noises, an FD technique combining the generalized CCA with the threshold-setting based on the randomized algorithm is proposed and applied to the simulated traction drive control system of high-speed trains. The achieved results show that the proposed method is able to improve the detection performance significantly in comparison with the standard generalized CCA-based FD method.

252 citations

Journal Article•DOI•
TL;DR: An extensible deep belief network (DBN) based fault diagnosis model is proposed and individual fault features in both spatial and temporal domains are extracted by DBN sub-networks, aided by the mutual information technology.

231 citations

Journal Article•DOI•
TL;DR: A systematic fault detection and isolation scheme is designed so that the whole large-scale process can be hierarchically monitored from the plant-wide level, unit block level, and variable level and the effectiveness of the proposed method is evaluated.
Abstract: In order to deal with the modeling and monitoring issue of large-scale industrial processes with big data, a distributed and parallel designed principal component analysis approach is proposed. To handle the high-dimensional process variables, the large-scale process is first decomposed into distributed blocks with a priori process knowledge. Afterward, in order to solve the modeling issue with large-scale data chunks in each block, a distributed and parallel data processing strategy is proposed based on the framework of MapReduce and then principal components are further extracted for each distributed block. With all these steps, statistical modeling of large-scale processes with big data can be established. Finally, a systematic fault detection and isolation scheme is designed so that the whole large-scale process can be hierarchically monitored from the plant-wide level, unit block level, and variable level. The effectiveness of the proposed method is evaluated through the Tennessee Eastman benchmark process.

221 citations

References
More filters
Journal Article•DOI•
Yoav Freund1, Robert E. Schapire1•
01 Aug 1997
TL;DR: The model studied can be interpreted as a broad, abstract extension of the well-studied on-line prediction model to a general decision-theoretic setting, and it is shown that the multiplicative weight-update Littlestone?Warmuth rule can be adapted to this model, yielding bounds that are slightly weaker in some cases, but applicable to a considerably more general class of learning problems.
Abstract: In the first part of the paper we consider the problem of dynamically apportioning resources among a set of options in a worst-case on-line framework. The model we study can be interpreted as a broad, abstract extension of the well-studied on-line prediction model to a general decision-theoretic setting. We show that the multiplicative weight-update Littlestone?Warmuth rule can be adapted to this model, yielding bounds that are slightly weaker in some cases, but applicable to a considerably more general class of learning problems. We show how the resulting learning algorithm can be applied to a variety of problems, including gambling, multiple-outcome prediction, repeated games, and prediction of points in Rn. In the second part of the paper we apply the multiplicative weight-update technique to derive a new boosting algorithm. This boosting algorithm does not require any prior knowledge about the performance of the weak learning algorithm. We also study generalizations of the new boosting algorithm to the problem of learning functions whose range, rather than being binary, is an arbitrary finite set or a bounded segment of the real line.

15,813 citations

Journal Article•DOI•
TL;DR: 2 general approaches that come highly recommended: maximum likelihood (ML) and Bayesian multiple imputation (MI) are presented and may eventually extend the ML and MI methods that currently represent the state of the art.
Abstract: Statistical procedures for missing data have vastly improved, yet misconception and unsound practice still abound. The authors frame the missing-data problem, review methods, offer advice, and raise issues that remain unresolved. They clear up common misunderstandings regarding the missing at random (MAR) concept. They summarize the evidence against older procedures and, with few exceptions, discourage their use. They present, in both technical and practical language, 2 general approaches that come highly recommended: maximum likelihood (ML) and Bayesian multiple imputation (MI). Newer developments are discussed, including some for dealing with missing data that are not MAR. Although not yet in the mainstream, these procedures may eventually extend the ML and MI methods that currently represent the state of the art.

10,568 citations

Journal Article•DOI•
TL;DR: A survey of contemporary techniques for outlier detection is introduced and their respective motivations are identified and distinguish their advantages and disadvantages in a comparative review.
Abstract: Outlier detection has been used for centuries to detect and, where appropriate, remove anomalous observations from data. Outliers arise due to mechanical faults, changes in system behaviour, fraudulent behaviour, human error, instrument error or simply through natural deviations in populations. Their detection can identify system faults and fraud before they escalate with potentially catastrophic consequences. It can identify errors and remove their contaminating effect on the data set and as such to purify the data for processing. The original outlier detection methods were arbitrary but now, principled and systematic techniques are used, drawn from the full gamut of Computer Science and Statistics. In this paper, we introduce a survey of contemporary techniques for outlier detection. We identify their respective motivations and distinguish their advantages and disadvantages in a comparative review.

3,235 citations

Journal Article•DOI•
TL;DR: This three part series of papers is to provide a systematic and comparative study of various diagnostic methods from different perspectives and broadly classify fault diagnosis methods into three general categories and review them in three parts.

2,263 citations

Journal Article•DOI•
TL;DR: This final part discusses fault diagnosis methods that are based on historic process knowledge that need to be addressed for the successful design and implementation of practical intelligent supervisory control systems for the process industries.

1,902 citations