Professional Documents
Culture Documents
net/publication/236891710
CITATIONS READS
8 965
3 authors, including:
Some of the authors of this publication are also working on these related projects:
Relative entropy based Bayesian design for accelerated degradation testing View project
All content following this page was uploaded by Linxia Liao on 23 May 2014.
The University of Cincinnati, NSF I/UCR Center for Intelligent Maintenance Systems,
Department of Mechanical Engineering, 560 Rhodes Hall,
Cincinnati OH 45221. Phone: 513-556-3412
Email: liaol@email.uc.edu haixia.wang@uc.edu jay.lee@uc.edu
Dr. Haixia Wang is a research associate in the NSF Industry/University Cooperative Research
Center (I/UCRC) for Intelligent Maintenance Systems (IMS) at the campus of Univ. of
Cincinnati. Her current research interest focuses on machinery and process prognostics and
health management, manufacturing process performance and quality improvement, and closed-
loop product life cycle service model studies. Haixia Wang received her B.S degree in
Mechanical Engineering from Shandong University at China, a Ph.D in Mechanical Engineering
from Southeast University at China, a M.S. and a Ph.D in Industrial and Systems Engineering
from the Univ. of Wisconsin-Madison at USA.
Dr. Jay Lee is Ohio Eminent Scholar and L.W. Scott Alter Chair Professor at the Univ. of
Cincinnati and is founding director of National Science Foundation (NSF) Industry/University
Cooperative Research Center (I/UCRC) on Intelligent Maintenance Systems (IMS
www.imscenter.net ) which is a multi-campus NSF Center of Excellence between the Univ. of
Cincinnati (lead institution), the Univ. of Michigan, and the Univ. of Missouri-Rolla in
partnerships with over 45 global companies including P&G, Toyota, GE Aviation, Boeing, AMD,
Caterpillar, Siemens, DaimlerChrysler, ETAS, Festo, Harley-Davidson, Honeywell, ITRI
(Taiwan), Komatsu, Omron, Samsung, Toshiba, Bosch, Parker Hannifin, BorgWarner, Spirit
Aerosystems, Nissan (Japan), Syncrude (Canada), and McKinsey & Company, etc. His current
research focuses on autonomic computing, embedded IT and smart prognostics technologies for
industrial and healthcare systems, design of self-maintenance and self-healing, systems, and
dominant design tools for product and service innovation. He also serves as honorary professor
and visiting professor for a number of institutions including, Cranfield Univ. in UK, Lulea Univ.
of Technology in Sweden, Shanghai Jiao Tong Univ. in China, Univ. of Manchester, City Univ.
of Hong Kong, and Hong Kong PolyU., and Harbin Institute of Technology (HIT) in China, etc.
He is also a Fellow of ASME, SME, as well as International Society of Engineering Asset
Management (ISEAM).
ABSTRACT
®
This paper presents a scalable Watchdog Agent -based toolbox approach for machine health prognostics. The toolbox
consists of modularized embedded algorithms for signal processing and feature extraction, performance assessment,
diagnostics and prognostics, which can be reconfigured for different machinery prognostic applications, and can be
extensible and adaptable to most real-world machine situations. A decision making technique, Quality Function
Deployment (QFD)-based tool selection method, is applied for the automatic selection of algorithms from the Watchdog
Agent® toolbox using multiple criteria. In addition, the architecture for the Watchdog Agent®-based real-time remote
machinery prognostics and health management, which incorporates remote and embedded predictive maintenance
technologies, is presented. An industrial case involving the automatic tool changer of a machine tool is presented to
illustrate how the Watchdog Agent® toolbox can be used in diverse scenarios.
Keywords: Watchdog Agent® toolbox, Machine health assessment, Prognostics
1. INTRODUCTION
The Watchdog Agent® toolbox enables one to Degradation assessment is conducted after the critical
quantitatively assess and predict performance properties of a process or machine are identified and
degradation levels of key product components, and to measured by sensors. It is expected that the
determine the root causes of failure [15-17], thus degradation process will alter the sensor readings that
making it possible to physically realize closed-loop are being fed into the Watchdog Agent®, and thus
product life cycle monitoring and management. enable it to assess and quantify the degradation by
quantitatively describing the corresponding change in The toolbox consists of a collection of state-of-
sensor signatures. In addition, a model of the process the-art machine learning algorithms and data
or piece of equipment that is being considered, or preprocessing tools, some of which have been used in
available application specific knowledge, can be used past projects, for predicting the degradation or
to aid the degradation process description, provided performance loss of devices, process, and systems. The
that such a model and/or such knowledge exist. The current tools employed in the signal processing and
prognostic function is realized through trending and feature extraction, performance assessment, diagnostics
statistical modeling of the observed process and prognostics modules of Watchdog Agent®
performance signatures and/or model parameters. functionality are summarized in Figure 3.
Signal processing and feature extraction storage capabilities, limited space, power consumption,
decompose multi-sensory data into a feature space, user preferences, etc., the performance assessment
which is related to the health assessment or diagnosis module of the Watchdog Agent® has been realized in
tasks. Health assessment is the evaluation of the the form of a modular, open architecture toolbox. The
overlap between the most recent feature space and that toolbox consists of different prognostics tools,
during normal product operation. This overlap is including neural network-based, time-series based,
continuously transformed into a Confidence Value (CV) wavelet-based and hybrid joint time-frequency
ranging from 0 to 1 (0-unacceptable and 1-normal) methods, etc., for predicting the degradation or
over time for evaluating the deviation of the most performance loss of devices, processes, products and
recent behavior from normal behavior or baseline. systems. The open architecture of the toolbox allows
Health diagnosis is the analysis of the patterns one to easily add new solutions to the performance
embedded in the data to determine what previous assessment modules, as well as to easily interchange
observed fault has occurred. Performance prediction different tools, depending on the needs of the
involves the extrapolation of the behavior of the application.
process signatures over time, and the prediction of their To enable rapid deployment, a Quality Function
behavior in the future. Deployment (QFD) based selection method had been
In order to facilitate the use of Watchdog Agent® developed to provide a general suggestion to aid in tool
in a wide variety of applications, with various selection; this is especially critical for those industry
requirements and limitations regarding the character of users who have little knowledge about the available
signals, available processing power, memory and algorithms. QFD was developed by Yoji Akao in the
1960s and provides a structured framework for as shown in Figure 4. In the problem definition and
concurrent engineering that ensures that the “voice of constraints step, all the system requirements and
the customer” is incorporated into all phases of product constraints, health monitoring and prognostics
development [18]. In the context of automatic tool requirements, trade studies, such as cost/benefit
selection, the QFD method is used to calculate the analyses, to determine the monitoring and prognostics
weight for each tool under the constraints of user- features are completed. To tackle a specific prognostics
defined condition priorities. The Analytical Hierarchy problem, appropriate tools are selected by the tool
Process (AHP) method is utilized to enhance the QFD selection method mentioned above, or new tools are
implementation by using a quantitative method. The developed for special situations. In the tool selection
AHP method is used to make an effective decision step, the parameters of the selected, or newly
based on pair-wise comparison of each alternative developed, tools need to be specified by training the
against various application conditions. More detail of models with historical data. The models will then be
this method is described in [19]. The AHP-enhanced programmed and tested by an evaluation procedure. If
QFD selection method gives a suggestion for the most satisfactory results are achieved, an in-house prototype
appropriate tool under different application conditions. of the system will be built and tested. The finalized
New tools will be continuously developed and system will be deployed on a commercially available
added to the modular, open architecture Watchdog hardware platform for industrial applications.
Agent® toolbox based on the development procedure,
Prototyping &
Program
testing
development
No No
Accepted Evaluation
Yes
Yes
Deployment
3.2.3 Feature Selection by Fisher’s Linear Figure 8 PCA Projection Results for Training Dataset
Discrimination Criterion and PCA
The features marked with stars in Figure 7 are those 3.3 Health Assessment
features that have the 5 largest Fish linear discriminate
values [23], which means those 5 features contain the 3.3.1 Introduction to the Technical Background
most of the separable information for the normal and
faulty datasets. Therefore, those 5 features are selected. Logistic Regression (LR)
The problem remains that it is difficult to decide how Logistic regression is a function that can easily
many features need to be chosen for the next step: represent the daily maintenance records as a
health assessment. dichotomous/binary problem. The goal of logistic
The Principle Component Analysis (PCA) method regression is to find the best fitting model to describe
is used to projects the original feature space to a lower the relationship between the categorical characteristics
dimensional feature space which contains 95% of the of a dependent variable and a set of independent
variation information of the original feature space. In variables [25]. The logistic function is
this application, the Karhunen-Loève Transform (KLT) 1 eg x
projects the N-dimension (N=16) feature space to an Px
M-dimension (M=2) feature space. The training data 1 e g x 1 eg x (8)
includes 40 normal datasets and 35 faulty datasets, and in which the logistic or logit model is
each of them is first decomposed by WPT. Since each P x
dataset has 16 energy components after the g x lo g
1 P x
decomposition, the input matrix X is 75×16 in size. (9)
Based on the dispersion and the magnitude of the 1 x1 2x 2 ... k xk
eigenvalues, M (M=2) of the largest eigenvalues are , in which g(x) is a linear combination of the
preserved and the corresponding eigenvectors are independent variables x1, x2, … , xk. and , are the
calculated. After the projection of the original feature parameters for the independent variables. The
space onto the two dimensions along with the selected maximum likelihood estimate method [8] is used to
two eigenvectors, the new feature space Y, at a size of find the parameter values that maximize the probability
75×2, is obtained. The projection result is shown in of the observed data, by iterative steps.
Figure 8. The circles indicate the projected normal
feature space and the crosses represent the projected Self-Organizing Maps (SOM)
faulty feature space. A comparison of the effect of SOM provides a way of representing
those two feature selection methods to the results of multidimensional feature space in a one or two-
health assessment is given in chapter 3.3. dimensional space, while preserving the topological
For the timing window signals (shown in Figure 6), properties of the input feature space. SOM is an
status 5 and status 13 (listed in Table 1), which unsupervised learning neural network which can
represent the time span of the tool topple down action organize itself according to the nature of the input data.
and the duration of the arm rotation action, are selected The input data vectors, which closely resemble each
as two independent features of the input to the logistic other, are located next to each other on the map after
regression algorithm. These two features are selected training. An n-dimensional input data space can be
because they are the most critical features to indicate denoted by
x [ x1 , x2 ,..., xn ]T the testing data into either the normal or faulty
(10) category. Logistic regression is a robust tool that can
The weight vector of each neuron j in the network easily represent the daily maintenance records as a
has the same dimension as the input space, and can be dichotomous problem; therefore it has been chosen as
represented by: the health assessment model.
wj [ w j1 , w j 2 ,..., w jn ]T , j 1, 2,..., m The training dataset mentioned above is used to
(11) train the logistic regression model. Feature space is
in which m is the number of neurons in the network. extracted from the raw data via WPA and the Fisher’s
The Best Machining Unit (BMU) in the SOM is the linear discrimination criterion is first used as the
neuron whose weight vector is the closest to the input feature selection method. After the training, the health
j xT w assessment model is achieved:
vector in the input space. The inner product can g ( x) 7.2932 0.7298 x1
be used as an analytical measure for the matching of x
2.0944 x2 1.9451x3 1.2929 x4 (15)
with wj. The Euclidean distance is a better and more
convenient measure criterion for the matching of x 1.7345 x5 2.1965 x6 3.9255 x7
with wj. The minimum distance defines the BMU, if 281 datasets are then used for validation. Figure
we define wc as the weight vector of the neuron that 9 shows the confidence value curve of the ATC. The
best matches the input vector x. The measurement can drop of the confidence value at cycle 250 was caused
then be represented by by a drop in the air pressure at that cycle.
|| x wc || min{|| x wj ||}, j 1,2,..., m PCA is also used as the feature extraction method
(12)
After the BMU is identified in the iterative and the same dataset is used to train the logistic
training process, the weight vectors and the topological regression model. After the training, the health
neighbors of the BMU are updated in order to move assessment model is achieved:
them closer to the input vector in the input space. The g ( x) 7.3238 0.5764 x1
(16)
following learning rule is applied: 0.1012 x2 0.2591x3 1.9164 x4
wj (t 1) wj (t) (t)hj,wc (t)(x wj (t)) The same 281 datasets are then used for
(13)
validation. Figure 10 shows the confidence value curve
h j , wc of the ATC. It also shows a drop in the confidence
in which denotes the topological neighborhood
kernel centered on the BMU wc. A typical choice for value at cycle 250, which was caused by a drop in the
the kernel function is the Gaussian function air pressure at that cycle.
h j , wc exp( d j , wc 2 / 2 2
)
(14)
d j , wc
in which is the lateral distance between the BMU
wc and neuron j. The parameter is the “effective
width” of the topological neighborhood [26]. (t ) is
the learning rate which monotonically decreases with
the training time. In the initial phase which lasts for a
given number of steps (e.g. first 1000 steps),
(t ) starts with a value that is close to 1, and it can be
linear, exponential, or inversely proportional to t [27].
During the fine-adjustment phase, which lasts for the
rest of the training, (t ) should keep small values
Figure 9 Health Assessment Results by LR using Fisher’s
over a long span of time. Criterion Feature Selection Method
3.3.2 Health Assessment by LR when Measurement
both in Normal Condition and Faulty Condition is
Available
After the feature spaces for both normal and
faulty conditions are obtained, the next step is to use
the feature space as the input of the health assessment
model to evaluate the health condition of the ATC. In
this case, both the normal and faulty data are made
available and the health assessment task is to classify
Figure 10 Health Assessment Results by LR using PCA Figure 11 Noise Test Results using Fisher’s Criterion
Feature Selection Method Feature Selection Method
Auto-regressive Moving Average (ARMA) function or activation function can be employed, which
An ARMA model consists of two parts, the is described as:
autoregressive (AR) part and the moving average (MA) y j f (s j) (24)
part. The AR (p) model can be represented by
p A popular representative of the transfer function
Zt i Zt i t
is the logistic function from the family of sigmoid
i 1 (20) functions, which is described as:
Z t , Z t 1 , Z t 2 ,...Z t 1
in which p
are deviations from f (s j)
(the mean about which the process varies), (25)
1 exp( s)
j
in which
Zt is the deviation from ,
neuron j and neuron i, E is the error function and K is a
constant proportionality. The learning algorithm will
i , i 1, 2,..., q update the weights of the network to match the outputs
are the parameters of the model and
with the desired target values in iterative steps; the
iteration stops when a certain criterion (such as
maximum iteration step, maximum iteration time,
mean square error, etc.) is met.