You are on page 1of 16

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/236891710

A Reconfigurable Watchdog Agent® for Machine Health Prognostics

Article  in  International Journal of COMADEM · July 2008

CITATIONS READS

8 965

3 authors, including:

Linxia Liao Jay Lee


GE Digital University of Cincinnati
35 PUBLICATIONS   722 CITATIONS    338 PUBLICATIONS   7,469 CITATIONS   

SEE PROFILE SEE PROFILE

Some of the authors of this publication are also working on these related projects:

Books on Industrial Big Data and CPS View project

Relative entropy based Bayesian design for accelerated degradation testing View project

All content following this page was uploaded by Linxia Liao on 23 May 2014.

The user has requested enhancement of the downloaded file.


Published in 2008, International Journal of COMADEM; 11(3):2.

A Reconfigurable Watchdog Agent® for Machine Health Prognostics


Linxia Liao Haixia Wang Jay Lee

The University of Cincinnati, NSF I/UCR Center for Intelligent Maintenance Systems,
Department of Mechanical Engineering, 560 Rhodes Hall,
Cincinnati OH 45221. Phone: 513-556-3412
Email: liaol@email.uc.edu haixia.wang@uc.edu jay.lee@uc.edu

Linxia Liao is currently a Ph.D. student in Department of Mechanical Engineering at University


of Cincinnati, US and he is also working as a research assistant at the National Science
Foundation (NSF) Industry/University Cooperative Research Center (I/UCRC) for Intelligent
Maintenance Systems (IMS www.imscenter.net). His current research area focuses on the Design
of Reconfigurable Platform for Embedded Health Monitoring and Prognostics of Machine Tool
Systems.

Dr. Haixia Wang is a research associate in the NSF Industry/University Cooperative Research
Center (I/UCRC) for Intelligent Maintenance Systems (IMS) at the campus of Univ. of
Cincinnati. Her current research interest focuses on machinery and process prognostics and
health management, manufacturing process performance and quality improvement, and closed-
loop product life cycle service model studies. Haixia Wang received her B.S degree in
Mechanical Engineering from Shandong University at China, a Ph.D in Mechanical Engineering
from Southeast University at China, a M.S. and a Ph.D in Industrial and Systems Engineering
from the Univ. of Wisconsin-Madison at USA.

Dr. Jay Lee is Ohio Eminent Scholar and L.W. Scott Alter Chair Professor at the Univ. of
Cincinnati and is founding director of National Science Foundation (NSF) Industry/University
Cooperative Research Center (I/UCRC) on Intelligent Maintenance Systems (IMS
www.imscenter.net ) which is a multi-campus NSF Center of Excellence between the Univ. of
Cincinnati (lead institution), the Univ. of Michigan, and the Univ. of Missouri-Rolla in
partnerships with over 45 global companies including P&G, Toyota, GE Aviation, Boeing, AMD,
Caterpillar, Siemens, DaimlerChrysler, ETAS, Festo, Harley-Davidson, Honeywell, ITRI
(Taiwan), Komatsu, Omron, Samsung, Toshiba, Bosch, Parker Hannifin, BorgWarner, Spirit
Aerosystems, Nissan (Japan), Syncrude (Canada), and McKinsey & Company, etc. His current
research focuses on autonomic computing, embedded IT and smart prognostics technologies for
industrial and healthcare systems, design of self-maintenance and self-healing, systems, and
dominant design tools for product and service innovation. He also serves as honorary professor
and visiting professor for a number of institutions including, Cranfield Univ. in UK, Lulea Univ.
of Technology in Sweden, Shanghai Jiao Tong Univ. in China, Univ. of Manchester, City Univ.
of Hong Kong, and Hong Kong PolyU., and Harbin Institute of Technology (HIT) in China, etc.
He is also a Fellow of ASME, SME, as well as International Society of Engineering Asset
Management (ISEAM).

ABSTRACT
®
This paper presents a scalable Watchdog Agent -based toolbox approach for machine health prognostics. The toolbox
consists of modularized embedded algorithms for signal processing and feature extraction, performance assessment,
diagnostics and prognostics, which can be reconfigured for different machinery prognostic applications, and can be
extensible and adaptable to most real-world machine situations. A decision making technique, Quality Function
Deployment (QFD)-based tool selection method, is applied for the automatic selection of algorithms from the Watchdog
Agent® toolbox using multiple criteria. In addition, the architecture for the Watchdog Agent®-based real-time remote
machinery prognostics and health management, which incorporates remote and embedded predictive maintenance
technologies, is presented. An industrial case involving the automatic tool changer of a machine tool is presented to
illustrate how the Watchdog Agent® toolbox can be used in diverse scenarios.
Keywords: Watchdog Agent® toolbox, Machine health assessment, Prognostics
1. INTRODUCTION

As machinery becomes more complex and model is involved in such applications. As an


sophisticated, its reliability becomes more crucial. The alternative, data-driven models, which do not
breakdown of one machine may result in the halt of the incorporate the use of any system knowledge, have
whole production line in a facility. Regular been proposed for machine fault detection and
maintenance becomes a necessity to keep machinery as diagnosis. For example, logistic regression is used for
reliable as possible. Looking back on the format of the elevator system degradation detection [8]; support
maintenance several approaches to conducting vector machines is proposed in [9] for motor system
maintenance exist such as reactive maintenance, diagnosis; an artificial neural network is used in chiller
preventive maintenance and predictive maintenance. fault detection and diagnosis in [10]; an Hidden
Today, most machine field services depend on sensor- Markov model is applied in [11] for tool wear
driven management systems that provide alerts, alarms monitoring tasks and so on. Data-driven methods do
and indicators of machine failure. Therefore, most not need prior knowledge of the system, although the
machine maintenance today is either purely reactive measurement under various working conditions is
(fixing or replacing equipment after it fails) or blindly usually necessary. Such methods are not feasible when
proactive (assuming a certain level of performance certain measurements are not available, as is usually
degradation, with no input from the machinery itself, the case for critical equipment. The hybrid approach
and servicing equipment on a routine schedule whether fuses the model-based information and sensor-based
service is actually needed or not); both scenarios could information, and takes advantage of both model-driven
be extremely wasteful. In addition, in the case of and data-driven approaches, through which more
reactive maintenance, it’s already too late to prevent reliable and accurate prognostic results can be
the failure when an alarm sounds. generated [12].
Rather than accepting the limitations of reactive However, a major breakthrough has not been
maintenance (or a “fail-and-fix” approach), world-class made since. In spite of the progresses in CBM, many
companies are moving forwards towards “predict-and- fundamental issues still remain. For example:
prevent” maintenance. A maintenance scheme, referred 1) Most research is conducted at the single
to as Condition Based Maintenance (CBM) [1], was equipment level, and no infrastructure exists for
developed by considering current degradation and its employing a real-time remote machinery
evolution over time. CBM methods and practices have diagnosis and prognosis system for maintenance;
been continuously improved for in recent decades to 2) Most developed prognostics approaches are
provide sufficient warning of an impending failure on a application or equipment specific. A generic and
particular piece of equipment, allowing that equipment scalable prognostic methodology or toolbox
to be maintained only when there is objective evidence doesn’t exist;
of an impending failure. In the past twenty years, a 3) Currently, methods are focused on solving the
significant number of health monitoring techniques, problem of failure detection. The need for tools
tools and prognostic methods have been developed for for system performance assessment and
machinery. Sensor fusion techniques [2, 3] are widely degradation prediction has not been well
used to make use of the data collected from the addressed;
physical assets of the machinery. The monitoring data, 4) Many developed prediction algorithms have been
which consists of the measurements related to machine demonstrated in a laboratory environment, but
performance, can include vibration, acoustic emissions, have yet to undergo validation in industry.
cutting force, temperature, humidity, pressure, To address the aforementioned unmet needs,
moisture, voltage, current, etc. To make the data Watchdog Agent®-based Intelligent Maintenance
understandable and interpretable, many signal Systems (IMS) has been presented by the IMS Center
processing and feature extraction algorithms [4-6] are [13] with a vision to develop a systematic approach in
used to extract the features that can best represent the advanced prognostics to enable products and systems
failure or degradation mechanisms. To build machinery to achieve near-zero breakdown reliability and
prognostics models, some researchers are working to performance.
build physical models for fault detection and diagnosis The rest of this paper is organized as follows:
for special systems, like chillers [7]. This kind of Section 2 presents a systematic Watchdog Agent®-
method has been proven effective in special cases; based approach for machinery health prognostics. The
although it is always not practical when accurate architecture for the Watchdog Agent®-based Real-time
knowledge of the whole system is not available. Remote Machinery Prognostics and Health
Moreover, it is not easy to apply the same model to Management (R2M-PHM) system is first presented.
other equipment; usually significant modification to the Then a description of how to deploy the Watchdog
Agent® toolbox (a collection of algorithms for determined and published to the internet through an
prognostics), and which algorithm to use, is presented. embedded web server.
Section 3 uses the automatic tool changer of the rotary The software system of the Watchdog Agent®-
machine as an example to illustrate how to select the based IMS platform consists of two parts: the
most appropriate algorithms for signal processing, embedded side software and the remote side software.
health assessment and performance prediction in The embedded side software is the software running on
different scenarios. Section 4 gives the summary and the Watchdog Agent® hardware platform, which
conclusion of this paper. includes a communication module, a command
analysis module, a task module, an algorithm module,
2. WATCHDOG AGENT®-BASED
a function module, and a DAQ module. The
INTELLIGENT MAINTENANCE SYSTEMS communication module is responsible for
Currently, most state-of-the-art manufacturing, communicating with the remote side via TCP/IP
mining, farming, and service machines (e.g., elevators) protocol. The command analysis module is used to
are actually quite “smart” in themselves. Many analyze different commands coming from the remote
sophisticated sensors and computerized components side. The task module includes multi-thread scheduling
are capable of delivering data concerning a machine’s and management. The algorithm module contains
status and performance. The problem is that little or no specific watchdog agent tools. The function module
practical use is made of most of this data. Although the has several auxiliary functions such as channel
necessary devices exist, a continuous and seamless configuration, security configuration, and email list and
flow of information throughout entire processes has so on. The DAQ module performs A/D conversion
been realized. Sometimes, this is because the available using either interrupt or software trigger to get data
data is not rendered in a useable, or instantly from different sensors. The remote side software is the
understandable, form. More often, no infrastructure software running on the remote computers. It is
exists for delivering the data over a network, or for implemented by ActiveX control technology and can
managing and analyzing the data, even if the devices be used as a component of the Internet Explorer
were networked. Browser. The remote side software is mainly composed
Watchdog Agent®-based Real-time Remote of a communication module and a user interface
Machinery Prognostics and Health Management (R2M- module. The communication module is used for
PHM) system has been recently developed by the IMS communicating with the embedded site via TCP/IP
Center. It focuses on developing innovative protocol. The user interface has different visualization
prognostics algorithms and tools, as well as remote and tools to present the equipment health information.
embedded predictive maintenance technologies to A four-layer infrastructure for the developed
predict and prevent machine failures. system setup is illustrated in Figure 1. The data
acquisition layer consists of multiple wireless sensors
2.1 Watchdog Agent®-based R2M-PHM which obtain data from multiple machines in the plant
Platform or even in the outside field. An Ethernet connection is
The reconfigurable and scalable Watchdog set up between the Watchdog Agent® platform and the
Agent®-based R2M-PHM platform being developed by controllers on the machine tools in order to collect
the IMS Center expands the well-known Open System useful controller signals. The Watchdog Agent®
Architecture for Condition-Based Maintenance (OSA- platform is located between the data acquisition layer
CBM) standard [14] by including real-time remote and the network layer. The machine data is processed
machinery diagnosis and prognosis systems and locally and converted into performance-related
embedded Watchdog Agent® technology. information before it is sent to the Internet. The
A “digital doctor” inspired by biological network layer will utilize either traditional Ethernet
perceptual systems and machine psychology theory, connections or wireless connections for
the Watchdog Agent® consists of embedded communication between the Watchdog Agent®
computational prognostic algorithms and a software platform and the Internet. A wireless sensor network
toolbox for predicting degradation of devices and access point is connected to the wireless router to send
systems. It is being built to be extensible and adaptable the vibration data to the Watchdog Agent® platform,
to most real-world machine situations. Once the health which is also connected to the wireless router to
related information is saved to the database, the publish the health information through the Internet. A
diagnostic and prognostic outputs of the Watchdog firewall is put between the Internet and the wireless
Agent®, mounted on all the machinery of interest, can router to provide secure communication and to protect
then used for maintenance decision making. The lean the Watchdog Agent® platform from outside malicious
and necessary information for maintenance can then be attacks. The application layer contains an application
server and an information database. The application
server controls the behavior of the Watchdog Agent®
platform and provides services between the enterprise
users’ requests and the information database. The
information database contains all the asset health
information including the health situation, historical
performance and basic information of the assets and so
on. The enterprise layer offers a user-friendly interface
and decision support tools for practitioners and
engineers to access the equipment information, and
make maintenance decisions via web-based human
machine interface.

Figure 1 System Architecture for R2M-PHM System


®
2.2 Watchdog Agent Toolbox for Multi-sensor into multiple-regime features by selecting appropriate
Performance Assessment and Prognostics algorithms for signal processing and feature extraction.
The Watchdog Agent® toolbox, with autonomic In the feature space, health indices are obtained by
computing capabilities, is able to convert critical statistically detecting the value shifts of features by
performance degradation data into health features and selecting appropriate algorithms for health
quantitatively assess their confidence value to further assessment/evaluation; and the future machine
predict trends so that proactive actions can be taken degradation tendency is predicted by applying
before potential failures occur. As illustrated in Figure appropriate performance prediction algorithms to the
2, data obtained from the embedded sensors on the health indices; a dynamic health feature radar chart,
machines, from the maintenance database and from which shows the health condition of the critical
manually input working conditions, are transformed components, is then presented for the users’ reference.

Figure 2 Watchdog Agent®: A Systematic Approach for Machine Health Prognostics

The Watchdog Agent® toolbox enables one to Degradation assessment is conducted after the critical
quantitatively assess and predict performance properties of a process or machine are identified and
degradation levels of key product components, and to measured by sensors. It is expected that the
determine the root causes of failure [15-17], thus degradation process will alter the sensor readings that
making it possible to physically realize closed-loop are being fed into the Watchdog Agent®, and thus
product life cycle monitoring and management. enable it to assess and quantify the degradation by
quantitatively describing the corresponding change in The toolbox consists of a collection of state-of-
sensor signatures. In addition, a model of the process the-art machine learning algorithms and data
or piece of equipment that is being considered, or preprocessing tools, some of which have been used in
available application specific knowledge, can be used past projects, for predicting the degradation or
to aid the degradation process description, provided performance loss of devices, process, and systems. The
that such a model and/or such knowledge exist. The current tools employed in the signal processing and
prognostic function is realized through trending and feature extraction, performance assessment, diagnostics
statistical modeling of the observed process and prognostics modules of Watchdog Agent®
performance signatures and/or model parameters. functionality are summarized in Figure 3.

Figure 3 Modular Structure of the Watchdog Agent® Toolbox

Signal processing and feature extraction storage capabilities, limited space, power consumption,
decompose multi-sensory data into a feature space, user preferences, etc., the performance assessment
which is related to the health assessment or diagnosis module of the Watchdog Agent® has been realized in
tasks. Health assessment is the evaluation of the the form of a modular, open architecture toolbox. The
overlap between the most recent feature space and that toolbox consists of different prognostics tools,
during normal product operation. This overlap is including neural network-based, time-series based,
continuously transformed into a Confidence Value (CV) wavelet-based and hybrid joint time-frequency
ranging from 0 to 1 (0-unacceptable and 1-normal) methods, etc., for predicting the degradation or
over time for evaluating the deviation of the most performance loss of devices, processes, products and
recent behavior from normal behavior or baseline. systems. The open architecture of the toolbox allows
Health diagnosis is the analysis of the patterns one to easily add new solutions to the performance
embedded in the data to determine what previous assessment modules, as well as to easily interchange
observed fault has occurred. Performance prediction different tools, depending on the needs of the
involves the extrapolation of the behavior of the application.
process signatures over time, and the prediction of their To enable rapid deployment, a Quality Function
behavior in the future. Deployment (QFD) based selection method had been
In order to facilitate the use of Watchdog Agent® developed to provide a general suggestion to aid in tool
in a wide variety of applications, with various selection; this is especially critical for those industry
requirements and limitations regarding the character of users who have little knowledge about the available
signals, available processing power, memory and algorithms. QFD was developed by Yoji Akao in the
1960s and provides a structured framework for as shown in Figure 4. In the problem definition and
concurrent engineering that ensures that the “voice of constraints step, all the system requirements and
the customer” is incorporated into all phases of product constraints, health monitoring and prognostics
development [18]. In the context of automatic tool requirements, trade studies, such as cost/benefit
selection, the QFD method is used to calculate the analyses, to determine the monitoring and prognostics
weight for each tool under the constraints of user- features are completed. To tackle a specific prognostics
defined condition priorities. The Analytical Hierarchy problem, appropriate tools are selected by the tool
Process (AHP) method is utilized to enhance the QFD selection method mentioned above, or new tools are
implementation by using a quantitative method. The developed for special situations. In the tool selection
AHP method is used to make an effective decision step, the parameters of the selected, or newly
based on pair-wise comparison of each alternative developed, tools need to be specified by training the
against various application conditions. More detail of models with historical data. The models will then be
this method is described in [19]. The AHP-enhanced programmed and tested by an evaluation procedure. If
QFD selection method gives a suggestion for the most satisfactory results are achieved, an in-house prototype
appropriate tool under different application conditions. of the system will be built and tested. The finalized
New tools will be continuously developed and system will be deployed on a commercially available
added to the modular, open architecture Watchdog hardware platform for industrial applications.
Agent® toolbox based on the development procedure,

Problem definition &


constraints

Tool selection Parameter & tool


selection

Prototyping &
Program
testing
development

No No
Accepted Evaluation
Yes
Yes
Deployment

Figure 4 Flowchart for Developing Watchdog Agent® Tools

Vibration signal analysis is a widely used method


3. INDUSTRIAL CASE STUDY
for detecting defective machine components. If the
Machining systems, such as Computer Numerical ATC develops a mechanical defect such as cam wear,
Control (CNC) machine tools, are used for independent tool drop or tool change position error, the vibration
parts machining, tooling fabrication and are therefore signal, which is taken when the ATC hits the spindle,
integrated into the manufacture of products. As the will deviate from the normal condition. Hence,
system become more complex and sophisticated, the vibration data is obtained to analyze the mechanical
reliability is increasingly crucial. The breakdown of degradation of the ATC. If the control logic or the
one component of the system may result in the halt of Programming Logic Control (PLC) performs
the whole production line in a facility. From machine abnormally, the timing of the control steps will be
tool manufacturers’ maintenance record, Automatic different from those during normal operation.
Tool Changer (ATC) is one of the components which Therefore, timing windows from the controller are also
fail the most. Therefore, the ATC is used as an monitored.
example to illustrate how to use the Watchdog Agent- For each machine tool, a Kistler 8141A1
based approach for prognostics in different situations. accelerometer is installed on the spindle housing to
obtain the vibration signals during the tool change
3.1 System Setup and Data Acquisition process. The raw data goes through the Kistler
5127B01 signal conditioning box to the Watchdog
Agent® platform by using Direct Memory Access In this case, timing windows from the controller is
(DMA) data acquisition technique. ATC timing widow also collected. In Figure 6, the arrows represent the
of each tool change cycle is also obtained from the eight steps of a tool changing process which are: Select
controller via an Ethernet connection to the Watchdog
tool; Tool topple down; Arm rotate 0°-> 65°; Unclamp
Agent® platform. The sampling rate for vibration is 10
kHz, and the sampling rate for controller signal is 75 tool; Arm pull down, rotate 180°, raise; Clamp tool;
Hz. Arm rotate 65°-> 0°; Tool return. In the steps, 15
Vibration signals, from both a normal condition and commands and statuses for the tool changing process
a faulty condition, are collected and shown in Figure 5. are monitored, as listed in Table 1.
It is difficult to determine which signal is indicative of
a normal condition and which one is a faulty condition
by just looking at the data.
Table 1.

Figure 6 Timing Widow from Controller

Figure 5 Raw Vibration Data

Table 1 Command and Status Description of Tool Changing Process


Index Name Type Description
1 TF-KP Command T-code finish keep
2 TCON Status Magazine count
3 ATCDN Command ATC holder down
4 ATCUP Command ATC holder up
5 MHDP Status ATC holder down
6 MHUP Status ATC holder up
7 ATCY Command M06 cycle run
8 ATCYS1 Command M06 arm turn 65 command
9 ATCYS3 Command M06 arm turn command
10 ATCYS5 Command M06 arm turn original command
11 PRS_5 Status Arm origin detect
12 PRS_6 Status Arm middle detect
13 PRS_7 Status Arm stop detect
14 SPCLSW Status Tool clamp
15 SPULSW Status Tool unclamp
2
3.2 Signal Processing and Feature Selection e j ,i c ij , k
k (5)
3.2.1 Introduction to the Technical Background
Fisher’s Linear Discrimination Criterion
Wavelet Packet Analysis (WPA) Fisher’s linear discrimination criterion [23] was
Usually, the vibration signals of defective used as a criterion function and is defined as
components are highly structured and can be grouped 2
into two categories: sustained defects and intermittent i, fl m, f l
defects [20]. For sustained defects, Fourier-based J f l i, m 2 2
analysis, which uses sinusoidal functions as base i, fl m, f l
(6)
functions, provides an ideal candidate for extraction of
these narrow-band signals [4]. For intermittent defects, for two-class problems where i , fl
, m , fl
are the mean
2
signals often demonstrate a non-stationary, transient
values of the lth feature, and fl, for class i and m. i , fl
,
nature [4]. Wavelet packet transform (WPT) provides a 2
very powerful tool for non-stationery signal analysis. are the variances of the lth feature, and fl for class
m , fl
Features can be extracted from signals that combine i and m respectively.
stationary and non-stationary characteristics with
arbitrary time-frequency resolution by using WPT. Principle Component Analysis (PCA)
WPT is an extension of wavelet transform (WT) that Principal Component Analysis (PCA) is one of the
provides complete level-by-level decomposition [21]. most commonly used statistical methods for reducing
The wavelet packets are particular linear dimensionality by transforming the original features
combinations of wavelets [22]. The wavelet packets into a new set of uncorrelated features. Karhunen-
inherit properties such as orthogonality, smoothness Loève Transform (KLT) is a linear dimensionality
and time-frequency localization from their selection procedure that is related to PCA, the goal of
corresponding wavelet functions. A wavelet packet is a which is to transform a given data set X of dimension
i
function with three indices, j ,k (t ) in which integers i, j, N to an alternative data set Y of a smaller dimension M
and k are the modulation or oscillation parameter, the (M<N). Equivalently, we are seeking to find the matrix
scale parameter, and the translation parameter Y which is the Karhunen-Loève transform of matrix X:
respectively. Y AT X
(7)
j
i
t 2 2 i j
2 t k in which AT is the Karhunen-Loève transform matrix.
j,k
i=1,2,.. (1) By choosing the eigenvectors corresponding to the M
The first wavelet is referred to as the mother largest eigenvalues of the correlation matrix of X, the
wavelet, or analyzing wavelet. In this application, mean square error (MSE) between the input X and its
Daubechies wavelet 4 (DB4, shown in figure 1), which projection is minimized.
is a kind of compactly supported wavelet, is used as the
mother wavelet. The following wavelets i for i = 2, 3.2.2 Signal Processing by WPA
3 … are obtained from the following recursive In this case, as we can see in Figure 5, the raw
relationships: vibration data is non-stationary as it contains two major
impacts. Hence, Wavelet Packet Transform (WPT),
2i i
t 2 h k 2t k (2) using a rich library of redundant bases with arbitrary
k time-frequency resolution, enables the extraction of
features from signals that combine non-stationary and
2i 1 i
t 2 g k 2t k (3) stationary characteristics [24]. Wavelet packet
k transform is used to decompose the raw data so the
in which h(k) and g(k) are the quadrature mirror filters energy of each node, at a resolution level of 4, can be
(QMF) associated with the predefined scaling function calculated. In Figure 7, the bars are the mean values of
and the mother wavelet function. the energy of each node from a set of samples and the
The wavelet packet coefficients of a function f can lines are the related variation of each node.
be computed by taking the inner product of the signal
and the particular basis function
c ij ,k f, i
j ,k t f t i
j ,k t dt
(4)
e j ,k
The wavelet packet node energy is defined as
the running status of the ATC. They are combined with
the features selected by each of the methods described
above as the feature space extracted from the raw data.
Faulty

Figure 7 Energy Distribution of Wavelet Decomposition


Components Normal

3.2.3 Feature Selection by Fisher’s Linear Figure 8 PCA Projection Results for Training Dataset
Discrimination Criterion and PCA
The features marked with stars in Figure 7 are those 3.3 Health Assessment
features that have the 5 largest Fish linear discriminate
values [23], which means those 5 features contain the 3.3.1 Introduction to the Technical Background
most of the separable information for the normal and
faulty datasets. Therefore, those 5 features are selected. Logistic Regression (LR)
The problem remains that it is difficult to decide how Logistic regression is a function that can easily
many features need to be chosen for the next step: represent the daily maintenance records as a
health assessment. dichotomous/binary problem. The goal of logistic
The Principle Component Analysis (PCA) method regression is to find the best fitting model to describe
is used to projects the original feature space to a lower the relationship between the categorical characteristics
dimensional feature space which contains 95% of the of a dependent variable and a set of independent
variation information of the original feature space. In variables [25]. The logistic function is
this application, the Karhunen-Loève Transform (KLT) 1 eg x
projects the N-dimension (N=16) feature space to an Px
M-dimension (M=2) feature space. The training data 1 e g x 1 eg x (8)
includes 40 normal datasets and 35 faulty datasets, and in which the logistic or logit model is
each of them is first decomposed by WPT. Since each P x
dataset has 16 energy components after the g x lo g
1 P x
decomposition, the input matrix X is 75×16 in size. (9)
Based on the dispersion and the magnitude of the 1 x1 2x 2 ... k xk

eigenvalues, M (M=2) of the largest eigenvalues are , in which g(x) is a linear combination of the
preserved and the corresponding eigenvectors are independent variables x1, x2, … , xk. and , are the
calculated. After the projection of the original feature parameters for the independent variables. The
space onto the two dimensions along with the selected maximum likelihood estimate method [8] is used to
two eigenvectors, the new feature space Y, at a size of find the parameter values that maximize the probability
75×2, is obtained. The projection result is shown in of the observed data, by iterative steps.
Figure 8. The circles indicate the projected normal
feature space and the crosses represent the projected Self-Organizing Maps (SOM)
faulty feature space. A comparison of the effect of SOM provides a way of representing
those two feature selection methods to the results of multidimensional feature space in a one or two-
health assessment is given in chapter 3.3. dimensional space, while preserving the topological
For the timing window signals (shown in Figure 6), properties of the input feature space. SOM is an
status 5 and status 13 (listed in Table 1), which unsupervised learning neural network which can
represent the time span of the tool topple down action organize itself according to the nature of the input data.
and the duration of the arm rotation action, are selected The input data vectors, which closely resemble each
as two independent features of the input to the logistic other, are located next to each other on the map after
regression algorithm. These two features are selected training. An n-dimensional input data space can be
because they are the most critical features to indicate denoted by
x [ x1 , x2 ,..., xn ]T the testing data into either the normal or faulty
(10) category. Logistic regression is a robust tool that can
The weight vector of each neuron j in the network easily represent the daily maintenance records as a
has the same dimension as the input space, and can be dichotomous problem; therefore it has been chosen as
represented by: the health assessment model.
wj [ w j1 , w j 2 ,..., w jn ]T , j 1, 2,..., m The training dataset mentioned above is used to
(11) train the logistic regression model. Feature space is
in which m is the number of neurons in the network. extracted from the raw data via WPA and the Fisher’s
The Best Machining Unit (BMU) in the SOM is the linear discrimination criterion is first used as the
neuron whose weight vector is the closest to the input feature selection method. After the training, the health
j xT w assessment model is achieved:
vector in the input space. The inner product can g ( x) 7.2932 0.7298 x1
be used as an analytical measure for the matching of x
2.0944 x2 1.9451x3 1.2929 x4 (15)
with wj. The Euclidean distance is a better and more
convenient measure criterion for the matching of x 1.7345 x5 2.1965 x6 3.9255 x7
with wj. The minimum distance defines the BMU, if 281 datasets are then used for validation. Figure
we define wc as the weight vector of the neuron that 9 shows the confidence value curve of the ATC. The
best matches the input vector x. The measurement can drop of the confidence value at cycle 250 was caused
then be represented by by a drop in the air pressure at that cycle.
|| x wc || min{|| x wj ||}, j 1,2,..., m PCA is also used as the feature extraction method
(12)
After the BMU is identified in the iterative and the same dataset is used to train the logistic
training process, the weight vectors and the topological regression model. After the training, the health
neighbors of the BMU are updated in order to move assessment model is achieved:
them closer to the input vector in the input space. The g ( x) 7.3238 0.5764 x1
(16)
following learning rule is applied: 0.1012 x2 0.2591x3 1.9164 x4
wj (t 1) wj (t) (t)hj,wc (t)(x wj (t)) The same 281 datasets are then used for
(13)
validation. Figure 10 shows the confidence value curve
h j , wc of the ATC. It also shows a drop in the confidence
in which denotes the topological neighborhood
kernel centered on the BMU wc. A typical choice for value at cycle 250, which was caused by a drop in the
the kernel function is the Gaussian function air pressure at that cycle.
h j , wc exp( d j , wc 2 / 2 2
)
(14)
d j , wc
in which is the lateral distance between the BMU
wc and neuron j. The parameter is the “effective
width” of the topological neighborhood [26]. (t ) is
the learning rate which monotonically decreases with
the training time. In the initial phase which lasts for a
given number of steps (e.g. first 1000 steps),
(t ) starts with a value that is close to 1, and it can be
linear, exponential, or inversely proportional to t [27].
During the fine-adjustment phase, which lasts for the
rest of the training, (t ) should keep small values
Figure 9 Health Assessment Results by LR using Fisher’s
over a long span of time. Criterion Feature Selection Method
3.3.2 Health Assessment by LR when Measurement
both in Normal Condition and Faulty Condition is
Available
After the feature spaces for both normal and
faulty conditions are obtained, the next step is to use
the feature space as the input of the health assessment
model to evaluate the health condition of the ATC. In
this case, both the normal and faulty data are made
available and the health assessment task is to classify
Figure 10 Health Assessment Results by LR using PCA Figure 11 Noise Test Results using Fisher’s Criterion
Feature Selection Method Feature Selection Method

3.3.3 Comparison of the Two Feature Selection


Methods in Health Assessment Model
In order to test the effect of the two feature
selection methods on the health assessment model, a
noise test is designed. Random Gaussian white noise is
added to the vibration data. The noise intensity is
defined by the following signal-to-noise ratio (SNR):
A
SNR 20 log10 S (dB)
AN (17)
in which AS and AN represent the root mean square
(RMS) value of the signal, and that of the noise,
respectively.
We use 50 datasets in normal condition as the
testing data. For each of the two feature selection Figure 12 Noise Test Results using PCA Feature Selection
methods, 6 experiments are carried out for no noise, Method
SNR=25dB, SNR=20dB, SNR=15dB, SNR=10dB, and
3.3.4 Health Assessment by SOM When Only
SNR=5dB, respectively. Figure 11 and Figure 12 show
Measurement in Normal Condition is Available
the test results by using Fisher’s linear discrimination
In most scenarios, only measurement of the
criterion and PCA methods. In Figure 11, we notice
normal operating conditions is available. SOM can
that by using Fisher’s criterion feature selection
give out a health index to evaluate the health condition
method, health conditions are correctly assessed when
when only normal measurement is available. For each
SNR>5dB, while the health assessment results become
input feature vector, a BMU can be found in the SOM.
very unstable when SNR=5dB. In Figure 12, by using
The distance between the input feature vector and the
PCA feature selection method, the health conditions
weight vector of the BMU, which can be defined as the
are correctly identified in all of the SNR levels tested.
minimum quantization error (MQE), actually indicates
The results show that the health assessment model is
how far away the input feature vector deviates from the
more stable by using PCA rather than Fisher’s criterion
normal operating state. Hence, the degradation trend
when the SNR is at low level.
can be measured by the trend of the MQE. The MQE is
defined as:
MQE VF VBMU (18)
in which VF is the input feature vector and VBMU is
the weight vector of the BMU.
As the MQE increases, the extent of the
degradation becomes more severe. A threshold
MaxMQE can be set as the maximum MQE that can be
expected, and the degradation can be indicated by
converting the MQE into a confidence value (CV) , , ,...
t t 1 t 2 t q
ranging from 0 to 1 by using the following equation: again are the error terms. To
cv ( MaxMQE MQE ) / MaxMQE (19) achieve greater flexibility in the fitting of the actual
time series, it is sometimes advantageous to include
The equation indicates that the MQE increases while both autoregressive and moving average terms in the
the CV decreases. model [28]. So an ARMA (p, q) model refers to a
model with p autoregressive terms and q moving
average terms, which can be written as:
p q
Zt i Zt i t i t i
i 1 (22) i 1

Optimized parameters of an ARMA (p, q) model can


be estimated by historical data. To check the adequacy
of the ARMA (p, q) model, an F-test statistical
hypothesis test method can be applied. (Pandit and Wu
gave the details of this method in [29].)

Neural Network (NN)


Figure 13 Health Assessment Results by SOM Neural network has its own specialty like non-
linear curve fitting and is also widely used in industrial
Data of the first 40 cycles, during which the ATC fields. A typical type of neural network consists of
is in normal condition, is used to train the SOM model. input layer, hidden layer and output layer. In some
After the training, the 281 cycles’ dataset is used for situations, the hidden layer contains multiple layers.
validation. The corresponding MQE values are If x1, x2, …, xn are defined as input neurons and
calculated and are then converted into a CV according y1, y2, …,yn are defined as hidden layer neurons, the
to equation (19). In Figure 13, a drop of the confidence mapping from the input layer to the output layer can be
value at cycle 250 is shown, which was caused by a defined as the following equations:
drop in the air pressure at that cycle. This result is n

consistent with the health assessment results achieved S j w x


ji j j
i 1 (23)
in section 3.3.2.
in which ji are the weights of connections between
3.4 Performance Prediction the input layer neurons and the hidden layer neurons;
3.4.1 Introduction to the Technical Background j is the bias of each input layer neuron. A transfer

Auto-regressive Moving Average (ARMA) function or activation function can be employed, which
An ARMA model consists of two parts, the is described as:
autoregressive (AR) part and the moving average (MA) y j f (s j) (24)
part. The AR (p) model can be represented by
p A popular representative of the transfer function
Zt i Zt i t
is the logistic function from the family of sigmoid
i 1 (20) functions, which is described as:
Z t , Z t 1 , Z t 2 ,...Z t 1
in which p
are deviations from f (s j)
(the mean about which the process varies), (25)
1 exp( s)
j

A back propagation algorithm [30] has been used


i , i 1, 2,..., p in many applications to train the neural network model.
are the parameters of the model
The weights will change according to the following
and t is the error term. The MA (q) model can be equation:
denoted by dE
q ji K (26)
d ji
Zt t i t i , i 1, 2,..., q
i 1 (21) in which ji is the weight of connections between

in which
Zt is the deviation from ,
neuron j and neuron i, E is the error function and K is a
constant proportionality. The learning algorithm will
i , i 1, 2,..., q update the weights of the network to match the outputs
are the parameters of the model and
with the desired target values in iterative steps; the
iteration stops when a certain criterion (such as
maximum iteration step, maximum iteration time,
mean square error, etc.) is met.

3.4.2 Performance Prediction by ARMA and NN


ARMA models and NN models have advantages
and disadvantages in performance prediction. ARMA
is applicable to linear time-invariant systems whose
performance features display stationary behavior, while
it is unfeasible for use in a non-linear or dynamic
process. Furthermore, since ARMA utilizes a small
amount of historical data, it cannot provide good long
term prediction. NN is an effective tool to predict
complex system behaviors which involve non-linear
behavior and unstable processes; however, currently
there is no standard method to determine the structure Figure 14 Performance Prediction Results by ARMA
of the network and sufficient computational resources Model
are a requirement. In this case, both ARMA models The fitted model is then used to predict the trend
and NN models are used to predict the degradation of the health index, which gives an assessment of the
trend of the ATC health condition. health condition of the ATC. It is of interest to predict
As mentioned prior, confidence values can be the degradation trend of the health index at the initial
used as health indices of the ATC. The degradation of defect stage in order to generate an alarm before the
the ATC can be clearly seen from the curve of the defects become severe. The confidence values from the
confidence value over time. In the experiment, the air latest 170 cycles are used to obtain the appropriate
pressure was adjusted to decrease gradually to cause ARMA model. The confidence values for 50
the ATC to function abnormally. WPA was used as the subsequent steps, starting from cycle 171, were
signal processing tool to extract the feature space from predicted by the fitted ARMA model. The results are
the raw data. PCA and SOM were then used as the shown in Figure 14. The star points are the predicted
feature selection and health assessment tools, results. The MSE is 0.0114. The results show that the
respectively. Finally, confidence values for 221 cycles degradation trend of the confidence value can be
were obtained as historical health index for predicted by using the ARMA model method.
performance prediction. The first 170 cycles’ health
indices are used as the historical data and the Performance Prediction by NN Model
performance from 30 steps ahead is predicted and In order to compare the predicted results with the
compared with the real health indices calculated from ARMA model, the same historical data, which consists
the measurement. of health indices from the first 170 cycles, is used for
performance prediction. The performance for the next
Performance Prediction by ARMA model 50 steps is predicted by using an NN model. 60
The historical confidence values from the first 170 samples were selected from the historical health
cycles were used to obtain an appropriate ARMA indices. A 3-layer neural network, which has 105
model. To determine adequate order for the ARMA neurons in the input layer and one neuron in the output
model, while achieving the smallest error at the same layer, was then trained with the historical data.
time, ARMA (2, 1) is a simple model to start with. The There is no specific guideline to decide the
orders of the ARMA model were then adjusted and the number of neurons in the hidden layer. A trial-and-
adequacy of the model judged by F-criterion, which is error method was used to choose an appropriate
a hypothesis test that measures the significance of the number for the hidden layer. Also due to the different
improvement in modeling by using two different orders. initial weights of the network, the final performance of
This is an iterative step that continues until an adequate the network varies for each experiment. Therefore, the
order for the ARMA model is achieved. number of neurons in the hidden layer increased from 3
to 12 in the experiment. 5 trials were carried out for
each hidden layer. The NN model which achieves the
minimum mean square error (MSE) was selected as the
final model for performance prediction. Finally, the
number of the neurons in the hidden layer was selected
as 7.
5. REFERENCES

1. Jardine, A.K.S., Lin, D., and Banjevic, D. (2006).


A review on machinery diagnostics and
prognostics implementing condition-based
maintenance, Mechanical Systems and Signal
Processing. Vol. 20. No. 7. 1483-1510.
2. Reichard, K.M., Van Dyke, M., and Maynard, K.
(2000). Application of sensor fusion and signal
classification techniques in a distributed
machinery condition monitoring system, in
Proceedings of SPIE - The International Society
for Optical Engineering.
3. Roemer, M.J., Kacprzynski, G.J., and Orsagh, R.F.
Figure 15 Performance Prediction Results by NN Model (2001). Assessment of data and knowledge fusion
The trained NN model was then used to predict the strategies for prognostics and health management,
confidence value of the next 50 cycles, starting from in IEEE Aerospace Conference Proceedings.
cycle 171. The results of which are shown in Figure 15.
4. Yen, G.G. and Lin, K.-C (2000). Wavelet packet
The star points are the predicted results. The MSE is
feature extraction for vibration monitoring,
0.0091. Compared with the ARMA model
Industrial Electronics, IEEE Transactions. Vol. 47.
(MSE=0.0114), NN model achieves smaller MSE to
No. 3. 650-667.
predict the confidence value of the future 50 cycles.
The experiment results show that both the ARMA 5. Kemerait, R.C. (1987). New Cepstral Approach
model and the NN model can predict the degradation for Prognostic Maintenance of Cyclic Machinery,
trend of the health index of the ATC. ARMA model in Proceedings of IEEE SOUTHEASTCON.
needs less computational resources and utilizes less
6. Goodenow, T., Hardman, W., and Karchnak, M.
historical data, which makes it appropriate for short
(2000). Acoustic emissions in broadband vibration
term prediction instead of long term prediction. As
as an indicator of bearing stress, in IEEE
Figure 14 shows, the confidence value drops
Aerospace Conference Proceedings.
significantly in the last 10 cycles, while the predicted
results of ARMA model still remain at a high level. 7. McIntosh, I.B.D., Mitchell, J.W., and Beckman,
The NN model needs more computational resources, W.A. (2000). Fault detection and diagnosis in
though it can take more historical data into chillers - Part I: model development and
consideration for performance prediction, which makes application, ASHRAE Transactions. Vol. 106.
it better for long term prediction. As shown in Figure
15, the NN model can achieve better prediction results 8. Yan, J. and Lee, J. (2005). Degradation assessment
than ARMA model as it captured the more severe and fault modes classification using logistic
degradation level of the last 10 cycles in this case. regression, Journal of Manufacturing Science and
Engineering, Transactions of the ASME. Vol. 127.
4. CONCLUSION No. 4. 912-914.
This paper presents a scalable and reconfigurable 9. Poyhonen, S., Jover, P., and Hyotyniemi, H.
Watchdog Agent® toolbox for machine health (2004). Signal processing of vibrations for
prognostics. The toolbox consists of algorithms for condition monitoring of an induction motor, in
signal processing, health assessment, performance International Symposium on Control,
prediction and diagnosis, which can be reconfigured Communications and Signal Processing, ISCCSP.
for different machinery prognostic applications. An
10. Gu, B., Wang, Z.Y., and Jing, B.Y. (2003). Fault
industrial case illustrates how to use different
detection and diagnosis in chillers using artificial
algorithms in the Watchdog Agent® toolbox for
neural networks, in Proceedings of the 2003
different tasks. The experiment results also show that
International Conference on Cryogenics and
the toolbox can be utilized to deal with various data
Refrigeration.
scenarios under different constraints. Furthermore, this
toolbox can be easily applied to the health monitoring 11. Atlas, L., Ostendorf, M., and Bernard, G.D. (2000).
and prognostics tasks for diverse machinery Hidden Markov models for monitoring machining
applications. tool-wear, in ICASSP, IEEE International
Conference on Acoustics, Speech and Signal 24. Sun, Z. and Chang, C.C. (2002). Structural
Processing - Proceedings. damage assessment based on wavelet packet
transform, Journal of Structural Engineering. Vol.
12. Hansen, R.J., Hall, D.L., and Kurtz, S.K. (1994).
128. No. 10. 1354-1361.
New approach to the challenge of machinery
prognostics, in American Society of Mechanical 25. Yan, J., Lee, J., and Koç, M. (2002). Predictive
Engineers. algorithm for machine degradation detection using
logistic regression, in 5th International Conference
13. http://www.imscenter.net.
on Managing innovations in Manufacturing,
14. Lebold, M., et al. (2002). A Framework for Next Proceedings of MIM.
Generation Machinery Monitoring and
26. Fei, Z., Ticlin, S., and Tao, H. (2005). Fault
Diagnostics, in Failure Prevention Technology to
diagnosis of motor bearing using self-organizing
Improve the Bottom Line, 56th Meeting of the
maps, in Proceedings of the Eighth International
Society for Machinery Failure Prevention
Conference on Electrical Machines and Systems.
Technology.
27. Kohonen, T. (1990). The self-organizing map,
15. Casoetto, N., et al. (2003). Multisensor process
Proceedings of the IEEE. Vol. 78. No. 9. 1464-
performance assessment through use of
1480.
autoregressive modeling and feature maps, Journal
of Manufacturing Systems. Vol. 22. No. 1. 64-72. 28. Tabbane, N., Tabbane, S., and Mehaoua, A. (2004).
Autoregressive, Moving Average and Mixed
16. Djurdjanovic, D., et al. (2000). Computerized
Autoregressive-Moving Average Process for
classification of temporomandibular joint sounds,
Forecasting Qos in Ad Hoc Networks for Real-
IEEE Transactions on Biomedical Engineering.
time Service Support, in Vehicular Technology
Vol. 47. No. 8. 977-984.
Conference.
17. Lee, J. (1996). Measurement of machine
29. Pandit, S.M. and Wu, S.M. (1993). Time Series
performance degradation using a neural network
and System Analysis with Applications, John
model, Computers in Industry. Vol. 30. No. 3.
Wiley, New York.
193-209.
30. Leonard, J.A. and Kramer, M.A. (1993).
18. Govers, C.P.M. (1996). What and how about
Diagnosing dynamic faults using modular neural
quality function deployment (QFD), International
nets, IEEE Expert. Vol. 8. No. 2. 44-53.
Journal of Production Economics. Vol. 46-47.
575-585.
19. Armacost, R.L., et al. (1994). AHP framework for
prioritizing customer requirements in QFD: An
industrialized housing application, IIE
Transactions (Institute of Industrial Engineers).
Vol. 26. No. 4. 72-79.
20. Paul, D. (1995). Condition monitoring and defect
diagnosis in manufacturing process using DDA
and wavelets, in Dep. Elect. Eng., Michigan
Technol. Univ.: Houghton, MI.
21. Goumas, S.K., Zervakis, M.E., and Stavrakakis,
G.S. (2002). Classification of washing machines
vibration signals using discrete wavelet analysis
for feature extraction, IEEE Transactions on
Instrumentation and Measurement. Vol. 51. No. 3.
497-508.
22. Theodoridis, S. and Koutroumbas (2003). Pattern
Recognition. 2 ed., Academic Press.
23. Duda, R.O., Hart, P.E., and Stork, D.G. (2001).
Pattern Classification, 2nd ed, John Wiley & Sons,
New York.

View publication stats

You might also like