Professional Documents
Culture Documents
Lec 7 Historybased Fault ANN
Lec 7 Historybased Fault ANN
Lec 7 Historybased Fault ANN
Farzaneh Abdollahi
Winter 2024
Example
Feature Extraction
Feature Selection
NN Classifier
Results
[3]
Data Collection
Feature Extraction
Feature Selection
Feature Selection
ANN-Based Approaches
▶ The output of BPNN is
XnH
out
(yi )k = σ ( wjout .xiH + bjout ), k = 1, ..., l
i=1
where (yi )k : the predicted output of the kth neuron in the output
layer, σ out : the activation function of the output layer, wjout , bjout :
respectively the weights and bias of the output layer
item The optimization objective of BPNN to be minimized are the
error between the output and the target
l
1X
min Ei = [(di )k − (yi )k ]2
w ,b 2
k=1
ANN-Based Approaches
▶ In Backward Path
▶ The training parameters w and b are updated by the gradient descent:
∂E ∂E
w ←w −η , b ←b−η
∂w ∂b
where η:the learning rate.
▶ The error gradient propagates backward from the output layer to the
input layer, and updates the training parameters layer by layer.
▶ Some other popular ANN-based approaches are RBF and Wavelet
NN
ANN-Based Approaches
[3]
Farzaneh Abdollahi Intelligent Fault Diagnosis Lecture 7 13/24
Outline Traditional ML for IFD Example
Example [4]
Feature Extraction
Feature Extraction
Feature Selection
▶ After the feature extraction, there are too many input features that
would require significant computational efforts to calculate
▶ It may result in low accuracy in the monitoring and fault diagnosis.
▶ Solution To remove the redundant or irrelevant information , the
data is mapped into a space of lower dimensionality by principal
component (PC) analysis (PCA).
▶ PCA is a multivariate technique that analyzes a data table in which
observations are described by several intercorrelated quantitative
dependent variables.
▶ By introducing a an orthogonal linear transformation, the important
information from the table is extracted and represent it as a set of
new orthogonal variables called PCs
▶ Therefore the data is transformed to a new coordinate system such
that the greatest variance by any projection of the data comes to lie
on the first coordinate (called the first PC), the second greatest
variance on the second coordinate, and so on
Farzaneh Abdollahi Intelligent Fault Diagnosis Lecture 7 17/24
Outline Traditional ML for IFD Example
NN Classifier
NN Classifier
NN Classifier
▶ MLP: For the second layer of the network, (MLP layer), proper
learning rule and transfer function should be selected for optimum
performance
▶ tanh is found the best choice for activation fcn
▶ Different learning rules are examined: Momentum (MOM), Conjugate
Gradient (CG), Quick Propagation (QP), Delta Bar Delta (DBD),
Levenberg–Marquardt (LM), and Step (STP)
▶ Step size, number of hodden layers, momentum and learning rate are
other parameters to be chosen properlly by trial and error
NN Classifier
▶ Final proposed design for NN:
▶ number of inputs: 6;
▶ stopping condition: 4500 epochs;
▶ number of hidden layers: 01
▶ error criterion: L2 norm;
▶ number of cluster centers(RBF hidden layer): 35;
▶ number of connection weights: 489.
▶ Hidden Layer :
▶ Transfer function: Tanh
▶ Step size: 0.3
▶ Momentum: 0.9
▶ Output Layer :
▶ Transfer function: Tanh
▶ Step size: 0.4
▶ Momentum: 0.5
Results
References I
References II