Download as pdf or txt
Download as pdf or txt
You are on page 1of 4

Method on Determining Feature Weight in

Case-Based Reasoning System


Baohua Yang Xiu Jin
College of Information and Computer College of Information and Computer
Anhui Agriculture University Anhui Agriculture University
Hefei Anhui, China Hefei Anhui, China

inadequate. BP neural network is essentially a non-linear


Abstract—At present, the determining method on feature input-output mapping, which can make up for the lack of
weight have the problems of complexity and one-sidedness inaccurate data and non-dynamic, through the study of
of determining in case-based reasoning system. In this known samples to obtain expert evaluation of the
paper, an integrated method based on BP Neural Network experience of the importance of knowledge and on the
and sensitivity analysis is studied. An algorithm based on BP target.
Neural Network and sensitivity analysis (BPNN-SA) is put So, for studying the dynamic relationships of these
forward, the network topology changes in accordance with features, in this paper,we at first construct model of BP
input nodes change. A red-seed watermelon case is given to Neural Network, which dynamic changes in the structure
illustrate this determining method. After the calculation and with input nodes change.Second,study on sensitivity
analysis, results show the new method is better than the analysis on new Neural Network, a integrated algorithm
current methods and effective for determining feature based on BP Neural Network and sensitivity analysis
weight. (BPNN-SA)is put forward. Finally, example is proved the
Keywords-Sensitivity analysis; Feature Weight; BP Neural effectiveness of the algorithm in red-seed watermelon
Network;Case-Based Reasoning case.

I. INTRODUCTION II. NEURAL NETWORK MODEL


Feature weight is one of the most important issues of 1) Network layer
case-based reasoning (CBR). It plays a significant role in BP network may contain different hidden layer, and
the decision-making and improving case retrieval [1]. So, the theories have proved that the BP Neural networks
whether the determining method on feature weight is which contain 2 layers network (including one hidden
reasonable or not is the key factor in case retrieval, which layer) can map any non-linear relation, when nodes
will decide the correctness and final reliability of the case number of the hidden layer is arbitrary.
method in decision-making. Studies on feature weight 2) Input and output layer dimension
have been investigated by many researchers. Input layer is used as buffer-storage, based on the
At present, AHP (Analytic Hierarchy Process, AHP) is need to solve problems and identify data. Dimensions of
quite commonly used method in determining the weights output layer depend on the two factors. One is the output
[2].But this approach require the expert higher score and data type. The other is the size of the required data types.
algorithm was complex. So, Professor Sun [3] uses rough We should minimize the system scale as possible in
sets to determine the weights. But the case of continuous design of input layer and output layer, so that the training
quantitative attribute call for discrimination, most have a time of network will be reduced.
discrete method of similarity measurement error case, and 3) Hidden layer nodes
cause data loss, resulting in case retrieval and case-based At present, there is not a good analytical, and the
reasoning results are inaccurate. As time and space number of nodes often determined by empirical or
change, to determine the weights are not in line with the conducting test themselves. Is generally believed that the
actual situation [4]. In particular, a number of complex number of nodes of hidden layer to solve the problem
nonlinear systems, the use of traditional algorithms to with the request, the number of input and output units
establish the weights of the learning mechanism seem have a direct relationship. In addition, too many hidden

978-1-4244-7161-4/10/$26.00 ©2010 IEEE


Authorized licensed use limited to: McGill University. Downloaded on February 16,2023 at 01:40:43 UTC from IEEE Xplore. Restrictions apply.
layer nodes will lead to the learning time is too long; and determine the size of sensitivity. First of all, get weights
too few hidden nodes, fault tolerant of poor identification vector {l1,l2,…ln} of features from trained neural network.
of samples without the ability to study low. Sensitive factor is Si:
The number of hidden layer is commonly Used 0 i
§ − | ·¸
hidden layer is one or two in the real application. Initial ¨ | p p
¨
¨
¦ L 0 ¸
¸
value of hidden layer nodes can also be is decided by the p ˄˅
s i = © ¹
n
following formula:
P0 is the predictive value of network output after
l= m+n +a ˄1˅ deleted characteristic i;P0 is predicted value without any
characteristics of the network deleted ;L is the training
l = 0.43mn + 0.12n2 + 2.54m + 0.77n + 0.35 + 0.51 ˄2˅
sample cases, n is the number of sample cases.
Where m is the number of output neurons, n is the Definition 3 In accordance with coefficient
number of input units, a is constant between 1 and 10. determined by weight-sensitive, then Wi is the feature
4) Selection of transfer function item weighting:
The way of neurons connect is decided by transfer
s
function, appropriate function can reduce distortion, w i = n
i

increase the supply of different fitting between neurons. ¦ j = 1


s j

The commonly used transfer functions are: ˄˅

Linear transfer function, the expression of function as The feature weighting value is 1 after standardization.
follows: According to the formula (6) and (7), calculations of the
§ 0 i characteristics of the weight.
¨ | p − p | ·¸
¨
¨
¦ L 0 ¸
¸ B. Theorem
p
si = © ¹
Theorem 1 Decision-making to the characteristics of
n ˄˅
the target data set composed of samples of the case library
S-shaped hyperbolic tangent transfer function, the
as a BP neural network learning samples for training, For
expression of function as follows:
Si and Sj , respectively, for the deletion of the first items i
si and j after the characteristics of the sensitive coefficient.
wi = n IF Si>Sj, the sensitivity of the feature i larger than the
¦
j =1
s j sensitivity of feature j, that is Wi>Wj.
 ˄˅ Theorem 2 Decision-making to the characteristics of
the target data set composed of samples of the case library
S-shaped logarithmic function, the expression of
as a BP neural network learning samples for training, Pi is
function as follows:
assumed as predictive value of network output after
0 i

lni =¦
p −p| | deleting feature i,P0 is assumed as predictive value of
L 0 network output without deleting any nodes, feature
p ˄˅ weight vector denoted as {l1,l2,…ln}.
0 i
III. SENSITIVITY ANALYSIS | p −p|
A. Definition
lni = ¦L 0

Definition 1 Sensitivity is the influence degree of


p ˄˅
possible beforehand change factors and
If lni>lnj,Then the feature i has more great influence on
superiority-inferiority of the decision-making. It can
test result than feature j, feature weight can be obtained in
indicate that the change in relevant factors influence
order each of the weight. According to relevance of
degree on decision-making goals, decision-making
network parameters to sensitivity, and metrics as a
scheme and other evaluation index.
parameter used to determine changes in network
Definition 2 Sensitivity [5] introduced into the
performance on the network size [6-7].
case-base, and defines the sensitivity coefficient to

Authorized licensed use limited to: McGill University. Downloaded on February 16,2023 at 01:40:43 UTC from IEEE Xplore. Restrictions apply.
IV. DETERMINING METHOD ON FEATURE WEIGHT Step6: Calculate network sensitivity coefficient after
A. Determining Method on Feature Weight Based deleting an input node, then in accordance with formula
on BP Neural Network and Sensitivity Analysis (7), calculate the feature weight.
Data sets are usually given in the form of tables,and Step7: Repeat to perform steps 6, according to
we call a case system, formulated as T=<U,A,V,,f>, different features of the case, we delete different input
where U is a set of finite and nonempty objects called the nodes from the network, that is, delete the various
universe,A is the feature set,V is the domain of the feature features of the case one by one, calculate their sensitivity
value,and f is the information function:f =U×AĺV.If the coefficients, and so all features weight are determined.
attribute set A is divided into condition feature set C and B. Cases Analysis
decision feature set D.The structure of cases accord with In this section, an example is used to illustrate the
people's habits and is good for the storage and neural analysis procedure proposed. In Anhui Agricultural
network training.For studying the dynamic relationships University red-seed watermelon horticulture 150 cases
of these features, in this paper,we construct model of BP are randomly selected as training data, 80 cases are
Neural Network,where feature set C are regarded as input chosen as testing data[8].It can be regarded as a set of
nodes of BPNN and feature set D are regarded as output attribute values of the feature of (feature value pairs)
nodes of BPNN.We analyse the degree of sensitivity of constitute a collection. In the case, there are many
the output neurons to the input neurons,when input characteristics can be described about red-seed
neurons are deleted one by one.As nodes reduce and watermelon. According to expert experience some
weights among neurons change,on the corresponding the features were concerned, such as fruit weight, seed size,
topology of the network may dynamic change. seed number, grain weight and seed weight and other
Then determining algorithm on feature weight based characteristics, as a division of the feature on different
on BP neural network and sensitivity analysis is types of red-seed watermelon kernel.
constructed as follows: In accordance with the above algorithm, by the
Step1˖Firstly, Select representative data as a sample training neural network model forecasting result is as
case data in case-based system; shown in Fig.1, which meets the requirements on
Step2˖ In order to reduce the effect resulting from the prediction accuracy.
different condition feature set, we normalize the
numerical attribute x into the interval [0, 1] with

xk −xmin
xk' =
xmax −xmin ˄˅
Step3˖Determine the network topology, input neurons
and output neurons are determined in different cases.
According to the formula (1) or (2) to determine the
Figure 1. The training neural network model forecasting results
number of hidden layer neurons;
Step4˖Select the learning parameters of BP Neural After the dynamic adjustment of neural network
Network; determine the network's transfer function and structure on testing data, by the training network different
training function, training steps and target errors. models of BP Neural Network prediction error curve are
Step5˖Training of the neural network to determine the shown in Fig2,where y1, y2, y3, y4, y5 representatives of
accuracy of the network to meet for the purpose of deletion fruit weight, seed size, seed number, grain weight
training the network. If the network does not meet the and seed weight. Between the real data and the
accuracy, then we go to Step 3 and repeat to perform steps forecasting data, which show the BPNN-SA model is
3 and step 4.Optimization of neural network parameters valid in precision of forecast.
and structure, adjust the training function,re-learning on
the network until meet the given accuracy. Otherwise, go
to Step 6.

Authorized licensed use limited to: McGill University. Downloaded on February 16,2023 at 01:40:43 UTC from IEEE Xplore. Restrictions apply.
The calculation results of weight are as shown in table
1. We can conclude that the feature of the seed size is the
smallest weight in all features. So the contribution of
weight on seed size is the smallest. Other features weights
followed by descending: fruit weight, seed number, grain
weight, seed weight. The results are in line with experts
and practitioners. That is to say, the feature with high
weight play important role in decision classification and
prediction.

Figure 2. Prediction error curve


TABLE 1 RESULTS OF ALGORITHMS

Feature Weights
Algorithm
fruit weight seed size seed number grain weight seed weight

BPNN-SA 0.07415 0.03214 0.07225 0.05640 0.05120

V. CONCLUSION REFERENCES
[1] Jiang Yan,Chen Rongqiu.Parameter Sensitivity Analysis Based on
Therefore, we can conclude that the algorithm of the Least Variance of Weights. Journal of Wuhan Automotive
BPNN-SA not only reflect the dynamic relationship Polytechnic University. vol.27,NO 1, Feb. 2005, pp.165-167.
[2] Zhu Yin,Meng Zhiyong,Kan Shuyu.Determination of Weight V
between features, but also calculate simply and overcome alue by AHP. Journal of Northern Jiaotong University. vol.23,NO
the one-sidedness. So it is accurate and effective. The 5, Oct. 1999,pp.119-122.
[3] Walczak B, Massart D L. Rough sets theory. Chemometrics and I
algorithms provide a basis for retrieval and re-use in ntelligent Laboratory Systems, vol.47,NO 1, Apr,1999, pp.1-16.
case-based reasoning system. Our future work includes the [4] SUN Huijun,WANG Xinhua.Determin ation of the Weight of
Evaluation Indexes with Artif icial Neural Network Method.
extension of determining method on feature weight. Journal of Shandong University of Science and Technology.
vol.20,NO 3, Sep. 2001,pp.84-86.
VI. ACKNOWLEDGEMENTS [5] Wang Wenjian.Improve of BP Neural Networks Model. Computer
Engineering and Design. vol.26,NO 11, Nov. 2000, pp.8-10.
The authors acknowledge support from the National [6] Zeng, X., & Yeung, D. S.Sensitivity analysis of multilayer
perceptron to input and weight perturbations. IEEE Transactions
Science Foundation of China (Grant No.30800663 and on Neural Networks, vol.12,NO.6, Dec. 2001, pp. 1358-1366.
[7] Zeng X,Yeung DS.Hidden neuron pruning of multilayer
Grant No.30971691). perceptrons using a quantified sensitivity measure.Neuro
computing, Vol. 69, Mar.2006, pp. 825-837.
[8] Yang Baohua, Li Shaowen, Fan Jianfeng. Distinction method for
red-seed watermelon quantitative trait based on rough set
.Agriculture Network Information, vol.6,Jun.2006, pp.38-39,58.

Authorized licensed use limited to: McGill University. Downloaded on February 16,2023 at 01:40:43 UTC from IEEE Xplore. Restrictions apply.

You might also like