Professional Documents
Culture Documents
Statistical Feature Selection of Narrowband RCS Sequence Based On Greedy Algorithm
Statistical Feature Selection of Narrowband RCS Sequence Based On Greedy Algorithm
contained in their echoes. Statistical features extracted from distance and inter class distance respectively. The results of
the Radar Cross Section (RCS) sequence can be used to target recognition are given in the end.
distinguish different space targets.
The number of RCS sequence samples is very large and 2. RCS statistical features
the samples are imperfect. In order to distinguish different
There are many factors that affect the measurement of
targets in their feature space, as many features are extracted
RCS, such as target attitude, radar wavelength and
as possible. This leads to high dimension of the original
polarization mode of radar wave etc. After a period of
feature space of RCS sequence. For the sake of decreasing the
observation time a sequence of RCS will be obtained. The
computation complexity and eliminating redundant features
statistics of RCS sequence contain a certain amount of
as well as obtaining good target recognition results, it is
information about the target. The statistical features of RCS
necessary to select RCS features in the high dimensional
sequence commonly used are as follows:
feature space. Search method select features that are useful
(1) Mean value: the arithmetic mean value of a sequence x ,
for target recognition according to a particular criterion. It can
be mainly divided into two categories: globally optimal i.e.,
N
search method and locally optimal search method. Exhaustion
method and B&B (Branch and Bound) method [1] are typical
x ¦x i
(1)
i 1
globally optimal search method. Exhaustion method finds the where N is the total number of the sequence;
globally optimal feature combination among the ergodic of (2) Median value: the number in the middle position of a
feature combinations. But if the dimension of feature space is sorted sequence x ' . That is to say, if the total number of
too high, it will be time-consuming and maybe impossible for the sequence N is odd,
Exhaustion method to find globally optimal feature
combination. Although the computation amount of B&B x median
(x N/2
x N / 2 1
)/2 (2)
method is less than Exhaustion method, B&B method if N is even,
requires the monotonicity of separation criterion. In fact, if x (x x )/2
N / 2 1
(3)
the separation criterion is monotonic in theory, the actual median N/2
___________________________________
978-1-4244-8443-0/11/$26.00 ©2011 IEEE
Authorized licensed use limited to: Indian Institute of Space Science And Technology. Downloaded on February 22,2022 at 10:26:55 UTC from IEEE Xplore. Restrictions apply.
(3) Truncated mean value: the arithmetic mean value of a Greedy algorithm is an efficient search method, and it is a
sorted sequence x ' with the first several numbers and good alternative for feature selection in high dimension
the last several numbers truncated, i.e., feature space. The basic idea of greedy algorithm for feature
N 1 k selection begins from an original feature combination, then
x tr ¦x i (4) generates a new feature combination from the adjacent area of
i k the original solution randomly, calculates the variation of cost
where k is the number of data to be truncated. function 'E for these two feature combinations, if 'E 0 , the
(4) Minimum value: the minimal value of a sequence, i.e., new feature combination is accepted, if 'E t 0 , the original
x (5) min{ x } feature combination is reserved.
min i
The steps of Greedy algorithm are as follows:
(5) Maximum value: the maximal value of a sequence, i.e., 1. Determine the range of dimension m to be [1, M] , where
x max
max{ x } i
(6) M is the total dimension of high dimensional feature
(6) Extreme difference: the difference between the space. Generates m integers within [1, M] which
maximum value and the minimum value, i.e., represent the original feature combination m , calculate 0
(7) Standard deviation: also known as Mean Square Error, 2. Regenerate m '(1 d m ' d M) integers within [1, M] to be the
i.e., new feature combination m , calculate the corresponding
1
N
1
¦x x (8) value of cost function E ( m ) . The variation between
2
S i
1
N 1 i 1
E ( m ) and E ( m ) is:
1 0
(8) Average absolute deviation:
'E E ( m ) E ( m ) (16)
1 0
1
P abs
N
¦ xi x (9) 3. If'E t 0 , the original feature combination is reserved, if
'E 0 , the new feature combination is accepted, then set
i
(14) Coefficient of kurtosis: measurement for peak value of of real RCS sequences.
sequence probability density function, i.e.,
2 4.1. Intra class distance
b k
B /B 4 2
(15)
where Bq ( q 2, 4) stands for second-order central Intra class distance stands for the degree of dispersivity
moment when q 2 and fourth-order central moment of samples of same class. Suppose that N kinds of targets are
when q 4 . to be recognized, and the i th sample of the l th kind of target
is denoted as xli (l 1, 2,! , N ; i 1, 2,! , N l ) , where N l is
3. Greedy algorithm the length of samples. Every sample xli LV D /
Authorized licensed use limited to: Indian Institute of Space Science And Technology. Downloaded on February 22,2022 at 10:26:55 UTC from IEEE Xplore. Restrictions apply.
GLPHQVLRQDOYHFWRULH xli ^ xli (1), xli (2),! , xli ( L)` Nk
1 Nl p ¦ p , (k 1,ĂĂ, N ) (25)
¦
k k ,l
m l
x , ( xi : l , l 1, 2, "" , N )
i
(18) N 1 l 1, l z k
N l i 1
For feature combination Z , the average degree of overlapping
The dispersivity degree of the samples of the l th kind target among the value of different kinds of RCS sequence features
is : is:
Nl N
1 1
Dl ¦ ( x 'li ml ) 2 ( x 'li : 'l , l 1, 2,"" , N ) (19) p
N
¦ p , (k 1,ĂĂ, N ) (26) z k
Nl i 1 k 1
and the total dispersivity degree of the sample set :l' , namely,
5. Experiment and result analysis
the total within-class distance is :
N
In this section, experiment on selecting statistical
D ¦ D , (l 1, 2,"" , N )
intra l
(20) features of real RCS sequence is carried out, and the results
l 1
Dinter ¦ Dlk . (23) randomly, and save these integers into a matrix as a
l , k 1, l z k
feature combination. Repeat this process for
N c ( N c C14m ) times, then eliminate the repeated feature
3.1.3. Purity parameter---- p
combinations;
Purity parameter represents the degree of overlapping (3) Set m m 1 ;
among different target features. Its value is within 0 ~ 1 , the (4) Repeat step (2) and step (3) until m 14 . The matrix
smaller the purity parameter the smaller the degree of with feature combination selected randomly is complete.
overlapping among different target features. (5) Calculate the value of cost function of all the feature
Suppose the RCS sequences of N kinds of targets have combinations in the matrix; choose the corresponding
feature combination of the smallest value of cost
been obtained, and the statistical features of RCS sequences
function.
also have been extracted. The degree of overlapping between
The flow chart of a single time experiment is as follows:
the k th kind of feature combination samples and the l th kind
of feature combination samples is:
Authorized licensed use limited to: Indian Institute of Space Science And Technology. Downloaded on February 22,2022 at 10:26:55 UTC from IEEE Xplore. Restrictions apply.
suboptimal feature combination through Greedy algorithm,
START
for the highest recognition rate is 72.72% which is obtained
through Exhaustion method.
initialize the index of
dimension˖m 1
5. Conclusions
In this paper, we aim to find a proper cost function to
generate m integers select optimal/suboptimal feature combination in high
dimension feature space. The efficient search method, Greedy
save them into a matrix algorithm, is adopted in selecting optimal/suboptimal
statistical feature combination of narrowband RCS sequence,
and the selected statistical feature combination is used for
no reach N c
space target recognition. Although there is limited
times? information about the space target contained in narrowband
RCS sequence, the greedy algorithm with the cost function
yes p can select the suboptimal feature combination to
m m 1 distinguish different space targets and obtain good target
recognition results. In fact, this new cost function can also be
no used to select features in other applications.
m 14?
yes
References
calculate the value of cost function [1] Zhaoqi Bian, Xuegong Zhang, Pattern Recognition (2nd
of all the feature combinations
Edition), Press of Tsinghua University: Beijing, 2000
[2] KuanChieh Huang, YauHwang Kuo, ICheng Yeh, “A
Novel Fitness Function in Genetic Algorithms to
choose the corresponding feature combination
of the smallest value of cost function Optimize Neural Networks for Imbalanced Data Sets”,
8th International Conference on Intelligent Systems
Design and Applications, 2008, pp.647-650.
END [3] Hai Deng, Polyphase code design for Orthogonal Netted
Radar systems, IEEE transactions on signal processing,
Fig.1. Flow chart of a single time experiment Vol.52, No.11, November.2004.
[4] Joonmin Gil, Chanmyung Kim, Younhee Han, “A
5.2. Results and analysis Greedy Algorithm to Extend the Lifetime of Randomly
Deployed Directional Sensor Networks”, 2010
With the selected optimal feature combination, the
Proceedings of the 5th International Conference on
nearest-neighbour fuzzy classifier [6] works on the 77
Ubiquitous Information Technologies and Applications
narrowband RCS sequences for test. The target recognition
(CUTE 2010), pp.1-5.
results are showed with the corresponding feature
[5] Jiawei Gao, Researeh and APPlications of Classifieation
combination in table 1 and table 2.
Algorithms in Imbalanced Data Sets, Shanxi Universit:
Shanxi, 2008.
Table 1. Optimal Feature Combination of the 1st Cost
[6] Xiankang Liu, Meiguo Gao, Xiongjun Fu, “A Nearest
Function and Results of Target Recognition
Neighbour Fuzzy Classifier for Radar Target
Feature Combination Recognition Rate
Recognition Using Combined Features”, 8th
12 27.27%
International Conference on Signal Processing, vol
1 12 49.35% 3,2006 .
1 28.57%
12 27.27%
1 12 49.35%
Authorized licensed use limited to: Indian Institute of Space Science And Technology. Downloaded on February 22,2022 at 10:26:55 UTC from IEEE Xplore. Restrictions apply.