Professional Documents
Culture Documents
Ijsrcsamsv 7 I 3 P 174
Ijsrcsamsv 7 I 3 P 174
Ijsrcsamsv 7 I 3 P 174
International Journal of Scientific Research in Computer Science Applications and Management Studies
Abstract— Wireless sensor network (WSN) generate a huge degrade the predictive performance of a classification model
amount of data in the form of a stream. Data classification of this [3]. Steam nature of most real world data including wireless
stream is often used to gather knowledge from these data sensor data poses new challenges for the existing classical
collected by the WSN. Most classification algorithm learns from classification techniques like CART [4], ID3 [5], C4.5 [6],
training data at the beginning to develop a classification model
IFN [7],. Methods for extracting patterns from continuous
and uses this model to classify future unclassified data. Most of
these models are unable to incorporate new knowledge at later streams of data are known as incremental (online) learning
stage. It is also seen that in most cases there is a need to algorithms [3]. In most data stream, a complete training data is
incorporate new knowledge as new concepts are discovered at not available at the initial learning step and new training
later stage. Therefore, classification algorithms used for such tuples are introduced to the system at later stage. An
data stream must be able to learn incrementally as new classified incremental classification algorithm should be able to learn
data is introduced, while maintaining its earlier knowledge. Also from such intermittent training data without reconstructing the
considering the limited battery life of sensor node and it’s large classification model.
deployment demands a distributed strategy to overcome the With recent advances in wireless sensor technologies,
limitations of sensor network both in terms of processing
wireless sensor networks are deployed for various purposes.
capability and network life time. To deal with this problem a
noble algorithm for distributed incremental learning of Wireless sensor networks can be used in many applications,
classification knowledge is introduced in this paper. Algorithm is such as wildlife monitoring [8], hazardous environment
based on Naïve Bayesian classification algorithm. Using exploration [10], military target tracking and surveillance [9],
simulated experiment on real data, incremental learning etc. A major constraint in the case of wireless sensor network
capability of the algorithm is presented. is its limited energy as sensors are run by batteries with
limited power and in most cases, replacing of these batteries is
Keywords— Datamining, Wireless Sensor Networks, not possible as they are deployed in large number and mostly
Classification, Naïve Bayesian, in hostile environment. In most of these application, sensor
node continuously sense the physical phenomenon, such as
I. INTRODUCTION temperature, humidity, light intensity and report the same to
Classification is a form of data analysis that extracts models the sink or the base station. Small amount of these data may
describing important data classes [1]. In the process of be used as training data with class label. Later such model can
classification models or classifier is constructed to predict be used to classify unseen data without class label. Since the
class labels. Data classification is a two steps process, the first initial training data is negligible as compared to the enormous
phase is learning step, where classification model is data generated by a wireless senor network, and physical
constructed and the later phase is classification step where phenomenon may change over time, new class label may be
new unknown data is used to predict the class labels. There added later or the concepts of existing class may change over
are many approaches to develop a classification model like time, there is often a need to introduce new training data to
decision tree induction, Bayesian classification methods, update the knowledge base without losing the previously
classification by back-propagation, support vector machines, learned knowledge. This process of generating and updating
nearest neighbor classifier etc. [1]. set of classification rules, without regenerating previous
Decision tree induction is one of the most common results (rules) when new data objects become available is
techniques to solve the classification problem [2]. Most known as incremental learning capability [11].
classification methods are based on the assumption that the Classification techniques can be used in WSN to save
data conforms to a stationary distribution. However, the real- energy of the sensor. Sensor nodes may transmit only the class
world data is usually collected over certain periods of time, label instead of the raw data, which can significantly reduce
ranging from seconds to years, and ignoring possible changes the amount of energy consumed based on the number of
in the underlying concept, also known as concept drift, may parameter each sensor is monitoring. Say for example, a
IJSRCSAMS
Volume 7, Issue 3 (May 2018) www.ijsrcsams.com
ISSN 2319 – 1953
International Journal of Scientific Research in Computer Science Applications and Management Studies
wireless sensor network responsible for monitoring (b) If Ak is continuous-valued, then continuous-valued
temperature, humidity and light need to send three messages attribute is typically assumed to have a Gaussian
for each parameter along with the with its epoch id and sensor distribution with a mean µ and standard deviation 𝜎,
id. Sending only class label will reduce the number of
message by at least 60%. Moreover, no further classification defined by
1 (𝑥 −𝜇 )2
of the sensor data is required at the base station. 𝑔 𝑥, 𝜇, 𝜎 =
−
𝑒 2𝜎 2
In this paper, we present a distributed classification 2𝜋𝜎
algorithm for WSN, where each sensor generate its local so that,
classifier based on a set of training tuple and passes these 𝑃 𝑥𝑘 𝐶𝑖 = 𝑔(𝑥𝑘 , 𝜇𝐶𝑖 , 𝜎𝐶𝑖 )
classifiers to its parent node which uses these information
from its child along with its own training data and produces an
For tuple X, 𝑃 𝑋 𝐶𝑖 𝑃(𝐶𝑖 ) is evaluated for each class Ci, The
enhanced classifier. Ultimately the root node or the sink
classifier predict the class label of tuple X is the class Ciif and
generated a global classifier. Algorithm is also able to handle
only if
future learning from new training data presented at later state.
𝑃 𝑋 𝐶𝑖 𝑃 𝐶𝑖 > 𝑃 𝑋 𝐶𝑗 𝑃(𝐶𝑗 )for1 ≤ 𝑗 ≤ 𝑚, 𝑗 ≠ 𝑖
II. NAÏVE BAYESIAN CLASSIFICATION To avoid the case of probability value of zero Laplacian
Naïve Bayesian Classification is based on Bayes’ theorem. correction or Laplace estimator is used.
In order to simplify computation, Naïve Bayesian classifiers III. NAÏVE BAYESIAN CLASSIFICATION FOR INCREMENTAL
assume that the effect of an attribute value on a given class is LEARNING
independent of the values of the other attributes. This
assumption is called class conditional independence. It has Performance of naïve Bayesian classification is found to be
been found that naïve Bayesian classifier is comparable to comparable with decision tree and neural network in some
decision tree and selected neural network classifier in terms of domain. It exhibit high accuracy in large databases and
its performance. theoretically minimum error rate is found.
Let D be a set of training tuple. Let each tuple be represented Naïve Bayesian classification can be used for incremental
by n-dimensional attributes vector X = {x1, x2, x3,…, xn} learning by updating counts for each attribute in case of
measurements made on n tuples be A1, A2, A3, … , An.Suppose, categorical data and updating sum of values and sum of
there are m classes, C1, C2, C3, …, Cm . Given a tuple, X, the squared values (mean and standard deviation can be calculated)
classifier will predict that X belongs to the class having the for continuous data. When new training data is arrived, it has
highest posterior probability, conditioned on X. That is, the to increment the count or recalculate the mean and standard
naïve Bayesian classifier predicts that tuple X belongs to the deviation for the existing and newly arrived data. In this way,
class Ci if and only if there is a no loss of the earlier information learned and also it
P(Ci|X) > P(Cj|X) for 1≤ i, j ≤ m, j ≠ i. can easy cope with the concept drift. Algorithm1 and
Thus, we maximize P(Ci|X). The class Ci for which P(Ci|X) is Algorithm 2 describe the Incremental Learning of Naïve
maximized is called the maximum posteriori hypothesis. By Bayesian classification for continuous valued attribute and
Bayes’ theorem, Classification process for Naïve Bayesian classification. To
𝑃 𝑋 𝐶𝑖 𝑃(𝐶𝑖 ) understand the working of Incremental classification, training
𝑃 𝐶𝑖 𝑋 = data in Table I and Table II can be used. Tuples in Table I are
𝑃(𝑋)
the initial training tuple while the tuples in Table II are the
P(X) is constant for all class and commonly assumed that all
training tuples which is introduced at later state.
classes are equally likely, i.e. P(C1 )=P(C2)=…=P(Cm).
Table II is the initial training data based on which
Therefore only P(X |Ci) must be maximized. Class prior
classification model is derived. Suppose a new tuple X arrives
probability can be estimated by using
after the initial training.
𝑃 𝐶𝑖 = 𝐶𝑖,𝐷 |/|𝐷| X= (Temperature = 20.09, Humidity = 48.89, Light = 549.9).
In Naïve Bayesian Classification, class conditional
independence is assumed and thus Prior probability P(Ci) of the each class can be computed
𝑛
using the training data,
𝑃 𝑋 𝐶𝑖 = 𝑃(𝑥𝑘 |𝐶𝑖 ) P(Class = 0) = 5/10 = 0.5
𝑘=1 P(Class = 1) = 5/10 = 0.5
= 𝑃 𝑥1 𝐶𝑖 × 𝑃 𝑥2 𝐶𝑖 × … × 𝑃(𝑥𝑘 |𝐶𝑖 )
P (Temperature=20.09 | Class = 0) = 𝑔(20.09, 𝜇𝐶0 , 𝜎𝐶0 )
In order to compute 𝑃 𝑋 𝐶𝑖 , we need to consider the
=0.172317845
following
P (Temperature=20.09 | Class = 1) =𝑔(20.09, 𝜇𝐶1 , 𝜎𝐶1 )
(a) If Akis categorical, then 𝑃(𝑥𝑘 |𝐶𝑖 ) is the number of
=0.133955904
tuples of class Ci in D having the value xk for Ak, P (Humidity = 48.89 |Class = 0) =𝑔(48.89, 𝜇𝐶0 , 𝜎𝐶0 )
divided by |Ci,D|, the number of tuples of class Ci in = 1.87765E-08
D.
IJSRCSAMS
Volume 7, Issue 3 (May 2018) www.ijsrcsams.com
ISSN 2319 – 1953
International Journal of Scientific Research in Computer Science Applications and Management Studies
IJSRCSAMS
Volume 7, Issue 3 (May 2018) www.ijsrcsams.com
ISSN 2319 – 1953
International Journal of Scientific Research in Computer Science Applications and Management Studies
V. EXPERIMENTAL RESULTS
In our simulated experiment, we used real dataset that has
been collected from 54 sensors deployed in the Intel Berkeley
Research lab between February 28th and April 5th, 2004
(http://db.csail.mit.edu/labdata/labdata.html). Data contains
about 2.3 million readings collected over the period. We used
Fig 3 Spanning tree of depth 5
the same location of the sensors and assumed that the sink
node is located at the center of the lab. As there is no
estimation of values or any pseudo data creation and all
50
information sensed by any sensor is preserved, it offers same
Millions
No. of message
VI. CONCLUSION
Use of distributed classification algorithm can preserve
energy of the sensors by reducing the message length of the
sensors. In this paper, we present a noble hierarchical
distributed classification algorithm based on naïve Bayesian
classification. Naïve Bayesian classification technique has
been found scalable with high accuracy and hence it is
suitable for wireless sensor networks as it generates a huge
stream of unbounded data. Moreover, incremental naïve
Bayesian classification can be used to ensure that new
knowledge can be easily incorporated without losing earlier
knowledge. In distributed hierarchical naïve Bayesian
classification, it further increases efficiency of the sensor
networks by increasing its network life time. Since this
algorithm classifies data at each sensor node and transmits
Fig 1 Spanning tree of depth 3. only the class label instead of the raw data, there is no need of
IJSRCSAMS
Volume 7, Issue 3 (May 2018) www.ijsrcsams.com
ISSN 2319 – 1953
International Journal of Scientific Research in Computer Science Applications and Management Studies
running classification algorithm at the base station. Simulated [11] Ning Shan and WojciechZiarko, ―Data-Based Acquisition and
Incremental Modification of Classification Rules,‖ Computational
experiments show that the messages transmitted reduces
Intelligence, Vol. 11(2), pp. 357–370, May 1995.
significantly by using this algorithm. One drawback of this [12] Yu Wang, ―An Incremental Classification Algorithm for Mining Data
algorithm is that naïve Bayesian classifier makes an with Feature Space Heterogeneity,‖ Mathematical Problems in
assumption of class conditional independence but practical Engineering, Feb 2014.
[13] Shaoning Pang, Seiichi Ozawa, and Nikola Kasabov, ―Incremental
scenario, dependencies between variables may exist.
Linear Discriminant Analysis for Classification of Data Streams,‖
IEEE transactions on systems, man, and cybernetics—part b, vol. 35(5),
pp. 905-914, Oct 2005.
REFERENCES [14] Seiichi Ozawa ;Shaoning Pang ; Nikola Kasabov, ―Incremental
[1] J Han, M Kamber, J Pei, Data Mining, Concepts and Techniques, 3rd Learning of Chunk Data for Online Pattern Classification Systems,‖
ed. Morgan Kaufmann, 2012, pp.327-391. IEEE Transactions on Neural Networks, vol.19(6), Jun 2008.
[2] Al-Hegami, Ahmed Sultan, ―Classical and Incremental Classification [15] XuCheng ;JiXu ; Jian Pei ; Jiangchuan Liu, ―Hierarchical distributed
in Data Mining Process,‖ International Journal of Computer Science data classification in wireless sensor networks,‖ IEEE 6th International
and Network Security, Vol. 7(12), pp. 179-187, Dec 2007. Conference on Mobile Adhoc and Sensor Systems MASS '09, Oct, 2009.
[3] M. Last, ―Online classification of nonstationary data streams,‖ Journal [16] Khushboo Sharma, ManishaRajpoot, and Lokesh Kumar Sharma,
of Intelligent Data Analysis, Vol. 6(2), pp. 129-147, Apr 2002. ―Nearest Neighbour Classification for Wireless Sensor Network Data,‖
[4] L. Breiman, J.H. Friedman, R.A. Olshen, & P.J. Stone, Classification International Journal of Computer Trends and Technology, vol. 2(2),
and Regression Trees, Wadsworth, 1984. 2011.
[5] J.R. Quinlan, ―Induction of Decision Trees,‖ Machine Learning, Vol. [17] Seidl T., Assent I., Kranen P., Krieger R., Herrmann J., ―Indexing
1(1), pp. 81-106, 1986. Density Models for Incremental Learning and Anytime Classification
[6] J. R. Quinlan, C4.5: Programs for Machine Learning, Morgan on Data Streams,‖ in Proceedings of 12th International Conference on
Kaufmann, 1993. Extending Database Technology (EDBT/ICDT 2009), pp. 311-322, Jan
[7] O. Maimon and M. Last, Knowledge Discovery and Data Mining, The 2009.
Info-Fuzzy Network (IFN) Methodology, Kluwer Academic Publishers, [18] Paul E. Utgoff, ―Incremental Induction of Decision Trees,‖ Machine
2000. Learning, vol. 4(2), pp 161–186, Nov 1989.
[8] Garcia-Sanchez, et. al. ―Wireless Sensor Network Deployment for [19] Charu C. Aggarwal, Jiawei Han, Jianyong Wang, Philip S.
Monitoring Wildlife Passages,‖ Sensors, Vol. 10(8), pp. 7236-7262, Yu, ―On demand classification of data streams,‖ in Proceedings of the
Aug 2010. tenth ACM SIGKDD international conference on Knowledge discovery
[9] T Bokarevaet. al., ―Wireless sensor networks for battlefield and data mining, pp. 503-508, Aug 2004.
surveillance,‖ in Proceedings of the Land Warfare Conference, Oct [20] http://db.csail.mit.edu/labdata/labdata.html
2006.
[10] G. Werner-Allen, et. al., ―Deploying a wireless sensor network on an
active volcano,‖ IEEE Internet Computing, Vol. 10 (2), pp 18–25, Mar-
Apr, 2006.
IJSRCSAMS
Volume 7, Issue 3 (May 2018) www.ijsrcsams.com