Professional Documents
Culture Documents
NILM 2017 Paper
NILM 2017 Paper
NILM 2017 Paper
Research Article
Nonintrusive Load Monitoring Based on Advanced Deep
Learning and Novel Signature
Received 5 May 2017; Revised 2 August 2017; Accepted 21 August 2017; Published 2 October 2017
Copyright © 2017 Jihyun Kim et al. This is an open access article distributed under the Creative Commons Attribution License,
which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Monitoring electricity consumption in the home is an important way to help reduce energy usage. Nonintrusive Load Monitoring
(NILM) is existing technique which helps us monitor electricity consumption effectively and costly. NILM is a promising approach
to obtain estimates of the electrical power consumption of individual appliances from aggregate measurements of voltage and/or
current in the distribution system. Among the previous studies, Hidden Markov Model (HMM) based models have been studied
very much. However, increasing appliances, multistate of appliances, and similar power consumption of appliances are three
big issues in NILM recently. In this paper, we address these problems through providing our contributions as follows. First, we
proposed state-of-the-art energy disaggregation based on Long Short-Term Memory Recurrent Neural Network (LSTM-RNN)
model and additional advanced deep learning. Second, we proposed a novel signature to improve classification performance
of the proposed model in multistate appliance case. We applied the proposed model on two datasets such as UK-DALE and
REDD. Via our experimental results, we have confirmed that our model outperforms the advanced model. Thus, we show that
our combination between advanced deep learning and novel signature can be a robust solution to overcome NILM’s issues and
improve the performance of load identification.
Oven element
long-range power pattern result, we can solve the previous
2
Refrigerator problems, which are mentioned in Section 3. We show
Oven element that the proposed model outperforms via the experimental
1 results with UK-DALE and REDD. Second one is a discovery
Refrigerator
of the novel signature. The proposed signature raises the
0 performance of classification for the multistate appliance. By
0 10 20 30 40 emphasizing the power variation when we train the model,
Time (min) the variation could be clearly trained. We show the efficiency
Figure 1: Basic concept of NILM [8]. via the validation experiments in Section 5.3.
This paper is organized the structure as follows. We
give problem statement from related works and proposed
solutions in Section 2. Section 3 describes two NILM datasets
current, and voltage (Figure 1). He showed how on-off events to be used in our experiments. In Section 5, we conduct two
were sufficient to characterize the use of some appliances. main experiments. The first experiment is about validating
Many researchers have studied this concept and improved the the expected effects of the novel signature and learning
NILM model. architecture. The second experiment is about measuring the
In order to disaggregate a power consumption based on overall performance and comparing with the existing models
NILM, we have to understand features of appliances. Hart and state of the art. Final section, we summarize our works
defined three types of appliance model according to the and provide the conclusion.
features. Three types are On/Off, Finite State Machine (FSM),
and Continuously Variable [8, 9]. 2. Related Work
The first type On/Off has binary states of power such as
a light and a toaster. Figure 2(a) shows the power pattern of 2.1. Signature Based Approach. There are two different kinds
light. When it is turned on, the power level is increased from of signature. The first signature is the steady state. It makes
10 to 12 watts. The appliances of type 1 can be easily classified use of steady-state features that are derived under the steady-
because of a clear feature. However, if two appliances which state operation of the appliances. Hart proposed the concept
have similar power consumption are operating, they could of NILM and power change method at the same time in his
not be distinguished. In this case, a classification model does paper [8]. In this method, real power (𝑃) and reactive power
not know which appliance is generating the current power. (𝑄) are used as the input signature. By computing changes of
Therefore, we need an additional factor to distinguish them. 𝑃 and 𝑄, the appliances are classified. Hart demonstrated how
The second type FSM has multiple states of power such different electrical appliances generated distinct power con-
as a lamp and a fan. The power pattern of lamp is shown in sumption signatures. He showed how on/off events were suf-
Figure 2(b). Three operation states are in lamp according to ficient to characterize the use of some appliances. The advan-
brightness. This kind of appliances can be modeled by FSM. tage of this method is that we can use a low sampling dataset
So it could be simple to classify a single appliance of type 2. and it is easy to identify the appliances having high power
However, when the multiple power consumption is summed, consumption. However, it is hard to classify the appliances
the classification is become complicated, because it is hard to having low power consumption and the multistate appliances
figure out that the current power is originated in the variable which are types 2 to 4 in Figure 2. In addition to that,
pattern. Therefore, we need to observe the pattern for a period the appliances having similar power consumption cannot be
of time in order to classify the appliances. classified. To improve NILM, other researches have attempted
The third type has a continuous changeable power pattern and proposed alternative signature identification techniques.
such as a washing machine and a light dimmer. In Figure 2(c), Najmeddine et al. and Figueiredo et al. used the current 𝐼
the power consumption is fluctuated while heating/washing and the voltage 𝑉 as the signature [11, 12]. They extracted
or rinsing/drying a laundry. Strictly speaking, it is a multiple the features such as Root Mean Square (RMS) of the cur-
state appliance but it cannot be modeled due to infinite rent and peak. These features are well suited for classifying the
middle states. To classify type 3 appliances, we have to appliances in the kitchen. However, the multistate appliances
understand the feature and observe the long-range pattern. cannot be classified as ever. One of the signatures extracted
Lastly, we define an additional type Always On. The appli- from high sampling rate data is the harmonic of current.
ances of type 4 are always operating except a special case. There are several researches acquiring the harmonics via
For example, a refrigerator in Figure 2(d) is operating consis- Fourier series [13–16]. The appliances of types 1 and 4 are
tently. The type 4 appliances could have a periodic pattern or a well classified by current harmonics. But it requires the
single pattern. According to circumstances, we can reduce the harmonics set of all combinations of the appliances. As
number of appliances when we train the classification model. increasing the number of appliances, the harmonics for
The main contributions in this work consist of two points. classification increase exponentially. This is not a practical
First one is a construction state-of-the-art NILM model. The approach and the memory problem could occur. There are
Computational Intelligence and Neuroscience 3
12 20
10
15 State 3
Power (watts)
Power (watts)
8
6 On 10 State 2
4 Off
Off
5 State 1
2
Off
0 0
0 100 200 300 400 500 600 700 800 900 0 500 1000 1500 2000 2500 3000 3500
Time (seconds) Time (seconds)
Type 1 Type 2
(a) (b)
2500 140
120
2000
Power (watts)
100
Power (watts)
Freezing
1500 Heat, wash 80
1000 60 Defrosting
Rinse, dry 40
500 Off 20
0 0
0 10 20 30 40 50 60 70 80 90 0 50 100 150 200 250 300 350 400
Time (seconds) Time (seconds)
Type 3 Type 4
(c) (d)
Figure 2: Four types of appliances. (a) Type 1 (On/Off): light; (b) type 2 (FSM): lamp; (c) type 3 (Continuously Variable): washing machine;
(d) type 4 (Always On): refrigerator.
fancy approaches for NILM research. Lam et al. proposed types of noise: on/off transient noise, steady-state line voltage
the appliance categorization method based on the shape of noise, and steady-state continuous noise. However, to use this
𝑉 − 𝐼 trajectory [17]. They categorized the appliances to eight method, we need to have knowledge about the power flow
groups with high accuracy. Gupta et al. tried to identify the such as the reactive power, the active power, and phase of
appliances based on the noise generated from the operation voltage relative to current. The method using the voltage noise
of appliance [18]. However, the noise is easy to be affected by is suitable for classification of the multistate appliances.
the environment. So it cannot be high efficient signature. In
addition, we need to equip the device for measuring the noise. 2.2. Learning Model Based Approach. Some useful temporal
Therefore, it cannot be practical method. graphical models, the variants HMM, are used for NILM.
The second signature is the transient state. This signature Zoha et al. proposed a solution using FHMM to recog-
is less overlapping in comparison with steady-state signa- nize appliance specific load patterns from the aggregated
tures. However, high sampling rate requirement in order to power measurements [25]. They defined five features that
capture the transients is the major disadvantage. Chang et are combinations of the power measurement such as the
al. showed that the power consumption while an appliance real power, reactive power, and voltage waveforms. They
is turning on can be calculated as a signature [19]. Five achieved the 𝑓-measure of 0.614 for five appliances when
years later, they identified the transient physical behavior they use multistate FHMM. Kolter and Johnson also used
of power by using wavelet transform [20]. Another way for FHMM for NILM [26]. They had collected their own dataset
classification, Leeb et al. proposed the method analyzing the called REDD which will be explained in the Section 4. The
spectral envelope via Short Time Fourier Transform (STFT) average accuracy was 47.7%. To improve the researches using
[21]. In this way, the method of using transient power as FHMM, a number of researchers have extended FHMM. Kim
a signature is suitable for classification of types 1 to 3. But et al. developed probabilistic models of appliance behavior
it cannot catch type 4. Norford and Leeb showed that the using variants of FHMM [27]. The variants are Factorial
shape of transient data can be a feature [22]. Cole and Albicki Hidden Semi-Markov Model (FHSMM), Conditional FHHH
used power spikes generated from transition states [23]. This (CFHMM), and Conditional FHSMM (CFHSMM). Figure 3
signature is efficient for classification but it is applied to shows the relationships between the variant FHMMs and
several specific appliances. Like this, the method of using their performance comparison.
the current transient when the appliances are turning on Kolter and Jaakkola used a combination of addi-
is suitable for classification of types 1 and 2. But it cannot tive FHMM and difference FHMM [28]. For inferencing,
catch types 3 and 4. Patel et al. sampled the voltage noise they proposed an Additive Fractional Approximate MAP
extracted from the transient event [24]. They defined three (AFAMAP) algorithm. They achieved the recall of 0.603 for 7
4 Computational Intelligence and Neuroscience
1.0
0.9
0.8
F-measure
0.7
Distribution
FHMM shape FHSMM
0.6
Additional
features
0.5
CFHMM CFHSMM 2 3 4 5 6 7 8
The number of appliances
Model
CFHSMM FHSMM
CFHMM FHMM
(a) (b)
Figure 3: Variants of FHMM and performance results [10]. (a) Relationships between the variant FHMMs; (b) performance comparison.
appliances. Parson el al. proposed the variant of the difference sample rate data for NILM, this problem might be
HMM and the extended Viterbi algorithm [29]. Zeifman solved due to catching a momentary change of power
proposed a Viterbi algorithm with Sparse Transitions (VAST) consumption. But there is a limitation for the real life
[30]. And he used Markov chain (MC) for NILM. because a high cost device is needed for collecting the
high sample rate data.
3. Problem Statement and Proposed Solution (iii) The third problem is a difficulty of classifying the
appliances having similar power consumption. If we
3.1. Problem Statement. The data for NILM is collected in a
use only power consumption as an observation of
timed sequence. Therefore, the models dealing with a sequen-
FHMM, the model could not distinguish the similar
tial data are used as an appliance classification model. The
power appliances.
models which are usually used are Factorial Hidden Markov
Model (FHMM) based on the Markov process and its variants We surveyed the basic models and the related works of
[32]. However, these models have three problems to classify NILM so far. However, there are some problems from the
the appliances. The problems are as follows: original basics of NILM. In this section, we summarized them
(i) The first problem is about the time complexity. In case into three.
of FHMM, the time complexity is 𝑂(𝑀𝑛𝑀+1 𝑇) where Problem 1. The researchers have used FHMM and the FHMM
𝑀 is the number of appliances and 𝑛 is the possible variants for NILM [25–27, 29]. As we know, the time
states of hidden unit and 𝑇 is the length of observation complexity of FHMM is 𝑂(𝑀𝑛𝑀+1 𝑇) where 𝑀 is the number
sequence. As the number of appliances is increased, of appliances and 𝑛 is the possible states of hidden unit and 𝑇
the time complexity is increased exponentially as is the length of observation sequence. Therefore, as increasing
well. This leads to reduction of the classification the number of appliances, the time complexity also increases
performance of the model. exponentially. As a result, the performance of classification
(ii) The second problem is about the difficulty of classify- model is dropped. We can easily notice that in Figure 3(b).
ing the multistate appliance. The multistate appliance
is the appliance which has the multiple state of Problem 2. The second problem is that FHMM and the
power consumption. To distinguish them, a long- FHMM variants are hard to classify type 2 to 4 appliances.
range pattern should be trained. However, the existing A long-term pattern is needed to learn but most models are
models are based on the first-order Markov process. based on the first-order Markov chain [35]. So these models
Therefore, the pattern could not be trained efficiently. predict the current state through the previous state. As a
In addition, a low sample rate data (e.g., a collected result, FHMM and the FHMM variants have a limitation
data by 1∼6 Hz sample rate) is not proper for clas- for classifying the multistate appliances. In addition, there is
sifying the multistate appliances. If we use a high another problem based on the signature perspective. In the
Computational Intelligence and Neuroscience 5
signature based approach, we can separate the low sampling 3.2. The Proposed Solution
rate signature and the high sampling rate signature. The
advantage of first signature is that we can easily collect the 3.2.1. Proposed Novel Signature. Sequence data selected from
data from a simple sensor. However, the low sampling rate real life might be power consumption sampled at 1 to 6 Hz.
signature was unsuitable for classifying types 2 to 4 in the Because of the cost of a sensor, there is a limitation to
most researches [8, 11, 12]. To solve this problem, we need collecting high sample rate data. To popularize NILM, we
more precise data and high sampling rate signature. By this have to use low sample rate data from such a sensor. In
signature, the researchers classified all types of appliances. Section 3, we acknowledge that it is hard to classify multistate
However, a specific device should be equipped in order to appliances (type 2, type 3, and type 4) when using low sample
get the signature. It means that the high cost is required rate data. To solve the problem, we propose a novel signature,
for application of the real life. This is not a good way for which is a key input feature. The main idea of the signature is
popularizing NILM. to separate the original signal, which is low sample rate data
(power consumption), using a reflection rate and to subtract
Problem 3. The last problem is the difficulty of classifying one variant power signal from the other variant power signal.
the appliances which have similar power consumption. In the We denote the difference as Δ 𝑝 which represents variation
previous models, the observation was only a series of power of the original signal. The purpose of Δ 𝑝 is to emphasize
consumption. HSMM deals with a duration of operation but the variation of multistate appliances by applying the same
how can we know the durations for each appliance? The importance with the original signal to a training model. The
durations are not stable and it is a very tiresome business result is that performance of the model for an appliance with
to extract the durations for each appliance. For example, in many variations may be better than when using only the
Figure 4, we extracted the patterns of air-conditioner from original signal.
one house. We can notice that the duration is different. It To calculate Δ 𝑝 , we need to generate the variant power
depends on a user’s behavior. signal. The purpose of using the variant power signal is to
reduce noise. An unintended noise can occur and it can have
Through summarizing the problems, we concluded that a negative influence on learning the signal pattern. We can
the problems from FHMM and the FHMM variants are regulate noise with the reflection rate. The reflection rate
the fundamental issues from the architecture of FHMM. ranges from 0 to 0.99. Power signals that increase or decrease
Therefore, we propose a deep learning based NILM model. gradually are reflected to the variant power signal, while
Also, we propose the novel low sampling signature in order to noise, which occurs in a moment, is not reflected well. This
apply to the real life. We have three challenges for successful lower reflection rate reduces the negative effect of momentary
application of deep learning. noise. However, if the reflection rate is too low, the original
power pattern may not be reflected. Thus, it is vital to set the
Challenge 1. Our model must be robust even though the number reflection rate properly. Through the experiment explained
of appliances is increased. in Section 5.3.1, we empirically found that 0.1 or 0.01 is the
The significant problem of NILM models based on proper reflection rate.
FHMM is that the performance is decreased as increasing Algorithm 1 shows how to generate the variant power
the number of appliances. This will be the biggest obstacle for signal. There are two inputs for the algorithm: 𝑃 is the
applying NILM to the real life. Therefore, the proposed model series of the original signal with the time length 𝑇 and 𝛼 is
should be robust even though the number of appliances is the reflection rate, which is a ratio for reflecting variation
increased. between the original signal and the variant power signal.
The output 𝑉𝑃 is the variant power signal generated by the
Challenge 2. The multistate appliance should be classified. algorithm. In line (1), the variant power signal is initialized to
zero. The generation process is given in lines (2) to (5). After
The problem when we use a low sample rate data is that it the variation at time 𝑖 is calculated in line (3), it is reflected to
is hard to classify the multistate appliances. This is caused by the variant signal at reflection rate 𝛼 in line (4). Figure 5 is an
the fact that various patterns of appliance cannot be reflected example of the variant power signal. Figure 5(a) is the series
efficiently to the main signal (total power consumption) due of the original signal with the time length 5000. Figure 5(b) is
to a wide sampling gap. Although it is type 1 appliance, the variant power signal when the reflection rate is 0.1. Note
the main signal could be a multistate signal because of that the pattern is smoother than the original. This indicates
overlapping signal for each appliance. So we must classify the that momentary noise has been reduced while long-range
multistate appliances. variations remain. Figure 5(c) is the variant power signal
when the reflection rate is 0.01. In this case, the original signal
Challenge 3. Although there are the appliances having the changes are more slowly reflected to the variant power signal.
similar power consumption, they should be classified. With two different variant power signals, we can calculate the
If power consumption is the same between two appli- variations in the original signal.
ances, they would be the same appliances. Except this case, The algorithm for the novel signature is in Algorithm 2.
the appliances having the similar power consumption can Two variant power signals are the input and the novel
slightly be different. For the accurate performance, we need signature representing the series of variation is the output.
to distinguish the similar appliances. Δ 𝑝 is calculated by subtracting the inputs. In the scale of the
6 Computational Intelligence and Neuroscience
3500
2500
2000
1500
1000
500
0
1
86
171
256
341
426
511
596
681
766
851
936
1021
1106
1191
1276
1361
1446
1531
1616
1701
1786
1871
1956
2041
2126
2211
2296
2381
2466
2551
2636
2721
2806
2891
2976
3061
3146
3231
3316
3401
3486
3571
3656
3741
3826
3911
3996
4081
4166
4251
4336
4421
4506
4591
4676
4761
4846
4931
5016
Time (sec)
(a)
3500
Power consumption (watt)
3000
2500
2000
1500
1000
500
0
1
39
77
115
153
191
229
267
305
343
381
419
457
495
533
571
609
647
685
723
761
799
837
875
913
951
989
1027
1065
1103
1141
1179
1217
1255
1293
1331
1369
1407
1445
1483
1521
1559
1597
1635
1673
1711
1749
1787
1825
1863
1901
1939
1977
2015
2053
2091
2129
2167
2205
2243
2281
2319
2357
2395
2433
2471
2509
2547
2585
Time (sec)
(b)
3000
Power consumption (watt)
2500
2000
1500
1000
500
0
1
132
263
394
525
656
787
918
1049
1180
1311
1442
1573
1704
1835
1966
2097
2228
2359
2490
2621
2752
2883
3014
3145
3276
3407
3538
3669
3800
3931
4062
4193
4324
4455
4586
4717
4848
4979
5110
5241
5372
5503
5634
5765
5896
6027
6158
6289
6420
6551
6682
6813
6944
7075
7206
7337
7468
7599
7730
7861
7992
8123
8254
8385
8516
8647
8778
8909
9040
9171
9302
9433
9564
Time (Sec)
(c)
Figure 4: Different durations of air-conditioner: (a) duration: 5016 sec; (b) duration: 2585 sec; (c) duration: 9564 sec.
original signal, small variations are easily ignored. However, if gradient problem. We call this model LSTM-RNN. There are
we give Δ 𝑝 the same weight as the original signal, these small two reasons for applying deep learning to NILM. The first
variations become more apparent than before. This effect of reason is that the time complexity was getting higher as the
the novel signature will be validated in Section 5.3.2. number of appliances was increased in the FHMM variants.
Figure 6 represents a computation of Δ 𝑝 . By Algorithm 1, The time complexity of FHMM is 𝑂(𝑀𝑛𝑀+1 𝑇) where 𝑀 is
the variant power signals (Figure 6(b)) are generated from the number of appliances and 𝑛 is the possible states of hidden
the original signal (Figure 6(a)). We can see that the patterns unit and 𝑇 is the length of observation sequence. As we can
of the variant power signal differ according to the reflection see, the number of appliances has a close relation with the
rates. The variant power signals in Figure 6(b) can be the time complexity. As a result, the performance was dropped
inputs for Algorithm 2. Finally, the series of variations, Δ 𝑝 , as the number of appliances was increased. In deep learning,
is generated by Algorithm 2. the major time complexity is originated from backpropa-
gation. Stochastic Gradient Decent (SGD) is usually used
3.2.2. Learning Architecture. In this section, we introduce for backpropagation [36]. As a result, the time complexity
how to apply deep learning to NILM. We choose RNN as of SGD is 𝑂(𝑚𝑑). In our model, the output dimension is
the learning algorithm. RNN can learn the sequential data the number of appliances. So the time complexity is not
such as the power consumption of household. And we apply increased exponentially in contrast with FHMM. If we use
LSTM to hidden layer in order to prohibit the vanishing the deep learning based NILM model, the model could be
Computational Intelligence and Neuroscience 7
4500 6000
4000
3500 5000
Power (watts)
Power (watts)
3000 4000
2500
2000 3000
1500 2000
1000
500 1000
0 0
0 1000 2000 3000 4000 5000 0 1000 2000 3000 4000 5000
Time (seconds) Time (seconds)
4000
3000
2000
1000
0
0 1000 2000 3000 4000 5000
Time (seconds)
3000 3000
2500 2500
Power (watts)
Power (watts)
2000 2000
1500 1500
1000 1000 Delta
500 500
0 0
0 2000 4000 6000 8000 10000 0 2000 4000 6000 8000 10000
Time (seconds) Time (seconds)
600
400
200
0
−200
−400
−600
0 2000 4000 6000 8000 10000
Time (seconds)
Delta(P)
(c) Pattern of Δ 𝑝
450
Power consumption (watt)
400
350
300
250
200
150
100
50
0
1
7
13
19
25
31
37
43
49
55
61
67
73
79
85
91
97
103
109
115
121
127
133
139
145
151
157
163
169
175
181
187
193
199
205
211
217
223
229
235
241
247
253
259
265
271
277
283
289
295
301
307
313
319
325
331
337
343
349
355
361
367
373
379
385
391
397
403
409
415
421
427
433
439
445
451
457
463
469
475
481
487
493
499
Time (sec)
ot−1 ot ot+1
Act as memory
ℎt−1 ℎt ℎt+1
xt−1 xt xt+1
Require: 𝑃 = 𝑝1 , 𝑝2 , . . . , 𝑝𝑇 , Reflection rate 𝛼 Require: VP1 = vp11 , vp12 , . . . , vp1𝑇 , VP2 = vp21 , vp22 , . . . , vp2𝑇
Ensure: VP = vp1 , vp2 , . . . , vp𝑇 Ensure: Δ 𝑝 = 𝛿1 , 𝛿2 , . . . , 𝛿𝑇
(1) VP ← 0 (1) for 𝑖 = 2 to T do
(2) for 𝑖 = 2 to T do (2) 𝛿𝑖 ← vp1𝑖 − vp2𝑖
(3) 𝑑 ← 𝑝𝑖 − vp𝑖−1 (3) end for
(4) vp𝑖 ← vp𝑖−1 + 𝛼 ⋅ 𝑑 (4) return Δ 𝑝
(5) end for
(6) return VP
Algorithm 2: Computing algorithm for the novel signature.
important thing when using LSTM-RNN model is how to means that we need to classify the multiple classes. Therefore,
choose the suitable factors of this model. Here we discover we set the boundary with assuming that the appliance is
four factors, which are considered much effective in our turned on when the output unit value is bigger than the
model. boundary. To do this, we change the softmax to the hyperbolic
tangent (tanh) as the activation function. By this way, we
Preprocessing Method. Generally, 𝑍-Score method and can classify the multiple appliances at the same time. The
Min–Max Scaling are considered as a normalization method sigmoid can also be the activation function. However, while
for deep learning. In deep learning, a model has better per- the output of sigmoid is not zero-centered, the tanh has the
formance when a distribution of input data is close to zero-centered output. As a result, we could set the static
Gaussian distribution. Min–Max Scaling is the most simple boundary when we use tanh as the activation function.
way to normalize the input data. It converts the scale in range
from 0 to 1 by using the minimum and maximum value. Optimizer. There are three optimizers for SGD. One of them
This method holds the original distribution but it cannot be is Adagrad. It decreases a learning rate when a signal is
the Gaussian distribution. 𝑍-Score method normalizes the frequent or the weight has a high gradient. In the opposite
input data by using the mean and standard deviation. This case, it increases the learning rate. However, this method has
method cannot hold the original distribution when the input a low performance due to decreasing the learning rate fast.
distribution is not the Gaussian. Nevertheless, it is import- RMSProb solves this problem by an exponential moving aver-
ant that the mean of input data is close to zero, because age. Adam is a combination of RMSProb and Momentum. It
if the input is all positive or negative, the weight could be is known as the most efficient method for SGD. Therefore we
updated in one way. In case of NILM, one of inputs, power use Adam as an optimizer.
consumption, is all positive value. Therefore, we use 𝑍-Score The inputs are the original signal (power consumption)
as the preprocessing method. We are going to take an experi- and Δ 𝑝 . The output unit represents the on/off state of each
ment about the preprocessing method in Section 5.4.1. appliance. By using the novel signature Δ 𝑝 , we can expect the
better performance for classification of multistate appliances.
Weight Initialization Method. Asymmetry is the most impor- We apply dropout between the input layer and the hidden
tant thing for the weight initialization. We can simply think layer in order to prevent overfitting. This is the simplest
that if the weights are initialized to zero, the update could be way to regularize the weights and is suitable for deep neural
clear. But soon we realize that this method is wrong. There networks like an RNN. We set the number of hidden units
are only bias values in forward propagation. The simplest way to be double the number of output units. A small number
to initialize the weights with asymmetry is selecting values of hidden units require more training epochs, whereas a
from normal distribution or uniform distribution. However, larger number of hidden units require more computation
this method has the problem that a variance of output time per epoch. We found that double the number of output
increases with an input dimension. To solve this problem, units is a proper number for the NILM model through
the variance of output needs to be divided by the square root heuristic experiments. Based on these methods, we construct
of input dimension. By this way, Glorot initialization and He the architecture of an LSTM-RNN for NILM (Figure 8).
initialization are commonly used for the weight initialization.
The simple equations of them are as follows:
4. Dataset Description
random (fanin , fanout ) There are several public datasets for NILM. Table 1 shows
𝑊glorot = ,
√fanin descriptions for them. Among them, we concentrated on two
(1) datasets such as REDD and UK-DALE because of proper
random (fanin , fanout ) for NILM. Beside the public dataset, we had collected 6
𝑊he = ,
√fanin /2 appliances and aggregate data during 5 months. We are going
to use our private dataset as well.
where fanin and fanout are the number of input and output The UK-DALE was released in 2014 at first and it has been
dimensions and 𝑟𝑎𝑛𝑑𝑜𝑚 is a function of random selection updated every year [38]. There are power data for 5 house-
in range from fanin to fanout from normal distribution or holds and each house has the different first measurement
uniform distribution. Both equations are similar but He date. House 1 includes 3.5-year power data for 52 appliances.
initialization selects the weight values in wide range relatively. The 16 kHz data will not be used in this paper. And houses 1,
In deep neural network like RNN, the weight can easily be 2, and 5 include 1-second data but there are no labeled data.
shrunk to nothing when the values are close to zero. This leads Therefore we will not use 1-second data as well. We give an
to vanishing gradient problem. Therefore, we use He initiali- explanation for the 6-second data. Table 2 shows the detailed
zation due to the wide range value. A related experiment will description of UK-DALE. House 1 has the large number
be conducted in Section 5.4.2. of appliances. As the number of appliances is increased, a
NILM model is hard to distinguish the appliances. But the
Activation Function. In general, the softmax is used as an disadvantage can be overcome because of much data. On the
activation function for multiclass classification [37]. It can other hand, houses 3 and 4 have the small number of meters.
select a one class among all classes. However, in NILM, many The difference between of them is that a meter of house 4
appliances could have been operated in the same time. It is shared. For example, “tv dvd digibox lamp” means that
10 Computational Intelligence and Neuroscience
Dataset Number of houses Duration per house Appliance sample frequency Aggregate sample frequency
REDD (2011) 6 3–19 days 3 sec 1 sec, 15 kHz
BLUED (2012) 1 8 days N/A 12 kHz
Smart (2012) 3 3 months 1 sec 1 sec
Tracebase(2012) N/A N/A 1–10 sec N/A
Sample (2013) 10 7 days 1 min 1 min
HES (2013) 251 1 or 12 months 2 or 10 min 2 or 10 min
AMPds (2013) 1 1 year 1 min 1 min
iAWE (2013) 1 73 days 1 or 6 sec 1 sec
UK-DALE (2016) 5 1–3.5 years 6 sec or 16 kHz 1 or 6 sec, 16 kHz
Activation tanh
Weight init. He’s initialization
Optimizer Adam
t= t+1
Applied LSTM
ℎ1 ℎ2 ℎ3 ℎm−2 ℎm−1 ℎm m= 2×o
Activation tanh
Weight init. He’s initialization
Regularization Dropout
P Δp
Preprocessing
Z-Score
the 4 appliances are using the same meter. When the NILM The problem of UK-DALE and REDD is the asynchro-
model is trained by a dataset, the shared meter may cause a nization between a total load and each appliance data due
confusion or make a distinguishable pattern. In Section 5, we to the different sample rate. So we have to synthesize each
can confirm the effect of the shared meter. appliance data for generating an output dataset. When we
The REDD is the first public dataset for NILM [26]. collect the data from the real household, noise can be
The major purpose of REDD is the standard dataset for included in the total load. However, the synthesized data is
benchmarking the NILM algorithms. In REDD, there are not affected on the noise. To solve the problem, we collect the
AC waveform data with sampling rate of 15 kHz. Therefore, data of six appliances for five months. The description of the
REDD can be used for each approach using the high or low private data is in Table 4.
sampling data. It is sampled from six different houses in Mas-
sachusetts of the United States. There are three categorized 5. Experiment
data, low freq, high freq, and high freq raw. We are going
to use low freq data in this paper. Each house’s dataset is 5.1. Experiment Description. The purpose of the experiment
composed of the main phases sampled 1 Hz and the individual is to satisfy three challenges defined. The challenges are as
data of each appliance sampled 3 or 4 Hz. Table 3 shows the follows.
detailed description of REDD. Each house has the appliance
in range from eleven to twenty-six. We renamed the same Challenge 1. Our model must be robust even though the
appliances like kitchen outlets1 and kitchen outlets2. number of appliances is increasing.
Computational Intelligence and Neuroscience 11
Challenge 2. The multistate appliance should be classified as Section 3.2.1, we expected that the novel signature could
well. classify the multistate appliances having many variations. To
validate this, we train two models having {𝑃} and {𝑃, Δ 𝑝 } as
Challenge 3. The similar power consumption in appliances inputs for each. And then we analyze the results.
should be classified. The second validation experiment is related to challenges
In this section, we take two validation experiments 1 and 3. This experiment consists of four subtests. In the
for the proposed signature and learning method and one first subtest, we are going to find an optimized preprocessing
performance measurement experiment. The first validation method. 𝑍-Score normalization and Min–Max Scaling are
experiment consists of two subtests. The first one is a heuristic popular methods for preprocessing. There is no rule which
approach test for the optimized reflection rate. In this test, one is proper for a specific situation. So we experiment them
we compute three different Δ 𝑝 by three different reflection for choosing the optimized method. The second subtest is
rates. By using Δ 𝑝 s, we train the models with the same about choice of the weight initialization methods. Among
condition. The second subtest is related to challenge 2. In the initialization methods, the Glorot initialization and the
12 Computational Intelligence and Neuroscience
Table 4: Private dataset. evaluating the NILM model. The formulas to compute the
Date of first
value of them are given by
House Submeters Appliances
measurement TP
Air-con., Recall = ,
TP + FN
dehumidifier,
1 2016-03-11 6 TP
TV, washer,
Precision = ,
toaster, oven TP + FP
(2)
TP + TN
Accuracy = ,
Positive Negative (TP + FP) + (FN + TN)
(predicted) (predicted)
Precision ∗ Recall
𝐹1-score = 2 ∗ .
Positive (Precision + Recall)
TP FN
(actual)
Recall is a ratio of the number of correct classifications
to the total number of actual positive instances. A meaning
Negative of that Recall is high and the appliances are well classified
FP TN
(actual)
by the NILM model when the actual instances are positive.
Precision is a ratio of the number of correct classifications to
Figure 9: Confusion matrix. the total number of predicted positive instances. A meaning
of that Precision is high and a probability of well classification
is high when the NILM model predicts the positive instances.
He initialization are well known for providing good per- Accuracy is a ratio of correct classification to the total test
formance. But it is not revealed that which one is better data. 𝐹1-𝑠𝑐𝑜𝑟𝑒 is the harmonic average of Recall and Precision.
or not. Therefore, we train two models with only changing
the initialization method and compare the result. In the 5.3. Proposed Signature Validation Experiment
third subtest, we compare the performance as the number of
appliances is increased. We could confirm that our model is 5.3.1. Optimum Reflection Rate. To find an optimum reflec-
robust even though the number of appliances is increased. In tion rate, we set three different reflection rates, 0.1, 0.01, and
the last subtest, we synthesize the sampled data of appliances 0.001, and compute three variant signals. For experiment, we
having the similar power consumption. Then, we confirm randomly select the 15 appliances from UK-DALE. By using
that the appliances are well classified or not. the variant signals, we calculate three Δ 𝑝 . Δ 𝑝 1 is a result
The last experiment is about measuring the overall per- of subtraction of the first signal and the second signal. Δ 𝑝 2
formance. We are going to use the UK-DALE and REDD is a result of subtraction of the first signal and the third
as the training/test data and measure the performance for signal. Δ 𝑝 3 is a result of subtraction of the second signal
each house in the dataset. Also, we take the same experiment and the third signal. Figure 10 shows the pattern of each Δ 𝑝 .
by FHMM and compare the result. All models used in the Intuitionally, Δ 𝑝 2 and Δ 𝑝 3 are biased. The values of each Δ 𝑝
experiment are implemented with Python programming and are detailed in Table 5. We can notice that the mean of Δ 𝑝 1
Theano library. is close to zero unlike the others. A variance and standard
deviation of Δ 𝑝 1 are almost in middle of the others. And the
5.2. Performance Metrics. Up to now we have generally scale of range of Δ 𝑝 1 and Δ 𝑝 2 is similar. We could know
assumed that the best way of measuring the performance of a which one is proper to NILM by a comparison experiment.
classifier is by its predictive accuracy, that is, the proportion Table 6 shows performance results. All performances are
of unseen instances it correctly classifies. However, predictive similar except that Δ 𝑝 1 is better than the others. Therefore,
accuracy on its own is not a reliable indicator of a classifier’s we will use (0.1, 0.01) as a reflection rate.
effectiveness. As well as the overall predictive accuracy on
unseen instances it is often helpful to see a breakdown of 5.3.2. Effect of the Novel Signature. In this section, we train
the classifier’s performance; it is a confusion matrix which is the two models having {𝑃} and {𝑃, Δ 𝑝 } as an inputs for each.
proposed by Kohavi and Provost, in 1998 [39]. We extracted 72-day data as a training dataset and 30-day data
Confusion matrix is described in Figure 9, including as a test dataset. Each model was trained 2000 epochs. As we
four categories. True positive (TP) is examples correctly explained in Section 4, the private dataset has six appliances.
labeled as positive. False positive (FP) refers to negative Actually, these appliances are not operating in the same time.
examples incorrectly labeled as positive. True negative (TN) However, in our work, we assumed that multiple appliances
corresponds to negatives correctly labeled as negative. Finally, could have been operated in the same time. Therefore, 26 = 64
false negative (FN) refers to positive examples incorrectly which is the number of combinations of all appliances. We
labeled as negative. represented the on/off state by the binary representation.
We assume that the positive state means that an appliance And we change the output unit values to one binary string.
is turned on. When the appliance is turned off, we regard After that, the string is converted to a decimal number
it as a negative state. Based on the confusion matrix, we that stands for the combination. A sequence of appliances is
can calculate 𝑅𝑒𝑐𝑎𝑙𝑙, 𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛, 𝐴𝑐𝑐𝑢𝑟𝑎𝑐𝑦, and 𝐹1-𝑠𝑐𝑜𝑟𝑒 for {Air-conditioner, Washer, Dehumidifier, Oven, TV, Toaster}.
Computational Intelligence and Neuroscience 13
3500 3500
3000 3000 Delta 2
2500 2500
Power (watts)
Power (watts)
2000 2000
1500 1500
1000 1000
500 500
0 0
−500 −500
−1000 −1000
0 5000 10000 15000 20000 0 5000 10000 15000 20000
Time (seconds) Time (seconds)
Delta 1
(a) (b)
3500
3000
2500
Power (watts)
2000
1500
1000
500
0
−500
−1000
0 5000 10000 15000 20000
Time (seconds)
Delta 3
(c)
Table 5: Statistic of Δ 𝑝 .
For example, if the washer and toaster are operating, the the power consumption, this kind of simple pattern could
binary string would be {010001} and the combination num- be learned slowly. To confirm this, we trained 2000 epochs
ber would be 17. more for the second model. In Table 7, the performance of
Figure 11 shows the classification ratio comparison dehumidifier is improved from 0.65 to 0.82 whereas the other
between the models. The 𝑥-axis is a combination and the 𝑦- performances are similar with before. Because the washer
axis is a classification ratio. We can easily notice that if we use has fluctuating Δ 𝑝 , the performance of the second model is
the proposed signature additionally, the more combinations about double. As a result, the proposed signature could be
could be classified. 30 combinations were classified when we very efficient factor for classifying the multistate appliances.
use the proposed signature whereas 11 combinations were
classified when we use only power consumption (not all 5.4. Deep Learning Model Validation
combinations are included in the dataset). Figure 12 shows
the performance of each appliance. The second model using 5.4.1. Optimum Preprocessing Method. 𝑍-Score normaliza-
{𝑃, Δ 𝑝 } has the better performance in all appliances except the tion and Min–Max Scaling are popular methods for prepro-
dehumidifier. Particularly, there are large gaps in the washer, cessing. However, it has not revealed which one is better. To
oven, and toaster. The operation time for the oven and toaster find an optimum method for NILM, we train the two different
is only 1.5 hours and 2.8 hours during 30 days. Even though models. The datasets for all models are the same. We extracted
the sample number is small, the second model has the better 23-day data from house 5 of UK-DALE. 17-day data is used as
performance relatively. a training dataset and 6-day data is used as a test dataset. Each
However, the dehumidifier performance of first model is model trains 1000 epochs.
higher than that of the second one. We see this performance Figure 14 shows the performance results of the models.
in Figure 13. The pattern of dehumidifier is simple. A fluctuat- We can notice that 𝑍-Score method has the better perfor-
ing Δ 𝑝 could be efficient for learning the pattern but the other mance than Min–Max Scaling. In NILM, one of the inputs is
case is not. Because the importance of Δ 𝑝 is the same with the power consumption. As we know, all power consumption
14 Computational Intelligence and Neuroscience
100
90
80
70
Classification ratio (%)
60
50
40
30
20
10
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63
Combinations
P Zero instances
(P, Δ) Below 10 instances
0.6 0.56
0.5 0.49
Washer 0.67
0.4
Dehumidifier 0.82 0.3 0.29
P
is positive. If all input data are the positive value, all weights
(P, Δ)
could be increased or decreased when the backpropagation
is processed. As a result, the dataset for NILM should be Figure 12: Classification performance of the appliances.
standardized. As expected in Section 3.2.2, 𝑍-Score method
is proper for preprocessing.
activation function used and is derived without explicitly
5.4.2. Optimum Weight Initialization Method. Among the considering the type of the distribution. As such, their
weight initialization methods introduced, the researchers theoretical conclusions hold for any type of distribution of
usually use Glorot initialization and He initialization. But the determined variance. To compare the weight initialization
they still have argued about the better method between two methods, we applied the same distribution (uniform distri-
methods. The methods use a similar theoretical analysis. They bution) to the methods. We extracted 71-day data from the
found a good variance for the distribution from which the private dataset. 50-day data are used as a training dataset and
initial parameters are drawn. This variance is adapted to the 21-day data are used as a test dataset. We train two models
Computational Intelligence and Neuroscience 15
800
700
600
Power consumption (watt)
500
400
300
200
100
0
1 9 17 25 33 41 49 57 65 73 81 89 97 105 113 121 129 137 145 153 161 169 177 185 193 201 209 217 225 233 241 249 257
Time (sec)
(a)
450
400
350
Power consumption (watt)
300
250
200
150
100
50
0
1
7
13
19
25
31
37
43
49
55
61
67
73
79
85
91
97
103
109
115
121
127
133
139
145
151
157
163
169
175
181
187
193
199
205
211
217
223
229
235
241
247
253
259
265
271
277
283
289
295
301
307
313
319
325
331
337
343
349
355
361
367
373
379
385
391
397
403
409
415
421
427
433
439
445
451
457
463
469
475
481
487
493
499
Time (sec)
(b)
Figure 13: The patterns of dehumidifier and washer: (a) dehumidifier; (b) washer.
120
100
Power consumption (watt)
80
60
40
20
0
1
3111
6221
9331
12441
15551
18661
21771
24881
27991
31101
34211
37321
40431
43541
46651
49761
52871
55981
59091
62201
65311
68421
71531
74641
77751
80861
83971
87081
90191
93301
96411
99521
102631
105741
108851
111961
115071
118181
121291
124401
127511
130621
133731
136841
139951
143061
146171
149281
152391
155501
158611
161721
164831
167941
Time (sec)
Table 12: Result of performance measurement for classification of 5.5. Performance Comparison with the Existing Models. We
similar power appliance. could confirm that our proposed signature and model are
efficient for NILM through the validation experiments. In this
Metric Similar power appliance based model
section, we would like to compare the performance with the
Precision 0.955 existing models. We take an experiment for all houses in UK-
Recall 0.948 DALE and REDD. Firstly, we train all houses of UK-DALE
Accuracy 0.957 and the result will be compared with FHMM. The parameters
𝐹1-Score 0.951 for UK-DALE experiments are in Table 15.
Table 16 shows the overall performance of UK-DALE
houses. The house having the lowest performance is the first
are well classified even though they are the similar power house. We proved that our model is affected by the data. There
appliances. are 53 appliances in the first house. The possibility of being
To validate our assumption, we collected the 11 numbers the appliance data decreasing the performance is higher than
of the similar power appliances and 3 numbers of the multi- the other houses. For example, the usage of battery charger
state appliances having a high power consumption from UK- and breadmaker is very rare. This kind of appliance is the
DALE and synthesized their data in Table 13. And the con- major reason for dropping the performance. We compare the
trol parameters are the same with the previous experiment performance with FHMM by implementing it. The result is
(Table 10). We train two models. The first one is the model shown in Table 17. We can notice that our model is outper-
using only similar power appliances. In the second model, forming FHMM. The performance is almost double in most
the multistate appliances are included additionally. As we houses.
expected, we can observe that the performance is going Secondly, we train all houses of REDD and the result will
down from the first model to the second model in Table 14. be compared with the existing models. The results are good
Although the performance is decreased, we can realize that. in all houses (see Table 18). We compare our model with
18 Computational Intelligence and Neuroscience
Table 13: A List of the similar power appliances with adding three multistate appliances.
Table 14: Result of performance measurement for classification of similar power appliances with adding three multistate appliances.
Metric Similar power appliance based model All appliance based model
Precision 0.955 0.896
Recall 0.948 0.863
Accuracy 0.957 0.911
𝐹1-Score 0.951 0.879
Table 15: Parameters for UK-DALE experiment. experimental results, we realized that our model overcomes
the existing problem in energy disaggregation. Via mea-
Parameter Method
surement the overall performance, we confirmed that our
Input 𝑃, Δ 𝑝 model outperformed the previous models. In particular, the
Time step 500 performance of house 4 in REDD dataset is double result
Num. of epochs 3000 compared to that of state of the art.
Cost function Mean Squared Error In future work, we will focus on a prediction of the power
Optimization Adam consumption of each appliance, because this problem is dif-
Preprocessing 𝑍-Score ferent in classification task. We cannot apply our currently
Weight init. He initialization model to prediction. Hence, we need to modify our model
to generative model. If we can do this, our model could be
Regularization Dropout
applied to gas or water disaggregation. In other words, we
construct the model which is used for the comprehensive
energy management at home.
three existing models such as FHMM, Additive FHMM, and
HieFHMM [28, 32]. Among them, HieFHMM is the state-
of-the-art model [34]. Due to few numbers of appliances, the
Abbreviations
existing models did not test house 2. Similar to UK-DALE NILM: Nonintrusive Load Monitoring
result, our model has the highest performance in all houses. FHMM: Factorial Hidden Markov Model
In case of house 4, our performance is double compared to GPU: Graphics Processing Unit
that of state of the art. We can confirm this in Table 19. UK-DALE: UK Domestic Appliance Level Electricity
REDD: Reference Energy Disaggregation Data
6. Conclusions HMM: Hidden Markov Model
FSM: Finite State Machine
In this paper, we considered advanced deep learning is new MP: Markov process
approach to build state-of-the-art NILM tool. Besides, we EM: Expectation Maximization
proposed a novel signature to overcome multistate appli- HSMM: Hidden Semi-Markov Model
ance issues in NILM. Furthermore, we give three-problem DHMM: Difference Hidden Markov Model
statement and then address them as efficiently. By this way, DBN: Deep Belief Network
we perform many experiments for validating the signature RBM: Restricted Boltzmann Machine
and the architecture of the proposed model. Through our AI: Artificial Intelligence
Computational Intelligence and Neuroscience 19
120
100
Power consumption (watt)
80
60
40
20
0
1
14
27
40
53
66
79
92
105
118
131
144
157
170
183
196
209
222
235
248
261
274
287
300
313
326
339
352
365
378
391
404
417
430
443
456
469
482
495
508
521
534
547
560
573
586
599
612
625
638
651
664
677
690
703
716
729
742
755
768
781
794
807
820
833
846
859
872
885
898
911
924
937
950
963
976
989
1002
1015
1028
1041
Time (sec)
Table 17: Performance comparison with FHMM (UK-DALE). LSTM: Long Short-Term Memory
PCA: Principal Component Analysis
House FHMM Our model
ReLU: Rectifier Linear Unit
1 0.304 0.764 SGD: Stochastic Gradient Decent
2 0.423 0.810 RMS: Root Mean Square
3 0.495 0.900 STFT: Short Time Fourier Transform
4 0.651 0.968 FHSMM: Factorial Hidden Semi-Markov Model
5 0.424 0.846 CFHMM: Conditional Factorial Hidden Markov
Model
CFHSMM: Conditional Factorial Hidden
Table 18: Overall performance of REDD.
Semi-Markov Model
House Precision Recall Accuracy 𝐹1-Score AFAMAP: Additive Fractional Approximate MAP
1 0.942 0.959 0.968 0.950 VAST: Viterbi Algorithm with Sparse Transitions
2 0.961 0.943 0.964 0.952
MC: Markov chain
CT: Current Transformer
3 0.829 0.840 0.920 0.835
BLUED: Building Level Fully Labeled Data Set for
4 0.846 0.867 0.909 0.856
Electricity Disaggregation
5 0.930 0.893 0.953 0.911 CMU: Carnegie Mellon University
6 0.898 0.954 0.943 0.925 UMASS: University of Massachusetts
HES: Household Electricity Use Study
AMPds: Almanac of Minutely Power Dataset
AAT: Automatic Alternative Text IIIT: Indraprastha Institute of Information
ANN: Artificial Neural Network Technology
tanh: Hyperbolic tangent iAWE: International Associations for Wind
MSE: Mean Squared Error Engineering
CE: Cross Entropy TP: True positive
GD: Gradient Descent FP: False positive
RNN: Recurrent Neural Network TN: True negative
BPTT: Backpropagation through Time FN: False negative
RTRL: Real Time Recurrent Learning HieFHMM: Hierarchical Factorial Hidden Markov
FFNN: Feed-Forward Neural Network Model
20 Computational Intelligence and Neuroscience
House FHMM [33] Additive FHMM [34] HieFHMM [34] Our model
1 0.450 0.749 0.854 0.950
3 0.590 0.619 0.834 0.835
4 0.430 0.417 0.424 0.856
5 0.500 0.795 0.796 0.911
6 0.440 0.391 0.820 0.925
supervised by the NIPA (National IT Industry Promotion IEEE Transactions on Consumer Electronics, vol. 53, no. 2, pp.
Agency). 653–660, 2007.
[18] S. Gupta, M. S. Reynolds, and S. N. Patel, “ElectriSense: Single-
point sensing using EMI for electrical event detection and class-
References ification in the home,” in Proceedings of the 12th International
[1] International Energy Agency, World Energy Outlook 2015. Conference on Ubiquitous Computing, UbiComp 2010, pp. 139–
White Paper, IEA, 2015. 148, dnk, September 2010.
[2] D. Larcher and J.-M. Tarascon, “Towards greener and more sus- [19] H. H. Chang, H. T. Yang, and C.-L. Lin, “Load identification
tainable batteries for electrical energy storage,” Nature Chem- in neural networks for a non-intrusive monitoring of industrial
istry, vol. 7, no. 1, pp. 19–29, 2015. electrical loads,” in Proceedings of the International Conference
on Computer Supported Cooperative Work in Design, pp. 664–
[3] International Electrotechnical Commission, Report to WTO 674, 2007.
TBT Committee. White Paper, IEC, 2010.
[20] H.-H. Chang, “Non-intrusive demand monitoring and load
[4] J. Lee, B. Bagheri, and H.-A. Kao, “A Cyber-Physical Systems identification for energy management systems based on tran-
architecture for Industry 4.0-based manufacturing systems,” sient feature analyses,” Energies, vol. 5, no. 11, pp. 4569–4589,
Manufacturing Letters, vol. 3, pp. 18–23, 2015. 2012.
[5] P. Zhao, S. Suryanarayanan, and M. G. Simoes, “An energy man- [21] S. B. Leeb, S. R. Shaw, and J. L. Kirtley, “Transient event detec-
agement system for building structures using a multi-agent tion in spectral envelope estimates for nonintrusive load moni-
decision-making control methodology,” IEEE Transactions on toring,” IEEE Transactions on Power Delivery, vol. 10, no. 3, pp.
Industry Applications, vol. 49, no. 1, pp. 322–330, 2013.
1200–1210, 1995.
[6] D. Niyato, L. Xiao, and P. Wang, “Machine-to-machine commu-
[22] L. K. Norford and S. B. Leeb, “Non-intrusive electrical load
nications for home energy management system in smart grid,”
monitoring in commercial buildings based on steady-state and
IEEE Communications Magazine, vol. 49, no. 4, pp. 53–59, 2011.
transient load-detection algorithms,” Energy and Buildings, vol.
[7] R. Ford and C. Church, Reducing domestic energy consumption 24, no. 1, pp. 51–64, 1996.
through behaviour modification [Phd. thesis], 2009.
[23] A. I. Cole and A. Albicki, “Data extraction for effective non-
[8] G. W. Hart, “Nonintrusive appliance load monitoring,” Proceed- intrusive identification of residential power loads,” in Proceed-
ings of the IEEE, vol. 80, no. 12, pp. 1870–1891, 1992. ings of the 1998 IEEE Instrumentation and Measurement Tech-
[9] S. Lee, B. Song, Y. Kwon, and J. Kim, “Non-intrusive Load Moni- nology Conference (IMTC ’98), pp. 812–815, May 1998.
toring for Home Energy Usage with Multiple Power States [24] S. N. Patel, T. Robertson, J. A. Kientz, M. S. Reynolds, and G. D.
Recognition,” in Proceedings of the Computer and Computing Abowd, “At the flick of a switch: detecting and classifying unique
Science 2015, pp. 282–289. electrical events on the residential power line (nominated
[10] H. Kim, M. Marwah, M. Arlitt, G. Lyon, and J. Han, “Unsuper- for the best paper award),” in In International Conference on
vised disaggregation of low frequency power measurements,” in Ubiquitous Computing, pp. 271–288, 2007.
Proceedings of the 11th SIAM International Conference on Data [25] A. Zoha, A. Gluhak, M. Nati, and M. A. Imran, “Low-power
Mining (SDM ’11), pp. 747–758, April 2011. appliance monitoring using Factorial Hidden Markov Models,”
[11] H. Najmeddine, K. El Khamlichi Drissi, C. Pasquier et al., “State in Proceedings of the 2013 IEEE 8th International Conference on
of art on load monitoring methods,” in Proceedings of the 2008 Intelligent Sensors, Sensor Networks and Information Processing:
IEEE 2nd International Power and Energy Conference, PECon Sensing the Future (ISSNIP ’13), pp. 527–532, April 2013.
2008, pp. 1256–1258, mys, December 2008. [26] Z. J. Kolter and M. J. Johnson, “Redd: A public data set for
[12] M. B. Figueiredo, A. De Almeida, and B. Ribeiro, “An experi- energy disaggregation research,” in Proceedings of the In Work-
mental study on electrical signature identification of non-intru- shop on Data Mining Applications in Sustainability (SIGKDD),
sive load monitoring (nilm) systems,” International Conference pp. 59–62, San Diego, CA, USA, 2007.
on Adaptive and Natural Computing Algorithms, pp. 31–40, 2011. [27] H. Kim, M. Marwah, M. Arlitt, G. Lyon, and J. Han, “Unsuper-
[13] K. Suzuki, S. Inagaki, T. Suzuki, H. Nakamura, and K. Ito, “Non- vised disaggregation of low frequency power measurements,” in
intrusive appliance load monitoring based on integer program- SDM, vol. 11, pp. 747–758, 2011.
ming,” in Proceedings of the SICE Annual Conference 2008 [28] J. Z. Kolter and T. Jaakkola, “Approximate inference in additive
- International Conference on Instrumentation, Control and factorial HMMs with application to energy disaggregation,”
Information Technology, pp. 2742–2747, August 2008. Journal of Machine Learning Research, vol. 22, pp. 1472–1482,
[14] A. Cole and A. Albicki, “Nonintrusive identification of electrical 2012.
loads in a three-phase environment based on harmonic con- [29] O. Parson, S. Ghosh, M. Weal, and A. Rogers, “Non-intrusive
tent,” in Proceedings of the IMTC/2000 - 17th IEEE Instrumenta- load monitoring using prior models of general appliance types,”
tion and Measurement Technology Conference ’Smart Connectiv- in Proceedings of the 26th AAAI Conference on Artificial Intelli-
ity: Integrating Measurement and Control’, pp. 24–29, May 2000. gence and the 24th Innovative Applications of Artificial Intelli-
[15] C. Laughman, K. Lee, R. Cox et al., “Power signature analysis,” gence Conference, pp. 356–362, July 2012.
IEEE Power and Energy Magazine, vol. 1, no. 2, pp. 56–63, 2003. [30] M. Zeifman, “Disaggregation of home energy display data
[16] J. Li, S. West, and G. Platt, “Power decomposition based on SVM using probabilistic approach,” IEEE Transactions on Consumer
regression,” in Proceedings of the 2012 International Conference Electronics, vol. 58, no. 1, pp. 23–31, 2012.
on Modelling, Identification and Control (ICMIC ’12), pp. 1195– [31] N. Batra, J. Kelly, O. Parson et al., “NILMTK: an open source
1199, June 2012. toolkit for non-intrusive load monitoring,” in Proceedings of the
[17] H. Y. Lam, G. S. K. Fung, and W. K. Lee, “A novel method to con- 5th ACM International Conference on Future Energy Systems (e-
struct taxonomy electrical appliances based on load signatures,” Energy ’14), pp. 265–276, ACM, Cambridge, UK, June 2014.
22 Computational Intelligence and Neuroscience
Advances in
Fuzzy
Systems
Modelling &
Simulation
in Engineering
Hindawi Publishing Corporation
Hindawi Publishing Corporation Volume 2014 http://www.hindawi.com Volume 2014
http://www.hindawi.com
International Journal of
Advances in Computer Games Advances in
Computer Engineering Technology Software Engineering
Hindawi Publishing Corporation Hindawi Publishing Corporation Hindawi Publishing Corporation Hindawi Publishing Corporation Hindawi Publishing Corporation
http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 201 http://www.hindawi.com Volume 2014
International Journal of
Reconfigurable
Computing