Download as pdf or txt
Download as pdf or txt
You are on page 1of 18

1

Predictive Analysis of COVID-19 Time-series Data


from Johns Hopkins University
Alireza M. Javid, Xinyue Liang, Arun Venkitaraman, and Saikat Chatterjee
School of Electrical Engineering and Computer Science
KTH Royal Institute of Technology, Sweden
{almj, xinyuel, arunv, sach}@kth.se
arXiv:2005.05060v3 [cs.LG] 22 May 2020

Abstract—We provide a predictive analysis of the spread TABLE I: Countries considered from JHU dataset.
of COVID-19, also known as SARS-CoV-2, using the dataset Countries considered in the analysis
made publicly available online by the Johns Hopkins University. Sweden
Our main objective is to provide predictions of the number of Denmark
infected people for different countries in the next 14 days. The Finland
predictive analysis is done using time-series data transformed Norway
on a logarithmic scale. We use two well-known methods for France
prediction: polynomial regression and neural network. As the Italy
number of training data for each country is limited, we use a Spain
single-layer neural network called the extreme learning machine UK
(ELM) to avoid over-fitting. Due to the non-stationary nature of China
India
the time-series, a sliding window approach is used to provide a Iran
more accurate prediction. USA
Index Terms—COVID-19, neural network, polynomial regres-
sion, extreme learning machine.
II. DATASET
I. G OAL The dataset from JHU contains the cumulative number
of cases reported daily for different countries. We base our
The COVID-19 pandemic has led to a massive global
analysis on 12 of the countries listed in Table I. For each
crisis, caused by the rapid spread rate and severe fatality,
country, we consider the time-series yn starting from the day
especially, among those with a weak immune system. In
when the first case was reported. Given the current day index
this work, we use the available COVID-19 time-series of the
n, we predict the number of cases for the day n + τ by
infected cases to build models for predicting the number of
considering as input the number of cases reported for the past
cases in the near future. In particular, given the time-series
w days, that is, for the days n − w + 1 to n.
till a particular day, we make predictions for the number
of cases in the next τ days, where τ ∈ {1, 3, 7, 14}. This
means that we predict for the next day, after 3 days, after III. A PPROACHES
7 days, and after 14 days. Our analysis is based on the We use data-driven prediction approaches without consider-
time-series data made publicly available on the COVID- ing any other aspect, for example, models of infectious disease
19 Dashboard by the Center for Systems Science and En- spread [2]. We apply two approaches to analyze the data to
gineering (CSSE) at the Johns Hopkins University (JHU) make predictions, or in other words, to learn the function f :
(https://systems.jhu.edu/research/public-health/ncov/) [1].
• Polynomial model approach: Simplest curve fit or approx-
Let yn denote the number of confirmed cases on the n-th
imation model, where the number of cases is approxi-
day of the time-series after start of the outbreak. Then, we
mated locally with polynomials − f is a polynomial.
have the following
• Neural network approach: A supervised learning ap-
• The input consists of the last n samples of the time-series proach that uses training data in the form of input-output
given by yn , [y1 , y2 , · · · , yn ]. pairs to learn a predictive model − f is a neural network.
• The predicted output is tn = ŷn+τ , τ ∈ {1, 3, 7, 14}.
We describe each approach in detail in the following subsec-
• Due to non-stationary nature of the time-series data, a
tions.
sliding window of size w is used over yn to make the
prediction, and w is found via cross-validation.
• The predictive function f ( · ) is modeled either by a A. Polynomial model
polynomial or a neural network, and is used to make the 1) Model: We model the expected value of yn as a third
prediction: degree polynomial function of the day number n:
ŷn+τ = f (yn )
f (n) = p0 + p1 n1 + p2 n2 + p3 n3
2

The set of coefficients {p0 , p1 , p2 , p3 } are learned using the IV. E XPERIMENTS
available training data. Given the highly non-stationary nature
A. With the available data till May 4, 2020
of the time-series, we consider local polynomial approxima-
tions of the signal over a window of w days, instead of using In this subsection, we make predictions based on the time-
all the data to estimate a single polynomial f ( · ) for the entire series data which currently is available until today May 4,
time-series. Thus, at the n-th day, we learn the corresponding 2020, for τ ∈ {1, 3, 7}. We estimate the number of cases for
polynomial f ( · ) using yn,w , [yn−w+1 , · · · , yn−1 , yn ]. the last 31 days of the countries in Table I. For each value
2) How the model is used: Once the polynomial is deter- of τ ∈ {1, 3, 7}, we compare the estimated number of cases
mined, we use it to predict for (n + τ )-th day as ŷn+τ with the true value yn+τ and report the estimation error
in percentage, i.e.,
ŷn+τ = f (n + τ )
|yn+τ − ŷn+τ |
For every polynomial regression model, we construct the Error = × 100. (2)
yn+τ
corresponding polynomial function f ( · ) by using yn,w as the
most recent input data of size w. The appropriate window size We carry out two sets of experiments for each of the two
w is found through cross-validation. approaches (polynomial and ELM) to examine their sensi-
tivity to the new arriving training samples. In the first set
of experiments, we implement cross-validation to find the
B. Neural networks hyperparameters without using the observed samples of the
1) Model: We use Extreme Learning Machine (ELM) as time-series as we proceed through 31 days span. In the second
the neural network model to avoid overfitting to the training set of experiments, we implement cross-validation in a daily
data. As the length of the time-series data for each country is manner as we observe new samples of the time-series. In the
limited, the number of training samples for the neural network latter setup, the window size w varied with respect to time
would be quite small, which can lead to severe overfitting to find the optimal hyperparameters as we proceed through
in large scale neural network such as deep neural networks time. We refer to this setup as ’ELM time-varying’ and ’Poly
(DNNs), convolutional neural networks (CNNs), etc. [3], [4]. time-varying’ in the rest of the manuscript.
ELM, on the other hand, is a single layer neural network which We first show the reported and estimated number of infec-
uses random weights in its first hidden layer [5]. The use of tion cases for Sweden by using ELM time-varying for different
random weights has gained popularity due to its simplicity τ ’s in Figure 1. For each τ , we estimate the number of cases
and effectiveness in training [6]–[8]. We now briefly describe up to τ days after which JHU data is collected. In our later
ELM. experiments, we show that ELM time-varying is typically more
Consider a dataset containing N samples of pair-wise P - accurate than the other three methods (polynomial, Poly time-
dimensional input data x ∈ RP and the corresponding Q- varying, and ELM). This better accuracy conforms to the non-
dimensional target vector t ∈ RQ as D = {(xn , tn )}N n=1 . We stationary behavior of the time-series data, or in other words
construct the feature vector as zn = g(Wxn ) ∈ Rh , where that the best model parameters change over time. Hence, the
h×P
• weight matrix W ∈ R is an instance of Normal result of ELM time-varying is shown explicitly for Sweden.
distribution, According to our experimental result, we predict that a total of
• h is the number of hidden neurons, and 23039, 23873, and 26184 people will be infected in Sweden
• activation function g( · ) is the rectified linear unit on May 5, May 7, and May 11, 2020, respectively.
(ReLU). Histograms of error percentage of the four methods are
To predict the target, we use a linear projection of feature shown in Figure 2 for different values of τ . The histograms
vector zn onto the target. Let the predicted target for the n- are calculated by using a nonparametric kernel-smoothing
th sample be Ozn . Note that O ∈ RQ×h . By using ℓ2 -norm distribution over the past 31 days for all 12 countries. The
regularization, we find the optimal solution for the following daily error percentage for each country in Table I is shown in
convex optimization problem Figures 7-15. Note that the reported error percentage of ELM
is averaged over 100 Monte Carlo trials. The average and the
X
N
standard deviation of the error over 31 days is reported (in
O⋆ = arg min ktn − Ozn k22 + λkOk2F , (1)
O
n=1
percentage) in the legend of each of the figures for all four
methods. It can be seen that daily cross-validation is crucial to
where k · kF denotes the Frobenius norm. Once the matrix O⋆ preserve a consistent performance through-out the pandemic,
is learned, the prediction for any new input x is now given by resulting in a more accurate estimate. In other words, the
t̂ = O⋆ g(Wx) variations of the time-series as n increases is significant
enough to change the statistics of the training and validation
2) How the model is used: When using ELM to predict set, which, in turn, leads to different optimal hyperparameters
the number of cases, we define xn = [yn−w+1 , ..., yn−1 , yn ]⊤ as the length of the time-series grows. It can also be seen
and tn = [yn+τ ]. Note that xn ∈ Rw and tn ∈ R. For a that ELM time-varying provides a more accurate estimate,
fixed τ ∈ {1, 3, 7, 14}, we use cross-validation to find the especially for large values of τ . Therefore, for the rest of
proper window size w, number of hidden neurons h, and the the experiments, we only focus on ELM time-varying as our
regularization hyperparameter λ. favored approach.
3

Another interesting observation is that the performance of Histograms of error percentage of the four methods are
ELM time-varying improves as n increases. This observation shown in Figure 6 for different values of τ . These experiments
verifies the general principle that neural networks typically verify that ELM time-varying is the most consistent approach
perform better as more data becomes available. We report the as the length of the time-series increases from May 12 to
average error percentage of ELM time-varying over the last 10 May 20. We report the average error percentage of ELM time-
days of the time-series in Table II. We see that as τ increases varying over the last 10 days of the time-series in Table IV. We
the estimation error increases. When τ = 7, ELM time-varying see that as τ increases the estimation error increases. When
works well for most of the countries. It does not perform well τ = 7, ELM time-varying works well for all of the countries
for France and India. This poor estimation for a few countries so we increase the prediction range to 14 days. We observe
could be due to a significant amount of noise in the time- that ELM time-varying fails to provide an accurate estimate
series data, even possibly caused by inaccurately reported daily for several countries such as France, India, Iran, and USA.
cases. This experiment shows that long-term prediction of the spread
COVID-19 can be investigated as an open problem. However,
by observing Tables II-IV, we expect that the performance of
B. With the available data till May 12, 2020
ELM time-varying to improve in the future as the number of
In this subsection, we repeat the prediction based on the training samples increases during the pandemic.
time-series data which is available until today May 12, 2020,
for τ ∈ {1, 3, 7}. In Subsection IV-A, we predicted the total V. C ONCLUSION
number of cases in Sweden on May 5, May 7, and May 11,
We studied the estimation capabilities of two well-known
2020. The reported number of cases on these days for Sweden
approaches to deal with the spread of the COVID-19 pan-
turned out to be 23216, 24623, and 26670, respectively, which
demic. We showed that a small-sized neural network such as
is in the similar range of error that is reported in Table II.
ELM provides a more consistent estimation compared to poly-
We show the reported and estimated number of infection nomial regression counterpart. We found that a daily update
cases for Sweden by using ELM time-varying for different of the model hyperparameters is of paramount importance to
τ ’s in Figure 3. For each τ , we estimate the number of cases achieve a stable prediction performance. The proposed models
up to τ days after which JHU data is collected. According currently use the only samples of the time-series data to predict
to our experimental result, we predict that a total of 27737, the future number of cases. A potential future direction to
28522, and 30841 people will be infected in Sweden on May improve the estimation accuracy is to incorporate constraints
13, May 15, and May 19, 2020, respectively. such as infectious disease spread model, non-pharmaceutical
Histograms of error percentage of the four methods are interventions, and authority policies [2].
shown in Figure 4 for different values of τ . These experiments
verify that ELM time-varying is the most consistent approach
R EFERENCES
as the length of the time-series increases from May 4 to May
12. We report the average error percentage of ELM time- [1] Ensheng Dong, Hongru Du, and Lauren Gardner, “An interactive web-
based dashboard to track covid-19 in real time,” The Lancet Infectious
varying over the last 10 days of the time-series in Table III. We Diseases, vol. 20, no. 5, pp. 533 – 534, 2020.
see that as τ increases the estimation error increases. When [2] Folkhlsomyndigheten, “Estimates of the peak-day and
τ = 7, ELM time-varying works well for all of the countries the number of infected individuals during the covid-19
outbreak in the stockholm region, sweden februaryapril
except India, even though the number of training samples has 2020,” Available at: https://folkhalsomyndigheten.se/publicerat-
increased compared to Subsection IV-A. material/publikationsarkiv/e/estimates-of-the-peak-day-and-the-number-
of-infected-individuals-during-the-covid-19-outbreak-in-the-stockholm-
region-sweden-february–april-2020/.
[3] Christian Szegedy, Alexander Toshev, and Dumitru Erhan, “Deep neural
C. With the available data till May 20, 2020 networks for object detection,” in Advances in Neural Information
In this subsection, we repeat the prediction based on the Processing Systems 26, C. J. C. Burges, L. Bottou, M. Welling, Z. Ghahra-
mani, and K. Q. Weinberger, Eds., pp. 2553–2561. 2013.
time-series data which is available until today May 20, 2020, [4] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton, “Imagenet
for τ ∈ {1, 7, 14}. In Subsection IV-B, we predicted the total classification with deep convolutional neural networks,” in Advances in
number of cases in Sweden on May 13, May 15, and May 19, Neural Information Processing Systems 25, pp. 1097–1105. 2012.
[5] G. Huang, G.-B. Huang, S. Song, and K. You, “Trends in extreme learning
2020. The reported number of cases on these days for Sweden machines: A review,” Neural Networks, vol. 61, no. Supplement C, pp.
turned out to be 27909, 29207, and 30799, respectively, which 32 – 48, 2015.
is in the similar range of prediction error that is reported in [6] R. Giryes, G. Sapiro, and A. M. Bronstein, “Deep neural networks with
random gaussian weights: A universal classification strategy?,” IEEE
Table III. Trans. Signal Process., vol. 64, no. 13, pp. 3444–3457, July 2016.
We increase the prediction range τ in this subsection and [7] Saikat Chatterjee, Alireza M. Javid, Mostafa Sadeghi, Shumpei Kikuta,
we show the reported and estimated number of infection cases Dong Liu, Partha P. Mitra, and Mikael Skoglund, “SSFN – self size-
estimating feed-forward network with low complexity, limited need for
for Sweden by using ELM time-varying for τ = 1, 7, and 14 human intervention, and consistent behaviour across trials,” arXiv preprint
in Figure 5. For each τ , we estimate the number of cases up arXiv:1905.07111, 2020.
to τ days after which JHU data is collected. According to our [8] A. M. Javid, A. Venkitaraman, M. Skoglund, and S. Chatterjee, “High-
dimensional neural feature using rectified linear unit and random matrix
experimental result, we predict that a total of 32032, 34702, instance,” in ICASSP 2020 - 2020 IEEE International Conference on
and 37188 people will be infected in Sweden on May 21, May Acoustics, Speech and Signal Processing (ICASSP), 2020.
27, and June 3, 2020, respectively.
4

TABLE II: Average estimation error in percentage (%) over the last 10 days for ELM time-varying. Update May 4.

Country Sweden Denmark Finland Norway France Italy Spain UK China India Iran USA

1 day prediction 0.9 0.5 0.9 0.2 0.8 0.1 0.5 0.3 0 0.7 0.1 0.4
3 days prediction 2.6 0.7 0.7 0.6 2 0.3 2.5 1.3 0 2.1 0.2 1.7
7 days prediction 2 4.8 2.2 1.2 18.2 1.1 3.1 3 0.2 8.8 0.6 4.9

104

2 Estimated
Reported

1.5
Cases

0.5

10 20 30 40 50 60 70 80 90
Apr 4, 2020 May 4, 2020
Day n
(a) τ = 1
4
10
Estimated
2 Reported

1.5
Cases

0.5

10 20 30 40 50 60 70 80 90
Apr 4, 2020 May 4, 2020
Day n
(b) τ = 3
4
10
2.5
Estimated
Reported
2
Cases

1.5

0.5

10 20 30 40 50 60 70 80 90 100
Apr 4, 2020 May 4, 2020
Day n
(c) τ = 7
Fig. 1: Reported and estimated cases after τ days over the last 31 days of Sweden for ELM time-varying. Update May 4.
5

7
ELM time-varying
6 ELM
Poly time-varying
Poly
5
Histogram

0 5 10 15 20 25 30
Error (%)
(a) τ = 1

3.5 ELM time-varying


ELM
3 Poly time-varying
Poly
2.5
Histogram

1.5

0.5

0 20 40 60 80 100 120 140 160 180 200


Error (%)
(b) τ = 3

20 ELM time-varying
ELM
Poly time-varying
Poly
15
Histogram

10

0
0 100 200 300 400 500 600 700 800 900 1000
Error (%)
(c) τ = 7

Fig. 2: Histogram of estimation error percentage over 31 days of all 12 countries for different values of τ for each of the ELM
and polynomial approaches. Update May 4.
6

TABLE III: Average estimation error in percentage (%) over the last 10 days for ELM time-varying. Update May 12.

Country Sweden Denmark Finland Norway France Italy Spain UK China India Iran USA

1 day prediction 0.7 0.2 0.6 0.1 0.5 0.1 0.3 0.3 0 0.9 0.3 0.2
3 days prediction 2.3 0.5 1 0.4 1.4 0.3 0.4 1.3 0 3.4 0.7 0.5
7 days prediction 1.8 1.2 1.4 0.8 3.6 0.7 1.3 3.8 0 10.9 2.9 2.3

104
2.5 Estimated
Reported
2
Cases

1.5

0.5

10 20 30 40 50 60 70 80 90 100
Apr 12, 2020 May 12, 2020
Day n
(a) τ = 1
4
10
2.5 Estimated
Reported
2
Cases

1.5
1
0.5

10 20 30 40 50 60 70 80 90 100
Apr 12, 2020 May 12, 2020
Day n
(b) τ = 3
4
10
3
Estimated
2.5 Reported

2
Cases

1.5
1
0.5

10 20 30 40 50 60 70 80 90 100 110
Apr 12, 2020 May 12, 2020
Day n
(c) τ = 7
Fig. 3: Reported and estimated cases after τ days over the last 31 days of Sweden for ELM time-varying. Update May 12.
7

ELM time-varying
8 ELM
Poly time-varying
Poly
6
Histogram

0
0 5 10 15 20 25 30
Error (%)
(a) τ = 1

14
ELM time-varying
12 ELM
Poly time-varying
10 Poly
Histogram

0
0 20 40 60 80 100 120 140 160 180 200
Error (%)
(b) τ = 3

4 ELM time-varying
ELM
3.5
Poly time-varying
3 Poly
Histogram

2.5
2
1.5
1
0.5

0 100 200 300 400 500 600 700 800 900 1000
Error (%)
(c) τ = 7

Fig. 4: Histogram of estimation error percentage over 31 days of all 12 countries for different values of τ for each of the ELM
and polynomial approaches. Update May 12.
8

TABLE IV: Average estimation error in percentage (%) over the last 10 days for ELM time-varying. Update May 20.

Country Sweden Denmark Finland Norway France Italy Spain UK China India Iran USA

1 day prediction 0.5 0.2 0.4 0.1 0.2 0.1 0.3 0.2 0 0.5 0.3 0.2
7 days prediction 1.6 1.4 1.4 0.3 1.4 0.3 1.4 1.2 0 2.9 3.6 1.2
14 days prediction 3.3 6 1.8 2.6 19.4 0.8 2.2 8.8 0.1 33.9 10.4 11.3

104
3 Estimated
2.5 Reported

2
Cases

1.5
1
0.5

10 20 30 40 50 60 70 80 90 100 110
Day n Apr 20, 2020 May 20, 2020

(a) τ = 1
104
Estimated
3
Reported
2.5
Cases

2
1.5
1
0.5

10 20 30 40 50 60 70 80 90 100 110
Day n Apr 20, 2020 May 20, 2020

(b) τ = 7
4
10
Estimated
3 Reported
Cases

20 40 60 80 100 120
Day n Apr 20, 2020 May 20, 2020

(c) τ = 14
Fig. 5: Reported and estimated cases after τ days over the last 31 days of Sweden for ELM time-varying. Update May 20.
9

ELM time-varying
4 ELM
Poly time-varying
Poly
3
Histogram

0 1 2 3 4 5 6 7 8 9 10
Error (%)
(a) τ = 1

5 ELM time-varying
ELM
Poly time-varying
4 Poly
Histogram

0 20 40 60 80 100 120 140 160 180 200


Error (%)
(b) τ = 7

5
ELM time-varying
ELM
4 Poly time-varying
Poly
Histogram

0
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Error (%)
(c) τ = 14

Fig. 6: Histogram of estimation error percentage over 31 days of all 12 countries for different values of τ for each of the ELM
and polynomial approaches. Update May 20.
10

14 ELM time-varying, avg = 1%, std = 1%


ELM, avg = 1.3%, std = 1.2%
12 Poly time-varying, avg = 3%, std = 3.4%
Poly, avg = 3.7%, std = 2.4%
10
Error (%)

65 70 75 80 85 90 95
Day n
(a) Sweden
4
ELM time-varying, avg = 0.7%, std = 0.6%
3.5 ELM, avg = 0.7%, std = 0.7%
Poly time-varying, avg = 1.5%, std = 1.1%
3 Poly, avg = 1.4%, std = 1%
2.5
Error (%)

1.5

0.5

40 45 50 55 60 65
Day n
(b) Denmark
14 ELM time-varying, avg = 1.6%, std = 1.9%
ELM, avg = 1.3%, std = 1.9%
12 Poly time-varying, avg = 3.5%, std = 3.3%
Poly, avg = 3.4%, std = 3%
10
Error (%)

70 75 80 85 90 95
Day n
(c) Finland

8 ELM time-varying, avg = 0.6%, std = 0.6%


ELM, avg = 0.5%, std = 0.5%
7 Poly time-varying, avg = 1.9%, std = 2.1%
Poly, avg = 2.1%, std = 2.1%
6
Error (%)

5
4
3
2
1

40 45 50 55 60 65
Day n
(d) Norway
Fig. 7: Daily error percentage of the last 31 days of 12 countries for ELM and polynomial regression. Here, τ = 1.
11

ELM time-varying, avg = 2.5%, std = 3.8%


ELM, avg = 2.4%, std = 3.2%
15 Poly time-varying, avg = 5.4%, std = 4.1%
Poly, avg = 5.8%, std = 4.8%
Error (%)

10

75 80 85 90 95 100
Day n
(a) France
3.5
ELM time-varying, avg = 0.3%, std = 0.2%
3 ELM, avg = 0.5%, std = 0.3%
Poly time-varying, avg = 0.8%, std = 0.7%
2.5 Poly, avg = 1%, std = 0.9%
Error (%)

1.5

0.5

65 70 75 80 85 90 95
Day n
(b) Italy
12
ELM time-varying, avg = 0.9%, std = 1.2%
ELM, avg = 1%, std = 1.2%
10 Poly time-varying, avg = 2.2%, std = 2.3%
Poly, avg = 5.2%, std = 3.3%
8
Error (%)

65 70 75 80 85 90
Day n
(c) Spain
12
ELM time-varying, avg = 1.4%, std = 1.7%
10 ELM, avg = 1%, std = 1.2%
Poly time-varying, avg = 3%, std = 3.1%
Poly, avg = 3.2%, std = 3.1%
8
Error (%)

65 70 75 80 85 90 95
Day n
(d) UK
Fig. 8: Daily error percentage of the last 31 days of 12 countries for ELM and polynomial regression. Here, τ = 1.
12

ELM time-varying, avg = 0%, std = 0.1%


0.4 ELM, avg = 0.1%, std = 0.1%
Poly time-varying, avg = 0.1%, std = 0.1%
Poly, avg = 0.1%, std = 0.1%
0.3
Error (%)

0.2

0.1

75 80 85 90 95 100
Day n
(a) China
40 ELM time-varying, avg = 2.7%, std = 4.2%
ELM, avg = 7.2%, std = 2.9%
35 Poly time-varying, avg = 8.9%, std = 9.7%
Poly, avg = 7.4%, std = 8.4%
30
Error (%)

25
20
15
10
5

70 75 80 85 90 95
Day n
(b) India
6
ELM time-varying, avg = 0.1%, std = 0.1%
ELM, avg = 0.1%, std = 0.1%
5 Poly time-varying, avg = 0.6%, std = 0.8%
Poly, avg = 1.6%, std = 1.9%
4
Error (%)

50 55 60 65 70 75
Day n
(c) Iran
25
ELM time-varying, avg = 0.5%, std = 0.4%
ELM, avg = 0.6%, std = 0.4%
20 Poly time-varying, avg = 2.7%, std = 3.6%
Poly, avg = 15.5%, std = 6.8%
Error (%)

15

10

75 80 85 90 95 100
Day n
(d) USA
Fig. 9: Daily error percentage of the last 31 days of 12 countries for ELM and polynomial regression. Here, τ = 1.
13

80 ELM time-varying, avg = 4.2%, std = 3.8%


ELM, avg = 5%, std = 2.8%
70 Poly time-varying, avg = 16.6%, std = 20.1%
60 Poly, avg = 8.2%, std = 9.7%
Error (%)

50
40
30
20
10

65 70 75 80 85 90 95
Day n
(a) Sweden
160 ELM time-varying, avg = 2.9%, std = 2.8%
ELM, avg = 13.2%, std = 6.6%
140 Poly time-varying, avg = 13.3%, std = 30.2%
Poly, avg = 12.1%, std = 10.9%
120
Error (%)

100
80
60
40
20

40 45 50 55 60 65
Day n
(b) Denmark
ELM time-varying, avg = 3.5%, std = 3.7%
80
ELM, avg = 3.4%, std = 3.5%
70 Poly time-varying, avg = 25.2%, std = 21.6%
Poly, avg = 29.8%, std = 22.4%
60
Error (%)

50
40
30
20
10

70 75 80 85 90 95
Day n
(c) Finland
100
ELM time-varying, avg = 1.1%, std = 0.8%
ELM, avg = 0.9%, std = 0.8%
80 Poly time-varying, avg = 10.5%, std = 20.5%
Poly, avg = 12.4%, std = 10.2%
Error (%)

60

40

20

40 45 50 55 60 65
Day n
(d) Norway
Fig. 10: Daily error percentage of the last 31 days of 12 countries for ELM and polynomial regression. Here, τ = 3.
14

160 ELM time-varying, avg = 7.1%, std = 6.2%


ELM, avg = 9.7%, std = 3%
140 Poly time-varying, avg = 26.1%, std = 31.5%
Poly, avg = 17.6%, std = 17.4%
120
Error (%)

100
80
60
40
20

75 80 85 90 95 100
Day n
(a) France
ELM time-varying, avg = 0.9%, std = 0.8%
120
ELM, avg = 6.7%, std = 11%
Poly time-varying, avg = 11.1%, std = 25.5%
100 Poly, avg = 2%, std = 2.2%
Error (%)

80

60

40

20

65 70 75 80 85 90 95
Day n
(b) Italy
70
ELM time-varying, avg = 2.2%, std = 1.8%
60 ELM, avg = 1.7%, std = 1.7%
Poly time-varying, avg = 12.4%, std = 14.1%
50 Poly, avg = 35.4%, std = 23.6%
Error (%)

40

30

20

10

65 70 75 80 85 90
Day n
(c) Spain
35 ELM time-varying, avg = 3.6%, std = 2.8%
ELM, avg = 21.8%, std = 5.4%
30 Poly time-varying, avg = 11.1%, std = 11%
Poly, avg = 11.5%, std = 8.7%
25
Error (%)

20

15

10

65 70 75 80 85 90 95
Day n
(d) UK
Fig. 11: Daily error percentage of the last 31 days of 12 countries for ELM and polynomial regression. Here, τ = 3.
15

0.8 ELM time-varying, avg = 0.1%, std = 0.1%


ELM, avg = 0.4%, std = 0.2%
0.7
Poly time-varying, avg = 0.2%, std = 0.2%
0.6 Poly, avg = 0.4%, std = 0.2%
Error (%)

0.5
0.4
0.3
0.2
0.1

75 80 85 90 95 100
Day n
(a) China
ELM time-varying, avg = 5.8%, std = 6.6%
ELM, avg = 14.5%, std = 8%
150 Poly time-varying, avg = 34.8%, std = 39.9%
Poly, avg = 34%, std = 21.6%
Error (%)

100

50

70 75 80 85 90 95
Day n
(b) India
60 ELM time-varying, avg = 0.9%, std = 1.9%
ELM, avg = 2%, std = 3%
50 Poly time-varying, avg = 8%, std = 13.9%
Poly, avg = 8.1%, std = 8.2%
40
Error (%)

30

20

10

50 55 60 65 70 75
Day n
(c) Iran
60
ELM time-varying, avg = 1.8%, std = 1.2%
ELM, avg = 2%, std = 1.2%
50 Poly time-varying, avg = 13.1%, std = 16.4%
Poly, avg = 17.3%, std = 17.6%
40
Error (%)

30

20

10

75 80 85 90 95 100
Day n
(d) USA
Fig. 12: Daily error percentage of the last 31 days of 12 countries for ELM and polynomial regression. Here, τ = 3.
16

400
ELM time-varying, avg = 7.2%, std = 6%
350 ELM, avg = 45.6%, std = 14.3%
Poly time-varying, avg = 105.7%, std = 121.1%
300 Poly, avg = 52.3%, std = 89.3%
Error (%)

250
200
150
100
50

65 70 75 80 85 90 95
Day n
(a) Sweden
1000
ELM time-varying, avg = 15.3%, std = 14.5%
ELM, avg = 18.5%, std = 16.1%
800 Poly time-varying, avg = 210.2%, std = 922%
Poly, avg = 2335%, std = 9655.6%
Error (%)

600

400

200

0
40 45 50 55 60 65
Day n
(b) Denmark
1000
ELM time-varying, avg = 12%, std = 15.6%
ELM, avg = 17.3%, std = 9.9%
800 Poly time-varying, avg = 177.6%, std = 574.6%
Poly, avg = 1868.6%, std = 5659.9%
Error (%)

600

400

200

0
70 75 80 85 90 95
Day n
(c) Finland
1000
ELM time-varying, avg = 4.8%, std = 5.5%
ELM, avg = 4.5%, std = 3.3%
800 Poly time-varying, avg = 124.8%, std = 487.1%
Poly, avg = 148.3%, std = 459.3%
Error (%)

600

400

200

0
40 45 50 55 60 65
Day n
(d) Norway
Fig. 13: Daily error percentage of the last 31 days of 12 countries for ELM and polynomial regression. Here, τ = 7.
17

ELM time-varying, avg = 18.4%, std = 11.6%


ELM, avg = 14.4%, std = 6.4%
Poly time-varying, avg = 34%, std = 27%
150 Poly, avg = 55.5%, std = 54.1%
Error (%)

100

50

75 80 85 90 95 100
Day n
(a) France
60 ELM time-varying, avg = 3.8%, std = 5.8%
ELM, avg = 1.6%, std = 1.2%
50 Poly time-varying, avg = 22.9%, std = 14.7%
Poly, avg = 12.6%, std = 13.6%
40
Error (%)

30

20

10

65 70 75 80 85 90 95
Day n
(b) Italy
120
ELM time-varying, avg = 10%, std = 12.3%
ELM, avg = 9.2%, std = 7.3%
100 Poly time-varying, avg = 40.4%, std = 27.2%
Poly, avg = 43.5%, std = 30.7%
80
Error (%)

60

40

20

65 70 75 80 85 90
Day n
(c) Spain
200
ELM time-varying, avg = 14.1%, std = 24%
ELM, avg = 169.8%, std = 27.8%
Poly time-varying, avg = 37%, std = 30.1%
150 Poly, avg = 27%, std = 17.1%
Error (%)

100

50

65 70 75 80 85 90 95
Day n
(d) UK
Fig. 14: Daily error percentage of the last 31 days of 12 countries for ELM and polynomial regression. Here, τ = 7.
18

3
ELM time-varying, avg = 0.3%, std = 0.2%
ELM, avg = 1%, std = 0.3%
2.5 Poly time-varying, avg = 1%, std = 0.7%
Poly, avg = 0.7%, std = 0.4%
2
Error (%)

1.5

0.5

75 80 85 90 95 100
Day n
(a) China
1000
ELM time-varying, avg = 25.9%, std = 27.5%
ELM, avg = 23.6%, std = 14.3%
800 Poly time-varying, avg = 161.2%, std = 609.8%
Poly, avg = 4342.9%, std = 12418%
Error (%)

600

400

200

0
70 75 80 85 90 95
Day n
(b) India
400
ELM time-varying, avg = 7.7%, std = 10.4%
350 ELM, avg = 12.9%, std = 5%
Poly time-varying, avg = 33.5%, std = 47.4%
300 Poly, avg = 35.2%, std = 71.8%
250
Error (%)

200

150

100

50

50 55 60 65 70 75
Day n
(c) Iran
350 ELM time-varying, avg = 24.1%, std = 69%
ELM, avg = 265%, std = 84.8%
300 Poly time-varying, avg = 34.5%, std = 31.2%
Poly, avg = 55.5%, std = 59.3%
250
Error (%)

200

150

100

50

75 80 85 90 95 100
Day n
(d) USA
Fig. 15: Daily error percentage of the last 31 days of 12 countries for ELM and polynomial regression. Here, τ = 7.

You might also like