Professional Documents
Culture Documents
1 s2.0 S0957417422016189 Main
1 s2.0 S0957417422016189 Main
1 s2.0 S0957417422016189 Main
Deep learning for downward longwave radiative flux forecasts in the Arctic
Dae-Hui Kim , Hyun Mee Kim *
a
Atmospheric Predictability and Data Assimilation Laboratory, Department of Atmospheric Sciences, Yonsei University, 50 Yonsei-ro, Seodaemun-gu, Seoul, 03722,
Republic of Korea
A R T I C L E I N F O A B S T R A C T
Keywords: Downward longwave radiative flux (LWD), a key factor affecting sea ice properties and warming (i.e., Arctic
Deep learning amplification) in the Arctic, has large uncertainties in numerical weather prediction (NWP) model simulations
Downward longwave radiative flux over the Arctic. LWD estimated in the European Centre for Medium-Range Weather Forecasts (ECMWF)’s fifth-
The Arctic
generation reanalysis (ERA5) underestimated the LWD observations at Hopen in Svalbard, Norway. Although
Reduction of forecast uncertainties
LWD underestimation in the ERA5 reanalysis with respect to observations was improved in 24 h forecasts using
the Polar Weather Research and Forecasting model (PWRF) without and with data assimilation (DA), 24 h LWD
forecasts using PWRF continue to underestimate LWD observations. To improve LWD estimation in the Arctic, a
deep learning post-processing model that corrects the bias of the LWD simulation was developed using con
volutional neural network and ERA5 reanalysis (2016–2019) as training data. By applying the trained deep
learning post-processing model to LWD from three independent datasets (i.e., ERA5 reanalysis data in 2020, 24 h
forecasts in 2020 using PWRF without and with DA), the time-averaged root mean square errors (RMSEs) of LWD
after deep learning post-processing were reduced by 17.62%, 14.98%, and 13.14%, respectively. Therefore, deep
learning reduces uncertainties in LWD simulations in the Arctic. The deep learning model trained with ERA5
reanalysis (2016–2019) was able to correct the bias in LWD simulation from the same type of independent data
(i.e., ERA5 reanalysis), as well as from different model type (i.e., PWRF forecasts without and with DA).
Therefore, when several NWP models simulate the same atmospheric phenomena, a deep learning model trained
with data from one NWP model can be applied to data from other NWP models to reduce uncertainties. Addi
tionally, deep learning can further improve forecasts with DA. Therefore, it is expected that the cost required to
generate training data will be reduced, and the efficiency of the deep learning model will increase.
1. Introduction Arctic Ocean in summer after sea ice melts, and the heat energy stored in
the Arctic Ocean is emitted to the atmosphere in early fall, which affects
For several decades, the extent of Arctic sea ice has decreased sea ice refreezing (Bintanja & Krikken, 2016; Boeke & Taylor, 2018).
(Comiso et al., 2008). The reduction in sea ice over the Arctic affects The relationship between SWD and sea ice is known as the sea ice-albedo
mid-latitude weather patterns (Semmler et al., 2012; Vihma, 2014) and feedback process. Because the sea ice-albedo feedback begins with the
the Arctic climate (Kumar et al., 2010; Boeke & Taylor, 2018; Jenkins & effect of LWD during winter and spring (Kapsch et al., 2016), the LWD
Dai, 2021). As Arctic sea ice becomes younger and thinner than it was in should be estimated accurately in numerical weather prediction (NWP)
the 1980s, the extent of sea ice is more sensitive to climate change models.
(Maslanik et al., 2007). The relationship between LWD and sea ice has been extensively
The sea ice extent variation is related to surface radiative fluxes studied. LWD anomalies in spring affect the time when sea ice begins to
(Döscher et al., 2014). When Arctic sea ice melts in summer, the onset melt (Huang et al., 2019; Urrego-Blanco et al., 2019) and the minimum
timing of sea ice melting is determined by downward longwave radia sea ice extent (Francis et al., 2005) in summer. Liu and Key (2014) found
tive flux (LWD) during winter and spring, because there is little to no that LWD anomalies in winter are associated with the extent of sea ice in
downward shortwave radiative flux (SWD) during winter and spring in the following summer. When heat energy is transferred from the Arctic
the Arctic. The SWD has little effect on the surface energy budget prior to Ocean to the atmosphere in early fall, the amount of LWD is important
the melting of sea ice (Kapsch et al., 2016). The SWD is absorbed into the for determining the speed at which sea ice refreezes (Kay & Gettelman,
* Corresponding author.
E-mail addresses: kdhui@yonsei.ac.kr (D.-H. Kim), khm@yonsei.ac.kr (H.M. Kim).
https://doi.org/10.1016/j.eswa.2022.118547
Received 25 January 2022; Received in revised form 18 July 2022; Accepted 12 August 2022
Available online 17 August 2022
0957-4174/© 2022 The Author(s). Published by Elsevier Ltd. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-
nc-nd/4.0/).
D.-H. Kim and H.M. Kim Expert Systems With Applications 210 (2022) 118547
2
D.-H. Kim and H.M. Kim Expert Systems With Applications 210 (2022) 118547
Zero padding, commonly used in CNN, adds zeros to the edges of two- model to test data with dropout (i.e., dropout ensemble). Scher and
dimensional data to accurately recognize features near the edges. A Messori (2021) used a deep learning model to predict 500 hPa geo
fully connected layer after the convolutional layers consisted of two potential height up to 120 h, and showed that the mean error of dropout
dense layers with the ReLU activation function, and dropout (ratio of ensembles for the geopotential height forecasts was smaller than the
0.5) (Srivastava et al., 2014) was applied between the two dense layers error of a deterministic result without dropout ensembles. However,
to prevent overfitting. Lakshminarayanan et al. (2017) found that the ensemble mean for the
After the CNN was applied to input1, one-dimensional data with 36 deep learning results after applying several trained models to test data
values were calculated (Fig. 1). Input2 was the LWD, and input3 was the was more accurate than the ensemble mean for those after applying a
mid- and high-cloud fractions simulated in the NWP model. The NWP trained model to test data with dropout (i.e., dropout ensemble).
models that produced input data for deep learning were explained in Therefore, in this study, we have used the dropout to produce eight deep
Section 2.1 in detail. Both input2 and input3 were 24-h time series with learning model ensembles instead of using the dropout when applying a
1-h intervals. The 36 values resulting from applying CNN to input1 were trained model to test data, and used the average of eight deep learning
concatenated with input2 and input3, resulting in one-dimensional data model results to calculate error statistics after deep learning.
with 84 values (Fig. 1). These 84 values passed two dense layers with a For all convolutional layers and dense layers in the deep learning
leaky ReLU activation function with a dropout (ratio of 0.5) between the model, normal (He et al., 2015) was used for the kernel initializer, and
two dense layers. Finally, the deep learning model calculated the output bias was initialized by zero. An Adam optimizer (Kingma & Ba, 2014)
with 24 values, which were the results of deep learning post-processing was used for minimization. If loss did not decrease for 20 epochs, then
for simulated LWD (i.e., input2). The output was a 24-h time series with the learning rate was reduced by 0.2 times. If the loss did not decrease
1-h intervals. The metrics (mean absolute error) and loss (mean squared for 25 epochs, training was completed.
error) were calculated for the post-processed LWD (i.e., output) using
observations as truth: 2.2. Model data for deep learning training and test
Mean absolute error
1∑ n ⃒ ⃒ (1) 2.2.1. ERA5 reanalysis data
=
n i=1
⃒LWDoutput − LWDobservation ⃒, The meteorological variables of ERA5 reanalysis 0.25◦ × 0.25◦
(Hersbach et al., 2020) from 2016 to 2019 with 1-h intervals were used
for deep learning training data. The NWP model to produce ERA5
1∑ n
Mean squared error = (LWDoutput − LWDobservation )2 , (2) reanalysis is Integrated Forecasting System (IFS) Cy41r2 (ECMWF,
n i=1
2016). ERA5 reanalysis data was downloaded from https://cds.climate.
copernicus.eu/cdsapp#!/dataset/reanalysis-era5-single-levels?
where n is 24, LWD output and LWD observation were LWD output of the
tab=overview.
deep learning model and LWD observations, respectively. The deep
The variables used for the training data were low-cloud fraction
learning model was trained to reduce loss.
(surface − 2 km), mid-cloud fraction (2–6 km), high-cloud fraction (6
When the training was finished, the deep learning model, which had
km − top of the atmosphere), 2 m temperature, latent and sensible heat
more than 200 epochs, metrics less than 0.07, and loss less than 0.0082,
flux, and LWD. These variables were used as inputs1, 2, and 3 in the deep
was used to evaluate the test data. Owing to dropout, the trained deep
learning model.
learning models performed differently every time. To evaluate the
The test data used to evaluate the trained deep learning model (i.e.,
general performance of the deep learning models, eight randomly
the deep learning test data) were meteorological variables of the ERA5
selected deep learning models were applied to the test data.
reanalysis in 2020. Note that test data period of 2020 is different from
The average performance of deep learning model ensembles could be
the training period. The variables in the deep learning test data were the
better than the performance of individual deep learning models because
same as those in the deep learning training data, because the test data
of statistical, computational, and representative factors associated with
should consist of the same variables as inputs1, 2, and 3 in the deep
uncertainties of individual deep learning calculations (Dietterich, 2000).
learning model.
Gal and Ghahramani (2016) introduced a method to produce ensembles
of deep learning results with low cost calculations by applying a trained
3
D.-H. Kim and H.M. Kim Expert Systems With Applications 210 (2022) 118547
2.2.2. PWRF forecast data DA). These data were used to verify whether the deep learning model
The forecast data of PWRF were also used as deep-learning test data could achieve further forecast improvements, even for PWRF forecasts
for the 2020 as the ERA5 test data. This was to evaluate the effectiveness using DA cycling.
of the deep learning model trained with ERA5 reanalysis data, even The PWRF forecasts with DA were produced using WRFDA v3.8
when it was used with test data produced by another NWP modeling three-dimensional variational DA (3DVAR, Barker et al., 2012) method.
system, PWRF. Conventional observations and satellite radiances (Advanced Micro
The PWRF v3.7.1 model (Hines & Bromwich, 2008) was used to wave Sounding Unit-A and Microwave Humidity Sounder) were assim
produce the PWRF forecast test data. In the PWRF, a Noah land surface ilated with ± 3 h assimilation window at each analysis time (i.e., 00, 06,
model was optimized for polar environments (Hines & Bromwich, 12, and 18 UTC). The background error covariance for 3DVAR was
2008), and the sea ice characteristics were considered (Bromwich et al., calculated based on differences between the 12 h and 24 h forecasts at
2009; Hines et al., 2015). Additionally, the Morrison 2-moment micro three-month intervals (January–March, April–June, July–September,
physics scheme was modified to reflect the characteristics of cloud and October–December) based on the National Meteorological Center
physics over the Arctic (Bromwich et al., 2009; Hines & Bromwich, (NMC) method (Parrish & Derber, 1992). The PWRF forecasts with DA
2017). The horizontal resolution of the PWRF was set to 9 km (103 × used the ERA5 reanalysis as the boundary condition. Similar to the
117 grid points), and the vertical layer was set to 50, with the model top PWRF forecasts without DA, the last 24 h forecasts from the 48 h fore
at 10 hPa. The experimental domain (Fig. 2) was set to minimize the casts were used as the test data time series with 1-h intervals, and the
influence of boundary condition on the simulation of meteorological variables in the PWRF test data with DA were the same as those in the
variables in Hopen. The physics schemes used were the rapid radiative ERA5 test data.
transfer model for GCMs (Iacono et al., 2008) for short- and long-wave
radiation parameterization, the optimized Noah land surface model
2.3. Observation data for loss calculation and evaluation
(Chen et al., 1996) for land surface parameterization, the Mellor-
Yamada-Nakanishi-Niino (MYNN) level-2.5 (Nakanishi & Niino, 2006)
The LWD observations at Hopen station (76.5097◦ N, 25.0133◦ E) in
for both planetary boundary layer and surface layer parameterizations,
Svalbard, Norway (https://klimaservicesenter.no) were used to calcu
the Grell–Devenyi ensemble (Grell & Dévényi, 2002) for cumulus
late the LWD errors in ERA5 reanalysis and those in PWRF forecasts, as
parameterization, and the Morrison 2-moment scheme (Morrison et al.,
well as the errors of post-processed LWD using the deep learning model.
2005) for microphysics parameterization.
The LWD observations exist at 1-h intervals from January 1, 2016 to
One type of test data generated using PWRF are forecasts without DA
December 31, 2020. The LWD observations from 2016 to 2019 were
from analysis-forecast cycling using PWRF (hereafter PWRF forecast
used to calculate the loss in training the deep learning model; the LWD
without DA). In this case, PWRF forecasts were produced using ERA5
observations from 2020 were used to evaluate the errors of simulated
reanalysis as the initial and boundary conditions. After producing 48 h
LWD in NWP models (i.e., ERA5 reanalysis, PWRF forecasts without and
forecasts at every 00 UTC, the last 24 h forecasts (i.e., from 25 h to 48 h)
with DA) and post-processed LWD by the trained deep learning model.
were used as the test data time series at 1-h intervals. The last 24 h
Observations outside ± 3 sigma were excluded from training the deep
forecasts from 48 h forecasts were used because the 24 h spin-up is
learning model as outliers. Missing observations during the training and
necessary to reflect characteristics of the Arctic surface on planetary
test periods (2016–2020) were not used for training and testing the deep
boundary layer development and to adjust the initial conditions to the
learning model. Fourteen observations were missing from 2016 to 2020.
hydrological cycle over the Arctic in PWRF (Kim et al., 2019). The
variables in the PWRF test data without DA were identical to those of the
ERA5 test data. 2.4. Pre-processing of data
Other test data generated using PWRF were forecasts with DA from
analysis-forecast cycling using PWRF (hereafter PWRF forecasts with Because excessively large or small observation values interfere with
deep learning model training and may cause incorrect learning results,
the observation values needed scaling for the deep learning model
training period. In addition to the LWD observations, the input data from
the ERA5 reanalysis needed scaling because the input data and obser
vations were compared with each other in the deep learning model. The
test data from the ERA5 reanalysis and PWRF forecasts without and with
DA also needed scaling to evaluate the deep learning model.
The min–max scaling for observations, ERA5 training data, ERA5 test
data, and PWRF forecast test data without and with DA is as follows:
x − xmin
xscaled = , (3)
xmax − xmin
where x is the value to be scaled, xscaled is the min–max scaling result for
x with a range from 0 to 1, and xmin and xmax are the minimum and
maximum values of x, respectively.
xmin and xmax were determined differently depending on the vari
ables. Because the LWD was used as the training data (i.e., input2) and
truth in the deep learning model, it was necessary to scale the two types
of data (i.e., input2 and truth) using the same minimum and maximum
values to compare the relative sizes of the two data during the deep
learning model training process. Therefore, for LWD, xmin and xmax were
determined from the datasets including the ERA5 reanalysis and ob
servations without outliers from 2016 to 2019. Variables such as 2 m
temperature, latent heat flux, sensible heat flux, and low, mid, and high
Fig. 2. Experimental domain of this study. The star indicates the Hopen site cloud fractions were used only as input data in the training process of the
where the downward longwave radiative flux was observed. deep learning model. For these variables, xmin and xmax were determined
4
D.-H. Kim and H.M. Kim Expert Systems With Applications 210 (2022) 118547
from the ERA5 reanalysis training data from 2016 to 2019. Table 2
The xmin and xmax for LWD were 145 W m− 2 and 370 W m− 2, Error statistics before applying deep learning post-processing model for bias
respectively. The xmin (xmax ) for 2 m temperature, latent heat flux, and corrections.
sensible heat flux were 245 K (285 K), 105 W m− 2 (-300 W m− 2), and 50 Mon. Bias [W m− 2] RMSE [W m− 2]
W m− 2 (-200 W m− 2), respectively. The low-, mid-, and high-cloud ERA5 PW PW_DA ERA5 PW PW_DA
fractions ranged between 0 and 1. Because the cloud fraction is often 1 14.40 9.62 8.73 32.84 33.85 32.46
relatively zero compared to other input variables, the xmin and xmax for 2 0.60 − 1.60 − 2.37 24.96 29.97 29.80
the low-, mid-, and high-cloud fractions were determined as − 0.01 and 3 − 6.48 − 6.25 − 6.09 25.37 29.60 29.24
4 − 0.51 0.39 0.91 27.84 28.18 28.44
1, respectively. This results in scaled cloud fractions that are greater
5 − 14.24 − 7.97 − 4.30 24.06 26.80 25.25
than zero, preventing the weighting in the deep learning model from 6 − 14.60 − 3.99 − 5.43 25.23 24.81 23.65
being multiplied by zero, which thereby reduces the possibility of hin 7 − 2.99 − 3.65 − 3.54 14.09 16.57 15.91
dered learning. 8 − 6.48 2.94 − 0.24 23.23 21.66 21.26
The xmin and xmax for each variable in the test data from the ERA5 9 − 12.60 − 0.15 − 0.87 24.56 27.41 27.38
10 − 14.56 1.38 0.47 26.42 25.71 27.19
reanalysis and PWRF forecasts without and with DA in 2020 were the
11 − 16.53 − 12.33 − 10.00 25.99 31.25 29.13
same as those for the training data. 12 –23.18 − 8.03 − 7.95 31.93 29.52 29.14
2020 − 8.10 − 2.47 − 2.56 25.54 27.11 26.57
5
D.-H. Kim and H.M. Kim Expert Systems With Applications 210 (2022) 118547
6
D.-H. Kim and H.M. Kim Expert Systems With Applications 210 (2022) 118547
Fig. 4. Histogram of the daily mean (a, c, and e) bias and (b, d, and f) root mean square error for the downward longwave radiative flux compared with observations
in 2020: simulated data before deep learning post-processing (red) and simulated data after deep learning post-processing (blue). (For interpretation of the references
to colour in this figure legend, the reader is referred to the web version of this article.)
7
D.-H. Kim and H.M. Kim Expert Systems With Applications 210 (2022) 118547
8
D.-H. Kim and H.M. Kim Expert Systems With Applications 210 (2022) 118547
Fig. 6. Monthly mean (a, c, and e) bias and (b, d, and f) root mean square error of the downward longwave radiative flux in each experiment based on observations
in 2020.
assimilation will be discussed in further studies. regions, and two-dimensional fields will be studied.
Applying the deep learning post-processing model to independent
test data resulted in reduced magnitude of bias and RMSE for LWD CRediT authorship contribution statement
simulations in all experiments. Therefore, in a situation where the same
weather phenomenon is simulated using multiple NWP models, the Dae-Hui Kim: Formal analysis, Investigation, Methodology, Soft
possibility that a deep learning model trained with existing high-quality ware, Validation, Visualization, Writing – original draft. Hyun Mee
global or regional reanalysis data can be directly applied to the forecasts Kim: Conceptualization, Formal analysis, Funding acquisition, Investi
of the local-to-regional scale NWP model was confirmed. Therefore, gation, Methodology, Project administration, Resources, Supervision,
when creating a deep learning model, the cost required to generate Validation, Writing – review & editing.
training data for a specific model can be reduced, thereby increasing the
effectiveness of the deep learning model. In the future, post-processing
of NWP model results through deep learning in other variables,
9
D.-H. Kim and H.M. Kim Expert Systems With Applications 210 (2022) 118547
10
D.-H. Kim and H.M. Kim Expert Systems With Applications 210 (2022) 118547
Huang, Y., Dong, X., Xi, B., Dolinar, E. K., Stanfield, R. E., & Qui, S. (2017). Quantifying Parrish, D. F., & Derber, J. C. (1992). The National Meteorological Center’s spectral
the Uncertainties of Reanalyzed Arctic Cloud and Radiation Properties Using statistical-interpolation analysis system. Monthly Weather Review, 120, 1747–1763.
Satellite Surface Observations. Journal of Climate, 30, 8007–8029. https://doi.org/ https://doi.org/10.1175/1520-0493(1992)120<1747:TNMCSS>2.0.CO;2
10.1175/JCLI-D-16-0722.1 Rasp, S., & Lerch, S. (2018). Neural Networks for Postprocessing Ensemble Weather
Iacono, M. J., Delamere, J. S., Mlawer, E. J., Shephard, M. W., Clough, S. A., & Forecasts. Monthly Weather Review, 146, 3885–3900. https://doi.org/10.1175/
Collins, W. D. (2008). Radiative forcing by long-lived greenhouse gases: Calculations MWR-D-18-0187.1
with the AER radiative transfer models. Journal of Geophysical Research, 113, Riihelä, A., Key, J. R., Meirink, J. F., Munneke, P. K., Palo, T., & Karlsson, K.-G. (2017).
D13103. https://doi.org/10.1029/2008JD009944 An intercomparison and validation of satellite-based surface radiative energy flux
Iacono, M. J., Mlawer, E. J., & Clough, S. A. (2000). Impact of an improved longwave estimates over the Arctic. J. Geophys. Res.-Atmos., 122, 4829–4848. https://doi.org/
radiation model, RRTM, on the energy budget and thermodynamic properties of the 10.1002/2016JD026443
NCAR community climate model, CCM3. Journal of Geophysical Research, 105, Sayeed, A., Lops, Y., Choi, Y., Jung, J., & Salman, A. K. (2021). Bias correcting and
14873–14890. https://doi.org/10.1029/2000JD900091 extending the PM forecast by CMAQ up to 7 days using deep convolutional neural
Jenkins, M., & Dai, A. (2021). The impact of sea-ice loss on Arctic climate feedbacks and networks. Atmospheric Environment, 253, Article 118376. https://doi.org/10.1016/j.
their role for Arctic amplification. Geophysical Research Letters 48, atmosenv.2021.118376
e2021GL094599. https://doi.org/10.1029/2021GL094599. Scher, S., & Messori, G. (2021). Ensemble Methods for Neural Network-Based Weather
Kapsch, M.-L., Graversen, R. G., Tjernström, M., & Bintanja, R. (2016). The Effect of Forecasts. Journal of Advances in Modeling Earth Systems13, e2020MS002331.
Downwelling Longwave and Shortwave Radiation on Arctic Summer Sea Ice. Journal https://doi.org/10.1029/2020MS002331.
of Climate, 29, 1143–1156. https://doi.org/10.1175/JCLI-D-15-0238.1 Semmler, T., McGrath, R., & Wang, S. (2012). The impact of Arctic sea ice on the Arctic
Kay, J. E., & Gettelman, A. (2009). Cloud influence on and response to seasonal Arctic sea energy budget and on the climate of the Northern mid-latitudes. Climate Dynamics,
ice loss. Journal of Geophysical Research, 114, D18204. https://doi.org/10.1029/ 39, 2675–2694. https://doi.org/10.1007/s00382-012-1353-9
2009JD011773 Sha, Y., Gagne, D. J., West, G., & Stull, R. (2020). Deep-Learning-Based Gridded
Kim, D.-H., Kim, H. M., & Hong, J. (2019). Evaluation of wind forecasts over Svalbard Downscaling of Surface Meteorological Variables in Complex Terrain. Part I: Daily
using the high-resolution Polar WRF with 3DVAR. Arctic, Antarctic, and Alpine Maximum and Minimum 2-m Temperature. Journal of Applied Meteorology and
Research, 51, 471–489. https://doi.org/10.1080/15230430.2019.1676939 Climatology., 59, 2057–2073. https://doi.org/10.1175/JAMC-D-20-0057.1
Kim, D.-H., & Kim, H. M. (2022). Effect of data assimilation in the Polar WRF with Shupe, M. D., & Intrieri, J. M. (2004). Cloud Radiative Forcing of the Arctic Surface: The
3DVAR on the prediction of radiation, heat flux, cloud, and near surface atmospheric Influence of Cloud Properties, Surface Albedo, and Solar Zenith Angle. Journal of
variables over Svalbard. Atmospheric Research, 272, Article 106155. https://doi.org/ Climate, 17, 616–628. https://doi.org/10.1175/1520-0442(2004)017<0616:
10.1016/j.atmosres.2022.106155 CRFOTA>2.0.CO;2
Kim, K.-Y., Kim, J.-Y., Kim, J., Yeo, S., Na, H., Hamlington, B. D., & Leben, R. R. (2019). Silber, I., Verlinde, J., Wang, S.-H., Bromwich, D. H., Fridlind, A. M., Cadeddu, M., …
Vertical Feedback Mechanism of Winter Arctic Amplification and Sea Ice Loss. Flynn, C. J. (2019). Cloud Influence on ERA5 and AMPS Surface Downwelling
Scientific Reports, 9, 1184. https://doi.org/10.1038/s41598-018-38109-x Longwave Radiation Biases in West Antarctica. Journal of Climate, 32, 7935–7949.
Kingma, D. P., & Ba, J. (2014). Adam: a method for stochastic optimization. arXiv https://doi.org/10.1175/JCLI-D-19-0149.1
preprint arXiv:1412.6980. https://arxiv.org/abs/1412.6980. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., & Salakhutdinov, R. (2014).
Kumar, A., Perlwitz, J., Eischeid, J., Quan, X., Xu, T., Zhang, T., … Wang, W. (2010). Dropout: A simple way to prevent neural networks from overfitting. Journal of
Contribution of sea ice loss to Arctic amplification. Geophysical Research Letters, 37, Machine Learning Research, 15, 1929–1958.
L21701. https://doi.org/10.1029/2010GL045022 Turton, J. V., Mölg, T., & Collier, E. (2020). High-resolution (1 km) Polar WRF output for
Lakshminarayanan, B., Pritzel, A., & Blundell, C. (2017). Simple and Scalable Predictive 79◦ N Glacier and the northeast of Greenland from 2014 to 2018. Earth System
Uncertainty Estimation using Deep Ensembles. Advances in Neural Information Science Data, 12, 1191–1202. https://doi.org/10.5194/essd-12-1191-2020
Processing Systems., 30, 6405–6416. https://proceedings.neurips.cc/paper/2017/ha Urrego-Blanco, J. R., Hunke, E. C., & Urban, N. (2019). Emergent Relationships Among
sh/9ef2ed4b7fd2c810847ffa5fa85bce38-Abstract.html. Sea Ice, Longwave Radiation, and the Beaufort High Circulation Exposed Through
LeCun, Y., & Bengio, Y. (1995). Convolutional networks for images, speech, and time series. Parameter Uncertainty Analysis. Journal of Geophysical Research: Oceans, 124,
The handbook of brain theory and neural networks (p. 3361 (10).). MIT Press. 9572–9589. https://doi.org/10.1029/2019JC014979
Liu, Y., & Key, J. R. (2014). Less winter cloud aids summer 2013 Arctic sea ice return van Wessem, J. M., Reijmer, C. H., Lenaerts, J. T. M., van de Berg, W. J., van den
from 2012 minimum. Environmental Research Letters, 9, Article 044002. https://doi. Broeke, M. R., & van Meijgaard, E. (2014). Updated cloud physics in a regional
org/10.1088/1748-9326/9/4/044002 atmospheric climate model improves the modelled surface energy balance of
Maslanik, J. A., Fowler, C., Stroeve, J., Drobot, S., Zwally, J., Yi, D., & Emery, W. (2007). Antarctica. Cryosphere, 8, 125–135. https://doi.org/10.5194/tc-8-125-2014
A younger, thinner Arctic ice cover: Increased potential for rapid extensive sea-ice Veldkamp, S., Whan, K., Dirksen, S., & Schmeits, M. (2021). Statistical Postprocessing of
loss. Geophysical Research Letters, 34, L24501. https://doi.org/10.1029/ Wind Speed Forecasts Using Convolutional Neural Networks. Monthly Weather
2007GL032043 Review, 149, 1141–1152. https://doi.org/10.1175/MWR-D-20-0219.1
Maturilli, M., Herber, A., & König-Langlo, G. (2015). Surface radiation climatology for Vihma, T. (2014). Effects of Arctic Sea Ice Decline on Weather and Climate: A Review.
Ny-Ålesund, Svalbard (78.9◦ N), basic observations for trend detection. Theoretical Surveys In Geophysics, 35, 1175–1214. https://doi.org/10.1007/s10712-014-9284-0
and Applied Climatology, 120, 331–339. https://doi.org/10.1007/s00704-014-1173- Walsh, J. E., & Chapman, W. L. (1998). Arctic Cloud-Radiation-Temperature Associations
4 in Observational Data and Atmospheric Reanalyses. Journal of Climate, 11,
Miller, N. B., Shupe, M. D., Cox, C. J., Noone, D., Persson, P. O. G., & Steffen, K. (2017). 3030–3045. https://doi.org/10.1175/1520-0442(1998)011<3030:ACRTAI>2.0.
Surface energy budget responses to radiative forcing at Summit, Greenland. CO;2
Cryosphere, 11, 497–516. https://doi.org/10.5194/tc-11-497-2017 Xu, W., Liu, P., Cheng, L., Zhou, Y., Xia, Q., Gong, Y., & Liu, Y. (2021). Multi-step wind
Morrison, H., Curry, J. A., & Khvorostyanov, V. I. (2005). A new double-moment speed prediction by combining a WRF simulation and an error correction strategy.
microphysics parameterization for application in cloud and climate models. Part I: Renewing Energy, 163, 772–782. https://doi.org/10.1016/j.renene.2020.09.032
Description. Journal of the Atmospheric Sciences, 62, 1665–1677. https://doi.org/ Yan, G., Jiao, Z.-H., Wang, T., & Mu, X. (2020). Modeling surface longwave radiation
10.1175/JAS3446.1 over high-relief terrain. Remote Sensing of Environment, 237, Article 111556. https://
Nakanishi, M., & Niino, H. (2006). An Improved Mellor-Yamada Level-3 Model: Its doi.org/10.1016/j.rse.2019.111556
Numerical Stability and Application to a Regional Prediction of Advection Fog. Zjavka, L. (2016). Numerical weather prediction revisions using the locally trained
Boundary-Layer Meteorology, 119, 397–407. https://doi.org/10.1007/s10546-005- differential polynomial network. Expert Systems with Applications, 44, 265–274.
9030-8 https://doi.org/10.1016/j.eswa.2015.08.057
Niemelä, S., Räisänen, P., & Savijärvi, H. (2001). Comparison of surface radiative flux
parameterizations Part I: Longwave radiation. Atmospheric Research, 58, 1–18.
https://doi.org/10.1016/S0169-8095(01)00084-9
11