Professional Documents
Culture Documents
This Version May No Longer Be Accessible
This Version May No Longer Be Accessible
This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which
this version may no longer be accessible.
arXiv:2103.08741v1 [eess.IV] 15 Mar 2021
2
70
Kappa AA
TABLE II
N UMBER OF L ABELED S AMPLES IN THE I NDIAN P INES , PAVIA U NIVERSITY, B OTSWANA , AND MUUFL G ULFPORT DATA S ETS .
• MVPCA [36]: A ranking-based band selection method lected according to entropies of the reconstructed bands.
that uses an eigenanalysis-based criterion to prioritize • DRL: Our proposed deep reinforcement learning model
spectral bands. for unsupervised hyperspectral band selection.
• ICA [37]: A band selection approach that compares mean 3) Classification Setting: We consider four commonly used
absolute independent component analysis (ICA) coeffi- classifiers in the remote sensing community to implement
cients of individual spectral bands and picks independent hyperspectral image classification. They are as follows:
ones including the maximum information. • k-NN: A k-nearest neighbors algorithm (the number of
• IE [38]: A ranking-based band selection algorithm in neighbors is set to 3).
which band priority is calculated based on information • RF: A random forest being made up of 200 decision trees.
entropy. • MLP: A multilayer perceptron that consists of three fully
• MEV-SFS [40]: A searching-based band selection method connected layers. The first two layers contain 256 units,
that combines maximum ellipsoid volume (MEV) [71] and their outputs are activated by Leaky RuLU. For the
method with sequential forward search (SFS). The MEV last layer, the number of units equals the number of
model deems an optimal band subset as a band combi- classes, and the used activation function is softmax. In
nation with the maximum volume. the learning phase, we select Adam as the optimizer and
• OPBS [40]: An accelerated version of MEV-SFS that define the loss function as categorical cross-entropy. The
takes advantage of a relationship between orthogonal learning rate is set to 0.0005, and the training epochs is
projections (OPs) and the ellipsoid volume of bands to 2000 for the purpose of sufficient learning.
find out an optimal band combination. 3
• SVM-RBF : A support vector machine (SVM) equipped
• WaLuDi [41]: A hierarchical clustering-based band se- with radial basis function (RBF) kernel. A five-fold cross-
lection method that uses Kullback-Leibler divergence as validation method is utilized to determine optimal hyper-
the criterion of the clustering algorithm. parameters, i.e., γ and C.
• E-FDPC [42]: A clustering-based band selection ap- For both the Indian Pines and the Botswana data sets, we
proach that makes use of an enhanced version of fast randomly select 10% samples from each class as training
density peak-based clustering (FDPC) [72] algorithm by instances, while the remaining are exploited to test models.
introducing an exponential learning rule and a parameter Regarding the Pavia University and MUUFL Gulfport data
to control the weight between local density and intra- sets, 1% samples per class are chosen randomly to build the
cluster similarity. training set, and all the other samples are utilized for the
• OCF [43]: A band selection method using an optimal purpose of testing. In order to know the stability of various
clustering algorithm that is capable of achieving optimal band selection models, final results are achieved by averaging
clustering results for an objective function with a care- 10 individual runs, and we report the mean and standard
fully designed constraint. deviation of performance metrics of different approaches over
• ASPS [44]: A clustering-based band selection method the 10 runs.
that exploits an adaptive subspace partition strategy.
• DARecNet [45]: An unsupervised convolutional neural C. Information Entropy or Correlation: Whose Call Is It in
network (CNN) for band selection tasks. It employs a Building The Reward Scheme?
dual-attention mechanism, i.e., spatial position attention Fig. 3 compares two instantiations of the reward scheme,
and channel attention, to learn to reconstruct hyperspec- namely information entropy and correlation coefficient (cf.
tral images. Once the network is trained, bands are se-
3 https://www.csie.ntu.edu.tw/∼cjlin/libsvmtools/
8
TABLE III
C OMPARISONS IN QUANTITATIVE METRICS AMONG DIFFERENT BAND SELECTION METHODS ON THE I NDIAN P INES DATA SET WITH 30 SELECTED
BANDS . W E REPORT THE MEAN AND STANDARD DEVIATION OF PERFORMANCE METRICS OF DIFFERENT APPROACHES OVER 10 INDIVIDUAL RUNS . T HE
BEST CLASSIFICATION PERFORMANCE IS HIGHLIGHTED IN BOLD .
TABLE IV
C OMPARISONS IN QUANTITATIVE METRICS AMONG DIFFERENT BAND SELECTION METHODS ON THE PAVIA U NIVERSITY DATA SET WITH 30 SELECTED
BANDS . W E REPORT THE MEAN AND STANDARD DEVIATION OF PERFORMANCE METRICS OF DIFFERENT APPROACHES OVER 10 INDIVIDUAL RUNS . T HE
BEST CLASSIFICATION PERFORMANCE IS HIGHLIGHTED IN BOLD .
TABLE V
C OMPARISONS IN QUANTITATIVE METRICS AMONG DIFFERENT BAND SELECTION METHODS ON THE B OTSWANA DATA SET WITH 30 SELECTED BANDS .
W E REPORT THE MEAN AND STANDARD DEVIATION OF PERFORMANCE METRICS OF DIFFERENT APPROACHES OVER 10 INDIVIDUAL RUNS . T HE BEST
CLASSIFICATION PERFORMANCE IS HIGHLIGHTED IN BOLD .
TABLE VI
C OMPARISONS IN QUANTITATIVE METRICS AMONG DIFFERENT BAND SELECTION METHODS ON THE MUUFL G ULFPORT DATA SET WITH 30 SELECTED
BANDS . W E REPORT THE MEAN AND STANDARD DEVIATION OF PERFORMANCE METRICS OF DIFFERENT APPROACHES OVER 10 INDIVIDUAL RUNS . T HE
BEST CLASSIFICATION PERFORMANCE IS HIGHLIGHTED IN BOLD .
70 76 79 81
65 71 74 76
60 66 69 71
55 61 64 66
50 56 59 61
45 51 54 56
5 10 20 30 40 50 60 5 10 20 30 40 50 60 5 10 20 30 40 50 60 5 10 20 30 40 50 60
MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS
OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF
ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands
Fig. 4. OA curves of different band selection methods on the Indian Pines data set. The x-axis indicates OA (%), and the y-axis indicates the number of
selected bands. (a) OA by k-NN. (b) OA by RF. (c) OA by MLP. (d) OA by SVM-RBF. All OAs are achieved by averaging 10 individual runs.
81 82 86 90
79 80 84 88
77 78 82 86
75 76 80 84
73 74 78 82
71 72 76 80
5 10 20 30 40 50 60 5 10 20 30 40 50 60 5 10 20 30 40 50 60 5 10 20 30 40 50 60
MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS
OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF
ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands
Fig. 5. OA curves of different band selection methods on the Pavia University data set. The x-axis indicates OA (%), and the y-axis indicates the number
of selected bands. (a) OA by k-NN. (b) OA by RF. (c) OA by MLP. (d) OA by SVM-RBF. All OAs are achieved by averaging 10 individual runs.
87 87 91 93
82 82 86 88
77 77 81 83
72 72 76 78
67 67 71 73
5 10 20 30 40 50 60 5 10 20 30 40 50 60 5 10 20 30 40 50 60 5 10 20 30 40 50 60
MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS
OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF
ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands
Fig. 6. OA curves of different band selection methods on the Botswana data set. The x-axis indicates OA (%), and the y-axis indicates the number of
selected bands. (a) OA by k-NN. (b) OA by RF. (c) OA by MLP. (d) OA by SVM-RBF. All OAs are achieved by averaging 10 individual runs.
80 81 84 84
78 79 82 82
76 77 80 80
74 75 78 78
72 73 76 76
70 71 74 74
5 10 20 30 40 50 60 5 10 20 30 40 50 60 5 10 20 30 40 50 60 5 10 20 30 40 50 60
MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS MVPCA ICA IE MEV-SFS
OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF OPBS WaLuDi E-FDPC OCF
ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands ASPS DARecNet DRL All Bands
Fig. 7. OA curves of different band selection methods on the MUUFL Gulfport data set. The x-axis indicates OA (%), and the y-axis indicates the number
of selected bands. (a) OA by k-NN. (b) OA by RF. (c) OA by MLP. (d) OA by SVM-RBF. All OAs are achieved by averaging 10 individual runs.
Section II-A), on the Pavia University data set. To quanti- can achieve higher OA, AA, and Kappa coefficient compared
tatively evaluate them, we make use of k-NN to perform clas- to the latter. Moreover, the computation cost of information
sification using spectral bands selected by models using these entropy is lower than that of correlation coefficient. Hence we
two schemes. From Fig. 3, it can be observed that the former choose information entropy as the reward scheme in our model
10
1 1
normalized spectral values
1 1
0 0
0 20 40 60 80 100 120 140 0 20 40 60
Selected bands Water Hippo grass Selected bands Trees Mostly grass
Floodplain grasses 1 Floodplain grasses 2 Reeds 1 Mixed ground surface Dirt/Sand Road
Riparian Firescar 2 Island interior Water Building shadow Buildings
Acacia woodlands Acacia shrublands Acacia grasslands Sidewalk Yellow curb Cloth panels
Short mopane Mixed mopane Exposed soils
Indian Pines
data set
tral image classification,” IEEE Transactions on Geoscience and Remote
MUUFL Gulfport
actions on Geoscience and Remote Sensing, vol. 51, no. 9, pp. 4800–
data set
4815, 2013.
[8] J. Li, P. R. Marpu, A. Plaza, J. M. Bioucas-Dias, and J. A. Benedikts-
son, “Generalized composite kernel framework for hyperspectral image
classification,” IEEE transactions on geoscience and remote sensing,
Pavia University color legend vol. 51, no. 9, pp. 4816–4829, 2013.
1 2 3 4 5 6 7 8 9 [9] L. Mou, P. Ghamisi, and X. X. Zhu, “Deep recurrent neural networks for
Botswana color legend hyperspectral image classification,” IEEE Transactions on Geoscience
1 2 3 4 5 6 7 8 9 10 11 12 13 14 and Remote Sensing, vol. 55, no. 7, pp. 3639–3655, 2017.
MUUFL Gulfport color legend [10] W. Li, G. Wu, F. Zhang, and Q. Du, “Hyperspectral image classification
1 2 3 4 5 6 7 8 9 10 11
using deep pixel-pair features,” IEEE Transactions on Geoscience and
Remote Sensing, vol. 55, no. 2, pp. 844–853, 2017.
Indian Pines color legend
[11] L. Mou, P. Ghamisi, and X. X. Zhu, “Unsupervised spectral–spatial
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
feature learning via deep residual conv–deconv network for hyperspec-
tral image classification,” IEEE Transactions on Geoscience and Remote
Fig. 12. Classification maps using 30 bands selected by the proposed DRL Sensing, vol. 56, no. 1, pp. 391–406, 2018.
model and an SVM-RBF classifier on the four data sets. [12] J. M. Haut, M. E. Paoletti, J. Plaza, A. Plaza, and J. Li, “Hyperspectral
image classification using random occlusion data augmentation,” IEEE
Geoscience and Remote Sensing Letters, vol. 16, no. 11, pp. 1751–1755,
to the learned policy. We conduct extensive experiments, and 2019.
[13] Z. Zhong, J. Li, Z. Luo, and M. Chapman, “Spectral-spatial residual
results show the effectiveness of our approach. Moreover, network for hyperspectral image classification: A 3-D deep learning
two instantiations of the reward scheme in Section II-A are framework,” IEEE Transactions on Geoscience and Remote Sensing,
quantitatively compared, and we believe that more alternatives vol. 56, no. 2, pp. 847–858, 2018.
[14] M. E. Paoletti, J. M. Haut, R. Fernández-Beltran, J. Plaza, A. J. Plaza,
are possible and may improve results. and F. Pla, “Deep pyramidal residual networks for spectral-spatial
In the future, several studies intend to be carried out. For hyperspectral image classification,” IEEE Transactions on Geoscience
example, combining deep reinforcement learning and some and Remote Sensing, vol. 57, no. 2, pp. 740–754, 2019.
heuristic band selection frameworks (e.g., the clustering-based [15] Y. Xu, L. Zhang, B. Du, and F. Zhang, “Spectral-spatial unified
networks for hyperspectral image classification,” IEEE Transactions on
method) is likely to offer better band selection solutions. Con- Geoscience and Remote Sensing, vol. 56, no. 10, pp. 5893–5909, 2018.
sidering that different classes may have different optimal band [16] W. Li and Q. Du, “Collaborative representation for hyperspectral
subsets (with a variable number of bands), how to determine anomaly detection,” IEEE Transactions on Geoscience and Remote
Sensing, vol. 53, no. 3, pp. 1463–1474, 2015.
the best band combination for each category is an interesting [17] Y. Zhang, B. Du, L. Zhang, and S. Wang, “A low-rank and sparse matrix
but challenging problem. A supervised deep reinforcement decomposition-based Mahalanobis distance method for hyperspectral
learning model may be able to provide insights. Moreover, anomaly detection,” IEEE Transactions on Geoscience and Remote
Sensing, vol. 54, no. 3, pp. 1376–1389, 2016.
we believe that deep reinforcement learning can be applied to [18] Y. Yuan, D. Ma, and Q. Wang, “Hyperspectral anomaly detection by
more remote sensing applications, such as multitemporal data graph pixel selection,” IEEE Transactions on Cybernetics, vol. 46,
analysis, visual reasoning in airborne or space-borne images, no. 12, pp. 3123–3134, 2016.
[19] X. Kang, X. Zhang, S. Li, K. Li, J. Li, and J. A. Benediktsson,
and other combinatorial optimization tasks in remote sensing. “Hyperspectral anomaly detection with attribute and edge-preserving
filters,” IEEE Transactions on Geoscience and Remote Sensing, vol. 55,
no. 10, pp. 5600–5611, 2017.
ACKNOWLEDGEMENT
[20] Z. Huang, L. Fang, and S. Li, “Subpixel-pixel-superpixel guided fusion
The authors would like to thank F. Zhang for helpful for hyperspectral anomaly detection,” IEEE Transactions on Geoscience
and Remote Sensing, DOI:10.1109/TGRS.2019.2961703.
discussions on this paper.
[21] L. Bruzzone and S. B. Serpico, “An iterative technique for the detection
of land-cover transitions in multitemporal remote-sensing images,” IEEE
R EFERENCES Transactions on Geoscience and Remote Sensing, vol. 35, no. 4, pp.
858–867, 1997.
[1] G. Camps-Valls, D. Tuia, L. Bruzzone, and J. A. Benediktsson, “Ad- [22] F. Bovolo and L. Bruzzone, “A theoretical framework for unsupervised
vances in hyperspectral image classification: Earth monitoring with change detection based on change vector analysis in the polar domain,”
statistical learning methods,” IEEE Signal Processing Magazine, vol. 31, IEEE Transactions on Geoscience and Remote Sensing, vol. 45, no. 1,
no. 1, pp. 45–54, 2014. pp. 218–236, 2007.
13
[23] F. Bovolo, S. Marchesi, and L. Bruzzone, “A framework for automatic [44] Q. Wang, Q. Li, and X. Li, “Hyperspectral band selection via adaptive
and unsupervised detection of multiple changes in multitemporal im- subspace partition strategy,” IEEE Journal of Selected Topics in Applied
ages,” IEEE Transactions on Geoscience and Remote Sensing, vol. 50, Earth Observations and Remote Sensing, vol. 12, no. 12, pp. 4940–4950,
no. 6, pp. 2196–2212, 2012. 2019.
[24] C. Wu, B. Du, and L. Zhang, “Slow feature analysis for change detection [45] S. K. Roy, S. Das, T. Song, and B. Chanda, “DARecNet-BS:
in multispectral imagery,” IEEE Transactions on Geoscience and Remote Unsupervised dual-attention reconstruction network for hyperspec-
Sensing, vol. 52, no. 5, pp. 2858–2874, 2014. tral band selection,” IEEE Geoscience and Remote Sensing Letters,
[25] F. Bovolo and L. Bruzzone, “The time variable in data fusion: A change DOI:10.1109/LGRS.2020.3013235.
detection perspective,” IEEE Geoscience and Remote Sensing Magazine, [46] J. Yin, Y. Wang, and Z. Zhao, “Optimal band selection for hyperspectral
vol. 3, no. 3, pp. 8–26, 2015. image classification based on inter-class separability,” in International
[26] M. Zanetti, F. Bovolo, and L. Bruzzone, “Rayleigh-Rice mixture param- Symposium on Photonics and Optoelectronics, 2010.
eter estimation via EM algorithm for change detection in multispectral [47] Y.-L. Chang, J.-N. Liu, Y.-L. Chen, W.-Y. Chang, T.-J. Hsieh, and
images,” IEEE Transactions on Image Processing, vol. 24, no. 12, pp. B. Huang, “Hyperspectral band selection based on parallel particle
5004–5016, 2015. swarm optimization and impurity function band prioritization schemes,”
[27] H. Lyu, H. Lu, and L. Mou, “Learning a transferable change rule from Journal of Applied Remote Sensing, vol. 8, no. 1, p. 084798, 2014.
a recurrent neural network for land cover change detection,” Remote [48] Q. Wang, J. Lin, and Y. Yuan, “Salient band selection for hyperspectral
Sensing, vol. 8, no. 6, p. 506, 2016. image classification via manifold ranking,” IEEE Transactions on Neural
[28] Q. Wang, Z. Yuan, Q. Du, and X. Li, “GETNET: A general end-to-end Networks and Learning Systems, vol. 27, no. 6, pp. 1279–1289, 2016.
2-D CNN framework for hyperspectral image change detection,” IEEE [49] W. Sun, L. Zhang, B. Du, W. Li, and Y. M. Lai, “Band selection using
Transactions on Geoscience and Remote Sensing, vol. 57, no. 1, pp. improved sparse subspace clustering for hyperspectral imagery classifi-
3–13, 2019. cation,” IEEE Journal of Selected Topics in Applied Earth Observations
[29] L. Mou, L. Bruzzone, and X. X. Zhu, “Learning spectral-spatial- and Remote Sensing, vol. 8, no. 6, pp. 2784–2797, 2015.
temporal features via a recurrent convolutional neural network for [50] W. Sun, J. Peng, G. Yang, and Q. Du, “Fast and latent low-rank
change detection in multispectral imagery,” IEEE Transactions on Geo- subspace clustering for hyperspectral band selection,” IEEE Transactions
science and Remote Sensing, vol. 57, no. 2, pp. 924–935, 2019. on Geoscience and Remote Sensing, DOI:10.1109/tgrs.2019.2959342.
[30] B. Du, L. Ru, C. Wu, and L. Zhang, “Unsupervised deep slow feature [51] S. Patra, P. Modi, and L. Bruzzone, “Hyperspectral band selection based
analysis for change detection in multi-temporal remote sensing images,” on rough set,” IEEE Transactions on Geoscience and Remote Sensing,
IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 12, vol. 53, no. 10, pp. 5495–5503, 2015.
pp. 9976–9992, 2019. [52] W. Sun and Q. Du, “Hyperspectral band selection: A review,” IEEE
[31] W. Zhao, L. Mou, J. Chen, Y. Bo, and W. J. Emery, “Incorporating Geoscience and Remote Sensing Magazine, vol. 7, no. 2, pp. 118–139,
metric learning and adversarial network for seasonal invariant change 2019.
detection,” IEEE Transactions on Geoscience and Remote Sensing, [53] C. Yang, L. Bruzzone, H. Zhao, Y. Tan, and R. Guan, “Superpixel-based
vol. 58, no. 4, pp. 2720–2731, 2020. unsupervised band selection for classification of hyperspectral images,”
[32] S. Saha, F. Bovolo, and L. Bruzzone, “Unsupervised deep change IEEE Transactions on Geoscience and Remote Sensing, vol. 56, no. 12,
vector analysis for multiple-change detection in VHR images,” IEEE pp. 7230–7245, 2018.
Transactions on Geoscience and Remote Sensing, vol. 57, no. 6, pp.
[54] X. Zhang, Z. Gao, L. Jiao, and H. Zhou, “Multifeature hyperspectral im-
3677–3693, 2019.
age classification with local and nonlocal spatial information via Markov
[33] S. Saha, L. Mou, C. Qiu, X. X. Zhu, F. Bovolo, and L. Bruzzone,
random field in semantic space,” IEEE Transactions on Geoscience and
“Unsupervised deep joint segmentation of multi-temporal high resolu-
Remote Sensing, vol. 56, no. 3, pp. 1409–1424, 2018.
tion images,” IEEE Transactions on Geoscience and Remote Sensing,
[55] J. Feng, J. Chen, Q. Sun, R. Shang, X. Cao, X. Zhang, and L. Jiao,
DOI:10.1109/TGRS.2020.2990640.
“Convolutional neural network based on bandwise-independent convo-
[34] J. Wang and C.-I. Chang, “Independent component analysis-based
lution and hard thresholding for hyperspectral band selection,” IEEE
dimensionality reduction with applications in hyperspectral image anal-
Transactions on Cybernetics, DOI:10.1109/TCYB.2020.3000725.
ysis,” IEEE Transactions on Geoscience and Remote Sensing, vol. 44,
no. 6, pp. 1586–1600, 2006. [56] M. Belkin and P. Niyogi, “Laplacian eigenmaps for dimensionality
[35] T. V. Bandos, L. Bruzzone, and G. Camps-Valls, “Classification of reduction and data representation,” Neural Computation, vol. 15, no. 6,
hyperspectral images with regularized linear discriminant analysis,” pp. 1373–1396, 2003.
IEEE Transactions on Geoscience and Remote Sensing, vol. 47, no. 3, [57] S. T. Roweis and L. K. Saul, “Nonlinear dimensionality reduction by
pp. 862–873, 2009. locally linear embedding,” Science, vol. 290, no. 5500, pp. 2323–2326,
[36] C.-I. Chang, Q. Du, T.-L. Sun, and M. L. G. Althouse, “A joint band 2000.
prioritization and band-decorrelation approach to band selection for [58] J. B. Tenenbaum, V. De Silva, and J. C. Langford, “A global geometric
hyperspectral image classification,” IEEE Transactions on Geoscience framework for nonlinear dimensionality reduction,” Science, vol. 290,
and Remote Sensing, vol. 37, no. 6, pp. 2631–2641, 1999. no. 5500, pp. 2319–2323, 2000.
[37] H. Du, H. Qi, X. Wang, R. Ramanath, and W. E. Snyder, “Band [59] V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G.
selection using independent component analysis for hyperspectral image Bellemare, A. Graves, M. A. Riedmiller, A. Fidjeland, G. Ostrovski,
processing,” in Applied Image Pattern Recognition Workshop (AIPRW), S. Petersen, C. Beattie, A. Sadik, I. Antonoglou, H. King, D. Kumaran,
2003. D. Wierstra, S. Legg, and D. Hassabis, “Human-level control through
[38] L. Xie, G. Li, L. Peng, Q. Chen, Y. Tan, and M. Xiao, “Band selection deep reinforcement learning,” Nature, vol. 518, no. 7540, pp. 529–533,
algorithm based on information entropy for hyperspectral image classi- 2015.
fication,” Journal of Applied Remote Sensing, vol. 11, no. 2, 026018, [60] J. C. Caicedo and S. Lazebnik, “Active object localization with deep
2017. reinforcement learning,” in IEEE International Conference on Computer
[39] C. Sheffield, “Selecting band combinations from multi spectral data,” Vision (ICCV), 2015.
Photogrammetric Engineering and Remote Sensing, vol. 58, no. 6, pp. [61] D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre, G. van den
681–687, 1985. Driessche, J. Schrittwieser, I. Antonoglou, V. Panneershelvam, M. Lanc-
[40] W. Zhang, X. Li, Y. Dou, and L. Zhao, “A geometry-based band tot, S. Dieleman, D. Grewe, J. Nham, N. Kalchbrenner, I. Sutskever,
selection approach for hyperspectral image analysis,” IEEE Transactions T. P. Lillicrap, M. Leach, K. Kavukcuoglu, T. Graepel, and D. Hassabis,
on Geoscience and Remote Sensing, vol. 56, no. 8, pp. 4318–4333, 2018. “Mastering the game of Go with deep neural networks and tree search,”
[41] A. Martı́nez-Usó, F. Pla, J. M. Sotoca, and P. Garcı́a-Sevilla, “Clustering- Nature, vol. 529, no. 7587, pp. 484–489, 2016.
based hyperspectral band selection using information measures,” IEEE [62] A. E. Sallab, M. Abdou, E. Perot, and S. Yogamani, “Deep reinforcement
Transactions on Geoscience and Remote Sensing, vol. 45, no. 12, pp. learning framework for autonomous driving,” Electronic Imaging, vol.
4158–4171, 2007. 2017, no. 19, pp. 70–76, 2017.
[42] S. Jia, G. Tang, J. Zhu, and Q. Li, “A novel ranking-based clustering [63] K. Shao, Z. Tang, Y. Zhu, N. Li, and D. Zhao, “A survey of deep
approach for hyperspectral band selection,” IEEE Transactions on Geo- reinforcement learning in video games,” arXiv:1912.10944.
science and Remote Sensing, vol. 54, no. 1, pp. 88–102, 2016. [64] R. Bellman, “A markovian decision process,” Journal of Mathematics
[43] Q. Wang, F. Zhang, and X. Li, “Optimal clustering framework for and Mechanics, vol. 6, no. 5, pp. 679–684, 1957.
hyperspectral band selection,” IEEE Transactions on Geoscience and [65] C. J. C. H. Watkins, “Learning from delayed rewards,” Ph.D. disserta-
Remote Sensing, vol. 56, no. 10, pp. 5910–5922, 2018. tion, University of Cambridge, 1989.
14