Professional Documents
Culture Documents
10.1007@978 981 13 7123 3
10.1007@978 981 13 7123 3
Songlin Sun
Meixia Fu
Lexi Xu Editors
Volume 550
Series Editors
Leopoldo Angrisani, Department of Electrical and Information Technologies Engineering, University of Napoli
Federico II, Napoli, Italy
Marco Arteaga, Departament de Control y Robótica, Universidad Nacional Autónoma de México, Coyoacán,
Mexico
Bijaya Ketan Panigrahi, Electrical Engineering, Indian Institute of Technology Delhi, New Delhi, Delhi, India
Samarjit Chakraborty, Fakultät für Elektrotechnik und Informationstechnik, TU München, München, Germany
Jiming Chen, Zhejiang University, Hangzhou, Zhejiang, China
Shanben Chen, Materials Science & Engineering, Shanghai Jiao Tong University, Shanghai, China
Tan Kay Chen, Department of Electrical and Computer Engineering, National University of Singapore,
Singapore, Singapore
Rüdiger Dillmann, Humanoids and Intelligent Systems Lab, Karlsruhe Institute for Technology, Karlsruhe,
Baden-Württemberg, Germany
Haibin Duan, Beijing University of Aeronautics and Astronautics, Beijing, China
Gianluigi Ferrari, Università di Parma, Parma, Italy
Manuel Ferre, Centre for Automation and Robotics CAR (UPM-CSIC), Universidad Politécnica de Madrid,
Madrid, Madrid, Spain
Sandra Hirche, Department of Electrical Engineering and Information Science, Technische Universität
München, München, Germany
Faryar Jabbari, Department of Mechanical and Aerospace Engineering, University of California, Irvine, CA,
USA
Limin Jia, State Key Laboratory of Rail Traffic Control and Safety, Beijing Jiaotong University, Beijing, China
Janusz Kacprzyk, Systems Research Institute, Polish Academy of Sciences, Warsaw, Poland
Alaa Khamis, German University in Egypt El Tagamoa El Khames, New Cairo City, Egypt
Torsten Kroeger, Stanford University, Stanford, CA, USA
Qilian Liang, Department of Electrical Engineering, University of Texas at Arlington, Arlington, TX, USA
Ferran Martin, Departament d’Enginyeria Electrònica, Universitat Autònoma de Barcelona, Bellaterra,
Barcelona, Spain
Tan Cher Ming, College of Engineering, Nanyang Technological University, Singapore, Singapore
Wolfgang Minker, Institute of Information Technology, University of Ulm, Ulm, Germany
Pradeep Misra, Department of Electrical Engineering, Wright State University, Dayton, OH, USA
Sebastian Möller, Quality and Usability Lab, TU Berlin, Berlin, Germany
Subhas Mukhopadhyay, School of Engineering & Advanced Technology, Massey University,
Palmerston North, Manawatu-Wanganui, New Zealand
Cun-Zheng Ning, Electrical Engineering, Arizona State University, Tempe, AZ, USA
Toyoaki Nishida, Graduate School of Informatics, Kyoto University, Kyoto, Kyoto, Japan
Federica Pascucci, Dipartimento di Ingegneria, Università degli Studi “Roma Tre”, Rome, Italy
Yong Qin, State Key Laboratory of Rail Traffic Control and Safety, Beijing Jiaotong University, Beijing, China
Gan Woon Seng, School of Electrical & Electronic Engineering, Nanyang Technological University,
Singapore, Singapore
Joachim Speidel, Institute of Telecommunications, Universität Stuttgart, Stuttgart, Baden-Württemberg,
Germany
Germano Veiga, Campus da FEUP, INESC Porto, Porto, Portugal
Haitao Wu, Academy of Opto-electronics, Chinese Academy of Sciences, Beijing, China
Junjie James Zhang, Charlotte, NC, USA
The book series Lecture Notes in Electrical Engineering (LNEE) publishes the latest developments
in Electrical Engineering - quickly, informally and in high quality. While original research
reported in proceedings and monographs has traditionally formed the core of LNEE, we also
encourage authors to submit books devoted to supporting student education and professional
training in the various fields and applications areas of electrical engineering. The series cover
classical and emerging topics concerning:
• Communication Engineering, Information Theory and Networks
• Electronics Engineering and Microelectronics
• Signal, Image and Speech Processing
• Wireless and Mobile Communication
• Circuits and Systems
• Energy Systems, Power Electronics and Electrical Machines
• Electro-optical Engineering
• Instrumentation Engineering
• Avionics Engineering
• Control Systems
• Internet-of-Things and Cybersecurity
• Biomedical Devices, MEMS and NEMS
For general information about this book series, comments or suggestions, please contact leontina.
dicecco@springer.com.
To submit a proposal or request further information, please contact the Publishing Editor in
your country:
China
Jasmine Dou, Associate Editor (jasmine.dou@springer.com)
India
Swati Meherishi, Executive Editor (swati.meherishi@springer.com)
Aninda Bose, Senior Editor (aninda.bose@springer.com)
Japan
Takeyuki Yonezawa, Editorial Director (takeyuki.yonezawa@springer.com)
South Korea
Smith (Ahram) Chae, Editor (smith.chae@springer.com)
Southeast Asia
Ramesh Nath Premnath, Editor (ramesh.premnath@springer.com)
USA, Canada:
Michael Luby, Senior Editor (michael.luby@springer.com)
All other Countries:
Leontina Di Cecco, Senior Editor (leontina.dicecco@springer.com)
Christoph Baumann, Executive Editor (christoph.baumann@springer.com)
** Indexing: The books of this series are submitted to ISI Proceedings, EI-Compendex,
SCOPUS, MetaPress, Web of Science and Springerlink **
Lexi Xu
Editors
123
Editors
Songlin Sun Meixia Fu
Beijing University of Posts Beijing University of Posts
and Telecommunications and Telecommunications
Beijing, China Beijing, China
Lexi Xu
China Unicom
Beijing, China
This Springer imprint is published by the registered company Springer Nature Singapore Pte Ltd.
The registered company address is: 152 Beach Road, #21-01/04 Gateway East, Singapore 189721,
Singapore
Preface
Songlin Sun
Yue Wang
ICSINC 2018 General Co-chairs
v
Organization
Committee Members
General Co-chairs
General Secretary
vii
viii Organization
Sponsors
Springer
China Unicom
HuaCeXinTong
Contents
ix
x Contents
Abstract. Lots of mobility data have been generated with the emergence of
smart devices and location-based services. The prediction of user mobility has
become a key factor driving the rapid development of many location applica-
tions. Location prediction has attracted more and more attention in various fields,
and many location prediction algorithms have been proposed. The data currently
used for researches has many problems such as data noise and redundancy. Many
researches directly used raw data and did not consider spatiotemporal charac-
teristics of historical data enough, which leads to low prediction accuracy. This
paper proposes a point-of-interest discovering algorithm, which fully considers
spatiotemporal characteristics of data. By combining the location preference of
users for location with the Hidden Markov Model (HMM), we propose LP-HMM
(Location Preference-based Hidden Markov Model), a location prediction model
based on location preference and HMM. The proposed model is compared with
other location prediction models driven by the massive and real mobile dataset
Geolife. The experiment results show that the prediction accuracy of the pro-
posed model can achieve 6.4% and 7% higher than Gaussian Mixture Model
(GMM) and traditional HMM respectively.
1 Introduction
cellular tower only when a mobile phone event occurs, resulting in the data to be sparse
in time and space. The coverage of smart traffic card data is small and the data can be
applied in specific scenarios merely. The shortcomings in smart traffic card data limit
its application in the research of moving object mobility patterns.
Studies of human mobility patterns show that the human moving trajectories have
high temporal and spatial regularity, so the potential visited location can be predicted by
human historical trajectory data. In order to exploit the common regularity of human
movement and build the prediction model of potential visited location, many scholars
have done some researches on human mobility pattern and location prediction in recent
years. The common location prediction algorithms include Bayesian Model, Neural
Network, Markov Model, Hidden Markov Model and Pattern Matching algorithm.
Hidden Markov Model (HMM) can reflect the random process of user movement well and
is suitable for short historical sequences. This paper proposes a prediction algorithm based
on HMM and moving object location preference which considers time characteristics of
historical trajectory data. We use a spatiotemporal threshold algorithm to transform tra-
jectory data into a series of user stay points and a POI Discovering algorithm to solve the
problem of low accuracy caused by long historical trajectory sequence.
2 Related Work
user’s prediction location. Literature [15] proposes the model of personal mobility data
by using Markov model. The paper compares the efficiency and accuracy on different
orders of Markov model. In [16], a hybrid Markov model that combines the user’s
location preference is proposed, the paper also consider to use the moving object data
with similar trajectories for prediction when the user’s prediction sequence never
appears in historical trajectory data. In [17], two prediction models based on HMM are
proposed. One is the “temporal and spatial prediction” model which is used to predict
regular moving trajectories and the other is the “next location” prediction model which
is suitable for limited trajectory points.
This section gives the methods and steps for modeling user mobility based on HMM
for addressing the problem of datasets and the low accuracy of prediction models.
In Eq. (3), psi indicates the initial probability of hidden state si, bsi ðoi Þ indicates
transmission probability from hidden state si to output state oi, and asi1 ;si indicates the
transition probability from hidden state si–1 to hidden state si.
Nt ðPOIA ! POIB Þ
Ppref ðPOIA ! POIB Þ ¼ ð4Þ
Nt ðPOIA Þ
this paper proposes Location Preference-based Hidden Markov Model (LP-HMM), the
prediction algorithm which combines the traditional HMM model and time-based
location preference model. The probability of the LP-HMM algorithm is calculated
by Eq. (5).
Where parameters b1 and b2 are the weights of the two generation probabilities,
b1 þ b2 ¼ 1, PðO; SjkÞ represents the probability of generating visited sequence, and
Ppref represents the time-based location preference probability.
4 Experiment
4.1 Dataset
This paper uses the public GPS dataset Geolife, which is composed of 182 users’ GPS
trajectory data collected by Microsoft Research Asia from 2008 to 2012. In the Geolife
dataset, a GPS trajectory is represented by a series of time-stamped points, each of
which contains latitude and longitude coordinates and timestamp. The complete dataset
contains 18,670 trajectories, making it an ideal choice for validating location prediction
methods. It can be seen from Fig. 1(a) that the distribution of user trajectory number is
very uneven, and the number of users with more than 100,000 trajectory points is the
most. Therefore, the dataset can be used to analyze the user’s mobility. We need to
select the users we need from all the users. In Fig. 1(b), it can be seen that the number
of users with a daily average of more than 2000 trajectory points is the most.
The distribution of user trajectory point The trajectory of user per day
>100 60
90-100
the number of trajectory point(10 3 )
80-90 50
70-80
40
60-70
users
50-60 30
40-50
30-40 20
20-30
10-20 10
1-10
0-1 0
<1 1-5 5-10 10-15 15-20 >20
0 20 40 60 The number of trajectory per day(101)
The number of users
(a) The distribution of user trajectory point (b) The trajectory of user per day
60
The number of users
50
50
40
40
30
30
20
20
10 10
0 0
1 2 3 4 5 6 7 8 9 10 11 12 0 1 2 3 4 5 6 7 8 9 10 >10
The number of cluster labels
The number of cluster labels
(a) The result in different Eps and same Minpts (b) The result in different Minpts and same Eps
Fig. 2. The cluster distribution of user trajectories in different Eps and Minpts
10 J. Huang et al.
TP
Accuracy Rate ¼ ð6Þ
TP þ FP
A multi-fold cross-validation prediction experiment is carried out on the users and the
experiment results are compared with the traditional HMM and GMM. The location
sequences are used in different prediction models. It can be seen from Fig. 3 that the
prediction accuracy of the hidden Markov model considering time factor is higher than
the traditional HMM and GMM. The factors that determine the location prediction
includes the space and time context. The proposed LP-HMM model considers the time
and space characteristics of the location information at the same time, while the HMM
and GMM only consider the conversion probability between the spatial locations,
which is equivalent to only utilizing the spatial characteristics of historical information.
LP-HMM improves the accuracy of prediction compared with HMM and GMM. When
the number of cluster labels exceeds a certain value, the accuracy of model prediction
decreases slightly, and GMM shows the superiority of prediction when the trajectory
sequences are longer. By analyzing the accuracy of the selected 20 users, it was found
that the increased accuracy predicted by the LP-HMM model can achieve 6.4% and
7%, respectively, compared to GMM and HMM.
LP-HMM: Location Preference-Based Hidden Markov Model 11
0.8
0.78
Accuracy 0.76
0.74
0.72
0.7
0.68
0.66
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
User
5 Conclusion
In this paper, we propose a method to discover POI sequences of moving objects using
GPS data. We use the POI sequences to build a model to predict the potential visited
locations of users, aiming at improving the prediction accuracy based on the HMM
prediction algorithm by considering the spatiotemporal characteristics of real datasets.
We represent a spatiotemporal threshold stay point extraction algorithm to extract stay
points. The algorithm mainly processes GPS points according to two thresholds: time
and distance. A POI discovering algorithm based on DBSCAN is proposed to obtain
the user’s POI and the sequence data which is favorable to the prediction model.
Finally, a prediction model LP-HMM combining HMM with temporal location pref-
erence is proposed based on HMM. Under the driving of the Geolife dataset, the
proposed model is compared with HMM and GMM to verify the validity of the model.
The results show that the prediction accuracy of the model is higher than that of HMM
and GMM.
References
1. Sungjun, L., Junseok, L., Jonghun, P.: Next place prediction based on spatiotemporal pattern
mining of mobile device logs. Sensors 16(2), 145–163 (2016)
2. Lin, Y., Huang, P.: Prefetching for mobile web album. Wirel. Commun. Mob. Comput. 16
(1), 18–28 (2016)
3. Yang, J., Qiao, Y., Zhang, X.: Characterizing user behavior in mobile internet. IEEE Trans.
Emerg. Top. Comput. 3(1), 95–106 (2015)
4. Qiao, Y., Cheng, Y., Yang, J., et al.: A mobility analytical framework for big mobile data in
densely populated area. IEEE Trans. Veh. Technol. 66(2), 1443–1455 (2017)
5. Wu, R., Luo, G., Yang, Q., Shao, J.: Individual moving preference and social interaction for
location prediction. IEEE Access 6(1), 10675–10687 (2018)
6. Lee, W.C., Ye, M.: Location-based social networks. In: Alhajj, R., Rokne, J. (eds.)
Encyclopedia of Social Network Analysis and Mining. Springer, Heidelberg (2014)
12 J. Huang et al.
7. Yi, X.: Research of Tourist Traffic Flow Characteristic Based on Phone Signaling Data.
Southeast University, Nanjing (2017)
8. Jiang, S., Ferreira, J., Gonzalez, M.C.: Activity-based human mobility patterns inferred from
mobile phone data: a case study of Singapore. IEEE Trans. Big Data 3(2), 208–2193 (2017)
9. Hu, Y., Zhu, X., Ma, G.: Location prediction model based on k-means algorithm. In: Bhatia,
S., Mishra, K., Tiwari, S., Singh, V. (eds.) Advances in Computer and Computational
Sciences, vol. 554, no. 1, pp. 681–687 (2018)
10. Kavak, H., Vernon-Bido, D., Padilla, J.J.: Fine-scale prediction of people’s home location
using social media footprints. In: International Conference on Social Computing,
Behavioral-Cultural Modeling and Prediction and Behavior Representation in Modeling
and Simulation, pp. 183–189. Springer, Heidelberg (2018)
11. Ikanovic, E.L., Mollgaard, A.: An alternative approach to the limits of predictability in
human mobility. EPJ Data Sci. 6(1), 6–12 (2017)
12. Wei, Z.: Research on Key Technologies of Moving Object Location Prediction. Nanjing
University of Aeronautics and Astronautics (2009)
13. Chen, N.C., Xie, W., Welsch, R.E., et al.: Comprehensive predictions of tourists’ next visit
location based on call detail records using machine learning and deep learning methods. In:
IEEE International Congress on Big Data. IEEE, Piscataway (2017)
14. Wu, F., Fu, K., Wang, Y., et al.: A spatial-temporal-semantic neural network algorithm for
position prediction on moving objects. Algorithms 10(2), 99–110 (2017)
15. Gambs, S., Killijian, M.O., del Prado Cortez, M.N.: Next place prediction using mobility
markov chains. In: Proceedings of the First Workshop on Measurement, Privacy and
Mobility. ACM (2012)
16. Qiao, W., Si, Z., Zhang, Y., et al.: A hybrid Markov-based model for human mobility
prediction. Neurocomputing 7(1), 278–290 (2017)
17. Lv, Q., Qiao, Y., Ansari, N., et al.: Big data driven hidden Markov model based individual
mobility prediction at points of interest. IEEE Trans. Veh. Technol. 66(6), 5204–5216
(2017)
Design and Performance Evaluation
of On-Board Computer Network Protocol
Stack Based on Spacecraft Operating System
Abstract. The spacecraft network protocol is the critical platform for the future
spacecraft cooperative mission. Especially at the level of operating system, it has
to support the network function to provide agile and reliable data transmission.
This paper proposes a new network protocol stack architecture based on the
Berkeley BSD’s TCP/IP network protocol stack of a practical spacecraft oper-
ating system. A thin glue layer is inserted between the operating system and the
protocol, and the performance testing results running on one real on-board
SPARC hardware platform shows that the highest IP layer throughout of
24.7Mbps can be obtained. This architecture has been used in some practical
spacecraft of China which has been proved to have high value of engineering.
1 Introduction
At present, most of the computers in the on-orbit spacecraft are interconnected directly
by buses [1]; however, network applications communicating through protocol has just
come out in the field of China’s spacecraft [2, 3] such as the third phase manned craft,
the space station and the third phase lunar exploration. In these projects, the connec-
tions between various computer systems are getting closer, and tasks to be conducted
are becoming more complicated. Thus traditional communication methods using buses
to connect various computers are no longer up to the standard for applications to work
well because of the low latency and transmission rate. Therefore, it is imperative
to utilize the network to realize the communication between various computer systems
[4, 5] to improve the performance.
Communication between the internal spacecraft computer systems have two main
types of data packets [6, 7]. One is the control data packets for transmitting control
commands between various modules, the other one is the ordinary data packets. On the
spacecraft, control packets are composed of a large number of small size packets.
Control packets are mainly control signals transmitting to another module which
requires high real-time data transmission, but the data should also be kept accurate and
integrated during transmission.
The TCP/IP protocol [8, 9] software is the base of the network which is conceptually
divided into 4 layers. When implementing a protocol stack of an actual spacecraft
system, the details of the specific hardware interface, OS features and interfaces, buffer
management, protocols, and error handling must be taken into account and therefore
the structure of protocol software is more complicated than the conceptual structure.
Specifically, to implement the TCP/IP protocol stack on a spacecraft computer [10, 11],
especially on a practical operating system SpaceOS [12], in addition to the logic of
each protocol of the protocol stack itself, the following aspects need to be taken into
account:
2.1 Interface
Three interfaces should be covered. As for the lower level interface, in order to support
all kinds of communication devices, the protocol stack should communicate with the
END driver [13] through the MUX [14] layer which constructs a unified END driver
model. For the middle level interface, the management of network tasks, a semaphore
must provide mutual exclusion and synchronization, coupled with switch interrupt
protection of critical data as well as timed interrupts, the allocation and release of the
timers and the protocol stack’s memory pool must be realized as well. As for the high
level interface, the application must use the socket to communicate with the protocol
stack.
The function call is applied in the entire process, which occurs in the context of
application layer tasks. In addition to the implementation of the protocol stack, there are
four tasks contained in the application layer: TFTP, FTP, Telnet, and HTTP to implement
the four protocols and the application logic. The interface between the partial protocol
and the SpaceOS is decoupled via a unified interface file which we call it the glue layer.
The code in the protocol part only uses the functions in the file and is isolated from
SpaceOS. Application layer programs all need to use the file system without exception.
Here in this architecture the YAFFS is used, which is a POSIX-compliant filesystem
designed for use with NAND and NOR flash. It is efficient, robust and also has low
resource use on the host system. There is a thin layer of file system interface between
YAFFS and the application, which also supports the control of the protocol stack.
GLUE Layer
Function call
Task
Management
Interrupt Time
Management Management
Task Scheduling
Message
Queue Timer
Management Interrupt
Memory
Management
The purpose of the protocol performance test is to verify whether the software system
can achieve the performance index proposed by the user, find the performance bot-
tleneck in the software system.
Target Board
Running SpaceOS+Protocol
Stack+Application,
2 Ethernet cards
Switch
Table 3. DCACHE off + ICACHE off (66 MHz, in RAM, IO read: 47619000, write: 26666650)
Packet size MAC DRV IP UDP TCP
64 6355000 5518800 934100 828900 1379700
164 10230400 8388900 1889350 1684500 2140000
264 11984050 9320950 2496650 2231050 2526750
364 13108550 9987450 2913000 2621700 2815250
464 13530800 10230600 3251550 2933250 2974850
564 13987200 10490400 3556050 3178900 3018800
Design and Performance Evaluation of On-Board Computer Network Protocol Stack 19
5 Conclusion
Taking into account the characteristics of large transmission volume, small data packet
and high real-time requirements in the current control data packet, this paper proposed
a new network protocol architecture that optimizes and updates the BSD protocol and
integrates it with the frequently used spacecraft SpaceOS, so as to meet the decent real-
time control required by the data packet of complex network spacecraft, thereby
realizing the data transmission on the basis of operating system. Noticeably, the
function and performance test result has been given on a practical computer target
board, and positive results have been achieved which proves that it can satisfy the
requirements of spacecraft network applications.
Acknowledgment. This work is supported in part by grants from the National Natural Science
Foundation of China (NSFC) under Grant Nos. 61632005 and 61502031.
References
1. Parkes, S.M., Armbruster, P.: SpaceWire: a spacecraft onboard network for real-time
communications. In: IEEE NPSS Real Time Conference, pp. 6–10. IEEE (2015)
2. Tan, C., Gu, Y.: Space Data Systems. China Science and Technology Press, Beijing (2004)
3. Yang, M., Guo, S., Sun, Z.: Spacecraft on-board computer control technology applications.
Aerosp. Control 23(2), 69–73 (2005)
4. Wolfgang, E., Ines, L., Reinhardt, W., et al.: Optical fibre grating strain sensor network for
X-38 spacecraft health monitoring. In: Proceedings of SPIE the International Society for
Optical Engineering (2000)
5. Alena, R., Nakamura, Y., Faber, N., et al.: Heterogeneous spacecraft networks: wireless
network technology assessment. In: IEEE Aerospace Conference, pp. 1–13. IEEE (2014)
6. Kitts, C.A.: A global spacecraft control network for spacecraft autonomy research. In:
Proceedings of Spaceops the Fourth International Symposium on Space Mission Operations
& Ground Data Systems, vol. 394, no. 394, pp. 16–20 (1996)
7. LAWRENCEB. TCP/IP Volume Detailed 1: Protocol. Machinery Industry Press, Beijing
(2017)
8. Tanenbaum, A.S.: Computer Networks, 4th edn. Tsinghua University Press, Beijing (2004)
9. Rose, M.: Structure and identification of management information for TCP/IP-based
internets. RFC 105(2), 97–110 (1990)
10. Partridge, C., Shepard, T.J.: TCP/IP performance over satellite links. IEEE Network 11(5),
44–49 (1997)
Design and Performance Evaluation of On-Board Computer Network Protocol Stack 21
11. Ghani, N., Dixit, S.: TCP/IP Enhancements for Satellite Networks. IEEE Press, Piscataway
(1999)
12. Qiao, L., Yang, M., Gu, B., Yang, H., Liu, B.: An embedded operating system design for the
lunar exploration rover. In: International Conference on Secure Software Integration &
Reliability Improvement Companion, pp. 160–165. IEEE (2011)
13. Kou, Y., Chen, H., Duan, X., et al.: Development and realization of END network driver
design in VxWorks systems. Comput. Measur. Control (2009)
14. Liang, H.: Methods study binding protocol based on VxWorks network system MUX layer.
Ind. Control Comput. (2012)
15. Ma, Q., Steenkiste, P.: Supporting dynamic inter-class resource sharing: a multi-class QoS
routing algorithm. In: INFOCOM 1999. Eighteenth Joint Conference of the IEEE Computer
and Communications Societies. Proceedings, vol. 2, pp. 649–660. IEEE (1999)
16. Wang, J.-G.: Based on VxWorks Embedded Real-Time System Design. Tsinghua University
Press, Beijing (2005)
17. Son, I., Kim, Y., Baek, S., Choi, J.: Improving the reliability and performance of the YAFFS
flash file system. IEICE Trans. Inf. Syst. 94(12), 2528–2532 (2011)
18. Kim, S., Cho, Y.: The Design and Implementation of Flash Cryptographic File System
Based on YAFFS. In: International Conference on Information Science and Security, pp. 62–
65. IEEE (2008)
19. Wei, L.I.: On-board Computer standard design and realization based on BM3803. Spacecr.
Eng. (2012)
20. SPARC International, Weaver, D.L., Gremond, T.: The SPARC architecture manual-version
9. In: SPARC Architecture Manual Version (1994)
Overview of Face Recognition Methods
1 Introduction
security, crime detection and so on. Therefore, face recognition technology is getting
more and more attention and has become a hot research field.
Face recognition is a routine task for human, but establishing a similar computer
system and constantly improving the accuracy is an ongoing research. The research on
face recognition can be traced back to 1950s. Up to now, researches have made fruitful
achievements in the research of computer-based face recognition methods. Especially
with the emergence of deep learning, face recognition technology has achieved a rapid
development. However, face recognition technology still faces many challenges, and it
is limited by the application environment. These limitations not only come from facial
expression, posture, position and size changes, as well as the effects of occlusion and
makeup, but also from the lighting, background changes. Face recognition technology
is also affected by data size and data distribution.
Face recognition techniques can basically be classified into two categories: feature
based methods and holistic methods. In feature based methods, some geometric fea-
tures like position and width of eyes, noses, and the thickness of eyebrow are extracted
to represent a face. On the contrary, in holistic approach, the whole face is taken as an
input to perform face recognition task. This paper will introduce some classical popular
face recognition methods in detail.
2 General Challenges
Although the emergence of deep learning has greatly promoted the development of face
recognition technology, face recognition technology still faces many challenges. There
is still a long way to go to improve the accuracy of the face recognition system in the
actual scenarios.
2.1 Illumination
Illumination is a long-standing problem in computer vision field, especially in face
recognition field. The shadow produced by illumination will enhance or weaken the
original facial features. If the illumination tends to vary, even the same individual is
captured by the same sensor with the same pose and expression, face recognition
system may make mistake to verify them to be different people.
2.2 Pose
Similar to the illumination problem, the pose problem is also a technical difficulty
needed to be solved in current face recognition research. Face recognition systems are
highly sensitive to pose variations. The pose problem involves facial changes caused by
head rotation in a three-dimensional vertical coordinate system. Deep rotation per-
pendicular to two directions of the image plane may cause partial loss of facial
information.
24 L. Fang et al.
2.3 Occlusion
Occlusion is a serious problem for face images acquisition in a non-cooperative situ-
ation. This problem is more obvious in the face images captured by video surveillance.
The monitoring object often wear glasses, hats and other ornaments, making the
captured face images incomplete, thus affecting the subsequent feature extraction and
recognition, and even lead to the failure of face recognition algorithm.
2.5 Data
Data is crucial to the performance of face recognition system. Data distribution and data
size do influence the performance of face recognition system. A research in [2] shows
that large amounts of data can improve the face recognition system’ performance. The
experimental results show that the performance of the system improves as the amounts
of data accumulates. However, too much data will have a long tail effect, and this effect
is caused by data imbalance, which means that some individuals have most facial
image instances, but some only have a little. When we continuously increase the
amount of data, the performance of the system may not increase or even be damaged.
In other words, adding the individuals with only a few facial image instances do not
help to improve the recognition performance and may harm the system’s performance.
3 Literature Review
The labeled faces in the wild home (LFW) dataset is intended to test the face recog-
nition systems’ performance and accuracy under an unconstrained environment. More
than 13000 images of faces are collected from web. Each face has been labeled with the
name of its owner. 1680 of the face images owner have two or more distinct image
instances in the dataset. It has become a standard dataset to evaluate the performance of
face recognition system. Other databases like LFPW (labeled face parts in the wild)
contains 3000 face images collected from web. The obtained images are detected by
face detection method and the positions of face in the images are obtained. Three
groups of 35 face feature points which are marked manually are also given. The
WDRef database contains 99773 images of 2995 individuals, and more than 2000
individuals have over 15 image instances in this database.
In order to push the accuracy limit on these datasets, large amounts of works have
been made. In this section, we will introduce some classic face recognition methods
which are widely used.
Overview of Face Recognition Methods 25
The main idea of AAM search method is based on the hypothesis that the difference
between the texture of the model and the region surrounded by the shape changes
linearly with the change of the parameters. The hypothesis is reasonable only in a
certain range of deviation, and its coverage is strongly dependent on the training set.
Therefore, the robustness of AAM algorithm is limited when it is used to search.
images from 16 individuals. This system achieved approximately 96% correct classi-
fication averaged over lighting variation, 85% correct averaged orientation variation,
and 64% correct averaged over size variation [8]. The disadvantages of Eigen face
methods are that the Eigen face is sensitive to lightening conditions and position of
head, and it is time consuming to find the eigenvalues and eigenvectors.
[13] achieved the highest accuracy, which has reached 99.63% on the LFW database,
and this accuracy even exceeded the ability of human eye.
Table 1. Comparisons between some face recognition methods based on deep learning
Method Performance on tasks
Pair-wise Rank- DIR (%) Verification Open-set
Accuracy (%) 1 (%) @ (%) @ Identification (%) @
FAR = 1% FAR = 0.1% Rank = 1,
FAR = 0.1%
IDL ensemble 99.77 98.03 95.8 99.41 92.09
model
IDL single model 99.68 97.60 94.12 99.11 89.08
FaceNet [13] 99.63 NA NA NA NA
DeepID3 [14] 99.53 96.0 81.40 NA NA
Face++ [2] 99.50 NA NA NA NA
Facebook [15] 98.37 82.5 61.9 NA NA
Learning from 97.73 NA NA 80.26 28.90
Scratch [16]
HighDimLBP [17] 95.17 NA NA 41.66 18.07
Table 1 shows the comparisons between some face recognition methods based on
deep learning. In this table, there are three evaluation protocols containing five tasks.
The first protocol was proposed by Huang et al. [18] and it was used to test the
recognition accuracy of 6000 face images pairs. The second protocol was proposed by
Best-Rowden et al. [19] and contained two tasks: a closed-set identification task and an
open-set identification task. The third protocol was proposed by Liao et al. [20] and
contained two tasks: a verification task and an open-set identification task.
This paper will introduce two face recognition systems based on deep learning.
Face++. The Face++ (Megvii Face Recognition System) was proposed by Zhou, Cao
and Yin [2]. The training set consists of 5 million face images collected from web. Four
regions of face images are used for feature extraction. The neural network consists of
10 layers and the last layer is a softmax layer. PCA algorithm is used to reduce the
dimension of the feature vectors extracted from the previous layers. And L2 distance is
used to represent the similarity of different face images.
After training, this system is tested on LFW, and it achieves 99.50% accuracy. Then
a new test set called Chinese ID (CHID) is collected offline and specializes on Chinese
people. Applying CHID test set to test this system in real scenarios, the true positive
rate is only 66% when keeping the false positive rate to 10 5 , and this performance
does not meet the requirement of real application.
After further research, it is found that data is crucial to the performance of the system.
Face recognition system faces three main challenges. First, unbalanced data do impact
the performance of the system. The data of the identities with only a few face image
instances do not work in a simple multi-class classification framework. Secondly, in real
test scenarios, the false positive rate of this system is quite low, but the true positive rate
Overview of Face Recognition Methods 29
is only 66%, and this performance does not meet the requirement. That is to say, very
low false positive rate can also influence the system’s performance. Finally, there are
some cross factors, such as occlusion, pose and age variation will affect the system’s
performance. These problems needed to be further investigated.
Two-Stage Approach From Baidu. Baidu has proposed a two-stage approach which
combines a deep convolutional neural network on multi-patch and deep metric learning
[9]. The same deep CNN structures are used to extract facial features centered on
different landmarks on face region. And the features extracted by each network is
simply concatenated together to form a high dimension feature. Since the high
dimensional feature is not efficient and consumes a lot of resources, so the next step is
to reduce the dimension of the feature. In this research, a metric learning supervised by
a triplet loss is used and the number of dimension is reduced to 128. The goal of metric
learning with triplet loss is to increase the L2 distance of different individuals and
shorten the L2 distance of the same individuals. After training, this system achieving
99.77% pair-wise verification accuracy on LFW.
In this research, the effect of data and multi-patch to the system’s performance are
investigated. The evaluation protocol is proposed by Huang et al. [18], that is to test the
accuracy of 6000 face pairs from LFW dataset. Table 2 shows the pair-wise error rate with
different amount of training data, and Table 3 shows the pair-wise error rate with different
number of patches. It is obvious that with the increase of the amount of training data, the
error rate of the system decreases, and with the increase of the number of patches, the error
rate of the system decreases, but somehow saturates after seven patches.
4 Application
Compared with other identification methods, face recognition has the advantages of
direct, friendly, convenient and robust. Its application field is gradually expending to
every field of daily life. On the one hand, it obviously improves work efficiency, on the
other hand, it is also secure and reliable. Its application prospect is very broad.
4.1 Payments
Face recognition has already been used in stores and ATMs, and customers can pay by
camera. That is to say, after the face recognition system confirms customers’ identities,
payment will success. This technology will be applied to online payment. Chinese
ecommerce firm Alibaba is planning to apply the face recognition system to payment
made over the Internet.
password. Innovative face recognition system could be especially useful for a company
or organization. Organizations can also use face recognition system to record atten-
dance records.
5 Conclusion
This paper introduces the definition and development of face recognition. General
challenges faced by face recognition system is discussed. And then we give description
to some classical popular face recognition methods in detail. At last, we indicate some
application fields of face recognition system. In short, even though the emergence of
deep learning greatly improves the performance of face recognition systems, this field
still needs to be further studied in order to improve the performance of face recognition
system in actual scenarios.
References
1. Wikipedia https://en.wikipedia.org/wiki/Facial_recognition_system. Accessed 18 October
2018
2. Zhou, E., Cao, Z., Yin, Q.: Naive-deep face recognition: touching the limit of LFW
benchmark or not? arXiv preprint arXiv:1501.04690 (2015)
Overview of Face Recognition Methods 31
3. Wang, W., Shan, S., Gao, W., Cao, B., Yin, B.: An improved active shape model for face
alignment. In: 2002 Proceedings of IEEE International Conference on Multimodal
Interfaces, pp. 523–528. IEEE (2002)
4. Cootes, T.F., Taylor, C.J., Cooper, D.H., et al.: Active shape models-their training and
application. Comput. Vis. Image Underst. 61(1), 38–59 (1995)
5. CSDN. https://blog.csdn.net/CCmilan/article/details/9424387. Accessed 8 October 2018
6. Edwards, G.J., Cootes, T.F., Taylor, C.J.: Face recognition using active appearance models.
In: Computer Vision — ECCV 1998. Springer, Berlin, Heidelberg, pp. 581–595 (1998)
7. Joshi, A.G, Deshpande, A.S.: Review of face recognition techniques. Int. J. Adv. Res.
Comput. Sci. Softw. Eng. 5, 71–75 (2015)
8. Turk, M.A., Pentland, A.P.: Face recognition using eigenfaces. In: Proceedings of IEEE
Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 1991,
pp. 586–591. IEEE (1991)
9. Liu, J., et al.: Targeting Ultimate Accuracy: Face Recognition via Deep Embedding (2015)
10. Huang, G.B., Lee, H., Learned-Miller, E.: Learning hierarchical representations for face
verification with convolutional deep belief networks. In: Computer Vision and Pattern
Recognition, vol. 157, pp. 2518–2525. IEEE (2012)
11. Taigman, Y., Yang, M., Ranzato, M., Wolf, L.: DeepFace: closing the gap to human-level
performance in face verification. In: IEEE Conference on Computer Vision and Pattern
Recognition, pp. 1701–1708. IEEE (2014)
12. Sun, Y., Wang, X., Tang, X.: Deep learning face representation from predicting 10,000
classes. In: Computer Vision and Pattern Recognition, pp. 1891–1898. IEEE (2014)
13. Schroff, F., Kalenichenko, D., Philbin, J.: FaceNet: a unified embedding for face recognition
and clustering, pp. 815–823 (2015)
14. Sun, Y., Liang, D., Wang, X., et al.: DeepID3: Face Recognition with Very Deep Neural
Networks. Computer Science (2015)
15. Taigman, Y., Yang, M., Ranzato, M., Wolf, L.: Web-scale training for face identification. In:
Computer Vision and Pattern Recognition, pp. 2746–2754. IEEE (2015)
16. Yi, D., Lei, Z., Liao, S., et al.: Learning face representation from scratch. In: Computer
Science (2014)
17. Chen, D., Cao, X., Wen, F., Sun, J.: Blessing of dimensionality: high-dimensional feature
and its efficient compression for face verification. In: 2013 IEEE Conference on Computer
Vision and Pattern Recognition (CVPR), pp. 3025–3032. IEEE (2013)
18. Huang, G.B., Ramesh, M., Berg, T., Learned-Miller, E.: Labeled faces in the wild: a
database for studying face recognition in unconstrained environments. Technical Report 07–
49, University of Massachusetts, Amherst, October 2007
19. Best-Rowden, L., Han, H., Otto, C., et al.: Unconstrained face recognition: identifying a
person of interest from a media collection. IEEE Trans. Inf. Forensics Secur. 9(12), 2144–
2157 (2014)
20. Liao, S., Lei, Z., Yi, D., et al.: A benchmark study of large-scale unconstrained face
recognition. In: IEEE International Joint Conference on Biometrics, pp. 1–8. IEEE (2014)
HAR-Net: Fusing Deep Representation
and Hand-Crafted Features for Human
Activity Recognition
Abstract. Wearable computing and context awareness are the focuses of study
in the field of artificial intelligence recently. One of the most appealing as well
as challenging applications is the Human Activity Recognition (HAR) utilizing
smart phones. Conventional HAR based on Support Vector Machine relies on
manually extracted features. This approach is time and energy consuming in
prediction due to the partial view toward which features to be extracted by
human. With the rise of deep learning, artificial intelligence has been making
progress toward being a mature technology. This paper proposes a new
approach based on deep learning called HAR-Net to address the HAR issue. The
study used the data collected by gyroscopes and acceleration sensors in android
smart phones. The HAR-Net fusing the hand-crafted features and high-level
features extracted from convolutional neural network to make prediction. The
performance of the proposed method was proved to be higher than the original
MC-SVM approach. The experimental results on the UCI dataset demonstrate
that fusing the two kinds of features can make up for the shortage of traditional
feature engineering and deep learning techniques.
1 Introduction
With the advent of big data era, it has been an urge for humans to find valuable
information. Among all types of applications addressing this issue, Human Activity
Recognition (HAR), which is applied frequently to the fields of human-computer
interaction, identification technology, and medical security, has been the one that is
commonly utilized. Currently, studies related to HAR is based on several approaches:
video-based HAR and wearable sensor-based HAR. The former approach of HAR is
based on the analysis and recognition videos captured by camera from moving human
bodies. This approach is mainly employed by the field of security monitoring, being
especially useful when monitoring anomalous behavior of the elderly and children.
Despite the multiple applications and benefits of video-based HAR, there are chal-
lenges that may hinder the development of the technology. Due to the demand for
storage resources and the limitations related to the deployment of cameras, the mon-
itoring of users’ whole-day activity state is impossible. This restriction causes the
considerable difficulty of popularization for the video-based approach.
Wearable sensor-based HAR, as an area of intensive interest, on the other hand, can
be achieved using wearable acceleration sensor, gyroscope, and heart rate sensor. The
potential of this approach is enabled by the portability and low cost of wearable
sensors. As an intelligent aided technology, sensor-based HAR has been employed into
commercialized devices as fitness tracker and fall detector. More socially valuable
applications, like providing dementia patients with Memory Prosthesis and proposing
suggestions for daily exercise which require a comprehensive understanding about past
and present human activity, are expected to be generalized. Whereas the traditional
approach of wearable-based HAR can be challenging to generalize as the result of the
requirement for specially designed sensors. In the 1950s, there were researchers related
to sensor-based HAR whose progress was retarded due to the immaturity of recognition
algorithm and unaffordable huge sized sensors. Decades of following advancement in
microelectronics and computer motivated the invention of MEMS that permitted the
micromotion and integration of sensors. Sensors now have attributes of high compu-
tational power, low cost, and tiny size. Meanwhile, the popularization of smart phone
enables people to enjoy an intelligent platform that incorporates amusement, com-
munication, and work, it counteracts the toughness faced by wearable-based HAR.
With the rise of artificial intelligence and wearable computing, plenty of sensors have
been integrated into smart phones. HAR based on smart phone built-in sensors has thus
become research hotspots. The portability and the ease to develop software on smart
phone makes real-time supervision of human body activity and generalization of smart
phone-based HAR applications feasible. Real-time supervision quantifies human
activity, raises humans’ self-awareness of health conditions, and offers reasonable
fitness tips, all of which make it a promising market with high value.
The structure of the remaining section is shown as subsequence: related work is
summarized in Sect. 2. Details of HAR-Net are illustrated in Sect. 3. Our model will be
evaluated in Sect. 4. Conclusions are made in Sect. 5.
2 Related Work
With the hardware development, sensor-based HAR has been commercially applied to
medical system and security. Information collected from sensors is a time series.
Different effective approaches to process time series have been proposed to act as the
foundation of HAR. The majority of focus in HAR study covers daily life, sport,
socialization, and health. Layering classification is achieved by precepting group
behavior using multiple sensors, while children’s socializing behaviors are successfully
identified through physiological sensors [1, 2]. Through using acceleration sensor,
researchers have achieved Parkinson patients’ recognition and detection of Down’s
Syndrome patients [3, 4]. Researchers in Stanford have utilized multi-model sensors
system to detect whether the elder is falling down [5]. Wearable sensors are also
applied to the recognition of newborns’ disorder related to cerebral stroke [6]. Sensor-
based HAR can be optimistic prospect when used to support vulnerable groups.
34 M. Dong et al.
Researchers have done study using an efficient Support Vector Machine algorithm
to compensate for computational loss [7]. In recent years, deep learning is rising
because of the big data. While conventional machine learning approach required
researchers to manually extract features, deep learning can accomplish end to end
learning, using original data as input without manual feature extraction. Deep neural
network can provide powerful feature representation by complex non-linear transfor-
mation. The mostly used networks are deep convolutional neural networks and LSTM
for video-based HAR. Researchers have shown the high competence of LSTM to
classify activity appear in video by analyzing Sports-1 M and UCF-101 data set [8].
Utilizing three wearable sensor data set to assess the performance on deep learning
framework, researchers are able to determine the effectiveness of deep learning on
extracting features [9]. Additionally, conventional machine learning approach has been
compared to sensor-based deep learning approach for human activity recognition [10].
When we use image and signals as input, convolutional neural networks perform
better than multilayer neural networks. A study has been done by researchers with two
dimensional images that are synthesized with 9 types of sensor data, processed by
Discrete Fourier Transformation (DFT), and recognized with convolutional neural
network [11]. Researchers have also adopted several convolutional neural networks
trained in parallel followed by max pooling, concatenated the activations, and fed them
into the fully connected layer. Convolutional neural networks can be advantageous
when applied to HAR because of the local dependency and scale invariance [12].
3 Methodology
For the wearable sensor-based human activity recognition task, we proposed a network
architecture called HAR-Net. The structure is illustrated in Fig. 1. The input of the
network will have a dimension of (1, 128, 9) with 9 channels, then will be processed by
the separative convolutional layer. This means that the 9 channels of the input will be
convoluted respectively by the same filters. For each channel of the input, it will be
processed in parallel by filters with size (1, 1), (5, 1), (9, 1), (13, 1), (17, 1), (21, 1), (25,
1), (29, 1), and (33, 1) respectively, each followed by a max pooling layer with size
(11, 1). The pattern of convolutional layer followed by a max pooling layer described
above will be repeated for four times consecutively. Hand-crafted features, including
time and frequency domain statistics, will be concatenated with the flattened output
from the convolutional layers. Then input into the first fully connected layer with 2048
neurons with relu activation function. The output of the first fully connected layer will
be input into the second layer with 64 neurons and tanh activation function. Softmax is
applied to the output of the second fully connected layer to classify different kinds of
activities. Adam optimizer with default parameters [13] was applied to the HAR-Net.
HAR-Net: Fusing Deep Representation and Hand-Crafted Features 35
The structure of the Inception Block is shown in Fig. 2. Figure 3 shows the
illustration of Separable Convolution. Separative convolution is chosen because of its
advantage of preventing the interference between different channels of input.
4 Experimental Results
4.1 HAR Dataset Description
The data adopted in this paper is a public domain dataset from UCI. The subject of the
experiment used to collect the data set has chosen to be a group of 30 people with age
between 19 to 48 years old. Each person was asked to carry out the six specified
Activities: walking downstairs, walking upstairs, sitting, laying, walking, and standing.
The data in the process was collected using build-in accelerometer and gyroscope in
smart phone at a constant sampling frequency of 50 Hz.
Table 1 shows the results of our model and other baselines on UCI dataset.
According to the study proposing the MC-SVM approach, among the 2,947 test set
samples, the overall accuracy of prediction can reach 96.0%. Despite the high accuracy
reached, there are potential improvements for the approach. According to the HAR-Net
proposed in this paper, convolutional neural network can automatically select the most
effective features for higher classification accuracy through utilizing loss function and
back propagation. Through applying different filters to the signal, the deep neural
network is able to detect similar features in signal waveform and generate distinctive
feature maps. When the network gets deeper, non-linearity of the neural network
increases. More abstract features are extracted by the neural network adaptively
through the automatic learning process. The advantage of the features extracted through
deep learning is the decrease of the amount of human’s subjective judgement about
which features to use. The HAR-Net combined the Hand-crafted features with the
learnt features by the deep neural network to provide more comprehensive perspectives
when making predictions. Regarding the prediction accuracy for the HAR-Net, the
HAR-Net: Fusing Deep Representation and Hand-Crafted Features 37
overall prediction accuracy reached 96.9% and outperformed the MC-SVM model by
0.9%, demonstrating the effectiveness of the HAR-Net.
Meanwhile, we compare the HAR-Net with the network that didn’t use the hand-
crafted features. The network without manually extracted statistics features may fail to
capture the extra effective representation and the performance is worse than the HAR-
Net. This can further prove the importance of adding the auxiliary feature engineering.
The confusion matrix of the HAR-Net is shown above in Table 2. Despite the
overall high accuracy, confusions happened when the model was used to predict
activities. From the confusion matrix above, a pattern of mild confusion between
walking, walking upstairs, and walking downstairs appears. A more serious confusion
between standing and sitting is discovered due to the large number of sitting that were
misidentified as standing. The confusion between those activities were caused by the
similarity in the signal’ frequency and amplitude.
For conventional convolutional layer, the output channel size is equal to the
number of filters applied. For each filter applied, the convolutions of the all the
channels are added together to generate a single channel output. The prediction
accuracy comparison is shown in Table 3.
Table 3. Comparison among the separable conv-based HAR-Net and conventional conv-based
HAR-Net
Approach Prediction accuracy
HAR-Net based on conventional convolution 95.2%
HAR-Net based on separable convolution 96.9%
5 Conclusions
In the study, a new approach utilizing Inception Block for convolutional neural net-
work with hand-crafted features was proposed to address Human Activity Recognition.
The HAR-Net for convolutional neural network can be an appealing option for real-
izing HAR in health care industry to take care of the vulnerable groups. Due to the
advantages of convolutional neural network’s local dependency and scale invariance
[12], it can be a potential application in the field of HAR. The HAR-Net outcompetes
the traditional MC-SVM approach because of the higher prediction accuracy resulted
from the extra self-adaptive features extracted by deep learning in the model. The study
confirmed that the combination of deep learning techniques and traditional feature
engineering can outcompete MC-SVM approach using the same public data set. The
result suggests an implication about the smartphone’s effect on recognizing human
activity. Future work will include testing with the Residual Network (Resnet) and
Recurrent Network to realize HAR utilizing more datasets and reduce the degree of
confusion between distinctive activities especially sitting and standing.
References
1. Hernandez, J., Riobo, I., Rozga, A., Abowd, G.D., Picard, R.W.: Using electrodermal
activity to recognize ease of engagement in children during social interactions. In: ACM
International Joint Conference on Pervasive & Ubiquitous Computing, vol. 48, pp. 307–317.
ACM, Geneva (2014)
2. Kjærgaard, M.B., Wirz, M., Roggen, D.: Detecting pedestrian flocks by fusion of multi-
modal sensors in mobile phones. In: ACM Conference on Ubiquitous Computing, pp. 240–
249. ACM, Geneva (2012)
3. Chengqing, Z.: Statistical Natural Language Processing. Tsinghua University Press, Beijing
(2008)
4. Xinqing, S.: A Brief Treatise on Computational Electromagnetics. Press of University of
Science and Technology of China, Beijing (2004)
5. Murata, S., Suzuki, M., Fujinami, K.: A wearable projector-based gait assistance system and
its application for elderly people. In: ACM International Joint Conference on Pervasive &
Ubiquitous Computing, pp. 143–152. ACM, Geneva (2013)
6. Fan, M., Gravem, D., Dan, M.C., Patterson, D.J.: Augmenting gesture recognition with
Erlang-Cox models to identify neurological disorders in premature babies. In: International
Conference on Ubiquitous Computing, pp. 411–420. ACM Geneva (2012)
7. Anguita, D., Ghio, A., Oneto, L., Parra, X., Reyes-Ortiz, J.L.: Energy efficient smartphone-
based activity recognition using fixed-point arithmetic. J. Univ. Comput. Sci. 19(9), 1295–
1314 (2013)
8. Ng, Y.H., Hausknecht, M., Vijayanarasimhan, S., Vinyals, O., Monga, R., Toderici, G.:
Beyond short snippets: deep networks for video classification. Comput. Vis. Pattern Recogn.
16(4), 4694–4702 (2015)
9. Plötz, T., Hammerla, N.Y., Olivier, P.: Feature learning for activity recognition in ubiquitous
computing. In: Proceedings of the 22nd International Joint Conference on Artificial
Intelligence (IJCAI), pp. 1729–1734. IJCAI, Barcelona, 16–22 (2011)
40 M. Dong et al.
10. Jiang, W., Yin, Z.: Human activity recognition using wearable sensors by deep
convolutional neural networks. In: Proceedings of the 23rd ACM International Conference
on Multimedia, pp. 1307–1310. ACM, Geneva (2015)
11. Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20(3), 273–297 (1995)
12. Zeng, M., Le, T.N., Yu, B., Mengshoel, O.J., Zhu, J., Wu, P.: Convolutional neural networks
for human activity recognition using mobile sensors. In: International Conference on Mobile
Computing, Applications and Services, pp. 197–205. IEEE, New York (2015)
13. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. CoRR abs/1412.6980.
http://arxiv.org/abs/1412.6980 (2014)
14. Anguita, D., Ghio, A., Oneto, L., Parra, X., Reyes-Ortiz, J.L.: A Public domain dataset for
human activity recognition using smartphones. In: 21st European Symposium on Artificial
Neural Networks, Computational Intelligence and Machine Learning, pp. 24–26. ESANN,
Bruges (2013)
Private-Preserving Analysis of Network Traffic
1 Introduction
It is necessary and useful to analyze network traffic, we can take measures to avoid
network congestion problems and reveal potential anomalies by the result of analysis.
But network providers usually know less about the real running status of the network.
There are much research on analysing network traffic. Cluster network traffic with
different methods is key technique of most research [1, 2]. Shadi and others opt netflow
to analyze data, and they cluster network traffic with hierarchical cluster algorithm,
eventually, they traverse a hierarchical cluster tree and get information about the status
of a network [3], their work is similar to us, but they don’t consider the hierarchy of IP
itself. Jeffrey and others apply unsupervised clustering algorithms to cluster network
traffic, then they compare the result with the result using AutoClass cluster method and
they find DBSCAN method has lower accuracy but produces better clusters [4]. Most of
these studies focus on traffic of specific Internet Service Provider (ISP), however, there
are usually more than three network service providers in a region by economic reasons.
Our research pay attention to coordinate network providers’ network data for analysis.
There is an obstacle to get a panoramic view of regional networks, as several
Internet Service Providers (ISPs) make up the whole network, each of them prefers not
to share their private data with other parties for commercial purposes, but they may
want to use data from all parties to miming for more valuable information. We
introduce secure multi-party computation (SMC) [5] into the field of data mining in
network traffic.
2 Preliminaries
2.1 Definitions
Certain parties are about to jointly compute a function f, each side of n participants
(P1, …, Pn) has its private input data, denoted as (v1, …, vn), all parties provide their
input data and jointly compute f (v1, …, vn), finally all sides get the correct result.
Assume that some adversary or participants try attack computing process f and except
for other parties’ privacy data. Secure multi-party computation (SMC) is a set of
protocols and algorithms to make sure that all participants can get correct result through
jointly compute f and keep vi secret [5].
Our research based on the semi-honest adversary model, which means corrupt
parties always follow the agreement of our protocol and try their best to derive other
sides’ private information.
In this section, we consider the scenario that each Internet Service provider (ISP) has its
own netstream data, we can mine for more accuracy and meaningful results from the
overall dataset. We apply secure multi-party computation to do jointly computation
without leaking each parties’ privacy.
We select netstream data as origin data. A netstream flow can be expressed as a
quintuple (source IP, source port, destination IP, destination port, volume). Raw
network packets are aggregate with pieces of tetrad (source IP, source port, destination
IP, destination port), packets that has the same tetrad are regarded as same network
flow during a period. Each flow has a volume attribute, which is defined as the sum
bytes of all packets that were aggregated in a flow. The problem can be described as
getting a union set of all service providers’ destination IPs from netstream data and get
the sum volume of each destination IP through all sharers, and how to cluster them. We
first discuss how to union destination IPs then calculate volume corresponding with
IPs, and continued to show how we cluster them as a clustering tree.
1
8IP1 ; IP2 2 M and IP1 6¼ IP2 ; for gievn k; e\
2k
Pr EKi1 ð. . .EKin ðIP1 Þ. . .Þ ¼ EKj1 . . .EKjn ðIP2 Þ. . . \e ð2Þ
44 Y. Tang and Y. Ma
Pohlig-Hellman encryption [9] satisfies the two equations above, which means that
each party encrypts an IP address in turn with its own secret key, and the encrypted
data can be decrypted by all parties using their private keys in any order, after all
parties decrypt the encrypted data, we can get origin secret IP. The encrypt example
shown in Fig. 1.
Fig. 1. Example of how to encrypt all parts’ destination IPs and then communicate to get every
parts’ IPs which are encrypted by every part
Through commutative encryption, we can obtain the union set of IPs owned by all
participants without revealing an IP’s source. However, it does reveal the number of
IPs that commonly exist in all parts, it can be reveal because encrypted IPs duplicate.
However, in our scenario, this leak does not affect the overall confidentiality.
Fig. 2. Example of how to get the sum volume of each party’s volume, which is corresponding
with an IP in unioned IPs set
Private-Preserving Analysis of Network Traffic 45
The secure add process of all IPs in the overall set can be done with parallel. Hence
the method can reduce the time of calculating and quite suit for calculate big scale data.
We construct tree_nodes from IPs and their corresponding volume, we use these
tree_nodes to form a clustering tree by algorithm 1. For more detail, each node of the
tree has an IP with network prefix and its corresponding volume, the volume of a node
is the sum value of all its children’s volume. We count number of nodes which have
different IP network prefix length vary from /0 to /31, clusters are normally distributed,
which means that most of the user’s traffic can be covered by the network. Figure 4
shows there are most clusters in network prefix of /24, and end users’ traffic of Internet
Service Providers (ISPs) or enterprises will be routed to the backbone router where we
gather those data, and will finally send to service providers, we can find the status of the
service provider’s services through our clustering method. To get more detailed
information, we traverse the clustering tree to get the top 10 clusters with the most
significant volume, and we can see 10 subnetworks with most massive traffic, in our
example, we analyze the clustering tree nodes of the network with /24 network prefix.
We can get the top 10 clusters with the highest traffic, as you can see from the Table 1,
according to the information from the table, network administrators can specify rea-
sonable network policies to ensure the optimal use of the network, it avoids looking for
reports from the vast and complex raw network traffic.
Private-Preserving Analysis of Network Traffic 47
4 Conclusion
clustering tree, the analysis of the tree can get conclusions that help network operators a
lot. In this paper, we validate our method by using a real Internet Data Center’s traffic
data, and our clustering method fits the data well. We believe that our work would help
network operators deepening the understanding of the overall network situation and
thus they can take more effective measures to ensure the rational use of networks and
avoid attacks.
Acknowledgement. The work in this paper is supported by the Joint Funds of National Natural
Science Foundation of China and Xinjiang (Project U1603261).
References
1. Snort. https://www.snort.org. Accessed 19 Oct 2018
2. Shafiq, M., Yu, X., Laghari, A.A., Yao, L., Karn, N.K., Abdessamia, F.: Network traffic
classification techniques and comparative analysis using machine learning algorithms. In:
2016 2nd IEEE International Conference on Computer and Communications (ICCC),
pp. 2451–2455. IEEE (2016)
3. Shadi, K., Natarajan, P., Dovrolis, C.: Hierarchical ip flow clustering. ACM SIGCOMM
Comput. Commun. Rev. 47(5), 48–53 (2017)
4. Erman, J., Arlitt, M., Mahanti, A.: Traffic classification using clustering algorithms. In:
Proceedings of the 2006 SIGCOMM Workshop on Mining Network Data, pp. 281–286.
ACM (2006)
5. Lindell, Y.: Secure multiparty computation for privacy preserving data mining. In:
Encyclopedia of Data Warehousing and Mining, pp. 1005–1009. IGI Global (2005)
6. Technical white paper for net stream. http://efile.huawei.com/tr/marketing/material/global/
products/enterprise_network/ce_switches/hw_353223. Accessed 19 Oct 2018
7. Rokach, L., Maimon, O.: Clustering methods. In: Data Mining and Knowledge Discovery
Handbook, pp. 321–352. Springer (2005)
8. Clifton, C., Kantarcioglu, M., Vaidya, J., Lin, X., Zhu, M.Y.: Tools for privacy preserving
distributed data mining. ACM SIGKDD Explor. Newsl. 4(2), 28–34 (2002)
9. Ly, J.T.: A serial version of the pohlig-hellman algorithm for computing discret logarithms.
Appl. Algebra Eng. Commun. Comput. 4(1), 77–80 (1993)
10. Beaver, D.: Efficient multiparty protocols using circuit randomization. In: Annual
International Cryptology Conference, pp. 420–432. Springer (1991)
11. Koga, H.: A simple secret sharing scheme using a key and its security against substitution of
shares. In: 2010 IEEE International Symposium on Information Theory Proceedings (ISIT),
pp. 2483–2487. IEEE (2010)
12. Shamir, A.: How to share a secret. Commun. ACM 22(11), 612–613 (1979)
Design of Fault Diagnosis System for Balise
Cable Based on Machine Learning
Abstract. Based on the research of the lineside electronic unit in the high-speed
rail system, the design of the fault diagnosis system based on machine learning-
based of balise cable is completed. For the cable fault characteristics, the impe-
dance method is used to analyze the impedance of the cable and the magnitude and
phase of the current is used for actual fault diagnosis. A FPGA-centric platform is
designed to calculate the magnitude and phase of the current generated in cable in
real time. Analysis by actual current data, the feasibility of the program has been
verified. For the diagnosis of fault characteristics, the way of the classifier is
adopted: Logistic Regression, and support vector machine (SVM). This paper
briefly introduces the implementation principles of the two models. The training
set and test set are constructed from the actual collected data. The two models are
trained and tested respectively. According to the accuracy of the test results and the
complexity of the model, the model suitable for fault diagnosis with FPGA is
selected.
1 Background
With the rapid development of China’s high-speed rail industry, the efficiency and
security of high-speed rail communication systems are becoming more and more
important. The lineside electronic unit is a pivotal part of the high-speed rail com-
munication system. It is responsible for transmitting the train control information of the
train control center to each balise, thus realizing the vehicle-to-ground communication
and ensuring the normal operation of the train. The cable is an indispensable com-
munication link between the lineside electronic and the balise, so it is essential to
monitor and diagnose the real-time fault of the cable. At the same time, with the rise of
artificial intelligence, there are more reliable analytical methods and theoretical models
for fault data. More and more machine learning is applied to the field of fault diagnosis,
which greatly reduces the difficulty of the theory and improves the reliability and
extensiveness of fault diagnosis [1]. The combination of fault feature analysis and
machine learning drives the development of fault diagnosis.
U
Z¼ ð1Þ
I
According to the TSI standard, the C6 signal, output of the lineside electronic unit
is a constant voltage source, which means the current is negatively correlated with the
impedance, and the magnitude and phase of the current can reflect the impedance state
of the current link. Collecting current signals and calculate the magnitude and phase by
using the hardware diagram on the left, the impedance map is shown as the right figure
(Fig. 1).
From the above distribution map, when the cable is in a different State, the
impedance is distributed in different intervals and linearly separable. Amplitude and
phase can be used as features of fault diagnosis, and these two features are trained to
establish a classification model.
In this paper, Logistic Regression, and SVM are used for training, and the per-
formance of the two classifiers is compared and analyzed.
Design of Fault Diagnosis System for Balise Cable 51
1
hðXÞ ¼ gðf ðXÞÞ ¼ ð2Þ
1 þ e wT X þ b
When the sample point is above the classification boundary, the function output is
greater than 0.5, and vice versa. Logarithmic loss function is used as loss function [5]:
The process of solving the Logistic Regression, model is the optimization process
of the following objective function:
1X N
min Cos tðhðxi Þ; yi Þ ð4Þ
h N i¼1
When the input data is above the classification boundary, the classifier output is
positive, otherwise it is negative. By introducing Lagrange Factor and SMO algorithm,
the following functions are optimized to get the classifier parameters [5]:
1 XN
min jjxjj2 þ C ni ; yi ðxT xi þ bÞ 1 ni ; ni 0; i ¼ 1; 2; . . .; N ð6Þ
2 i¼1
classification. Any two types of data to build a classifier, and a total of three classifiers
are needed. When predicting, vote according to the prediction result of each classifier,
the label with the highest number of votes is the prediction result of the input sample.
From the perspective of Prediction precision, Logistic Regression, and SVM have
high precision, and meet the accuracy requirements of the system. For linearly sepa-
rable data sets, the model can solve the classification boundary with the same kind
parameters. During the training process, the SVM needs to find the support vector,
which takes longer than the Logistic Regression, and the space occupied during the
Design of Fault Diagnosis System for Balise Cable 53
training process also multiplies. Logistic Regression, training takes a very short time
and is not prone to over-fitting. It is more suitable for training and forecasting of large
amounts of data. Therefore, the system uses Logistic Regression, to classify fault types.
4 Conclusion
In this paper, the transponder transmission system in the high-speed rail system is taken
as the research background. The C6 signal is used flexibly, and the fault characteristics
of the balise cable are extracted by simple and efficient impedance method. The
classification algorithm is used for training and prediction, and the balise cable fault
diagnosis system is completed. At present, the system has been widely used in high-
speed rail systems and has achieved good results. However, due to the actual appli-
cation, the operating environment, cable characteristics, board characteristics and other
aspects are more complicated, the accumulation of data is not comprehensive enough,
and compatibility needs to be improved.
References
1. Kawai, T., Takinami, N., Chino, T.: A new approach to cable fault location using fiber optic
technology. IEEE Trans. Power Delivery 10(10), 85–91 (1995)
2. Wang, L.: Design and implementation of transponder fault cable detection system. Beijing
University of Posts and Telecommunications (2017)
3. Shen, Y., Liu, C.: Circuit Analysis. People’s Posts and Telecommunications Press, Beijing
(2004)
4. Harrington, P.: Machine Learning in Reality. People’s Posts and Telecommunications Press,
Beijing (2017)
5. Introduction to Support Vector Machines [EB/OL], July. https://blog.csdn.net/v_july_v/
article/details/7624837
6. Li, H.: Statistical Learning Method. Tsinghua University Press, Beijing (2017)
7. Zhou, Z.: Machine Learning. Tsinghua University Press, Beijing (2017)
Design and Implementation of ARQ
Mechanism in High-Speed Data Acquisition
System
1 Introduction
4.5M
t1
T represents the time which the train can transmit data between the two balise,
which is, the maximum allowable processing time for a single transmission; t1 is the
maximum processing time required for the acquisition system to communicate a balise
data with the host computer. In normal operation, it is necessary to guarantee:
t1 T ð1Þ
The ADC data needs to be into a FIFO first, so it is later about 1:77 105 s when
the train just passed the balise.
Suppose the train travels at a speed of V m=s ð14:29 km=h V 350 km=hÞ.
5 ðmÞ 5
T¼ 1:77 105 ðsÞ ðsÞ ð2Þ
Vðm=sÞ V
Further analysis of the time required for the T1: The sending end uses FPGA as the
master, and the network card can send and receive data at the same time. So ignoring
56 X. Li and J. Lv
the time spent on the sending side, the time required for T1 only contains 2 parts: The
time to send data as tsend , and the time to retransmit data as tresend .
t1 ¼ tsend þ tresend ð3Þ
The sampling rate of the ADC data in the acquisition system is 120.58 MB/s.
Suppose the amount of data sent by FPGA be P,
The Gigabit Ethernet NIC (Network Interface Card) has a maximum transmission
and reception rate of 125 MB/s, which is currently 57.97 MB/s. The length of read data
is 1026 Bytes and the period is 17700 ns, so the transmission rate is (Fig. 2):
1026 ðByteÞ
Vsend ¼ ¼ 57:97 ðMB/sÞ ð5Þ
17700 ðnsÞ
Fig. 2. The format of the data packet sent by the acquisition system to the host computer
Since the packet loss rate is positively correlated with the number of packets P, it is
assumed that the packet loss rate is
1068
Pf ðPÞ 1:08f ðPÞ
tresend ¼ 1024 ¼ ð7Þ
Vsend V
1:08ð1 þ f ðPÞÞ
t1 ¼ tsend þ tresend ¼ ð8Þ
V
It can be seen that sending a packet of data, only when all of it is lost and
retransmitted over 3 times still cannot be received, the time may not be enough to cause
timeout, which is impossible in practice. Therefore, the retransmission method can
theoretically solve the problem of packet loss in high-speed acquisition systems.
The data analysis center reads the buffered data and sends it to the data transmission
module. If the packet is lost, it receives the retransmission packet from the PC. First
store the retransmission information to the local ram. After the normal data has been
sent, read from the ram for parsing. Re-read data based on the parsed row and column
address to resend data.
The data transmission module completes the parameter configuration of the net-
work card, receives the data sent by the data analysis center for framing, controls the
transmission of the Gigabit Ethernet IP core to the host computer and parse retrans-
mission data in real time and send to the data analysis center.
4 Performance Testing
The theoretical simulation of the three basic retransmission methods is shown in Fig. 5.
The packet loss rate is set to be 40% similar to the actual acquisition system packet loss
rate. In Fig. 5, the horizontal axis represents different propagation delays, and the
vertical axis represents the normalized throughput efficiency. It can be seen that the SR-
ARQ is 1.5 times more efficient than the other two retransmission methods. So the
implementation of the SR-ARQ mode is in line with actual needs.
The resource occupancy of the software is shown in Fig. 6. The logic unit occu-
pancy rate is 59%, and the storage unit utilization rate is 39%, which satisfies the
design requirements.
In order to further verify the feasibility of the design and test the performance, the
acquisition system communicates with the host computer through the Gigabit Ethernet
direct connection.
1000 independent experiments were performed on the data transmission, and the
number of packages per group was 2,203. Count the number of packets lost during
each transmission with or without retransmission, and plot the statistics according to
the test results as shown in Fig. 7.
Fig. 7. A statistical comparison chart of the number of packets lost per transmission with or
without retransmission
60 X. Li and J. Lv
In Fig. 7, the horizontal axis represents the serial number of the number of tests,
and the vertical axis represents the number of lost packets of a complete transmission.
When there is no retransmission, the maximum number of lost packets reaches about
900 packets, the maximum packet loss ratio is 40.8%, and the packet loss rate is 95%.
If the retransmission is not performed, the data cannot be used normally. After the
retransmission mechanism is increased, the maximum number of packet loss is 11
packets, the maximum packet loss ratio is 0.005%, and the packet loss rate is 0.012%,
which is in line with the demand.
5 Conclusion
References
1. Kai, L: Ethernet-based data acquisition system implemented on FPGA. Wuhan University of
Science and Technology (2009)
2. Ning, Y., Guo, Z., Shen, S., Peng, B.: Design of data acquisition and storage system based
on the FPGA. Procedia Eng. (2012)
3. Zhong, G.: Design of high speed data acquisition system based on FPGA. Nanjing
University (2013)
4. Wu, C., Xu, J., Jiang, J.: Research and implementation of gigabit ether-net interface based on
FPGA. Mod. Electron. Tech. 41(09), 1–5 (2018)
5. Shi, P.: Design of FPGA bidirectional data transmission system based on Gigabit Ethernet.
Xidian University (2014)
6. Yang, J.: Research on ARQ Mechanism in MAC Layer of 802.16 m System. University of
Electronic Science and Technology of China (2012)
7. Yi, Q.: Design and FPGA Implementation of Ten Gigabit Ethernet MAC Controller. IEEE
Beijing Section, Global Union Academy of Science and Technology, Chongqing Global Union
Academy of Science and Technology, Chongqing Geeks Education Technology Co., Ltd.
Proceedings of 2017 IEEE 2nd Advanced Information Technology, Electronic and Automation
Control Conference (IAEAC 2017). IEEE Beijing Section, Global Union Academy of Science
and Technology, Chongqing Global Union Academy of Science and Technology, Chongqing
Geeks Education Technology Co., Ltd:IEEE BEIJING SECTION (2017)
8. Xia, Y.: 10 Gigabit ethernet TCP frame data registration algorithm based on FPGA. In:
Proceedings of 2016 2nd International Conference on Mechanical, Electronic and
Information Technology Engineering (ICMITE 2016) (2016)
9. Zhong, Z.: Top-down FPGA design. Electronic Products World (1998)
10. Zou, L., Tian, L.: Design and implementation of gigabit network data acquisition system
based on FPGA. Digital Technol. Appl. 136 (2011)
Application of Hilbert Huang Transform
in Car Signal Demodulation
1 Preface
Railways are an important infrastructure of the country and an important support for the
national economy. Railway signal equipment is an important part of the railway sys-
tem, shouldering the heavy responsibility of ensuring the safe operation of the railway
and improving the efficiency of train transportation [1, 2]. The Car Signal system is an
important subsystem of the China Train Operation Control System (CTCS) [3–5]. It
receives the ground driving commands sent by the track circuit through electromag-
netic induction. After parsing the instruction, the car signal system will display the state
of the front signal lamp, which can reduce the pressure of the driver’s look-out signal
lamp. What’s more, it will sent the instruction to the on-board safety computer, which
has the functions of automatic speed limit and emergency parking, so as to avoid
accidents caused by human errors.
Hilbert Huang Transform (HHT) is a signal processing method based on instan-
taneous frequency proposed by Huang et al. Compared with traditional signal pro-
cessing methods, it can analyze non-stationary nonlinear signals, and is adaptive and
suitable. Processing of mutational signals and other characteristics. China’s current
track circuit system is numerous. This paper takes the ZPW2000 track circuit, which is
widely used at present, as an example, and briefly analyzes its signal characteristics.
Aiming at the problem of degraded quality degradation in the supporting locomotive
ATP
System
Car Signal
Car Signal
System Mainframe
Car Signal
Power
TCR
Interface
Receiving Receiving
coils 1 coils 2
frequency and other information of signals. It is suitable for non-stationary and non-
linear signals. Hilbert Huang Transform can be divided into three parts: Hilbert
transform, empirical mode decomposition and Hilbert spectrum [8–10].
Hilbert Transform
The Hilbert transform is defined as follows: the original signal is x(t), and the Hilbert
transform is ^xðtÞ
Z 1
1 xð t Þ 1
^xðtÞ ¼ ds ¼ xðtÞ ð3Þ
p 1 t s pt
Therefore, the Hilbert transform can be considered as the result of a filter whose
unit response is hðtÞ ¼ pt1 . That is, the Hilbert transform has a positive frequency
hysteresis of 90°, and a negative frequency of 90°. The analytic signal of x(t) can be
defined as
^xðtÞ
uðt) ¼ arctan ð6Þ
xðtÞ
(a) For the original signal xðtÞ, find all the local maximum points and all the local
minimum points. The cubic spline interpolation method is used to fit the maxi-
mum points and the minimum points, and the upper and lower envelopes are
obtained.
(b) The average value of the upper and lower envelopes is calculated as m1(t), make
h1(t) = x(t) – m1(t).
(c) If h1(t) satisfies the IMF condition, we get c1(t) = h1(t); if not, we take h1(t) as a
new signal, repeat steps (a), (b). Get the average of the upper and lower envelopes
as m11(t), make h11(t) = h1(t) – m11(t), and continue to judge whether h11(t)
satisfies the IMF condition. If not, we continue to repeat steps (a), (b) until
screening. Select the k times to get h1k(t) satisfying the condition or satisfying the
termination condition, and get c1(t) = h1k(t).
(d) Based on the first IMF component c1(t), the residual component r1(t) = x(t) – c1(t)
is obtained. When rn(t) satisfies the stopping condition, the decomposition ends.
At this time, the signal has been decomposed into several IMFs components and
one rn(t) remainder.
Through these steps, we can get
Xn
xðtÞ ¼ j¼1
Cj ðtÞ þ rn ðtÞ ð2Þ
In this paper will introduce the concrete steps of demodulation algorithm and validates
it by MATLAB simulation.
(1) Simulate to generate 2CPFSK signal
Based on the theoretical analysis of the second chapter, the 2CPFSK signal is
generated by the method of integral, and the sampling rate of analog signal is
16384 Hz.
(2) Signal preprocessing and EMD decomposition
First, the signal is preprocessed through a strap-pass filter with a pass band of
1500 Hz to 2800 Hz. This is mainly to filter out 50 Hz frequency interference and
its high harmonic to extract the effective spectrum of car signal. Next, decomposes
66 H. Xie and J. Lv
Fig. 2. (a) EMD results, (b) instantaneous frequency, (c) Normalization processing
5 Peroration
This paper introduces the application of HHT algorithm in car signal demodulation.
Taking ZPW2000A track circuit signal as an example, the algorithm is simulated and
verified. This paper summarizes the structure of car signal and the characteristics of
track circuit, introduces the principle and implementation of HHT algorithm in detail.
Through the MATLAB simulation, it has proved that the HHT demodulation algorithm
can greatly improve the demodulation quality of car signal and ensure the safety of
train operation.
References
1. Wang, R.: Foundation of Railway Signal Operation. Railway Publishing House, Beijing
(2008)
2. Chao, Q.L.: Track Circuit and Electrification. Railway Publishing House, Beijing (2008)
3. China Railway Signal History Editorial Committee: Electrified Railway Signal Equipment.
Railway Publishing House, China (2016)
4. Dong, Y.: Interval signal control and Train Operation Control System. Railway Publishing
House, Beijing (2008)
5. Liu, X., Zhang, L., Liu, S.: Research and implementation of ZPW2000 shift frequency signal
detection technology. Instrument technology and sensor, pp. 82–85, December 2014
6. Gai, Q., Zhang, H., Xu, X.: Adaptive frequency multiresolution analysis of Hibert Huang
transform. Electron. J. 33(3), 563–566 (2005)
7. Zhang, J.: Adjacent line interference of type ZPW2000 track circuit. Railway Technol.
Innov. 1, 53–55 (2009)
8. Qiong, G.: Application analysis of ZPW2000A track circuit for passenger dedicated line.
Railway Commun. Signal Eng. 6, 85–87 (2012)
9. Huang, N.E., et al.: The empirical mode decomposition and the Hilbert spectrum for non-
linear and non-stationary time series analysis. Proc. R. Soc. Lond. 454, 903–995 (1998)
10. Huang, N.E., et al.: A confidence limit for the empirical mode decomposition and Hilbert
spectral analysis. Proc. R. Soc. Lond. 459, 2317–2345 (2003)
Human Action Recognition: A Survey
1 Introduction
Recently, computer vision has been a popular branch in artificial intelligence, which
gives a critical influence on human-machine interactions, intelligent vision surveil-
lance, automated driving and other applications based on vision. Most of the infor-
mation is obtained through human vision as well as machine. Human action
recognition is inescapable for a machine to understand the human behavior in computer
vision. The definition of human action has many versions in psychology, sociology and
philosophy. In computer vision, human action aims at the external performance in daily
life for humans [1–3], which means that machine could understand what human do
Feature
Video Sequences Processing Model Action Labels
Extraction
(a)
Action Recognition
(b)
Fig. 1. The two architecture of human action recognition system. (a) shows a methodology
based on presentations, which always extract the feature firstly and then recognize the actions.
(b) presents another methodology, which completes future extraction and action recognition
simultaneously in one model.
through itself way. According to the complexity of actions [3], human activity is di
vided broadly into four levels, including gesture action, interaction and group activity.
The research in this field is more close to real-sceneries, which is in the stage of
actions and interactions.
In this work, two methodologies based on presentations and deep learning are
presented, as shown in Fig. 1(a) and (b). First, many previous-distinctive work are
adopted in Sect. 2. Then we descript 4 public datasets in detail in Sect. 3. Besides, we
compare the mentioned methods based on these public datasets. At last, we summary
the previous work and propose new challenges for future work in this field in
conclusion.
2 Previous Work
Several techniques for human action recognition have been reported, the method based
on presentations [4–6], deep learning [7–10]. We introduce these two methodologies
separately. Figure 2 shows the development of human action recognition over time.
Human action
Signal
recognition
Image Skeleton
time
2005 2014 2018
time
2007
Space-time shape [13] 2016
TSN [16]
2008
LBP+HMM [12] 2016
key frame + ST [17]
2010
Survey [14]
2018
Video Coding [18]
2018
ATW CNN [19]
input of matching. [6] detected all locations through 3-D space-time volume and
matched with similar dynamic behaviors. In [11, 12], Hidden Markov Models (HMMs)
were applied for the variations for sequences in an observation. [13] applied a 2D
shapes to space-time action recognition. [2] provided a detailed overview of previous
work in human action recognition.
[16] presented a key volume mining framework to decrease calculation and complexity
for human action recognition. [17] applied the compressed video theory to action
recognition to decrease the redundant information among frames. [18] presented an
attention-based temporal weighted (ATW) CNN to process spatial RGB images for
object recognition, optical flow images for motion information and optical warped flow
images for motion information in three streams, separately.
Fig. 4. The structure of two-stream ConvNet for human action recognition [8].
For skeleton, the input of the proposed model is skeleton information. In [19], a
hierarchical recurrent neural network (RNN) was proposed to recognize actions. This
paper divided the whole skeleton into five parts fed to five subnets. With the layer
increasing, the features from subnets were fused to higher payers. In [20], geometric
relation modeling among skeletons was applied to RNN-based action recognition. [21]
presented a spatial-temporal graph convolution (ST-GCN) structure on skeleton
sequences, as shown in Fig. 5. [22] showed a deep progressive reinforcement learning
(RL) method skeleton-based for selecting key frames form original video sequences
and a graph-based CNN to deal with key frames for human action recognition.
Human Action Recognition: A Survey 73
3 Datasets Description
With the development of computer vision, the understanding on images and videos has
attracted more attention from researchers. The most important part in this field is
datasets, which has tremendous effect on the action recognition performance for real-
world application scenarios. There are 4 datasets that are used widely in this field will
be introduced in the following.
3.1 KTH
KTH was published by Christian Schuldt in 2004 [23], which has 6 human action
classes, as shown in Fig. 6. In this dataset, each class consists of 400 clips
approximately.
3.2 HMDB51
HMDB51 was published by Kuehne in 2011 [24], which has 51 action classes with 7000
annotated clips. Each class includes 101 clips at least. Some examples of HMDB51 are
shown in Fig. 7. All the videos are from different movies and YouTube videos.
3.3 UCF101
UCF101 was published by Khurram in 2012 [25]. It is a large dataset including 101
human action classes, some examples are shown in Fig. 8. The time span is over 27 h
and each class has about 130 clips. All of the videos are from YouTube, which is more
challenging and diverse than the existing datasets [23, 24].
3.4 Kinetics-600
Kinectics-400 was published by Will Kay in 2017 [26], which is used in ActivityNet
from 2017. Kinectics-400 is a sub-set of Kinectics-600 generated from YouTube
videos, which includes 600 human action classes with at least 600 video clips for each
class. Six examples of 600 classes human actions are showed in Fig. 9. This dataset is
about consist of 500,000 video clips and each clip lasts around 10 s. Up to now,
Kinectics-600 is the largest dataset, which not only includes single class but also
human-human interactions.
5 Conclusion
In this paper, we presented a comprehensive survey for human action recognition. All
proposed methods have achieved great performance on public datasets. Considering the
practical sceneries, there are many challenges that needs the researchers to be dealt with
in the future. For example, all of 4 datasets have one label for each video. However,
there are not one class for single person in real-life, like running with listening, talking
with shaking hands. And there are also not one class for several persons in one video,
like one is drinking and others are talking. We would create new dataset according to
real-life condition.
References
1. Moeslund, T.B., Granum, E.: A survey of computer vision-based human motion capture.
Comput. Vis. Image Underst. 81(3), 231–268 (2001)
2. Poppe, R.: A survey on vision-based human action recognition. Image Vis. Comput. 28(6),
976–990 (2010)
3. Vishwakarma, S.: A survey on activity recognition and behavior understanding in video
surveillance. Vis. Comput. 29(10), 983–1009 (2013)
4. Bobick, A.F., Davis, J.W.: The recognition of human movement using temporal templates.
IEEE Computer Society (2001)
5. Laptev, I., Lindeberg, T.: On space-time interest points. Int. J. Comput. Vision 64(2–3),
107–123 (2005)
6. Shechtman, E., Irani, M.: Space-time behavior based correlation. In: IEEE Computer Society
Conference on Computer Vision and Pattern Recognition, pp. 405–412. IEEE Computer
Society (2005)
Human Action Recognition: A Survey 77
7. Xu, W., Yang, M., et al.: 3D convolutional neural networks for human action recognition.
IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 221–231 (2013)
8. Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition in
videos. In: International Conference on Neural Information Processing Systems, pp. 568–
576. MIT Press (2014)
9. Wang, L., Qiao, Y., Tang, X.: Action recognition with trajectory-pooled deep-convolutional
descriptors. In: Computer Vision and Pattern Recognition, pp. 4305–4314. IEEE (2015)
10. Varol, G., Laptev, I., Schmid, C.: Long-term temporal convolutions for action recognition.
IEEE Trans. Pattern Anal. Mach. Intell. 40, 1510–1517 (2018)
11. Natarajan, P., Nevatia, R.: Coupled hidden semi Markov models for activity recognition. In:
IEEE Workshop on Motion and Video Computing, p. 10. IEEE Computer Society (2007)
12. Kellokumpu, V., Zhao, G., Pietikäinen, M.: Human activity recognition using a dynamic
texture based method. In: British Machine Vision Conference, Leeds. DBLP, September
2008
13. Gorelick, L., Blank, M., Shechtman, E., et al.: Actions as space-time shapes. IEEE Trans.
Pattern Anal. Mach. Intell. 29(12), 2247–2253 (2007)
14. Donahue, J., Hendricks, L.A., Rohrbach, M., et al.: Long-term recurrent convolutional
networks for visual recognition and description. IEEE Trans. Pattern Anal. Mach. Intell. 39
(4), 677–691 (2014)
15. Wang, L., Xiong, Y., Wang, Z., et al.: Temporal segment networks: towards good practices
for deep action recognition. In: European Conference on Computer Vision, pp. 20–36.
Springer, Cham (2016)
16. Zhu, W., Hu, J., Sun, G., et al.: A key volume mining deep framework for action
recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1991–
1999. IEEE Computer Society (2016)
17. Wu, C.Y., Zaheer, M., Hu, H., et al.: Compressed video action recognition. In: Computer
Vision and Pattern Recognition. IEEE (2018)
18. Zang, J., Wang, L., Liu, Z., et al.: Attention-based temporal weighted convolutional neural
network for action recognition. In: IFIP International Conference on Artificial Intelligence
Applications and Innovations, pp. 97–108. Springer, Cham (2018)
19. Du, Y., Wang, W., Wang, L.: Hierarchical recurrent neural network for skeleton based action
recognition. In: Computer Vision and Pattern Recognition, pp. 1110–1118. IEEE (2015)
20. Zhang, S., Xiao, J., Liu, X., et al.: Fusing geometric features for skeleton-based action
recognition using multilayer LSTM networks. IEEE Trans. Multimed. 20, 2330–2343 (2018)
21. Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-
based action recognition (2018)
22. Tang, Y., Tian, Y., Lu, J., et al.: Deep progressive reinforcement learning for skeleton-based
action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 5323–5332 (2018)
23. Schuldt, C., Laptev, I., Caputo, B.: Recognizing human actions: a local SVM approach. In:
Proceedings of the 17th International Conference on Pattern Recognition, ICPR 2004, vol. 3,
pp. 32–36. IEEE (2004)
24. Kuehne, H., Jhuang, H., Garrote, E., et al.: HMDB: a large video database for human motion
recognition. In: IEEE International Conference on Computer Vision (ICCV), pp. 2556–
2563. IEEE (2011)
25. Soomro, K., Zamir, A.R., Shah, M.: UCF101: a dataset of 101 human actions classes from
videos in the wild. arXiv preprint arXiv:1212.0402 (2012)
26. Kay, W., Carreira, J., Simonyan, K., et al.: The kinetics human action video dataset. arXiv
preprint arXiv:1705.06950 (2017)
Genetic Algorithm for Base Station ON/OFF
Optimization with Fast Coverage Estimation
and Probability Scaling for Green
Communications
1 Introduction
Various methods of base station ON/OFF strategies have been proposed in green
communications. Oh et al. [8] proposed an dynamic base station ON/OFF strategy
algorithm to reduce the power consumption in wireless cellular networks. The proce-
dure is to turn off the base station which has the least effect on the network by
considering the additional influence of the adjacent base stations. Zhou et al. [9] solved
the concerned base station energy saving problem in green communications and pre-
sented an algorithm to dynamically turn off certain base stations when the network
traffic is low. Son et al. [10] designed a theoretical framework to save the energy
consumption of base station, which consisted of the dynamic BS ON/OFF strategy and
the user association problem. The optimization also emphasized the total cost mini-
mization but neglecting the quality of service. Bousia et al. [11] developed a switch
base station ON/OFF algorithm green communications, which adequately considered
the information of the distance between the neighbor user equipment and their asso-
ciated base station. The ON/OFF algorithm mainly lied on reducing the power con-
sumption of the communication networks by minimizing the power utilization. Bousia
et al. [12] presented a dynamic base station switch ON/OFF strategy to solve the
problem of the inefficient use of the base station power as an improvement of the
previous work. Kim et al. [13] formulated an energy minimization problem to balance
the power consumption and the revenue of the cellular networks. They also proposed
an algorithm for both base station ON/OFF strategy and user terminal association in
heterogeneous networks. Oh et al. [14] researched a dynamic base station ON/OFF
strategy system to reduce the power consumption of base station considering the
temporal characteristic of the user terminals. The existing approaches of network
configurations are mainly ignore the user terminal number and the quality of service.
In this paper, we propose a novel probability aware genetic algorithm for base
station ON/OFF strategy in green communications. Our contributions mainly lie in two
folds. We introduce an optimization problem to maximize the user terminal coverage
ratio with a given number of ON state base stations. Our contributions mainly lie in two
folds. We formulate an optimization problem to maximize the user terminal coverage
ratio with a given number of activated base station. We propose a novel genetic
algorithm to optimize the ON/OFF strategy with fast coverage estimation, in which the
scaling and selection operators are carefully designed to take the probability distribu-
tion of the estimated coverage ratio into account. Experiments have been conducted to
prove the proposed algorithm for the network configuration for green communication.
The rest of the paper is organized as follows. Section 2 formulates the problem
models and Sect. 3 proposes the novel probability aware genetic algorithm. Section 4
shows the experiments and discusses the results. Finally Sect. 5 concludes this paper.
80 Y. Ren et al.
This section introduces the network scenario and formulates the optimization problem.
We can turn ON/OFF the base stations artificially according to the variations of the
users’ locations.
The ON/OFF state strategy can be expressed as
Note that the strategy G is an n-length vector in which the elements are the states of
the base stations. An ON/OFF strategy represents a set of the activated base stations’.
There are m user terminals distributed in region R. Let the user terminal set be
U = {U1, U2,. .., Um}. The illustration of the network deployment is shown in Fig. 1.
To simplify the coverage problem, we assume that if user terminal Uj is in the range
of a circle whose center is base station Bi with a predefined radius RadiusBase, the user
terminal Uj is covered by base station Bi. Covered user terminals have good com-
munication performance and satisfied quality of service, and vise versa. The illustra-
tions of the covered user terminal and the uncovered user terminal are demonstrated in
Fig. 2.
Genetic Algorithm for Base Station ON/OFF Optimization 81
Fig. 2. Illustrations of the covered user terminal and the uncovered user terminal.
^ is the set of covered user terminals, and |X| is the cardinality of set X.
where U
max
C ðSB1 ; . . .; SBn Þ ð4Þ
SB1 ; . . .; SBn
X
n
s:t: SB1 ¼ ^n:
i¼1
where k is the population size, that is, the number of the state strategies in a population.
82 Y. Ren et al.
Now the optimization goal is to find the optimal individual Gl to maximize the
region coverage ratio with a given number of the on state base stations.
^ X
M
^ ðGl Þ ¼ M ¼ 1
C I Ui0 is covered in Gl ð7Þ
M M i¼1
where M ^ is the number of the covered user terminals among the sampled user termi-
nals? Note that the coverage here is not the accurate coverage, but an roughly estimated
one which only calculated by M user terminals.
Assume that the coverage status of any user terminal, say Uk obeys the Bernoulli
distribution with the covered probability P(Uk is covered) ¼ C. Here C is the accurate
coverage ratio and is treated as a parameter to be estimated. Denote the probability
density function of Bernoulli distribution for user terminal Uk as
C B is covered;
f ðxÞ ¼ Cx ð1 CÞ1x ¼ ð8Þ
1 C otherwise:
Thus, the mean value and the variance for the stochastic variable I Ui0 is
covered in Gl Þ is given by
X
1
E I Ui0 is covered in Gl ¼ xi f ðxÞ ¼ 0 þ C ¼ C; ð9Þ
i¼0
X1
Var IðUi0 is covered in Gl Þ ¼ ^ C
ðxi CÞ2 f ðxÞ ¼ 1 C ^ ð10Þ
i¼0
Genetic Algorithm for Base Station ON/OFF Optimization 83
XM
E C^ ðGl Þ ¼ 1 C ¼ C; ð11Þ
M i¼1
X
M
C ð1 C Þ
Var C^ ðGl Þ ¼ 1 C ð1 C Þ ¼ : ð12Þ
2
M i¼1 M
Note the variance of the estimated coverage ratio is reduced if the sample number
M is increased.
Fig. 3. Illustrations of the covered user terminal and the uncovered user terminal.
In order to calculate the fitness function, we first select a scaling baseline constant
THP. Now we check the probability Fitl that the estimation value is greater than or
equal to THP.
XP
MTH
M
Fitl ¼ P C^ ðGl Þ THP ¼ 1 Ci ð1 C ÞMi ð13Þ
i¼0
i
84 Y. Ren et al.
The larger the fitness function is, the better the individual is.
Selection operation is an important way to generate new population by selecting the
good performance one into next generation. Denote the selection probability for
individual Gl as
Fitl
l ¼ Pk
ProSel ð14Þ
n¼1 Fitl
The original base station location map and original user terminal coverage map are
illustrated in Fig. 6. The blue dots are the on state base stations in Fig. 6(a). We
randomly generate an ON/OFF base station strategy and the on state base stations are
disperse in region R. The red region is the covered region in Fig. 6(b).
86 Y. Ren et al.
Fig. 6. The base station location and the user terminal coverage map in the initial state.
The base station location map and user terminal coverage map after the opti-
mization of the probability aware genetic algorithm are illustrated in Fig. 6. We can see
from Fig. 7(a) that the on state base station distribute more even. More even distri-
bution makes larger coverage ratio. Thus, the algorithm is valid in the base station
ON/OFF strategy. We can see from Fig. 7(b) that the user terminal coverage area is
larger than the initial one. Therefore, the algorithm is accurate to maximize the number
of the user terminal, that is, the user terminal coverage, whose quality of service meets
the required threshold in green communications.
Fig. 7. The base station location map and the user terminal coverage map of the optimization of
the probability aware genetic algorithm after 50 iterations.
The relationship of the coverage ratio and the computation load, i.e., the product of
iterations and samples, of genetic algorithm is illustrated in Fig. 8. If the We can see
from Fig. 8 that the convergence rate of the proposed genetic algorithm is fast even
with a small sample number M. In the same time span, we can run 50 iterations for
M = 100 but only 10 iterations for M = 500. Note that if M = m, i.e., we sample all the
user terminals, our algorithm degenerates into a naive one, which still in its first
iteration in the same time span. Therefore, compared to the naive one without fast
coverage estimation, the proposed algorithm converges faster and is more efficient.
Genetic Algorithm for Base Station ON/OFF Optimization 87
Fig. 8. The relationship between the maximum coverage and the iterations of probability aware
genetic algorithm.
5 Conclusion
Minimizing the power consumption while maximizing the quality of service has
become a mainstream problem in green communications. The existing approaches of
network configurations are mainly ignore the user terminal number and the quality of
service. Our contributions mainly lie in two folds. We introduce an optimization
problem to maximize the user terminal coverage ratio with limited on state base station
number. We not only consider the power consumption by limiting the on state base
station number, but also concern the quality of service by maximizing the user terminal
number. Moreover, we propose a novel probability aware genetic algorithm to optimize
the ON/OFF strategy. The coverage fitness function is converted into a probability one
with less user terminals. Experiments have been conducted to prove the proposed
algorithm for the network configuration.
References
1. Mahapatra, R., Nijsure, Y., Kaddoum, G., Hassan, N.U., Yuen, C.: Energy efficiency
tradeoff mechanism towards wireless green communication: a survey. IEEE Commun. Surv.
Tutorials 18(1), 686–705 (2016)
2. Abrol, A., Jha, R.K.: Power optimization in 5G networks: a step towards GrEEn
communication. IEEE Access 4, 1355–1374 (2016)
3. Gandotra, P., Jha, R.K., Jain, S.: Green communication in next generation cellular networks:
a survey. IEEE Access 5, 11727–11758 (2017)
4. Luo, C., Guo, S., Guo, S., Yang, L.T., Min, G., Xie, X.: Green communication in energy
renewable wireless mesh networks: routing, rate control, and power allocation. IEEE Trans.
Parallel Distrib. Syst. 25(12), 3211–3220 (2014)
88 Y. Ren et al.
5. Son, K., Kim, H., Yi, Y., Krishnamachari, B., Wu, J., Rangan, S., Zhang, H.: Toward
energy–efficient operation of base stations in cellular wireless networks. In: Green
Communications: Theoretical Fundamentals, Algorithms, and Applications (2012)
6. Wu, Q., Li, G.Y., Chen, W., Ng, D.W.K., Schober, R.: An overview of sustainable green 5G
networks. IEEE Wirel. Commun. 24(4), 72–80 (2017)
7. Hasan, Z., Boostanimehr, H., Bhargava, V.K.: Green cellular networks: a survey, some
research issues and challenges. IEEE Commun. Surv. Tutor. 13(4), 524–540 (2011)
8. Oh, E., Son, K., Krishnamachari, B.: Dynamic base station switching–on/off strategies for
green cellular networks. IEEE Trans. Wirel. Commun. 12(5), 2126–2136 (2013)
9. Zhou, S., Gong, J., Yang, Z., Niu, Z., Yang, P.: Green mobile access network with dynamic
base station energy saving. In: ACM MobiCom, vol. 9, no. 262, pp. 10–12 (2009)
10. Son, K., Kim, H., Yi, Y., Krishnamachari, B.: Base station operation and user association
mechanisms for energy-delay tradeoffs in green cellular networks. IEEE J. Sel. Areas
Commun. 29(8), 1525–1536 (2011)
11. Bousia, A., Antonopoulos, A., Alonso, L., Verikoukis, C.: “Green” distance-aware base
station sleeping algorithm in LTE– Advanced. In: IEEE International Conference on
Communications (ICC), pp. 1347–1351 (2012)
12. Bousia, A., Kartsakli, E., Alonso, L., Verikoukis, C.: Dynamic energy efficient distance-
aware base station switch on/off scheme for LTE– Advanced. In: Global Communications
Conference (GLOBECOM), pp. 1532–1537. IEEE (2012)
13. Kim, S., Choi, S., Lee, B.G.: A joint algorithm for base station operation and user
association in heterogeneous networks. IEEE Commun. Lett. 17(8), 1552–1555 (2013)
14. Oh, E., Krishnamachari, B.: Energy savings through dynamic base station switching in
cellular wireless access networks. In: GLOBECOM, vol. 2010, pp. 1–5 (2010)
Satellites and Remote Sensing
Inter-layer Link Design
for Resource-Constrained Double-Layered
Satellite Network
1 Introduction
Since double-layered satellite networks (DLSN) can integrate the strong points of both
low-orbit satellites and high-orbit satellites, it has attracted much attention in the
research field [1]. Due to the relative motion of low-orbit satellites and high-orbit
satellites in the DLSN, they cannot be always visible. Therefore, the inter-layer links
(ILL) of the DLSN need to be dynamically switched, thus causing satellite network
topology changing dynamically. Since the inter-satellite link within a layer can generally
be fixed by selecting an appropriate constellation configuration [2], the dynamical
topology of DLSN is mainly attributed to the continuous switching of the ILLs.
Currently, the academic community has carried out some research work on the
inter-layer topological dynamics of DLSN, and mainly uses some optimal strategies to
establish ILL, such as the shortest distance, the longest connection time, and the
maximum resource utilization. Currently, the most popular ILL establishment strategy
is based on the longest connection time strategy, that is, low-orbit satellites will choose
a high-orbit satellite which has the longest predictable connection time to establish ILL
[3]. Each time the ILL changes, a new topology snapshot will be generated. Since this
strategy does not consider snapshot optimization issues, it will generate a large amount
of snapshots and the durations of snapshots are mostly short. Zhou et al. [4] and Long
et al. [5] proposed a snapshot optimization method based on snapshots merging, which
can greatly reduce the snapshots amount, while at the same time, increase the snapshots
duration.
Wu et al. [6] propose a concentrated link establishment strategy for ILLs of
LEO/MEO satellite networks. The authors force the ILLs to be reestablished at the
same time to reduce the topological dynamics. For example, for concentrated time
strategy, each LEO satellite will independently find the MEO satellite that can provide
the longest coverage time to establish ILL, and the next actual reconstruction time of all
LEO ILLs is the minimum of all LEO theoretical reconstruction time. Because the
reconstruction of the laser inter-satellite link takes a long time, the centralized ILL
reconstruction strategy of DLSN will cause the MEO layer and the LEO layer to be
isolated during the link reconstruction. Therefore, Li et al. [7] proposed a two-step
synchronous handover scheme for laser ILLs. The authors divide the ILLs into two
groups which perform ILL switching alternately. On the basis of ensuring the con-
nectivity of satellites in two layers, the topological dynamics of the inter-layer can be
reduced. Shi et al. [8] propose a traffic aware ILL selection method by considering both
the flow situation of the associated nodes and the connection time of the associated
nodes and obtain balanced flow distribution among high-orbit satellites.
The above ILL establishment strategy is low-orbit-centric and assumes that the
number of ILLs of high-orbit satellite is not limited. However, actual DLSNs are
generally resource-constrained, such as weight and power. Thus the number of ILLs
can be equipped on one satellite is limited. Therefore, traditional ILL establishment
strategy is hard to apply in practice [9].
In this paper, we first propose an ILL establishment strategy which is high-orbit-
centric. We assume that high-orbit satellites have only one ILL and use this as a
precondition to establish ILLs. Then, the system snapshot performances of two ILL
establishment strategies under two typical DLSNs constellation configurations are
simulated and analyzed. Finally, further research trends are pointed out.
The typical architecture of DLSN is shown in Fig. 1. Since the inter-satellite links
(ISL) in the same layer can be fixed by choosing an appropriate constellation config-
uration [2], the topology of low-orbit satellite constellation or high-orbit satellite
constellation alone can be regarded as constant. However, due to the relative motion of
high-orbit satellites and low-orbit satellites, the ILLs have to be dynamically switched.
Inter-layer Link Design 93
Then, it can be summarized that the topological dynamics of DLSN are mainly
attributed to the dynamic switching of the ILLs.
The traditional ILL establishment strategy is generally based on the longest cov-
erage time strategy [3], where, the low-orbit satellite will choose a high-orbit satellite
with longest predicted connection time in all visible high-orbit satellites until the high-
orbit satellite can no longer be visible. Then, the low-orbit satellite will continue to use
the same longest coverage time strategy to select the next high-orbit satellite to
establish ILL.
The traditional ILL establishment strategy is low-orbit-centric, that is, all low-orbit
satellites can establish ILLs with high-orbit satellites. Since the high-orbit satellites is
generally less than low-orbit satellites, high-orbit satellites generally need to establish
multiple ILLs. Taking the LEO/MEO constellation selected in [9] for example, the MEO
needs to establish 10 ILLs at most, 2 ILLs at least, 6 ILLs on average. This poses very
high requirements for the MEO satellite platform, since it should be equipped with many
ILLs. Therefore, traditional ILL establishment strategy is hard to apply in practice.
We propose an ILL establishment strategy which is high-orbit-centric based on the
premise that each high-orbit satellite can only establish one ILL. The ILL establishment
strategy which is high-orbit-centric is described as follows: the high-orbit satellite will
choose a low-orbit satellite with longest predicted visible time in all visible low-orbit
satellites until the low-orbit satellite can no longer cover itself. Then, the high-orbit
satellite will continue to use the same longest visible time strategy to select the next
low-orbit satellite to set up ILL.
The ILL establishment strategy which is high-orbit-centric can guarantee that each
high-orbit satellite only establishes one ILL, and because there are generally fewer
high-orbit satellites than low-orbit satellites, it cannot guarantee that all low-orbit
satellites can establish ILL.
94 H. Yan et al.
3 Simulation Analysis
This section analyzes the ILL topology through simulation. The simulation time is 24 h
and we choose the simulation step as 1 s. The simulation analysis is focusing on two
typical DLSNs, one is LEO/MEO and the other is IGSO/MEO. For each type of
DLSNs, traditional ILL establishment strategy which is low-orbit-centric and the
proposed ILL establishment strategy which is high-orbit-centric are used to establish
ILL. The snapshots are all merged using the method proposed by Zhou-Long.
3.1 LEO/MEO
The constellation configuration of the LEO/MEO DLSN selected is shown in Table 1,
in which LEO adopts the Celesti constellation and MEO adopts the ICO constellation.
Figure 2 shows the system snapshot sequence generated by two ILL establishment
strategies. Figure 2(a) shows the simulation result of traditional ILL establishment
strategy, and Fig. 2(b) shows the simulation result of proposed ILL establishment strat-
egy. From Fig. 2, it can be seen that the system snapshot generated by the proposed
Fig. 2. System snapshot sequence generated by two ILL establishment strategies (a) low-orbit-
centric, (b) high-orbit-centric
Inter-layer Link Design 95
strategy is less than traditional strategy. The system snapshot duration generated by the
proposed strategy is also larger than that of the traditional strategy. This is due to the
reason that since there are fewer ILLs generated by the proposed strategy, the inter-layer
topology may be less dynamic which is demonstrated by fewer number and larger
duration of system snapshot.
The number and duration of system snapshots generated by two ILL establishment
strategies are shown in Table 2. The system snapshot generated by the proposed
strategy has a 22% reduction in number compared to the traditional strategy. In terms
of time, the shortest snapshot duration is reduced, the longest snapshot duration is
increased to 3.1 times and the average snapshot duration increased by 28%.
From Fig. 4, it can be seen that the proposed ILL establishment strategy will make
the ILLs of LEO satellite idle for certain periods of time, i.e., it cannot establish an ILL
with the MEO. Moreover, the ILL of some LEO satellites are always in an idle state,
i.e., these LEO satellites can never establish an ILL with MEO satellites.
3.2 IGSO/MEO
The constellation configuration of the IGSO/MEO DLSN selected is shown in Table 3,
in which MEO adopts the walker 24/3/1 constellation and IGSO adopts a constellation
with the same satellite sub-point track.
Figure 5 shows the system snapshot sequence generated by two ILL establishment
strategies. Figure 5(a) shows the simulation result of traditional ILL establishment
strategy, and Fig. 5(b) shows the simulation result of proposed ILL establishment strat-
egy. Similar to the LEO/MEO constellation, it can also be seen from Fig. 6 that the system
snapshot generated by the proposed strategy is less and longer than that of traditional
strategy.
Inter-layer Link Design 97
Fig. 5. System snapshot sequence generated by two ILL establishment strategies (a) low-orbit-
centric, (b) high-orbit-centric
The number and duration of system snapshots generated by two ILL establishment
strategies are shown in Table 4. The system snapshot generated by the proposed
strategy has a 29% reduction in number compared to the traditional strategy. In terms
of time, the shortest snapshot duration is increased by 112 times, the longest snapshot is
increased to 4 times and the average snapshot duration increased by 33%.
98 H. Yan et al.
3.3 Discussion
From the above simulation results we can see that since the number of high-orbit
satellites is generally smaller than that of low-orbit satellites, low-orbit satellites will
not establish ILL with high-orbit satellites at certain time intervals. Moreover, some
low-orbit satellites can never establish ILLs with high-orbit satellites.
Inter-layer Link Design 99
Therefore, in the future, we need to study how to make more efficient use of ILLs of
low-orbit satellite. For example, we can make all low-orbit satellites alternately establish
ILLs with high-orbit satellites to prevent some low-orbit satellites ILLs from being idle;
we can also make the idle low-orbit satellites ILLs establish intra-layer links and
increase the topology connectivity of low-orbit layer; or, we can just make some low-
orbit satellites equipped with less ILLs to saving satellite construction cost, and so on.
4 Conclusions
References
1. Qi, X., Ma, J., Wu, D., Liu, L., Hu, S.: A survey of routing techniques for satellite networks.
J. Commun. Inf. Networks 1, 66–85 (2016)
2. Werner, M., Frings, J., Wauquiez, F., Maral, G.: Topological design, routing and capacity
dimensioning for ISL networks in broadband LEO satellite systems. Int. J. Satell. Commun.
Netw. 19, 499–527 (2001)
3. Chen, C., Eylem E., Akyildiz, I.F.: Satellite grouping and routing protocol for LEO/MEO
satellite IP networks. In: Proceedings of the 5th ACM international workshop on Wireless
mobile multimedia, pp. 109–116. ACM (2002)
4. Zhou, Y., Sun, F., Zhang, B.: A novel QoS routing protocol for LEO and MEO satellite
networks. Int. J. Satell. Commun. Network. 25, 603–617 (2007)
5. Long, F., Xiong, N., Vasilakos, A.V., Yang, L.T., Sun, F.: A sustainable heuristic QoS
routing algorithm for pervasive multi-layered satellite wireless networks. Wirel. Netw. 16,
1657–1673 (2010)
6. Wu, T., Wu, S.: Performance analysis of the inter-layer inter-satellite link establishment
strategies in two-tier LEO/MEO satellite networks. J. Electron. Inf. Technol. 30, 67–71 (2008)
7. Li, Y., Wu, J., Zhao, S., Deng, B., Miu, X.: A two-step synchronous handover scheme of
optical inter-orbit links in LEO and MEO satellite network. Acta Electronica Sinica 45, 762–
768 (2017)
8. Shi, W., Gao, D., Zhou, H., Xu, Q., Foh, C.H.: Traffic aware inter-layer contact selection for
multi-layer satellite terrestrial network. In: GLOBECOM 2017, IEEE Global Communica-
tions Conference, pp. 1–7 (2017)
9. Yan, H., Guo, J., Wang, X., Zhang, Y., Sun, Y.: Topology analysis of inter-layer links for
LEO/MEO double-layered satellite networks. In: International Conference on Space
Information Network, pp. 145–158. Springer, Singapore (2018)
Safety Design for Customer Furniture
Instrument in Satellite
Abstract. More and More piggy-back units are required in recently spacecraft
mission due to the willingness of space industry development of customer. The
risk caused by such kind unit shall be considered in all projects, and normally
the units which are mounted on board are required to ensure the safety of
satellite. However, the risks shall be considered at system level. These risks are
analyzed at system level, and some measures are taken in a typical project which
include the power supply safety, communication safety and system safe mode
design. All these methods are validated in ground test. The satellite was laun-
ched in 2018 and the performance can meet the requirement. The safety design
to CFI is successful and all the measures can be a reference to other similar
mission.
1 Introduction
The spacecraft design related with its safety issue is a key element during the whole
process which had indicated in the national standard program [1] and also issued in the
earlier spacecraft design [2, 3].
As aerospace technology keeps developing during past years, the usage of satellite
is very important for many countries and companies; however, the satellite design and
manufacture process became more attractive for those end users. In fact, the satellite
design and manufacture process is quite comprehensive so that will cause a huge risk
when they start from a spacecraft development directly. A general way is to develop a
unit of spacecraft to gather experience at the initial stage before they start the whole
satellite design and manufacture. The piggy-back unit requirement from customer is
more and more frequency in spacecraft development mission. The risk shall be rec-
ognized and the measures shall be taken since design phase to eliminate the fault in
mission level, even in case of piggy-back unit abnormal. A lot of research have been
done and most of them are focused on how to design this kind new equipment, [4]
especially in energy interface design [5, 6] and signal interface design [7].
A customer furnished instrument (CFI) mission is received from customer in some
project. The power safety, communication safety and safe mode design are designed at
spacecraft level in this paper.
2 Design Method
2.1 Description to CFI
The CFI unit which is used for data process on board was designed, developed in-
house, tested by customer and supplied to the satellite contractor in line with the
contract. The design, material, manufacture and process can be validated in orbit
through this unit. It was required to provide necessary resource which include power,
communication and so on. The core requirement of design is safety of the satellite and
no operation of CFI will cause any degradation in the performance of the satellite.
The CFI unit is separated in 2nd payload section of power distribution design, and the
power on and off are controlled by power distribution unit (PDU) directly. There is no
other unit is included in this 2nd payload power section. In normal case, the power on
and off are controlled by tele-command through on-board computer (OBC). However,
the power distribution unit will monitor the current and it can also be powered off by
power distribution unit directly. A fuse is designed in power distribution unit, and it
will ensure the shortcut fault will not affect the other part of satellite. At the same time,
another fuse is required inside of the CFI as a redundancy design. The permitted current
of the fuse inside the unit is a little lower than the outside one.
The CFI unit is powered by 30 V power bus and the power interface is shown in
Fig. 1. Two fuses in parallel are designed for overcurrent protection. The rated power
consumption of CFI is 4 W and the rated current is 0.135 A. The working current of
chosen fuse can meet the requirement of derating design. The 30 V power bus is
powered by power control unit and the maximum current is 13 A for this circuit. In case
of the shortcut occurred in CFI, the instantaneous shortcut current will be 22 A when
the other units are working in normal case. The fuse can be fused in this case and the
abnormal will not affect the other unit of satellite.
102 T. Tan et al.
nd
Current of 22nd
Current of
Payload
Payload SecƟon
SecƟon
30V CFI
CFI Main
Main Power
Power Supply
Supply ++
30V Power
Power Bus
Bus ++
CFI
CFI backup
backup Power
Power Supply
Supply ++
nd
Voltage of 22nd
Voltage of
Payload
Payload SecƟon
SecƟon
30V
30V Power
Power Bus
Bus -- CFI
CFI Main
Main Power
Power Supply
Supply --
CFI
CFI backup
backup Power
Power Supply
Supply --
4 Communication Design
The OBC provides dual redundant CAN 2.0 bus interface link with the CFI-EnD unit
for sending and receiving data streams. CFI unit connect to a separated payload CAN
bus and only this unit is connected with satellite through this CAN bus. 4 OC com-
mands to power on/off CFI-EnD unit are designed for this unit (Fig. 2).
OBC
CAN Bus A
Payload
CAN Bus
CAN Bus B
5 Grounding Requirements
A CFI current safe mode and a CFI communication safe mode are designed in system
level. The healthy status is monitored and the safe mode will be implemented auto-
matically by on board computer (OBC) through the telemetry. The PDU will monitor
the current of 2nd power section also.
The power supply subsystem is most important part in the safety design, so this
status is monitored by both PDU and OBC. The current will be collected every second
by PDU and a threshold value (4 A) is designed. Once the current is exceeded, the tele-
command of power off CFI will be send by PDU directly. This safe mode is tested
during PDU acceptance test. It is also tested in satellite level.
As to the communication safe mode, the OBC will communicate with the CFI each
second to collect the telemetries. The safe mode will be trigged when 20 times con-
tinuous communication abnormal occurred. The CFI will be powered off by OBC
automatically. This safe mode is tested during the interface test between CFI and OBC.
It is also tested in satellite level.
7 Conclusion
Based on the CFI embarked requirement, the power supply design, communication
design, grounding design and satellite safe mode design are considered in satellite
development. All these methods are validated in ground test. The satellite was launched
in 2018 and the performance can meet the requirement. The safety design to CFI is
successful and all the measures can be a reference to other similar mission.
References
1. Guide to satellite reliability design. QJ 2172A-2005
2. Tang, B.: Satellite safety design. Spacecr. Eng. 1–5 (1994)
3. Jin, L., Wu, D.: Satellite system reliability and safety review. Spacecr. Environ. Eng. 500–504
(2010)
4. Du, Z., Zhen, G., Dong, X.: Safety and reliability design of ignition controller. Comput.
Measur. Control (2013)
5. Yu, L., Ren, X.: Design of sapcecraft power system. Chin. J. Power Sources (2013)
6. Yu, L., Ren, X.: Study on DC/DC structure analysis for spacecraft. Electron. Product Reliab.
Environ. Test. (2010)
7. Chen, X., Wang, Y.: The analysis and conceiving for the grounding architecture of satellite.
Aerosp. Control (2011)
Dynamic Path Planning Algorithm Based
on an Optimization Model
1 Preface
The study of unmanned surface vessels (USVs) began during World War II [1]. USVs
have to function in more complex sea conditions [2]. This causes application of
autonomous navigation which contain generate obstacle maps and obtain path points
via path planning. The common path planning for USV are as follows:
After constructing map, a graph search is often performed to search the optimal
trajectory. A* [3] is a heuristic optimal search algorithm that employs a heuristic
function to guide the nodes close to the target point to obtain a preferential expansion
opportunity. In [4], A first-responder path planning under uncertain moving obstacles
based on algorithm A* is proposed. Zuo proposed a new hierarchical path planning
method for navigation in complex environments [5]. This method has all the future path
information known after the planner execution and before the vehicle movement [6].
However, the number of nodes is excessive and exponentially increases in the
dimension. The Morphin algorithm [7] is a path planning method that is based on
limited environmental information. This method is typically employed to understand
only local environmental information and global goals. Considering the dynamic
performance of the algorithm, algorithm based on the optimization model is gradually
applied to motion planning, with the development of optimization theory in recent
years [8, 9]. Besides, potential field method is a prevalent method due to its small
computational load; however, it may be trapped in the local optimum [10]. Enhanced
discrete particle swarm optimization path planning is proposed for vision-based surface
inspection [11]. Adaptive algorithm was proposed to estimate the scalar field over a
region of interest [12]. Devaurs developed sampling-based approaches combining
RRT⁄ and T-RRT [13]. The algorithm is fast, guarantees the optimum but may not be
solved.
Two problems need to be solved for planning in a marine task: (1) without con-
sidering obstacle’s moving; (2) without considering USV dynamic constraints which
cause path fitting effect is poor. They are not both considered meanwhile in planning
before. For this reason, this paper proposes a path planning algorithm that incorporates
time dimension and considers the kinematics constraints of an USV. The algorithm
abstracts the path planning into a nonlinear optimization model and outputs the
sequence data of the planned position and speed of USV.
Based on the distance between the USV and an obstacle, a nonlinear optimization
model is established with some limits, such as speed and acceleration limitation, dis-
tance from obstacle, finally obtain the optimized pose sequence.
Ok;t including the position ½xk þ vk t sin hk ; yk þ vk t cos hk (obstacle position ðx; yÞ,
velocity ðvÞ, angle ðhÞ at 0 time) and the expansion distance dk;t . The accuracy is lower
after a period of time, the dynamic obstacle may appear in a larger range. The 0 time
and the t time space obstacle is shown in Fig. 1(a),(b).
As the obstacle may move with time, this paper establishes the xyt space to
represent the position of the obstacle as shown in Fig. 2. The horizontal plane is
xy(position), and the vertical axis is t(time). Stationary obstacle is the cylinder which is
perpendicular to the plane. Dynamic obstacles position shifts at certain velocity, as an
inclined cylinder. For obstacles with lower certainty, the inflated circle increases with
time. The cylinder’s upper base is larger than the lower base.
Setting Optimization Variables. The location at the discrete time t is
st ¼ ½xt ; yt 2 R2 ð3Þ
X
n
V ðsÞ ¼ min ½ðxt xe Þ2 þ ðyt ye Þ2 ð4Þ
s
t¼m
ðxe ; ye Þ is the end. The Euclidean distance between the USV and the end which is
calculated from time m to the final time n is calculated to minimize.
Define the time interval of adjacent values in USV and obstacle position sequence
is Dt. The conditions to be satisfied are
(1) The starting point of the USV is s1 ¼ ss , which is known.
(2) b 2 C is the yaw angle of the USV in the map coordinate system,
x 1 xt
bt ¼ arctan ytt þ
þ 1 yt
.
(3) The distance between the USV and the k-th obstacle at the position st minus the
expansion distance of the k-th obstacle at time t needs to be greater than 0.:
(4) The speed falls within a certain range. vlt is the forward speed at time t, vlmin ; vlmax
are the limit, correspondingly; vrt is the steering speed at time t, vrmin ; vrmax are the
limit, correspondingly.
(5) Similarly, the acceleration falls within a certain range. alt is forward acceleration at
time t; almin ; almax are the limit, correspondingly. art is the angular acceleration at
time t; armin ; armax are the limit, correspondingly.
108 J. Zhang et al.
(6) Due to the existence of obstacles, the model is a typical nonlinear non-convex
optimization model (The objective function is nonlinear, and the feasible region is
the position coordinate definition domain, not connected region), and its local
optimal point is related to the number of obstacles. KKT condition explain that the
dimension of the H matrix to be solved in the process of solving the Lagrange
function is equivalent to the dimension of the independent variable of the original
optimization. Thus, direct solving of this type of model is difficult.
To improve efficiency, the original problem is approximate to a least squares
optimization problem to solve the inequality constrained problem. The method
employs the first derivative to approximate the H matrix. The sparsity of the matrix is
utilized to improve the efficiency of the solution. For example, the inequality con-
straints can be approximate to a least squares penalty function.
2
DðOk;t ; st Þ ¼ distðOk;t ; st Þ dk;t 0 ! r0 minf0; DðOk;t ; st Þg2 ð14Þ
r0 denotes the penalty factor, which represents the weight relative to other penalty
items. The greater r0 , the more focused is the objective to satisfy this items. For
convenience, the penalty factor is omitted in the latter formula.
The objective function to be optimized is
8
>
> Pn P
k¼m;t¼n
>
> V ð sÞ ¼ ½ðx - x Þ 2
þ ðy - y Þ 2
þ jjmin 0; DðOk;t ; st Þ jj22 þ Fvp þ Fap
>
>
t e t e
>
> t¼p k¼0;t¼0
>
> tP
¼n tP¼n
>
> Fvp ¼ jjminf0; vlmin vlt gjj2 þ 2
jjminf0; vlmax vlt gjj22
>
>
>
>
< t¼0
tP
¼n
t¼0
tP¼n
þ jjminf0; vrmin vrt gjj22 þ jjminf0; vrmax - vrt gjj22
>
>
>
>
t¼0 t¼0
>
> P
t¼n P
t¼n
> 2
jjminf0; almax alt gjj22
> Fap ¼
>
>
jjminf0; almin alt gjj2 þ
>
> t¼0 t¼0
>
> tP
¼n tP
¼n
>
> þ jjminf0; armin art gjj2 þ 2
jjminf0; armax - art gjj22
:
t¼0 t¼0
ð15Þ
Dynamic Path Planning Algorithm Based on an Optimization Model 109
X
n X
k¼m;t¼n
VðsÞ ¼ ½ðxt - xe Þ2 þ ðyt - ye Þ2 þ jjmin 0; DðOk;t ; st Þ jj22 þ Fvp þ Fap
t¼p k¼0;t¼0
X
n X
n X
k¼m;t¼n
¼ ðxt - xe Þ2 þ ðyt - ye Þ2 þ jjmin 0; DðOk;t ; st Þ jj22 þ Fvp þ Fap
t¼p t¼p k¼0;t¼0
X
t¼n X
t¼n
Fvp ¼ jjminf0; vlmin vlt gjj22 þ jjminf0; vlmax vlt gjj22
t¼0 t¼0
Xt¼n X
t¼n
þ jjminf0; vrmin vrt gjj22 þ jjminf0; vrmax - vrt gjj22
t¼0 t¼0
X
t¼n X
t¼n
Fap ¼ jjminf0; almin alt gjj22 þ jjminf0; almax alt gjj22
t¼0 t¼0
X
t¼n Xt¼n
þ jjminf0; armin art gjj22 þ jjminf0; armax - art gjj22
t¼0 t¼0
ð18Þ
X
t;j
VðsÞ ¼ e2t;j ðsÞ ð19Þ
110 J. Zhang et al.
where j is the j summation of VðsÞ. The first-order Taylor expansion of et;j ðsÞ at
^
S ¼ fst jt ¼ 1; 2; . . .; ng is performed:
^ ^ ^
et;j ðs þ DsÞ et;j ðs Þ þ Jt;j Ds S ¼ fst jt ¼ 1; 2; . . .; ng ð20Þ
^
where Jt;j is the Jacobian matrix of et;j ðsÞ at, and e2t;j ðs þ DsÞ is
^ ^ ^
e2t;j ðs þ DsÞ ¼ et;j ðs þ DsÞT et;j ðs þ DsÞ
^ ^
ðet;j ðs Þ þ Jt;j DsÞT ðet;j ðs Þ þ Jt;j DsÞ
^ ^ ^
¼ et;j ðs ÞT et;j ðs Þ þ 2 et;j ðs ÞT Jt;j Ds þ DsT Jt;j
T
Jt;j Ds ð21Þ
|fflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflffl{zfflfflfflfflfflffl} |ffl{zffl}
ct;j bt;j Hi;j
^
The approximate value of Vðs þ DsÞ is obtained as follow.
^
X
t¼n;j¼2
^
X
t¼n;j¼9
^
Vðs þ DsÞ ¼ e2t;j ðs þ DsÞ þ e2t;j ðs þ DsÞ
t¼p;j¼1 t¼0;j¼0
X
t¼n;j¼2 X
t¼n;j¼9
ct;j þ 2bt;j Ds þ DsT Hi;j Ds þ ct;j þ 2bt;j Ds þ DsT Hi;j Ds
t¼p;j¼1 t¼0;j¼0
¼ c þ 2bDs þ Ds HDs T
ð22Þ
P P P ^
where c ¼ ct;j , b ¼ bt;j , and H ¼ Ht;j . Thus, the derivative of Vðs þ DsÞ is 0,
and we can obtain the following linear equation:
HDs ¼ b ð23Þ
^
To obtain the Ds that corresponds to the minimum Vðs þ DsÞ, the minimum
optimal solution of approximate V (s) is obtained:
s ¼ s þ Ds
^
ð24Þ
The Gauss-Newton algorithm updates the H matrix and the b matrix in each iter-
ation. The iteration variation Ds is obtained from Eq. (21) and the optimal solution s is
updated from Eq. (22). In each iteration, the optimal solution s is considered to be the
^
initial solution S ¼ fst jt ¼ 1; 2; . . .; ng in the last iteration until the given termination
condition is satisfied.
Dynamic Path Planning Algorithm Based on an Optimization Model 111
In the Levenberg-Marquart algorithm which can effectively reduce the problem that
the simple Gauss-Newton method may not converge [14], the damping coefficient is
added to place Ds within a certain range, which prevents Ds from being too large to cause
inaccurate approximate results and affects the convergence effect. Unlike Eq. (20), the
Levenberg-Marquart algorithm solves a linear equation with a damping coefficient.
ðH þ kIÞDs ¼ b ð25Þ
where k is a damping coefficient. The larger k, the smaller Ds . This is very useful for
controlling step size. The idea of the Levenberg-Marquart algorithm is to dynamically
control the damping factor. In each iteration, we compare the difference between the
approximate model and the actual function. If the difference is small, we reduce k.
Otherwise, we increase k to re-solve until Ds is sufficiently small. The initial value of
the damping coefficient k is closely related to the state of the coefficient matrix H [14,
15], which can be expressed as
k ¼ s maxfðHÞii g; i ¼ 0; . . .; n 1 ð26Þ
3 Experiment
ship represent the origin of the world coordinate system (0, 0), and the destination
coordinates are (0,230). Specify the Y axis direction b ¼ 0. Three obstacles
Dk ¼ ½xk ; yk ; vk ; hk ; dk;t T are inserted in the clockwise direction: D1 ¼ ½50; 230; 3:5;
215; 20T , D2 ¼ ½0; 110; 0; 0; 20T , D3 ¼ ½30; 60; 3; 255; 20T , and dk ¼ 15(k = 1,2,3).
This paper’s algorithm is shown in Fig. 3. While A* algorithm is shown in Fig. 4. The
blue point is the output locus which start from left and end from right. Circles are
obstacles. Arrow indicates the obstacles direction and speed. The A* algorithm
assumes that the obstacle is stationary; thus, the resulting path will coincide with the
obstacle. The algorithm proposed in this paper considers the location of the obstacle at
multiple times and the kinematics constraints of the USV. Thus, a better degree of
fitting is achieved.
Comparison Test 3. Path planning based on Morphin [7] is shown in Fig. 6, while the
effect of our algorithm in the same environment is shown in Fig. 7. The environment is
same as above test. Morphin algorithm can quickly calculate the path, because its
model is simple, the computational amount is very small, and it can satisfy the real-time
(100 Hz can be achieved in actual navigation). However, the dynamic nature of
Dynamic Path Planning Algorithm Based on an Optimization Model 113
Fig. 7. (a) Effect of this paper’s algorithm (flat) (b, c) this paper’s algorithm (space) (Green line:
path. Start: left. End: right. Arrow direction & length: obstacles direction & speed)
obstacles is also not taken into account, and the planned path cannot be dynamic, so the
path accept from Morphin algorithm is longer than this paper’s algorithm. On the other
side, The path obtained by Morphin algorithm doesn’t connect start point and end
point. This may result in a path that appears to be circumnavigated on a larger scale.
Fig. 8. Real experiment (requires 0.2 s). Blue line: planned path; Hollow triangle: dynamic
obstacle
4 Conclusion
Compared with the three algorithms, we find that traditional path planning algorithm
assumes that all obstacles are stationary. In Marine applications, USVs and dynamic
obstacles usually have large inertia, that meanings changing their state is slow and
difficult. If only the current position of the obstacle is considered for path planning, the
dynamic obstacle may move to the planned path at the next moment, A collision may
occur. Or there is a shorter path because of obstacle’s moving away but not find. On the
other hand, planning should consider the USV dynamic constraints, such as the speed
and acceleration limit. Otherwise, USVs may not fit the trajectory. The experimental
results show that this algorithm which considers time dimension and USV dynamic
constraints has more advantages.
114 J. Zhang et al.
References
1. Motwani, A.: A survey of uninhabited surface vehicles. Marine and Industrial Dynamic
Analysis, School of Marine Science and Engineering, Plymouth University (2012)
2. Mooney, D.: Metric path planning. robotics notes (2009). http://clipper.ship.edu/djmoon/
robotics/robotics-notes.html
3. Yoo, S.-J., Park, J.-H., Kim, S.-H., Shrestha, A.: Flying path optimization in UAV-assisted
IoT sensor networks. ICT Exp. 2(3), 140–144 (2016)
4. Wang, Z., Zlatanova, S., Oosterom, P.V.: Path planning for first responders in the presence
of moving obstacles with uncertain boundaries. IEEE Trans. Intell. Transp. 99, 1–11 (2017)
5. Zuo, L., Guo, Q., Xu, X., Fu, H.: A hierarchical path planning approach based on a* and
least-squares policy iteration for mobile USVs. Neurocomputing 170(C), 257–266 (2015)
6. Shum, A., Morris, K., Khajepour, A.: Direction-dependent optimal path planning for
autonomous vehicles. USV. Autonom. Syst. 70, 202–214 (2015)
7. Simmons, R., Henriksen, L., Chrisman, L.: Obstacle avoidance and safeguarding for a lunar
rover. Proc AIAA Forum Adv. Develop. Space Robot. (1996)
8. Rösmann, C., et al.: Integrated online trajectory planning and optimization in distinctive
topologies. USVics and Autonomous Systems (2016)
9. Rösmann, C., et al.: Timed-elastic-bands for time optimal point-to-point nonlinear model
predictive control. European Control Conference (ECC), pp. 3352–3357 (2015)
10. Wu, B., Wen, Y., Huang, Y., Zhu, M.: Research of Unmanned Surface Vessel (USV) path
planning algorithm based on ArcGIS. In: Proceedings of ICTIS 2013, pp. 2125–2134. ASCE
(2013)
11. Phung, M.D., Cong, H.Q., Dinh, T.H., Ha, Q.: Enhanced discrete particle swarm
optimization path planning for UAV vision-based surface inspection. Autom. Constr. 81,
25–33 (2017)
12. Cui, R., Li, Y., Yan, W.: Mutual information-based multi-AUV path planning for scalar field
sampling using multidimensional RRT*. IEEE TSMC. 46(7), 993–1004 (2016)
13. Devaurs, D., Simon, T., Corts, J.: Optimal path planning in complex cost spaces with
sampling-based algorithms. IEEE Trans. Autom. Sci. Eng. 13(2), 415–424 (2016)
14. Madsen, K., Nielsen, H.B., Tingleff, O.: Optimization with Constraints. 2nd Edn. (2004)
15. Madsen, K., Nielsen, H.B., Tingleff, O.: Methods for Non-Linear Least Squares Problems.
2nd Edn. (2004)
16. Kummerle, R., Grisetti, G,, et al.: G2o: a general framework for graph optimization. In:
IEEE International Conference on Robotics and Automation (ICRA), pp. 3607–3613 (2011)
Research on Autonomous Health Management
and Reconstruction Technology of Satellite
1 Introduction
fault diagnosis of complex systems, which attracted great attention from the aviation
and spaceflight field including National Aeronautics and Space Administration
(NASA), Jet Propulsion Laboratory (JPL), Ames Research Center, DLR and European
Space Agency (ESA), etc. Meanwhile, early applications mainly include various types
of satellites, autonomous spacecraft and landers, such as Deep Space series, the
Phoenix lander, etc. As a result, Europe and America are far ahead in the field of
spacecraft management and fault diagnosis.
Generally, the development of foreign spacecraft autonomous health management
technology can be divided into five stages: reliability analysis, fault diagnostics and
prognostics, integrated diagnostics and system monitoring, integrated system fault
prediction and health management. Its origins can date back to the birth of reliability
theory, environmental testing and system testing, and quality methods in the 1950s and
1960s. In the 1970s, methods for diagnosing sources and causes of failures emerged,
and the development of prediction and health management technologies was beginning
to emerge. Integrated diagnostics technique became popular in the field of space and
aeronautics in U.S. in 1980s. Inspired by this, NASA proposed the concept of
flight vehicle health monitoring on the 1990s, and ultimately proposed the concept of
Integrated Systems Health Management (ISHM), which can support the high perfor-
mance requirements of the new generation of Reusable Launch Vehicles (RLVs), in
2000s.
One of typical ISHM method is Complex System Integrated Health Management
(CSIHM). CSIHM has entered the engineering practice stage and has been fully
applied and verified in projects such as X-33, X34 and X-37 in present. Among these
projects, many organization and companies, such as NASA, Lockheed Martin, pro-
posed two specific techniques: Integrated Vehicle Health Management (IVHM) system,
and Prognostic and Health Management (PHM) system [1–4]. Meanwhile, with the
development of IVHM, many diagnostic reasoning tools have emerged, including
Livingstone, BEAM and SHINE, etc.
Compared with U.S., Europe has also conducted in-depth research on spacecraft
fault diagnosis technology, and developed practical fault diagnosis systems. Many
scientists have researched and developed a variety of knowledge-based fault man-
agement and fault diagnosis systems for different tasks, such as environmental control
and life support system (ECLSSD) for Columbus module.
Russia also has advanced technology in fault diagnosis technology. Boris Katorgin
et al. developed the health monitoring and life prediction system for cryogenic
liquid-propellant rocket engines. Georgy Vasilchenko et al. developed track real-time
monitoring system that provides visualization information for astronauts of Buran
spacecraft.
In conclusion, the development of ISHM is still in its infancy, and a system with
full functionality of ISHM does not exist yet.
Research on Autonomous Health Management and Reconstruction Technology 117
The self-management of the existing spacecraft in China is mainly based on the data
management system that can only take some emergency measures for the failure, which
is a safety-oriented design. Common ones include autonomous energy management,
autonomous thermal control, and measurement and control component management.
The method adopted is to collect the security status information of the entire satellite
system by using the data management subsystem or the service subsystem. According
to the preset control strategy, it will controls and manages the key equipment that
affects the safety of the whole satellite system. The specific method is to design a safety
threshold for some key telemetry parameters. The software analyzes the telemetry
parameters collected by the software and compares them with the corresponding
thresholds. If the value exceeds the threshold, the fault is considered to occur, and the
predetermined safety instruction sequence is executed autonomously, such as close the
payload.
By analyzing the requirements of autonomous health management of satellites such
as the moth series satellites, the Mars Explorer Yinghuo-1 and the small satellite Trial
5, we summarizes the principles of satellite safety mode design in China:
(1) The system carries out safety mode management independently.
(2) Detect and identify faults as quickly as possible, and take necessary measures to
isolate the faults in order to prevent the expansion of the scope.
(3) To partly carry out reconfiguration operation of the system to ensure the con-
tinuous operation of the task.
Generally, each model mainly carries out the design of security mode control
strategy aiming at energy conservation, including the central computer security mode,
attitude security mode, power security mode, BAPTA blocking and switching safety
mode, bus communication abnormal security mode and so on. Specific examples are
shown as follow:
(1) Central computer security mode: satellite due to loss of programmable area
instructions when the central computer is reset or cut off. The working state of the
load is unknown. If the short-term payload equipment can’t shut down after
starting, and the short-term work becomes long-term work, it will inevitably affect
the safety of the whole satellite power supply.
(2) State Safety Mode: if the satellite’s attitude is abnormal, or the satellite can’t
imagine normally, or the satellite can’t face the sun normally, the satellite may
lose energy, so it is necessary to shut down the payload equipment to keep the
satellite in a stable low-power working state.
(3) Power Safety Mode: when the power supply system is extremely difficult, it will
affect the normal operation of the satellite, and be incapable of working. So it is
necessary to remove unnecessary loads as much as possible, so that the whole
satellite is in the lowest power consumption mode which can ensure the normal
TT&C of the satellite.
118 L. Yang et al.
Although the above measures have been verified on-orbit, it is sure that all fault
diagnosis and treatment are based on a small number of measured parameters moni-
toring and threshold judgment. Moreover, the fault type is also known in advance.
Thus, a fault diagnosis method with comprehensiveness and depth are yet to be solved.
In addition, in order to improve the on-orbit operation management ability of
spacecraft, the on-orbit fault diagnosis method of spacecraft is studied. And the fault
diagnosis system is established to improve the ability of diagnosis and location when
the system fails and improve the efficiency of operation management.
Through the analysis of the above investigation and comparison of the domestic and
foreign research, and application of autonomous health management technology, it is
not difficult to find that there is a big gap between China and foreign countries. The
domestic gap is mainly reflected in the following aspects:
(1) The research work started late, the theoretical basis is weak and engineering
experience is few.
(2) The existing satellite health management methods mainly rely on the ground
system, which is not conducive to timely and effective handing of unexpected
problems and emergency faults in orbit, and it’s liable to cause the expansion of
on-orbit faults, or even the risk of the failure of the entire satellite.
(3) Ground is usually monitored by the traditional way of edge measurement and
alarm. Fault diagnosis relies on manual work. Satellite long tube has many items
of manual monitoring, so it is impossible to carry out on-orbit fault early warning
and fast fault diagnosis. Generally speaking, the rescue level after the failure does
not really involve autonomous health management and fault prediction.
The existing satellite fault diagnosis expert system is basically in the ground test
type, and there are few practical applications; even if there are engineering applications,
the real-time effect of autonomous fault diagnosis and processing can’t be achieved.
The health management of satellite can be divided into two parts. The first part focuses
on the health management of the design, testing and verification process. Its purpose is
to ensure that the delivered products meets the technical requirements and works
properly. Therefore, we need to find out the problems and hidden dangers in the system
through testing and verification. The second part focuses on the health management of
on-orbit satellites. By analyzing the telemetric data and observed data of on-orbit
satellites, we can not only evaluate the health status of the system, but also discover and
even predict fault of the system [5, 6].
Research on Autonomous Health Management and Reconstruction Technology 119
Based on the concept above, autonomous health management and system recon-
struction is divided into several levels:
Status Detection: Through effective technical methods, various parameters closely
related to the working state of each component of the measured object are detected.
Then, a multi-parameters model of the operating state of the system is established. On
this basis, it is possible to further analyze the development and trend of the working
state of each component based on the detected data, and make valuable judgments to
diagnose the fault that has occurred, and even predict the impending fault.
Fault Diagnosis: By analyzing the characteristic signals reflecting the state of the
system and the information recorded in the system log under a specific working
environment, we can identify whether the system is in working order, or whether the
system structure is degraded or deteriorated. Furthermore, we can isolate and locate the
fault based on the relationship between system failure and actual reflection.
Failure Prediction: By pre-detecting the state of the system, or detecting the state
when the system completes the task, we can get performance indicators that reflect the
normal operation of the system. Then, we can provide an early warning to the system
based on the existing records.
Health Assessment and Generates Decision: Because it is impossible to complete
equipment replacement and maintenance for on-orbit satellites, health assessment
method for on-orbit satellite systems is completely different from the classical method.
Thus, we need to conduct a comprehensive analysis based on the existing data, and use
a variety of methods to evaluate the health of the system, and finally give reasons and
suggestions for equipment or sub-system with faults or troubles. Moreover, we can
estimate the life of the equipment to provide reference for task decision.
In conclusion, the autonomous health management and system reconstruction
means of satellite is a method apply in the period of design, verification and on-orbit
operation. It can accurately detect the fault information and identify the cause, location,
type and degree of the fault. Then, we can take corresponding measures to isolate,
reconstruct and recover the fault [7]. By this way, the risk of system failure will be
reduced. Even the satellite is working properly, and the reliability of the satellite is
improved. Furthermore, by analyzing the system logs, sensor information and envi-
ronmental information, we can predict faults of satellites and accurately assess the
health of the satellites so that we can detect system performance degradation and faults
in time. Finally, we can provide assistance for management works of on-orbit satellites,
and even extend the working life of satellites.
In fact, the autonomous health management and system reconstruction means are
committed to solve the following problems:
(1) Fault detection and diagnosis: It can detect the faults of satellites in real-time, and
quickly diagnose and locate the faults.
(2) Real-time fault self-management: According to the specific faults that being
detected and diagnosed, evaluating its severity and development trend, and give
timely countermeasures for the fault. Moreover, completing the system recon-
struction autonomously according to the available resources and task
120 L. Yang et al.
Therefore, there are a variety of methods to meet the above needs. On one hand, we
can establish an automatic judgment system for in-obit satellite status, quickly and
intuitively display the working status of the on-orbit satellite, detect hidden dangers in
the early stage of fault indication. On the other hand, we can establish the system which
can quickly locate and provide correct fault countermeasures for satellite design and
maintenance personnel when the fault occurs. Both of them are important methods to
ensure the healthy and reliable operation and extend the life of on-orbit satellites.
In addition, the development of military satellites in the future will gradually be
transformed from satellites used for research to satellites with equipment. In this case,
requirements of rapid satellite development, rapid equipment and reduced development
costs, will demand better results on the comprehensive support links for satellite
development. From the existing satellite development process we can find that, on the
one hand, due to the satellite product design and testing are completely separated into
independent links, resulting in complex ground testing, system fault location, and
troubleshooting, the entire testing process takes up a lot of satellite development time.
On the other hand, stand-alone test, sub-system test and whole-satellite test are inde-
pendent. However, none of the test support equipment is uniform, and the models
cannot be universal. At the same time, the product has no self-test capability and
completely relies on external test equipment, resulting in huge cost of test equipment.
For the above problems, the development requirements of satellite autonomous
health management technologies mainly include the following aspects:
(1) The requirements of task continuity and success: In order to meet the application
requirements, the satellite needs to ensure the timeliness, accuracy and continuity
of the system work firstly. By this way, the satellite will have the abilities of
timely discover faults, realize redundancy management, and improve the relia-
bility and success of the mission. Secondly, the satellite could detect concealed
faults, and then notify the upper-level system in time to take measures to avoid
occurrence of them and ultimately improve the reliability and security of the
mission. Thirdly, the satellite can record information when the state of system has
changed. Then it should analyze and predict the fault trend, and remind the
ground station to take preventive measures. Finally, through system fault detec-
tion, isolation and autonomy, the satellite will have the ability to build and
enhance the continuity of the task.
(2) The requirements of reliability and security: Due to the increased complexity and
design life of the satellite, the environmental effects of the space will cause
degradation of the performance parameters of the satellite-related components,
resulting in abnormal and faults in the satellite. If anomalies and faults for on-orbit
satellite cannot be detected in time, or corresponding control measures are taken,
it will bring serious consequences to safety of the satellite, and even invalidate the
entire satellite.
(3) Timeliness requirements for troubleshooting: With the increase of new types of
satellites, the existing management mode and fault handling methods cannot be
used for on-orbit failure warning and rapid fault diagnosis. The processing time of
satellite anomalies is too long, which may cause faults to expand and even cause
catastrophic consequences.
122 L. Yang et al.
(4) Further improvement of fault prediction capability: Through fault prediction and
health management technologies, the focus of satellite operation management will
shift forward, and give a warning of potential damage timely. Through the
analysis of satellite itself, or through the control of the ground station, hidden
troubles for the satellite can be avoided timely. Further, it can even avoid unre-
coverable damage that will affect system reliability.
(5) The requirements of usability evaluation for the system: Satellite users need to
have a comprehensive of the health and life of satellites when establish project or
performance evaluation. By this way, users can acquire an effective assessment of
the availability of satellite systems. This is also one of the research priorities of
failure prediction and health management.
(6) Improve test efficiency and reduce test costs: By analysis the test design
technology of the system and improving the test design level of the system, the
real-time detection method during system operation can quickly find faults, locate,
improve test efficiency, reduce system test time, and reduce external test equip-
ment and manpower requirements. Reducing costs will help the future develop-
ment of military satellites gradually transform from satellites of research to
satellites with equipment.
6 Conclusion
Autonomous health management capability is one of the capabilities that satellites need
to focus on development and improvement in the future. It is an important guarantee for
the safe and stable operation of on-orbit satellites. This paper investigates the current
research of autonomous health management technologies in the aerospace field at home
and abroad, and analyzes the gaps at home and abroad. Aiming at the needs of the
future satellite intelligent autonomous, we analysis the task requirements of health
management in the two stages of satellite development and satellite operation in detail.
By this way, we finally provide a guideline for the in-depth study of satellite intelligent
self-management.
References
1. Chang, Q., Yuan, S.: Overview of integrated vehicle health management (IVHM) technology
and development. Syst. Eng. Electron. 31(11), 2652–2657 (2009)
2. Datta, K., Jize, N., et al.: An IVHM systems analysis & optimization process. In: Aerospace
Conference, pp. 3706–3716 (2004)
3. Shi, W., Sun, Y., Wang, Z., et al.: A study of PHM system and its fault forecasting model.
Fire Control Command Control 34(1), 29–35 (2009)
4. Zeng, S., Michael, G., Wu, J., et al.: Status and perspectives of prognostics and health
management technologies. Acta Aeronautica ET Astronautica Sinica 26(5), 626–632 (2005)
Research on Autonomous Health Management and Reconstruction Technology 123
5. Gordon, B., Patterson, A.A.: Review of system health state determination methods. In: First
Space Exploration Conference: Continuing the Voyage of Discovery, pp. 1–13 (2005)
6. Ke, L., Kaiwen, D., Rui, D., et al.: Autonomous on-orbit health management architecture and
key technologies for manned spacecrafts. Manned Spaceflight 20(2), 116–121 (2014)
7. Felke, T., Gadden, G., Miller, D., et al.: Architectures for integrated vehicle health
management. In: IEEE Aerospace Conference 2010, Atlanta, Georgia, pp. 20–22 (2010)
The Design and Implementation
of Digital Satellite Simulator
Abstract. Digital Satellite Simulator (DSS) is a system which can realize high
precision simulation of each subsystem on board, verify and drill the flight
process in orbit. It can generate the original Telemetry (TM) frames according
the frame format of the real satellite, execute and respond the Telecommand
(TC) correctly, provide the real TM, TC, and external data interface for the
TT&C ground test system. DSS can be used to verify the correctness of the
TT&C ground test system hardware function, verify the correctness of TC
sequence which is used for mission planning in orbit, verify the correctness of
attitude in orbit and orbital maneuver, verify the authenticity and reliability of
the software operating environment on board, provide a real training and
operating platform for the satellite designers and operators, this platform can
help users to deal with the FDIR in orbit much more better in the exercise.
Function, structure, and realization method of DSS are introduced in this doc-
ument from the angle of software engineering. Generality and mass production
requirement has been considered in this design, the product development mode
has been formed. This DSS has been applied in many projects such as VRSS-1,
VRSS-2, PRSS-1 and so on. Some suggestions are also proposed for the DSS
follow-up technological development.
1 Foreword
With the development of the spacecraft, more and more spacecraft development
company have begun to pay attention to the application and technology development of
satellite simulator. The subsystem lever simulation platform could be built by the
simulation of units. The system lever simulation could be built by the simulation of
subsystems. Unit equipments could be connected to subsystem lever simulation plat-
form for subsystem lever design and function testing, subsystem equipments could be
connected to system lever platform for system lever design and function testing in the
future. In this way, product design defects can be exposed as early as possible in the
product development stage, it can also reduce the development cost of product
re-upgrade caused by functional abnormality after product delivery [1] and provide a
more realistic environment for the product joint testing. Compared with the conditions
for the expensive spacecraft equipment joint testing, the satellite simulator will greatly
reduce the cost of product development, improve the efficiency and shorten the
development cycle [2].
The satellite simulator can also realize the simulation and verification of in-orbit
flight mission, failure mode and effects analysis, as well as the functional verification of
ground control system, providing a more authentic training platform for in-orbit
operation [3].
2 System Design
CAN Bus
( Protocol)
PSS AOCS DTS PLD
PCU PDU AOCC ALTU DTS Camera
Lower Lower Lower Lower Lower Lower
computer computer computer computer computer computer
LAN ( Protocol )
Track
Monitoring
SCC 3D Tool Maintenance Simulate Model
Console Tool Database
3.4 3D Monitoring
3D monitoring displays various working modes and the flight attitude of the satellite in
3D mode. Functions of the 3D monitoring software include:
(1) To support online and offline mode.
(2) To support satellite perspective, earth perspective, sun perspective. To show the
relative position of the sun, earth and satellite.
(3) To display orbit track and the subastral points.
(4) To display working modes of the satellite.
(5) To display orbit maneuver of the satellite.
(6) To display satellite and ground station transmission.
5 Conclusion
Digital Satellite Simulator can realize high precision simulation of each subsystem,
help users to conduct in-orbit operation and training. System testing and equipment
acceptance test can be performed in the virtual simulation environment. This helps
designers expose the design flaws on the onboard software as early as possible.
References
1. Zheng, A., Zhou, J., Ma, Y.: Development of Chang’E-1 operational simulator. J. Beijing
Univ. Aeronaut. Astronaut. 39(1), 57–61 (2013)
2. Aghili, F., Namvar, M., Vukovich, G.: Satellite simulator with a hydraulic manipulator. In:
Proceedings of the 2006 IEEE International Conference on Robotics and Automation,
pp. 3886–3892 (2006)
3. Yuan, X.: Key research programs of LEO communication Microsat Constellation. In: CAS
Shanghai Institute of Microsystem and Information Technology (2002)
4. Chen, Q., Yu, S., Guan, H.: Design and implementation of satellite failure simulation system
based on satellite simulator. J. Acad. Equip. Command Technol. 22(5), 63–67 (2011)
5. Yang, H., Zhang, F., Liu, J.: Design of digital simulator platform for satellite’s in orbit faults.
J. Spacecr. Eng. 21(4), 21–25 (2012)
6. Wu, Z., Xiong, X., Song, Q.: Object-oriented Design of Satellite Simulators. J. Spacecr.
TT&C Technol. 27(2), 13–16 (2008)
7. Li, X., Xu, X., Chen, J.: Design of generic infrastructure for telecommunication satellite
simulators. J. Spacecr. TT&C Technol. 30(3), 1–5 (2011)
8. Cui, W., Yang, H., Yang, J.: Design and implementation of simulation test platform for
satellite subsystems. Comput. Measur. Control 23(10), 3264–3266 (2015)
9. Zhao, G., Zheng, B., Liang, J.: Design of universal simulator for small satellite. J. Sci.
Instrum. 30(6), 357–360 (2009)
10. He, Z., Xu, X., Zhao, Q.: Design of thermal control subsystem for flying satellite simulator.
Journal of Spacecraft Engineering 20(1), 82–87 (2011)
Research on Target Recognition
Technology of Satellite Remote Sensing
Image Based on Neural Network
1 Introduction
Remote Sensing technology is an important means for human beings to carry out
scientific investigation of their living environment. In 1972, the United States launched
its first earth resource satellite Landsat-1, since then, remote sensing technology has
become a new and high technology developed by major powers in the world. Remote
sensing satellites with multi-source sensors can obtain the spatial and physical infor-
mation of various targets in the earth’s sphere, such as vegetation, lakes, oceans and
atmosphere, and present them in the form of two-dimensional image. Geological
information contained in remote sensing images can be used in land resources survey,
urban development space monitoring, disaster assessment and early warning and other
fields, which are closely related to the development of national economy.
With the rapid development of remote sensing technology, the number of remote
sensing satellites is increasing, the resolution of remote sensing satellite image is
getting higher and higher, and the image coverage is getting wider and wider. Remote
sensing images show the development trend of “big data”, single satellite daily data
increased at TB level, national data center archive data reaches PB level. Relying
entirely on manual translation of images requires a lot of manpower and material
resources, and the timeliness is not good, this can not meet the requirements of con-
tinuous upgrading applications. This brings urgent demand for automatic processing
and intelligent application of massive remote sensing images.
In recent years of research, deep learning method has achieved remarkable results in
target recognition of natural images. Especially based on Deep Convolution Neural
Network, the target feature can be automatically extracted and analyzed, and the target can
be accurately identified. The remote sensing image translation framework constructed by
convolution neural network can obtain the type information and location information of
the target. Therefore, this depth learning translation method has a wide range of appli-
cation prospects in the task of satellite remote sensing image translation [1, 2].
This kind of translation method can extract the features of the target independently
and form an “end-to-end” remote sensing image processing system, which improves
the intelligence of remote sensing image processing. It can not only save a lot of
manpower in the complicated remote sensing image interpretation work, but also
effectively improve the processing speed of optical remote sensing image, while
ensuring the timeliness and accuracy of information acquisition.
In this paper, a method of remote sensing image target recognition based on depth
neural network is proposed. Typical frame Faster R-CNN is used as the basic frame,
and the image augmentation is used to improve the accuracy and generalization ability
of the neural network model. The proposed method is verified by ship recognition and
playground recognition cases. The results show that the proposed method can be used
in remote sensing image recognition. Under high speed and automatic processing, the
accuracy of target recognition is better than 75%.
In recent years, with the rapid development of in-depth learning technology, it has been
paid more and more attention on speech signal processing, image processing, pattern
recognition and other fields. Its research value and application value have been fully
affirmed by academia and industry. In deep learning, Convolutional Neural Network is
a feed-forward neural network, its artificial neurons can respond to some of the sensory
units within the converge, especially suitable for large-scale image processing. CNN
includes input layer, convolutional layer, pooling layer, activation layer and other
important network units.
• Normalization: the variance is normalized to the same range, reducing the inter-
ference caused by differences in the range of values of each dimension.
• PCA/Albefaction: using PCA to achieve dimensionality reduction; albefaction is
normalization of amplitude on each characteristic axis of data.
The effect diagram of mean value and normalization are shown in Fig. 1.
Convolution Layer
Digital image is a two-dimensional discrete signal. Convolution operation for digital
image is to use convolution kernel (convolution template) to slide on the image,
multiply the pixel gray value to the image point with the corresponding convolution
kernel value, then add all the multiplied values as the gray value to the pixel in the
image corresponding to the middle pixel of the convolution kernel, and finally slide all
the image process.
Image convolution is shown in Fig. 2.
Pooling Layer
The Pooling layer follows the convolution layer, which is used to compress data and
parameters and reduce over fitting. The pooling layer has the following functions:
(1) Feature invariant; (2) Feature dimensionality reductional; (3) Avoiding over fitting.
As pooling processing has the function of feature aggregation, the introduction of
pooling layer can reduce the risk of network falling into over-fitting and ensure that
neural network has strong generalization ability.
The working instruction of pooling layer is shown in Fig. 3.
Activation Layer
The role of the activation layer is to incorporate non-linear factors, because the linear
model is not good at expressing practical problems. In convolution neural network,
when convolution filtering is performed on an image, each pixel is assigned a weight,
this assignment is linear. However, in practical applications, the target characteristics
are not necessarily linearly separable. Therefore, the nonlinear model can be simulated
better by introducing an activation layer into convolution neural network and utilizing
the nonlinear factors produced by the activation layer.
Commonly used activation functions should be nonlinear and continuous differ-
entiable. In addition, they should have the following properties:
• Larger unsaturated zone: when there is a saturated interval, if the system opti-
mization enters the interval and the gradient approximates to zero, the learning of
the network will stop.
• Monotonicity: when the activation function is monotonic, the error function of the
single-layer neural network is convex, which is beneficial to the optimization
process.
• Approximately linear at the origin: when the weight is initialized to a random
number near zero, the gradient of the activation function is large in the approximate
linear interval, which can speed up the iteration of the network.
According to the above characteristics, the commonly used activation functions are:
Sigmoid function, Tanh function, ReLU function and etc.
Research on Target Recognition Technology 135
1 X
yðxÞ aL ð xÞ2
C¼ ð1Þ
2n x
Among them, C means cost function, n means total sample size, x means sample,
y means actual value, a means input value.
Cross entropy loss is expressed as:
1X
C¼ ½y ln a þ ð1 yÞ lnð1 aÞ ð2Þ
n n
The higher the value of the cost function, the faster the network weight is adjusted,
and the training efficiency is much higher than that of the square error function.
Remote sensing image automatic target recognition technology has great research value
and application value [5, 6]. For example, in military applications, automatic identi-
fication of strategic objectives such as aircraft, ships and oil depots is one of the key
link to improve military reconnaissance capability. On the civilian side, automatic
identification and statistics of specific targets such as stadiums, supermarkets and
bowling alleys will greatly improve the level of urban regulation.
136 Q. Zhang et al.
ðx; yÞ ! ðW x; yÞ ð3Þ
4 Recognition Result
Two cases of ship target recognition and motion field recognition are used to verify the
performance of the proposed method based on neural network for remote sensing
image target recognition.
5 Conclusion
References
1. Yu, D., Deng, L.: Deep learning and its applications to signal and information processing.
Signal Process. Mag. 28(1), 145–154 (2011)
2. Lu, C., Tang, X.: Surpassing human-level face verification performance on LFW with
GaussianFace. CoRR: abs/1404.3840 (2014)
3. Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional
neural networks. Adv. Neural Inf. Process. Syst. Stateline: NIPS 2012, 1097–1105 (2012)
4. Girshick, R., Donahue, J., Darrell, T., et al.: Rich feature hierarchies for accurate object
detection and semantic segmentation. In: Proceedings of the IEEE Conference oil Computer
Vision and Pattern Recognition, pp. 580–587. IEEE Press, Piscataway (2014)
5. Chen, T.Y., Chen, J.Y., Zhao, H.P.: Ecognition-based ship detection on optical remote
sensing images. Radio Eng. 43(11), 11–13 (2013). (in Chinese)
6. Wang, Y.Q., Ma, L., Tian, Y.: State-of-the–art of ship detection and recognition in optical
remotely sensed imagery. Acta Automatica Sinica 37(9), 1029–1039 (2011). (in Chinese)
Research and Implementation
of Automatic Test Technology
for Power Modules in Aerospace
Abstract. The increasing variety and quantity of power modules for aerospace
causes difficulty and complication for their testing. This paper studies the testing
technology, develop the testing software and design with hardware universal
interface to realize an automatic testing technology with data collection and
interpretation, waveform automatic storage, command automatic switch, output
channel automatic switch. The technology not only increases the testing efficacy
by more than 100%, but also the risk factors caused by manual operation during
the test are eliminated and the reliability and safety of the test system are
improved. The technology is suitable for testing all current power modules.
1 Introduction
In the field of aerospace, the aerospace power module is responsible for providing
stable voltage and current for subsystems of aerospace equipment such as computer
systems, navigation systems, and communication systems. As a basic component, it is
used in a variety of aerospace projects and is directly connected to electrical equipment.
Its quality and reliability are one of the basic conditions for ensuring the completion of
space missions. According to statistics, 60 to 100 power modules are generally required
on a common satellite. In recent years, with the in-depth development of manned
spaceflight engineering and deep space exploration technology, the wide application of
distributed power supply and the increase of payload, the types of power required by
aircraft are more and more complicated, and the requirements for performance indi-
cators of various aspects of power supply are also stricter. Therefore, the test of the
electrical performance indicators of the power module is more arduous. Through the
research of power module test technology, a fully automated one-button test technol-
ogy is provided to replace the traditional manual operation mode, which can eliminate
the wiring and measurement errors caused by human factors, improve test efficiency
and save labor cost.
2 System Design
System plan diagram as shown below, the system is mainly composed of three parts:
the instrument required for testing, the hardware interface and the test program. The test
system uses the GPIB interface card to closely link several program-controlled test
equipment with the industrial computer and realizes the operation and control of the
instrument through the industrial computer. During the test, the tester connects the
equipment and cables, operates the test program, starts the test project, the system will
automatically execute the test, and automatically save the test results (see Fig. 1).
Secondary
Hardware Software test
power supply
equipments itmes
for test
DC regulated
Efficiency
power supply
and stability
Device test
initialization
Pull off and
Electronic load
on/off test
Equipment
calibration
Input ripple
Oscilloscope surge
6014A
Power Function
module test
34922 Board
Telemetry
impedance
Data collector
34921 Board
34980A Waveform
storage Output
34938 Board dynamic
characters
Command Data
interpretation Output
power overcurrent
&
Overcurrent
Oscilloscope recovery
Industrial
3032A
GPIB to USB computer
3 Industrial Computer
By setting the power supply overvoltage protection function, the subsequent control
voltage can be limited to the voltage range allowed by the tested product, preventing
the power supply output from overvoltage due to software programming errors and
manual operation errors during the test.
Input Overcurrent Protection Design: The positive and negative reverse connection
of the power supply is the most likely mis-operation of the power supply product
testing process, which is a low chromatographic quality problem. In order to prevent
the reverse connection operation, it is necessary to design the power-on program. The
power-on mode is to set the load to no load, set the input current to 0.2A (according to
the actual no-load power consumption of the product), and then the input voltage starts
from 0 V. The 1 V step is continuously energized, so that even if there is a reverse
connection operation, the primary power supply will over-current protection when the
input voltage is low, which will not cause damage to the product and improve the
reliability of the test system.
7.8 8.0
≥80
≤100
1.2 times
9.48V
12.75
≤10
≤20
≤0.38
≤0.38
≤5
0 5
≤5
≤500
Comparison with before and after using the test system (see Fig. 3)
5 Summary
In summary, using the automatic test system to complete the test work can reduce the
labor intensity and save the test time by more than 95%, which can effectively reduce
the cost, improve the quality of the test, improve the profitability of the enterprise, and
enhance the enterprise’s core competitiveness. At present, the system has been suc-
cessfully applied to the testing of all kinds of module power products produced by our
company, and the application prospect is broad. The follow-up will continue to
transform and upgrade the test system, focusing on the comparison of the test data of
different production stages of the same module power supply, which will form a
database to make the application of the test system reach a higher level and give full
play to the power of automation equipment and digital technology. To make greater
contributions to the smooth development of spacecraft products and the development of
space industry.
References
1. Beshr, H.M., et al.: Modelling of a residential solar standalone power system. In: Proceeding
of the 1st International Nuclear and Renewable Energy Conference (2010)
2. Bhandari, R., Stadler, I.: Electrification using solar photovoltaic systems in Nepal. Appl.
Energy J. 88, 458–465 (2011)
3. Dambrine, G., Cappy, A., Heliodore, F., Playez, E.: A new method for determining the FET
small-signal equivalent circuit. IEEE Trans. Microw. Theory Tech. 36(7), 1151–1159 (1988)
4. Gu, W.B., Wang, C.Y., Liaw, B.Y.: Numerical modeling of coupled electrochemical and
transport processes in lead-acid batteries. J. Electrochem. Soc. 144(6), 2053–2061 (1997)
5. Vandersmissen, R., Ruythooren, W., Das, J., Germain, M., Xiao, D., Schreurs, D.: Transfer
from MHEMT to GaN HEMT technology: Devices and integration. In: Int. Compound
Semicond. Manuf. Technol. Conf., pp. 237–240, April 2005
6. Lee, J.-H., Yoon, H.-S., Park, C.-S., Park, H.-M.: Ultra low noise characteristics of
AlGaAs/InGaAs/GaAs pseudomorphic HEMT’s with wide head T-Shaped gate. IEEE
Electron Device Lett. 16(6), 271–273 (1995)
7. Liu, W.: “FET high-frequency properties” in Fundamentals of III-V Devices:
HBTs MESFETS and HFETs/HEMTs, pp. 422–429. Wiley, New York (1999)
Design and Implementation of User-Oriented
On-Board Mission Management System
for Remote Sensing Satellite
Abstract. In order to improve the operation, efficiency and reduce the project
cost of remote sensing satellites, this paper proposes a user-oriented hierarchical
and distributed on-board mission management system (OMMS) design and
implementation. Combined with user requirements for remote sensing satellite
applications, the typical mission mode of the OMMS and the simple and flexible
user interface are designed. The mission mode implementation adopts the
method based on sub-mission sequence, mission mode construction rules and
system configuration database to implement the OMMS software. The actual
test shows that the system can greatly reduce the amount of mission instruction
data upload by users, shorten the task preparation time and reduce the project
cost.
1 Preface
In order to facilitate users, improve the efficiency of the usage of the remote sensing
satellite, and save cost, the paper proposes a user-oriented design and implementation
method of the OMMS. The mission management system can provide flexible and
efficient control interfaces while shielding the internal design details of the satellites
[1]. The missions can be achieved merely through transmitting mission related satellite
parameters, which greatly improves the operation efficiency of the satellite.
The satellite system mission layer is the core of the mission management system. It
is responsible for constructing the sub-mission for the lower subsystems, according to
the user demands and the system configuration database. Each subsystem may execute
their mission separately accordingly. The safety monitoring modules is responsible for
judging the rationality of user mission demands and monitoring the execution of each
level to ensure satellite security during mission execution. The system configuration
database is used to store the system optional configuration, such as the working status
and mode of each subsystem.
6 Conclusion
This paper is based on the mission construction rules, system configuration database
and sub-sequence design to implement the remote sensing satellite OMMS. Compared
with the traditional mission operation mode, the actual test data indicates that the
mission instruction data uploaded by the user is reduced to less than 5% of the tra-
ditional mode. The time for the user to build the ground control system and the system
development workload are also greatly reduced, greatly improving the satellite’s use
efficiency.
The implementation of the OMMS lays solid foundation for its subsequent
development to the intelligent networked direction. The subsequent OMMS will
implement on-board autonomous mission planning, constellation mission planning and
constellation mission assignment, based on the functions of self-feedback mission
planning for autonomous image processing on the satellite will provide users with a
more flexible, powerful and intelligent OMMS, which will further enhance the mission
capability of the remote sensing satellite and reduce the project costs.
Design and Implementation of User-Oriented 153
References
1. Bonnet, J., Gleizesy, M-P., Kaddoumy, E., Rainjonneau, S., Flandin, G.: Multi-satellite
mission planning using a self-adaptive multi-agent system. In: 2015 IEEE 9th International
Conference on Self-Adaptive and Self-Organizing Systems, pp. 1949–3673. IEEE (2015)
2. Pan, Y., Chi, Z.M., Rao, Q.L., Sun, K.P., Liu, Y.N.: Modeling and simulation of mission
planning problem for remote sensing satellite imaging. In: MATEC Web of Conferences, vol.
179, p. 03024 (2018)
3. Lic. De Elia, Estefanía: Master in Emergency Early Warning and Response Space
Applications. Septiembre (2010)
4. Donati, A.: Infusion of innovative technologies for mission operations. Acta Astronaut. 67,
1132–1137 (2010)
5. European Cooperation for Space Standardization (ECSS): Space project Management, Project
planning and implementation. ECSS-M-ST-10C Rev. 1, 6 March 2000
Extraction of Salient Region
Based on Visual Perception
1 Introduction
on this basis. However, the drawback is that the space-based visual attention model
focuses too much on the location information of the focus, ignores the integrity of the
imaging subject, and the computational structure is usually complex. In order to
simplify the computational structure and enhance the integrity of the target, a visual
attention model based on transform domain is proposed. In 2007, literature [5] put
forward Spectral Residual (SR) significant detection method based on Fourier trans-
form for the first time by using frequency domain information of natural image scale
invariance characteristics. Later, the literature [6] proposed the Phase spectrum
Quaternion Fourier Transform (PQFT) model. In 2009, Achanta et al. [7] proposed IG
algorithm based on frequency tuning, which has stronger detection ability than SR
algorithm, but the algorithm overemphasizes image edge contour and texture, which is
not conducive to the next step of processing. Although these methods can show good
detection ability in experiments, they do not have biological theoretical support.
From the perspective of biological vision, a visual attention model based on
transformation domain is put forward in this paper to extract the salient region of
images. The image characteristic graph in multi-scale is obtained by using gaussian
pyramid decomposition, difference of gaussian (DoG) filtering and PCT transforma-
tion. Through the optimal adaptive segmentation threshold, the regular focus window is
effectively constructed. Meanwhile, experiments were carried out on the image library
of significant objects provided by Microsoft research Asia (MSRA). Experiments show
that this method can adapt to the change of the position and size of the subject and
overcome the interference of background.
Where, Gðx; y; rÞ stands for the scaled variable gaussian function, as defined below.
r is the scale coordinate, and represents the convolution operation.
1 ðx2 þ y2 Þ=2r2
Gðx; y; rÞ ¼ e ð2Þ
2pr2
The commonly used 5x5 gaussian filtering template is as follows.
2 3
1 4 6 4 1
6 4 16 24 16 47
1 6 7
w¼ 6 6 24 36 24 67 ð3Þ
256 64 4 16 24
7
16 45
1 4 6 4 1
The original image is taken as the 0-level image. First, the image is processed by
gaussian smoothing with a low-pass filter, and then the image is sampled to obtain the
image of the next level. Three sub-images are selected as input images for the next
step. Then, the sub-image is convolved with the gaussian function with different
standard deviations.
1 ðx2 þ y2 Þ=2r2
Gr ðx; yÞ ¼ e ð5Þ
2pr2
Two images in the adjacent gaussian scale space are performed by central-
peripheral difference operation, namely the subtraction operation. Feature detection of
the image can be carried out on a certain scale to obtain DoG response value image.
g1 ðx; yÞ g2 ðx; yÞ ¼ Gr1 I ðx; yÞ Gr2 I ðx; yÞ ¼ ðGr1 Gr2 Þ I ðx; yÞ ð6Þ
In the above equation, r1 [ r2 . When the standard deviation of the two gaussian
functions r1 =r2 ¼ 1:6, DoG filter is widely used for edge detection. Its performance is
similar to that of gauss Laplace filter, but the calculation quantity is far lower than that
of gauss Laplace transform. And when the ratio of the two approaches 5, it approxi-
mates the response of the ganglion cells in the retina, simulating the central - peripheral
antagonistic properties of the human eye.
Extraction of Salient Region Based on Visual Perception 157
1 X
M 1 X
N 1
ð2x þ 1Þup ð2y þ 1Þvp
Cðu; vÞ ¼ pffiffiffiffiffiffiffiffi aðuÞaðvÞ f ðx; yÞ cos cos ð8Þ
MN x¼0 y¼0
2M 2N
The formation of visual saliency is mainly due to the lateral inhibition between
similar neurons in the V1 region of the primary visual cortex [10]. The so-called lateral
inhibition is that when a neuron detects the same visual characteristics as the neigh-
boring neurons, the excitability of the neuron will inhibit the excitability of the neigh-
boring neurons and reduce the frequency of its pulse distribution. Instead, the frequency
of the pulses is increased. Therefore, the frequency of distribution of corresponding
neurons is higher at positions where are more prominent of the object, such as the edges
and contour. Lateral inhibition helps the eye separate objects from the background.
The DCT coefficient can reflect the statistical characteristics of similar visual fea-
tures in the visual space. The higher the coefficient, the higher the probability of the
occurrence of the visual features corresponding to the frequency and direction in the
space. Discrete pulse cosine transformation (PCT) extracts the symbolic information of
DCT coefficient and sets the amplitude uniformly to 1. By this operation, the visual
features with high probability are suppressed and that with low probability are
improved. From the perspective of visual neurons, this process simulates the lateral
inhibition of similar neurons in the V1 region.
For a given image I, the calculation process of PCT is
P ¼ signðdct2ðIÞÞ ð10Þ
F ¼ absðidct2ðPÞÞ ð11Þ
from Fig. 1(b), the two-dimensional DCT transformation of natural images has the
characteristic of “energy concentration”. The low-frequency information is mainly
concentrated in the upper left corner of the DCT coefficient matrix, while the high-
frequency information corresponding to the image edge information of the space
domain is concentrated in the lower right corner of the coefficient matrix, with the
coefficient amplitude close to 0. PCT transformation only retains the symbol of DCT
coefficient, namely the phase information, and sets the amplitude to the uniform value.
Through this operation, the low-frequency components of the image are suppressed and
the high-frequency components are amplified, which is helpful to extract the edge
information of the image spatial domain, as shown in Fig. 1(c).
(a) Original image (b) DCT coefficient distribution (c) PCT transformation
Where, Fi ðx; yÞ is the feature graph to be fused, and i is the number of feature
graphs.
In order to distinguish the target subject from the background and extract the salient
area, filtering and binarization should be performed on the fused feature map. Adaptive
threshold method is adopted. The filtering formula and threshold calculation formula
are as follows.
1:5 Xm1 X
n1
T¼ Fsaliency ðx; yÞ ð15Þ
m n x¼0 y¼0
1 X
WIx ¼ F ði; jÞ ð17Þ
NIx ði;jÞ2I
x
W ¼ maxfWIx g x ¼ 0; 1; 2 ð18Þ
Where, NIx is the amount of pixels contained in the target region Ix ðx ¼ 0; 1; 2. . .Þ,
and F ði; jÞ is the pixel value of the corresponding characteristic graph.
Through the above steps, the interested region of human eyes can be extracted as the
only focus window. Taking an actual image as an example, the image processing
results of each calculation stage of the visual attention model are illustrated, as shown
in Fig. 2. The pixel number of the source image is 512768. After the gaussian
pyramid decomposition, three sub-images of different scales are selected for DoG
filtering, with the size of 256384, 128192 and 6496 respectively. Then, PCT
transformation was performed for the sub-images after multi-scale gaussian difference
filtering, and the larger two images were sampled to the size of 6496, and weighted
fusion was performed. Finally, gaussian low-pass filtering and adaptive threshold
binarization are performed on the fused feature map, and then sampled to the size of the
source image. The unique regular rectangular window of the image is constructed by
using MBR method and feature weight method. The window size constructed in this
example is 247 by 399.
To test and verify the extraction effect of salient regions, the experiments are
conducted on the salient object image library provided by Microsoft research Asia
(MSRA). The gallery contains 5000 images, including people, scenes, etc., whose
salient regions were manually marked by nine volunteers according to their subjective
intentions. Here are some of the results. Figure 3(a) is the original image, Fig. 3(b) is
the extracted feature map, Fig. 3(c) is obtained after binarization, Fig. 3(d) is the
regular window constructed by MBR method, and Fig. 3(e) is the salient object
manually marked. From the experimental results, it can be seen that the focusing
window algorithm proposed can effectively separate the subject and background
regions, extract the regions that human eyes are interested in. The focusing window
extracted automatically (as shown in Fig. 3(d)), basically contains the salient target
manually marked (as shown in Fig. 3(e)).
In order to evaluate the focus window construction method proposed in this paper
effectively and objectively, regional precision and recall are adopted, which are
respectively defined as
Precision ¼ ðS \ AÞ=S
ð19Þ
Recall ¼ ðS \ AÞ=A
Where, S is the salient region automatically extracted, and A is the salient region
manually marked in the reference set, \ shows the intersection.
The precision reflects the proportion of the correct salient region in the salient
region automatically extracted, while the recall reflects how much of the salient area
manually marked is detected by the algorithm. There is a certain mutual restriction
relationship between the two, expressed by F-measure parameters, and the calculation
formula is as follows.
1 þ b2 Precision Recall
F-measure ¼ ð20Þ
b2 Precision þ Recall
It can be seen from the above figure that the method proposed is superior to the
classic Itti method in terms of the precision, recall and F-measure. When b2 ¼ 0:3, the
value of F-measure of proposed method can reach up to 0.816, while that of Itti is lower
than 0.6. Moreover, the Itti algorithm needs to extract color, brightness and directional
features at different scales, which requires a large amount of calculation. However, this
algorithm only needs grayscale information and the computing time is greatly reduced.
Extraction of Salient Region Based on Visual Perception 163
4 Conclusion
References
1. Marr, D.: Visual information processing: the structure and creation of visual representations.
Philos. Trans. R. Soc. Lond. 290(1038), 199–218 (1980)
2. Treisman, A.M., Gelade, G.: A feature integration theory of attention. Cogn. Psychol. 12(1),
97–136 (1980)
3. Koch, C., Ullman, S.: Shifts in selective visual attention: towards the underlying neural
circuitry. Hum. Neurobiol. 4(4), 219–227 (1985)
4. Itti, L., Koch, C., Niebur. E.: A model of saliency-based visual attention for rapid scene
analysis. IEEE Comput. Soc. (1998)
5. Hou, X., Zhang, L.: Saliency detection: a spectral residual approach. In: IEEE Conference on
Computer Vision and Pattern Recognition, 2007. CVPR 2007, pp. 1–8. IEEE (2007)
6. Guo, C., Zhang, L.: A novel multiresolution spatiotemporal saliency detection model and its
applications in image and video compression. IEEE Trans. Image Process. 19(1), 185–198
(2009)
7. Achanta, R., Hemami, S., Estrada, F., et al.: Frequency-tuned salient region detection. In:
IEEE Conference on Computer Vision and Pattern Recognition, 2009. CVPR 2009,
pp. 1597–1604. IEEE (2009)
8. Yu, Y., Wang, B., Zhang, L.: Pulse discrete cosine transform for saliency-based visual
attention. In: International Conference on Development and Learning, pp. 1–6. IEEE (2009)
9. Zhang, X.: Computational models and applications of the retinal color vision. University of
Electronic Science and technology of China. (2017)
10. Li, S.: Research on visual perception based spatial gamut mapping. Tianjin University
(2016)
11. Xiong, W., Xu, Y., Cui, Y., et al.: Geometric feature extraction of ship in high-resolution
synthetic aperture radar images. Acta Photonica Sinica 47(1), 49–58 (2018)
Thermal Design to Meet Stringent
Temperature Gradient/Stability Requirements
of Space Camera’s Tube
2 Thermal Environment
External heat flux in the inner surface of space camera’s tube is obtained through
theoretical analysis, which includes direct solar heat flux, albedo heat flux and earth
infrared heat flux [2, 3].
The orbit beta angle (b), angle between sunlight and orbital plane, is from 17.2° to
27°. a, argument of apogee, is from 115° to 245°. The depth of solar radiation in the
tube, H, is calculated with the following formula [4].
H ¼ D= tan U
ð1Þ
cos U ¼ maxð0; cosbcosa)
Hmax ¼ 0:44D:
166 Y. Li et al.
Q2, albedo heat flux reaching on the inner surface of space camera’s tube, is
calculated as follow.
2
RE
Q2 ¼ S q maxð0; cos b cos aÞ X12 ð2Þ
RE þ h
where X12 is the angle factor between inner surface of tube and the virtual surface of the
light inlet. q is albedo factor. S, the intensity of sunlight reaching Earth, varies
approximately 3.5%, depending on Earth’s distance from the sun. At summer solstice,
the value of S is equal to 1322 W/m2. At winter solstice, the value of S is equal to
1414 W/m2. RE is Earth’s radius. h is distance from orbital plane to the surface of
Earth. h1 is depth from camera entrance (Fig. 2).
light inlet
h1=0.4D
h1=0.6D
330 h1=0.8D
300 h1=D
270
albedo heat flux Q2/ W/m2
240
210
180
150
120
90
60
30
0
0 50 100 150 200 250 300 350
α/ °
Q3, Earth infrared heat flux flow on the inner surface of space camera’s tube, is
calculated as follow [5].
2
RE
Q3 ¼ S ð1 qÞ X12 ð3Þ
RE þ h
The variations in +Z Earth-emitted IR are much less severe than the variations in
albedo. Q3 is depend on X12. The depth from camera entrance is increases and Earth-
emitted IR decreases rapidly (Table 1).
Thermal Design to Meet Stringent Temperature 167
The baffle is attached to tube at 24 points by screws and heat insulation pads are
installed on each point. However, heat conduction between the baffle and tube is
considerable. The tube views space (3K). The upper part of the tube has more heat
dissipation to space than the lower part. The orbital heat flux absorbed by tube in the
circumferential and axial directions is not uniform. At the same time, albedo heat flux
will fluctuate periodically. It is a challenge to meet the stringent thermal requirements.
The tube is cold biased, and active thermal control system is used to meet the thermal
requirements. System design based on multi-level insulation and graded heating is
proposed. Passive thermal control strategies, such as enhanced thermal insulation
design and optimum design of the baffle length, are used to reduce the sensitivity of
space camera’s tube to the external thermal environment. An active thermal control
system is taken to solve the temperature control problem for multi-zone coupling with
each other.
The external surfaces of baffle are covered with MLI. The baffle is thermally
insulated from tube by layered polyimide gaskets to increase contact thermal resistance
(Fig. 3).
Four heater loops are mounted on the baffle near the 100 mm of tube. The tem-
perature set point for baffle’s heaters should be designed so that its peak temperature is
lower than the temperature set point for tube’s heaters in order to avoid heat transfer
from baffle to tube. Therefore, heater controller set points for heaters are set to 12 °C.
The heating power is 10 W for one loop.
Td Proportional u
Heater
heater OBJECT
output
controller
Tm
Temperature
measurement
The required heater time is calculated by the control algorithm. The control period
applied is 10 s. The resolution is better than 1 s.
A thermal balance test was undertaken to validate the thermal design and to provide
measured data for improving the thermal model. Thermal test results are shown in
Table 2 and Fig. 6. The thermal balance tests show that:
(1) The temperature of tube is 19.29–20.12 °C, both their radial and axial temperature
gradient are less than 0.57 °C, their temperature fluctuation during imaging are
less than 0.08 °C.
(2) The periodic average heating power for tube is 18.2 W in cold case test.
(3) The analysis results are close to the thermal balance tests, indicating that thermal
analysis model is correct. The thermal control system can meet the stringent
thermal requirements.
5 Conclusions
The thermal environment of the space camera’s tube is analyzed and the formula of
orbital heat fluxes for tube is given subsequently. The tube is cold biased, and active
thermal control system is used to meet the thermal requirements. Enhanced thermal
insulation design and optimum design of the baffle length, are used to reduce the
sensitivity of space camera’s tube to the external thermal environment. An active
thermal control system is taken to solve the temperature control problem for multi-zone
coupling with each other. In-orbit analysis of the last 3 years shows that, the axial
and circumferential temperature gradient of tube are less than 0.5 °C. The temperature
fluctuation are less than 0.33 °C/3 month. Periodic average heating power for tube
is 16.7 W.
172 Y. Li et al.
References
1. Wang, X., Guo, C., Hu, Y.: Design and verification for front mirror-body structure of on-axis
three mirror anastigmatic space camera. Acta Photonica Sinica (2011)
2. Gilmore, D.G.: Spacecraft Thermal Control Handbook. The Aerospace Press, California
(2002)
3. Hou, Z., Hu, J.: Foundation and Application of Spacecraft Thermal Control Technology.
China Technology Press, Beijing (2007)
4. Zhang, R.: Satellite Orbit Attitude Dynamics and Control. Astronautic Press, Beijing (1998)
5. Ning, X., Zhang, J., et al.: Extreme external heat flux analytical model for inclined-orbit
hexahedral satellite. J. Astronaut. (2008)
6. Tong, Y., Li, G., Geng, L.: A review on precise temperature control technology for spacecraft.
Spacecraft Recovery Remote Sens. (2016)
Application Design of Virtual Assembly
Simulation Technology for Installing Cables
on Biaxial Drive Mechanism
1 Introduction
With the development of technology, virtual assembly has been widely used in
spacecraft assembly integration. It has achieved application results in three dimensional
space interference analysis, ergonomics analysis, assembly mode selection, mechanical
grounding supporting equipment (MGSE) design, site adaptability etc. [1]. It has
guided the specific realization of process design, process planning and MGSE design.
Virtual simulation technology is the combination of computer technology and virtual
reality technology. Virtual environment and virtual products are constructed to simulate
and control the real entities. Virtual simulation technology is used to verify the cor-
rectness and rationality of the whole spacecraft and component-level assembly process
and assembly routes [2]. Digital visualization technology is used to test the assembly
sequence and assembly route, evaluate the product whether can be assembled, optimize
the process design, and further improve the design quality and extend the process
expression technique [3].
© Springer Nature Singapore Pte Ltd. 2019
S. Sun et al. (Eds.): ICSINC 2018, LNEE 550, pp. 173–179, 2019.
https://doi.org/10.1007/978-981-13-7123-3_21
174 C. Yang et al.
In this project, the emphasis is laid on laying and routing of thermal control cables.
Therefore, the diameter, minimum turning radius and elastic performance of cables
should be clear. Meanwhile, for additional heating plates, thermistors, platinum
resistors and so on, it is necessary to model matching and revising and positioning with
the size specified in the design documents, so as to complete the building of the
simulation analysis environment.
1) the layout of the thermistor wires through 2) the layout of backup-heater the wire
the bracket to the cable beam through the bracket to the cable bundle
3) the layout of main-heater wires through 4) the inner heater loop of the sleeve is con-
the bracket to the flange of the sleeve nected in series
5) the sleeve is connected in series with the 6) the layout of heating circuit wires after
external heater circuit connection in series
7) the layout of platinum resistance wires is 8) the layout of dehumidifier wires through the
through the biaxial bracket biaxial bracket
9) welding plugs for dehumidification wires 10) stick the thermistor outside of the cable
near the bracket bundle and layout relevant wires
Fig. 3. The simulation and analysis process of layout cables on azimuth axis
178 C. Yang et al.
In the process of virtually simulate cable installing, we should pay attention to cable
layout, cable lines fixed mode and interference feasibility analysis, operational feasi-
bility analysis, cable routes optimization, etc. [5, 6]. Taking the feasibility analysis of
the cable through hole in bracket as an example, we premise the diameter of the single
thermal control wire to be 0.8 mm, then we can learn that no more than 19 wires can go
through the same hole together. However, there are 15 wires are designed to go through
the hole. We layout wires carefully, then we can analysis that the minimum distance
between wires are less than 0.5 mm. Considering the requirements of process pro-
tection in actual implementation process, the thermal wires will also be wrapped with
many layers of 3M tape for protection, which will increase the diameter of the wire.
Therefore, it is necessary to optimize the layout of wire by separating the harness to
different holes or changing the path. The simulation analysis is shown in Fig. 4.
5 Concluding Remarks
In this paper, the flexible cable laying of two-axis drive mechanism is studied. Set up
simulation environment should be according to the design requirements to layout
cables and plan the routes, simulate and analyze the wire installing process. In the
course of the study, we can identify the problems which may appear during design
phase or operating phase. We can optimize the cable layout effectively to ensure the
accuracy and enforceability of the process design. So it can effectively reduce the
Application Design of Virtual Assembly Simulation Technology 179
failure rate of assembly defects and products, reduce the risk of product development,
and ensure the quality of product assembly. The successful application of virtual
assembly simulation in the cable laying of biaxial drive mechanism of the two-
generation data relay satellite can provide reference and guidance for the assembly
process design for other spacecraft.
References
1. Yang, Z., Li, L., et al.: 3D digital assembly process for rocket body cabin. Aerosp. Mater.
Technol. 4, 85–91 (2016)
2. Yuan, C.: Assembly process simulation of rocket components and on-site visualization
technology application. Xi’an University of Electronic Science and Technology (2015)
3. Feng, W., Zhang, Y., et al.: The virtual simulation technology in assembly process design of
china’s lunar exploration project II. Spacecraft Environ. Eng. 6, 326–331 (2014)
4. Zhou, R., Zhang, H., et al.: Virtual assembly process simulation of Bogie based on DELMIA.
Mach. Electron. 2, 47–50 (2015)
5. Shen, Z., Zheng, H., et al.: Application of digital assembly simulation technology for
complicated vehicle based on DELMIA. Missile Space Launch Technol. 6, 42–45 (2015)
6. Ritchie, J.M., Dewar, R.G., Simmons, J.E.L.: The generation and practical use of plans for
manual assembly using immersive virtual reality. In: Proceedings of the Institution of
Mechanical Engineers (Part B), vol. 213, pp. 461–474 (1999)
7. Li, T., Han, F., et al.: Virtual assembly simulation of catheter based on DELMIA. Mech.
Manuf. 5, 74–77 (2014)
Data Analysis and Research Based on Satellite
Micro-vibration Disturbance Test
Yang Gao1(&), Qiang Wang2, Fei-hu Liu3, Lu Cao3, and Wei Cheng3
1
Beijing Institute of Spacecraft System Engineering, Beijing 100094, China
35794664@qq.com
2
Beijing Institute of Spacecraft Environment Engineering,
Beijing 100094, China
3
School of Aeronautical Science and Engineering,
Beihang University, Beijing 100191, China
1 Introduction
With the increasing resolution of remote sensing satellites, the requirements of satellite
jitter and attitude stability are becoming more and more important. The main
influencing factors include external force interference, attitude maneuver and micro-
vibration of satellite moving parts (such as momentum wheel, CMG etc.) [1, 2]. The
micro-vibration generated by these moving parts is always present, and has a great
influence on the satellite. Therefore, the suppression of micro-vibration becomes a
necessary measure to ensure the image quality [3–5]. If the disturbance characteristics
of the disturbance source are known in the development stage of the product com-
ponents, the corresponding isolation and vibration reduction measures can be taken in
the structural design stage [6–10].
During the working process of a certain type of CMG, the disturbance character-
istics of the CMG and the effectiveness of vibration isolation scheme are still unclear.
Therefore,the corresponding micro-vibration test and analysis evaluation are required.
This paper introduces the test scheme of a certain type of CMG micro-vibration test,
analyzes the test results of rigid installation state and isolator installation state and
evaluates the effectiveness of vibration isolation scheme.
2 Test Scheme
The platform can measure six components of vibration force (three axial forces, two
bending moments and one torque). The main performance indexes are shown in
Table 1.
When the disturbance source is installed and working on the platform, there will be
eight voltage signal outputs. The calibration process is needed to determine the
182 Y. Gao et al.
In the above formula, C is used to make the force system of hammering force
(F) equivalent to the force system at the geometric center point O(F0 Þ. UðxÞ816 is the
output voltage of the platform; WðxÞ68 is the calibration matrix. Then there are the
following formulas:
If the output voltage is UðxÞ81 , the six-component forces produced by the mea-
sured disturbance source are as follows:
Fig. 3. The schematic diagram of the test under the condition of isolator installation
Under the working condition of CMG, the speed of the high-speed shaft is stable at
7000 rpm, and the low-speed frame rotates around the low-speed shaft at a certain
speed. Therefore, we should carry out the test when the low-speed frame is at different
positions (0°–330°, every 30°, for a total of 12 positions). After the stabilization, the
micro-vibration test is performed. The test conditions are shown in Table 2.
Table 2. The test conditions (The high-speed shaft: 7000 rpm; The low-speed frame: 12
different positions.)
Number of test conditions Description of test conditions
I CMG is rigidly mounted on the force measuring platform
II CMG is mounted on the force measuring platform via isolators
3 Data Analysis
3.1 The Data of Background Noise
After setting up the test system, LMS and the control equipment of CMG boot up. At
this moment, the high-speed rotor part is not activated. The data collected by LMS is
considered as the background noise of the system. Figure 4 shows the background
noise diagram of Fx in time domain. From Fig. 4, the amplitude of background noise of
Fx is within 0.35N, and the root mean square value in time domain is 0.14N. The root
mean square values of the residual force and torque are all below 0.14.
184 Y. Gao et al.
Table 3. The mean values of the root mean square values (Band: 0–400 Hz; Position: 0°–330°)
Conditions Fx (N) Fy (N) Fz (N) Mx (Nm) My (Nm) Mz (Nm)
Condition I 6.06 11.95 1.13 0.80 0.44 4.81
Condition II 1.11 1.11 0.66 0.45 0.12 0.62
Table 4. The mean values of the peak values at the fundamental frequency (Position: 0°–330°)
Conditions Fx (N) Fy (N) Fz (N) Mx (Nm) My (Nm) Mz (Nm)
Condition I 4.08 11.60 0.76 0.22 0.41 4.67
Condition II 0.09 0.06 0.03 0.02 0.01 0.12
4 Conclusion
The laboratory carries out the micro-vibration test of a CMG in two installation states:
rigid installation and isolator installation. Then we obtain the data of background noise
and the data in CMG working status.
(1) Compared with the data of background noise, the data in CMG working status is
increased by at least one order of magnitude. Thus the test result is considered
valid.
(2) Comparing the data of two installation states: the rigid installation state and the
isolator installation state, there is a significant decline in both the root mean square
value in time domain and the peak value in frequency domain, especially the peak
value at the fundamental frequency decreases most obviously.
Therefore, it is considered that the vibration isolation effect of the isolator is
obvious, and the vibration isolation effect at fundamental frequency is the best.
Data Analysis and Research 187
References
1. Wu, Y., Xie, Y.-c., Yan, T.-f., Fang, G.-q., Jiao, A.C.: Micro-vibration Testing Technology
for Space Station Control Moment Gyroscope. Equip. Environ. Eng. 94–99 (2018)
2. Zhang, Z.-h., Yang, L., Pang, S.-w.: High-precision mechanical environment analysis of
micro-vibration of spacecraft. Spacecraft Environ. Eng. 528–534 (2009)
3. Wang, G.-y., Zhou, D.-q., Zhao, Y.: Data analysis of on-orbit micro-vibration measurement
by remote sensing satellite. Astronaut. J. 261–267 (2015). Author, F., Author, S., Author, T.:
Book title. 2nd edn. Publisher, Location (1999)
4. Lambert, S.G., Casey, W.L.: Laser Communications in Space. Artech House, Boston (1995)
5. KATZMAN, M.: Laser Satellite Communications, p. 250. Prentice-Hall Inc., Englewood
Cliffs (1987)
6. Wu, D.-y., Li, G., Lu, M.: Noise measurement and analysis of 1000Nms control moment
gyroscope. Space Control Technol. Appl. 31–55 (2012)
7. Zhou, D.-q., Cao, R., Zhao, Y.: Measurement and analysis of micro-vibration in orbit by
remote sensing satellites. Spacecraft Environ. Eng. 627–630 (2013)
8. Zhou, J.-p.: General conception of space station project in China. Manned Spaceflight 1–10
(2013)
9. Wang, T.-m., Wang, H., Li, H.-y.: Space station common orbit vehicle deployment for
supply mission. Manned Space 583–596 (2017)
10. François, D., Mark, W., Stephen, A.: New facility for micro-vibration measurements ESA
reaction wheel characterisation facility. In: Proceedings of the 12th European Conference on
Space Structures, Materials & Environmental Testing, Noordwijk (2012)
Spacecraft Automation Test Procedure
and System Design
Abstract. In view of the fact that the existing manned spacecraft electrical
measurement process cannot be streamlined and automated, it consumes a
considerable amount of human resources and time costs, the manned spacecraft
automated test program and system were designed. Through the analysis and
design of the format, function and processing flow of the automated test pro-
gram, the basic data resources for spacecraft electrical survey are provided, and
the automatic test system is designed to realize the automation and process of
the spacecraft electrical test cycle. It is suitable for different series spacecraft
electrical survey, effectively reduce the manual operation of the testers and the
number of personnel, improve the test efficiency, meet the high-intensity ground
test, and efficiently complete the high-density manned spacecraft launch
mission.
1 Introduction
system to form an efficient running test procedure. Thales Alenia Space has imple-
mented fully automated testing of payloads. China’s aerospace technology department
introduced the idea of automation testing to improve the automation and intelligence of
test systems. Compared with foreign countries, the domestic research on spacecraft
automation test is still in its infancy, mainly summarizing the characteristics of
spacecraft automation test language [3–8], a valid test executable digital resource has
not yet been formed. In order to improve the research capability of China’s spacecraft
and catch up with international advanced testing technology, it is extremely urgent to
develop spacecraft automation testing.
The test procedure is composed of a single piece of text and organized by time. Each
field is separated by a tab, and each row represents a specific operation. Each operation
may include multiple sub-operations according to different opcodes, and each sub-
operation may include multiple operation parameters. The definition of the test lan-
guage begins with a step number and ends with a specific field. By designing the format
and function of the automated test language, the spacecraft automation testing can be
streamlined and fully cycled. Table 1 shows the automation test procedure format
(Tables 2 and 3).
Step number: The serial number of the operation of the line is generally incre-
mented in order from the beginning of 1.
Time type: 0 means execution in absolute time (relative to the first step), 1 means
execution in relative time (relative to the previous step),
Time value: The time of execution of the operation, absolute time, relative time are
expressed in two ways, one is “number” and the unit is second; one is “day: hour:
minute: second”, such as “00: 01:03:14” is equivalent to “3794 s”.
Opcode: The content and process of subsequent sub-operations are defined. Dif-
ferent opcodes have different number of sub-operations and can be a combination of
multiple sub-operations.
Subaction: In general, sub-operations can be divided into judging telemetry
parameters, test program control, sending instructions, and combinations thereof.
Judging telemetry parameters support:
1. The bitwise operation “bitwise AND”, “bitwise OR” is supported in the conditional
judgment;
2. The conditional judgment supports the comparison operators <, >, <=, >=,
==, *=, which means less than, greater than, no greater than, no less than, equal,
not equal;
3. The conditional judgment supports the numerical operators +, –, *, /, ^, and %,
which respectively represent addition, subtraction, multiplication, division, expo-
nential power, and modulo;
Spacecraft Automation Test Procedure and System Design 191
Test procedure control operation contents can indicate start (EX), pause (HOLD),
delay execution (WAIT), stop (QUIT), inter-procedure jump (JUMPTO), in-procedure
jump (GOTO), program link (LINKTO), Conditional judgment (IF…ELSE), etc.
(Table 5).
The spacecraft automation testing business process is shown in Fig. 1. It realizes the
automation of the three stages of test preparation, test execution and test evaluation,
and realizes the automation of the whole test process [9].
Test preparation: Through the modular design idea, the test project design is carried
out according to the test outline, the test project is organized, and the test rules and test
procedures are automatically generated.
Test execution: The test procedure generated during the test preparation phase is
input, the test procedure is executed, and the unattended self-monitoring is performed
in real time. When an abnormality occurs, the program jumps autonomously according
to the program design.
Test evaluation: The spacecraft and subsystem are used as the structure to evaluate the
implementation of each test procedure and the execution of the test procedure, and the
detailed test evaluation report is automatically generated according to the test rules. The
tester’s phased test coverage and test intensity analysis are shown in schematic form.
The automation test system consists of a control terminal, an execution terminal, a
database, and a display terminal. The automation test system consists of a control
terminal, an execution terminal, a database, and a display terminal.
The execution end of the automation test system extracts the corresponding test
resources from the automation test database, the operator operates the control terminal,
controls the execution terminal to run the automation test program, and the execution
terminal broadcasts the test process information to the local area network, and the
database records the test process information to provide a number for the test
evaluation.
The tester monitors the test process through the display terminal.
The automation test system and other components of the spacecraft integrated test
system work together to complete the automation test of the spacecraft, as shown in
Fig. 2.
Spacecraft Automation Test Procedure and System Design 193
The automation test procedure and system are applied to the electrical test of a certain
type of spacecraft, covering electrical performance test, mechanical test, thermal test,
launch field test, etc. The test program execution accuracy is 100%, and the command
is sent correctly and on time.
After the test procedure is applied, the instructions are not manually sent, the test
task is programmed, the test implementation is standardized, the manual operation is
reduced by about 60%, the test time is saved by about 30%, and the accuracy of the test
operation is improved.
According to statistics, in the same series of models, 201 of the 225 test procedures
can be directly inherited, and the proportion of use is close to 90%, which can effec-
tively reduce the number of test personnel and achieve better application results
(Table 6).
5 Conclusion
The automation test procedures and applications provide the basic data resources for
the spacecraft electrical test, so that the spacecraft test is no longer an independent
individual and discrete instructions, with process and test procedures control capabil-
ities, which can effectively reduce the manual operation of testers and manual inter-
vention. At the same time, integrate test resources, reduce unnecessary duplication of
work, effectively reduce testers, adapt to high-intensity testing, and efficiently complete
high-density spacecraft launch missions.
References
1. Wang, Q.: Eletrical Test Technology of Spacecraft. Chinese S&T Press, Beijing (2007)
2. Li, X., Wei, C., Zhang, W., Xia, Q., Li, T.: On-orbit maintainability design system for
manned spacecraft. J. Syst. Eng. Electron. 38(1), 84–89 (2016)
3. Britton, K.J., Schaible, D.M.: Spacecraft Testing Programs: Adding Value to the Systems
Engineering Process, Kennedy Space Center, Langley Research Center, 29 March 2011
4. Ma, S., Yu, D.: Automated Test Language and system for Spacecraft Test. National Defence
Industry Press, Beijing (2011)
5. Zhu, W., Wang, J., Li, D.: Modeling for spacecraft automated test and design of automated
test language. J. Electron. Meas. Instrum. 118–121 (2009)
6. Sun, B., Ma, S., Yu, D.: Spacecraft automatic test and spacecraft test language. J. Beijing
Univ. Aeronaut. Astronaut. 35(11), 1375–1378 (2009)
7. Dang, J., Yu, J., Zhou, J.: Automatic Measurement Technology of Integrated Test System for
FY-2 Meteorological Satellite, Aerospace Shanghai, pp. 72–77 (2005)
8. Zhu, W., Wang, J., Li, D.: Design of spacecraft automated test software based on directed
graph modeling. Comput. Eng. Des. 31(8), 1702–1705 (2010)
9. He, Y., Pan, S., Li, H., et al.: Design and application of automatic test system for manned
spacecraft. Comput. Meas. Control 23(10), 3258–3263 (2015)
Analysis of Wave Propagation in Functionally
Graded Material Annular Sector Plates
Using Curved-Boundary Legendre Spectral
Finite Elements
Teng Wang(&)
1 Introduction
Functionally graded materials (FGMs) are increasingly used in the fields of aerospace,
civil and automotive engineering. Generally, FGMs are made of two kinds of materials
to satisfy specify structural performance demands [1]. Compared to traditional com-
posite, the significant property of FGMs is that the spatial coordinate dependent
material properties can overcome the interfacial problem typical for mostly layered
composite structures. However, the FGM structures are often subjected to corrosion,
impact or fatigue under severe environment condition. The damage detection of those
structures is critical for maintaining their status. Recently, the guided-wave based
damage detection has been is proposed as a very potential way to identify and location
damage in structures [2]. Understanding the wave propagation behavior of wave is very
important for building an effective damage detection strategy.
In the past decades, dynamics analysis of FGM structures becomes necessary and
significant. A number of investigations have been done to study the wave propagation
behavior in the FGM structures by analytic methods and numerical methods.
Numerical methods are proposed to analyze the wave propagation in FGM struc-
tures with complex geometries recently. Among many numerical methods, the finite
element method (FEM) and the spectral finite element method (SFEM) are two popular
tools. To analyze wave propagation in FGM structures with numerical methods,
material simplification has to be done to consider the continuous varying of material
properties [3]. However, the studies in [4] show that only the continuous material
model can give more accurate and detailed results for FGM structures.
Simulations of wave propagation in FGM structures with finite element method can
be found in [4–6]. Exact numerical simulation of wave propagation in structures always
costs a lot of time and memory space. The spectral finite element method has attracted
high interest in recent years, which can improve computational efficiency and reduce
computation time. The time-domain spectral finite element (SEM) is proposed by
Patera in the 1984 [7]. A Chebyshev spectral plate element for wave propagation in
isotropic structures is developed in [8].
Curved structures are widely used in Engineering. To the author’s best knowing,
wave propagation in curved component has not been paid much attention. In this paper,
elastic wave propagation in a curved FGM waveguide is investigated. To achieve more
accurate geometry approximation, a curved-edge time-domain spectral element is
developed. A quadratic Lagrange function is to conduct the geometry shape transform
while higher orders Lagrange function is to interpolate the displacement field. The
continuous material model is involved to model FGMs.
2 Motion Equation
€ þ C Q_ þ KQ ¼ F;
MQ ð1Þ
Like the classic FEM, the spectral finite element method follows the similar pro-
cedure. Then the establishment of the two-dimensional spectral finite element method
is presented in detail.
Analysis of Wave Propagation in Functionally Graded Material 197
The whole structure is made of functionally graded material. Outside and inside
phases are pure Ceramic and Chrome. Properties between them change along the
thickness. Establish a one-dimensional system d along the thick direction. Property
values f ðdÞ of the middle material, including elastic modulus, Poisson ratio and density
can be determined by the volume fraction. The volume fraction is a power law function
with exponent v.
X
n X
n
x¼ u k xk ; y¼ uk yk ð3Þ
k¼1 k¼1
ð5Þ
4 Numerical Validation
To validate the efficiency of the proposed element, an FGM plate excited by a time-
load is investigated, shown in Fig. 3. In this paper, the FGM plate excited is simplified
to 2D plane strain problem. The continuous material model in Legendre spectral finite
element is implemented to solve the 2D wave propagation problem. The results are
validated by the traditional FEM, operating in ABAQUS. In FEM analysis, the
structure is divided into many layers and in each layer, material properties are regarded
as constant value, which lies on the average of each layer.
The 2D model is shown in Fig. 3 with the thickness of 0.02 m and a finite length of
3 m. The material property is changing linearly in the thick direction where grads
parameter. Phases of top and bottom are Aluminum and Zirconia where elastic moduli
are 70 GPa, 151 GPa and densities are 2700 kg/m3, 3000 kg/m3 and Passion ratio is
200 T. Wang
the 0.3 constantly. The point impact is 5 periods sinusoid with the central frequency of
50 K modulated by the Hanning window function.
The elements are 2 80 and there are 7 7 nodes distributed in each element for
SFEM. The elements are 16 1200 and element type is 4-node bilinear plane strain
element. To catch the wave responses accurately, an time increment of 1 10–8 s is
needed. Displacements responses of measuring points located at x = 1.5 m, y = 0.02 m
are recorded. As shown in Fig. 4, a very good agreement with FEM solution is
observed. The results demonstrate that the introduced continuous material model for
FGM is validated.
Fig. 5. Displacement field at t ¼ 280 ls. (a, b) Displacement of X, (c, d) Displacement of Y; (a,
c) Solved by curved element and (b, d) solved by straight element.
The results show that using curved edge elements is very necessary in wave
propagation analysis of the FGM structures with curved boundaries.
Fig. 6. Responses of the measuring point at a = 90°, Ro = 1.02 m, with varied grads parameters.
202 T. Wang
Effects on dispersion are investigated through the phase velocity. Figure 7 shows the
phase velocity of S0 mode oscillation and A0 mode oscillation. The dispersion results
demonstrate that with the increase of the material grads parameters, amplitude of the phase
velocity decreases. The level of the dispersion with varied powers does not change
obviously.
Fig. 7. (a) Phase velocity of S0 mode oscillation and (b) phase velocity of A0 mode oscillation.
6 Conclusion
In order to investigate the wave propagation in the functionally graded material mul-
tiple waveguides with complex boundary, a curved-edge Legendre spectral element
involved a continuous material model, is introduced. In the new element, quartic
Lagrange function is implemented to complete the coordinate transform from the real
structure element to standard element. Considering continuously spatial changing of the
material property inside the element, determine the property value in the integral node
through the volume fraction function of the coordinates for accurate modeling of FGM.
Then numerical analysis on a finite FGM plate with a point impact validates the
continuous material model, compared to the FEM results. And wave propagation in the
multiple guides is studied. Wave field, time-domain response and dispersion of phase
velocity are presented. Comparison between curved SFE and straight SFE is done.
Effects of difference material grads parameters on wave propagation behavior are
investigated.
It is validated that the quartic Lagrange function for shape transform and the
continuous material model inside the Legendre spectral finite element is very necessary
and efficient when analyzing wave propagation in FGM multiple waveguides structure.
Results show that the FGM multiple guides structure loaded an impact holds two
different vibration mode S0 and A0. A0 mode oscillation disperses obviously while
S0 mode oscillation does not. A further study on material grads parameter show that as
the power v increases, amplitude of wave response increases and phase velocity des-
cends. It is suggested that the stability of S0 mode oscillation is preferred to be chosen
for damage detection based on guided wave. The future work can be extended to the
crack monitoring in FGM structures or material characterization.
Analysis of Wave Propagation in Functionally Graded Material 203
References
1. Markworth, A.J., Ramesh, K.S., Parks Jr., W.P.: Modelling studies applied to functionally
graded materials. J. Mater. Sci. 30(9), 2183–2193 (1995)
2. Birman, V., Byrd, L.W.: Modeling and analysis of functionally graded materials and
structures. Appl. Mech. Rev. 60(5), 195–216 (22 pages) (2007). https://doi.org/10.1115/1.
2777164
3. Chen, W.Q., Wang, H.M., Bao, R.H.: On calculating dispersion curves of waves in a
functionally graded elastic plate. Compos. Struct. 81(2), 233–242 (2007)
4. Berezovski, A., Engelbrecht, J., Maugin, G.A.: Numerical simulation of two-dimensional
wave propagation in functionally graded materials. Eur. J. Mech. A. Solids 22(2), 257–265
(2003)
5. Cho, J.R., Ha, D.Y.: Averaging and finite-element discretization approaches in the numerical
analysis of functionally graded materials. Mater. Sci. Eng. A 302(2), 187–196 (2001)
6. Patera, A.T.: A spectral element method for fluid dynamics: laminar flow in a channel
expansion. J. Comput. Phys. 54(3), 468–488 (1984)
7. Żak, A.: A novel formulation of a spectral plate element for wave propagation in isotropic
structures. Finite Elem. Anal. Des. 45(10), 650–658 (2009)
8. Dauksher, W., Emery, A.F.: Accuracy in modeling the acoustic wave equation with
Chebyshev spectral finite elements. Finite Elem. Anal. Des. 26(2), 115–128 (1997)
Research on Design of Quality Management
Module in Spacecraft Assembly MES
Qiang Wang(&)
1 Introduction
With the rapid growth of China’s aerospace business, a large number of major Aero-
space missions with high-quality and high-efficiency support have put forward new
requirements and challenges to the development of spacecraft AIT projects. However,
it is difficult to meet the requirements for core competency and leapfrog development
of spacecraft AIT projects only by expanding the scale of scientific research and
production and increasing the cost of scientific research investment. Therefore, it is
necessary to continuously deepen and improve the model of spacecraft AIT project to
meet the needs of development, while systematically plan to seek breakthroughs,
practice and explore new models, fundamentally enhance basic capabilities, and pro-
mote the sound and rapid development of the aerospace industry.
For the high-risk aerospace industry itself, the improvement of quality management
capability is particularly prominent and urgent. In recent years, information technology
has been integrated into the quality management of aerospace enterprises, and has
become an important technical way to upgrade and optimize the quality management
capability [1]. Based on the analysis and elaboration of the characteristics of spacecraft
assembly, this thesis puts forward some suggestions for improving the design of quality
management module in spacecraft assembly MES (Manufacturing Execution System).
This thesis mainly studies the design idea of quality management function in spacecraft
assembly MES, so it does not discuss the specific overall MES framework and
information technology.
© Springer Nature Singapore Pte Ltd. 2019
S. Sun et al. (Eds.): ICSINC 2018, LNEE 550, pp. 204–211, 2019.
https://doi.org/10.1007/978-981-13-7123-3_25
Research on Design of Quality Management Module 205
The assembly work of spacecraft model is the terminal part of spacecraft development.
(1) Spacecraft assembly has the characteristics of small batch. Most of the platform
spacecrafts are manufactured at the same time in a small number. The engineering
experience of previous models has strong reference significance for spacecrafts in
research.
(2) Spacecraft assembly work requires a tight schedule and high-quality requirements.
Spacecraft development is mostly a national key project, which is of great sig-
nificance. Once the progress delays and quality problems occur as the terminal
link, the launch and use of spacecraft will be directly affected.
(3) Under the management system of the combination of R&D and production in
spacecraft assembly, it has strong dynamic characteristics in the process of
assembly implementation, including demand change, design change, process
route, resource conflict, single product change and other factors.
(4) Spacecraft assembly involves many fields and units. There may be no adminis-
trative subordinate relationship between units. Cooperative working environment
is complex, and conflict coordination and problem handling is complex.
knowledge of the shared spacecraft assembly to the development and cooperation units,
strengthens the main environment of spacecraft assembly quality management, trans-
mits the quality management requirements, rules and workflow in specific problems,
gradually eliminates differences in understanding quality management with collabo-
rative units, strengthens the identity of the collaboration units for the quality control of
the spacecraft assembly, creates the foundation for multi-unit collaboration and efficient
deployment of spacecraft assembly, and avoids the difficulty in handling and technical
coordination caused by the difference of quality management perception, and time
delay, which will affect the progress and quality of spacecraft assembly.
Quality-Knowledge-Base Establishment
Quality-Knowledge-Base is an evolving dynamic system. The acquisition of new
knowledge needs to be completed in many times of spacecraft assembly practice. The
existing engineering experience, quality control measures, technical specifications and
quality requirements need to be verified, evaluated and improved through many times
of spacecraft assembly practice. Therefore, quality information is the source and carrier
Research on Design of Quality Management Module 209
A special quality management system and a set of system oriented to the whole
process of model development should be established. On this basis, a unified quality
management process technology interface and management interface should be
established to eliminate the differences in understanding quality management modes,
habits and even quality management among different units, and to establish quality
management adapted to multi-enterprise environment. The mechanism of coordination
is used to achieve quality information interaction and development collaboration in all
aspects of spacecraft development.
5 Conclusion
Aiming at the characteristics of spacecraft total quality management, this thesis studies
the design of quality management module in spacecraft assembly MES. The spacecraft
assembly design system, the integrated quality management environment taking MES as
the cores, quality-knowledge-base support platform, as well as the design of quality
management platform for tooling are emphatically discussed. These designs can promote
the improvement of quality management capability and quality management mode.
References
1. Tang, X., Duan, G., Du, F.: Implementing Technologies of Quality Information System in
Manufacturing Enterprises. National Defense Industry Press, Beijing (2009)
2. Feng, X., Zhou, L., et al.: Study of supporting platform in bi-base based integrated quality
management system. Eng. J. Wuhan Univ. 39(2), 67–69 (2006)
3. Han, X., Li, M., et al.: Design and key technologies of spacecraft assembly MES.
Aerosp. Manuf. Technol 2015(4), 55–59 (2015)
4. Baujut, J.F., Laureillard, P.: A co-operation framework for product-process integration in
engineering design. Des. Stud. 23(6), 497–513 (2002)
5. Yuan, F., Xu, D., et al.: Reserach and realization of the quality management system based on
case reasoning. Sci. Technol. Eng. 8(9), 2502–2506 (2008)
Research on Design of Quality Management Module 211
6. Shanbao, Y., Wei, H.: The construction and implement of quality management system which
faced to the manufacture process of aerospace product. Aerosp. Manuf. Technol. 2012(6), 19–
23 (2012)
7. Bingru, Y.: Knowledge Discovery Theory Based on Internal Mechanism and Its Application.
Electronic Industry Press, Beijing (2009)
The Design and Implementation of Secure
Cloud Desktop System
1 Background
PCs with Windows XP have been widely used in enterprises, and confidential data
generated by users during the work is stored on the single hard disk of the PC. Thus the
following problems occur:
(1) Poor data reliability: Most staff may have access to confidential information and a
large number of confidential documents are stored locally on the PC. Once key
business data is lost due to physical or logical damage of the hard disk, it may
cause incalculable loss. Moreover, there are uncertainties in the maintenance time
and the, which seriously affects the user’s work.
(2) Low data security: It is required for the system with confidential information to
prevent external illegal intrusion and illegal data stealing by internal personnel.
Although the PC implements multiple security protection methods such as host
monitoring and three-in-one, a large amount of confidential data is scattered and
stored in the user’s local PC. There is still chance that users maliciously destroy
security protection or physically dismantle the local hard disk.
(3) High-cost maintenance: Owing to different PC models and configuration, the user
terminal administrator needs to completely install, configure, and manage the
operation system, application software, security protection program, and patch
update after testing and verifying deployment of multiple PC configurations,
which is labor-consuming. Meanwhile, it may further increase the support cost
since technical personnel often need to go to the filed to solve the problem due to
the low degree of standardization.
(4) Limited access: Since the data is stored locally on the PC, the users need to need
to temporarily collect, sort, and copy it when they discuss plans in the conference
room, develop the design in the designing room, or compare the data in the
laboratory. The data may be lost or incomplete during transmission, which makes
it difficult for users to work outside of office.
Based on the above problems, it is urgent to innovate and change the traditional PC
usage model. With mainstream virtualization technology [1] and customized security
design to build a secure cloud desktop [2], it shall realize centralized storage of user data
and unified release management of desktop application, to meet the needs of users’ daily
office use, effectively control the output of user data, enhance the security of the user’s
desktop system, and improve operational management and maintenance efficiency.
On the virtualization platform: the underlying server cluster and storage devices
form the physical layer of hardware to provide a secure and stable computing, network,
and storage environment for the upper layer. By adopting the leading virtualization
technology as the underlying platform for the virtual desktop, it enables a single
physical server to be run by multiple virtual desktops to form a virtual desktop resource
pool. After virtual desktop management and access server clusters are deployed, it can
be uniformly managed with terminal session access and desktop state, to effectively
solve the tight coupling problem among operation system application, and hardware.
All data is stored on a centralized storage device, and data is accessed through the FC
switch between the server and the storage.
On the Internet, switches, firewalls, and link transmission protection devices are
deployed to provide a secure and reliable data transmission link for data transmission
between the terminal and the virtualization platform [5].
It can be modified through either an existing PC or a customized thin client ter-
minal. PC can be modified by replacing the hard disk with an electronic disk. It shall
deploy a simplified embedded operating system, install a USB Key driver, add a
domain, install a lock screen program, etc.; The thin client terminal performs safe
reinforced cutting based on customized operating system to provide secure and stable
operating environment for users.
The Design and Implementation of Secure Cloud Desktop System 215
Physical Security
The operation environment of the host machine in virtualization platform, blade
workstation, and centralized storage device is stored in the data center computer room.
It adopts physical protection methods such as the national secret access control system
and video surveillance system, equipped with UPS power supply. With a strict man-
agement system and computer room inspection system, it ensures that all equipment is
physically and safely controllable.
Operational Safety
The storage system is used to implement centralized storage of virtual desktop data and
platform management data. To ensure the security of virtual desktop images and
templates, a unified backup and recovery policy is implemented to support recovery
through templates, snapshots, and migration modes, which ensures the integrity of
virtual machine image and user data.
Information Security
In terms of information security and confidentiality, a series of security measures are
taken.
Identity authentication adopts “domain user” and “USB Key” authentication, to
ensure the legal identity of users in the desktop virtualization environment.
The secure cloud desktop system performs security auditing by recording logging
of user virtual desktop access behaviors, including login, startup, shutdown, restart,
maintenance, status query, etc.; Auditing Administrator operations includes virtualized
resource management, update maintenance, storage resource allocation, data backup
and recovery operations.
Computer virus protection shall be conducted by deploying anti-virus software on
the virtual desktop, and selecting an anti-virus version optimized for the virtualized
environment to prevent the virtual machine from simultaneously scanning for disk
viruses and upgrading the anti-virus storm caused by the client side.
User data protection shall be realized by encrypting and protecting with encryption
technology in the centralized storage device. The user data shall be effectively read and
written after the identity of the user is recognized through identity authentication.
Regarding transmission network security, the transmission network between the
user terminal and the back-end virtual desktop, uses the USB Key certificate to
establish an encrypted network transmission channel, to ensure data encryption and
integrity protection between the terminal and the virtual desktop, and effectively pre-
vent data information from being intercepted or analyzed.
In terms of storage environment security, the terminal does not store user data in the
desktop virtualization environment. All data is stored in centralized storage. Commands
and image information are transmitted between the terminal and the virtual desktop, so
that the storage environment shall be secured.
216 H. Pan et al.
Terminal Security
From the prospective of the convenience of management and the security of usage, the
terminal in the secure cloud desktop adopts software and hardware design, and multiple
security protection mechanisms to ensure its security. The security zone of the terminal
device is separately divided by devices such as firewalls and switches. The data of the
confidential information system is encrypted and saved to the back-end centralized
storage without data in the terminal.
4 Summary
References
1. Wang, L.Z., Laszewski, G., Younge, A., et al.: Cloud computing: a perspective study. New
Gener. Comput. 28(2), 137–146 (2010)
2. Yang, C.: Big Data and cloud computing: innovation opportunities and challenges. Int.
J. Digit. Earth 1, 13–53 (2016)
3. Netto, M., Calheiros, R., Rodrigues, E., et al.: HPC cloud for scientific and business
applications: taxonomy, vision, and research challenges. ACM Comput. Surv. 51(1), 8:1–8:29
(2018)
4. Luis, M.V., Luis, R.M., Juan, C., et al.: A break in the clouds: towards a cloud definition.
ACM SIGCOMM Comput. Commun. Rev. 39(1), 50–55 (2009)
5. Chris, R.: A break in the cloud: the reality of cloud computing. In: EABR & TLC
Conferences Proceedings, pp. 1–5 (2009)
6. Von Laszewski, G., et al.: Comparison of multiple cloud frameworks. In: IEEE 5th
International Conference on Cloud Computing (CLOUD) (2012)
7. Kostantos, K., et al.: OPEN-source IaaS fit for purpose: a comparison between OpenNebula
and OpenStack. Int. J. Electron. Bus. Manage. 11, 3 (2013)
Research on Satellite Power Subsystem
Anomaly Detection Technology Based
on Health Baseline
Abstract. There is a huge amount of data in the satellite power subsystem, and
the information and knowledge related to the operating status of the power
subsystem in these data is an important for the abnormal state detection of the
power subsystem. In order to make data mining of satellite power subsystem and
set the reference for its anomaly detection, a satellite power subsystem anomaly
detection technology based on the health baseline is proposed. According to the
core operating conditions of the satellite power subsystem, health baselines are
constructed respectively, and the effectiveness of the proposed method is veri-
fied by the telemetry data of the satellite power subsystem. With robustness, the
test results show that anomaly detection of the onboard power subsystem can be
achieved based on the constructed health baseline.
1 Introduction
The satellite power sub-system is the satellite subsystem used for the whole star power
supply and distribution and power connection between equipment. The health status of
the satellite power subsystem largely determines whether the satellite can operate
normally. All studies on the anomaly detection of satellite power subsystem are of great
significance.
In the internal data stream of the satellite power subsystem, there are massive multi-
source heterogeneous data such as design data, on-orbit telemetry data, historical
telemetry data and ground test data. The information and knowledge about the health
status of the power subsystem in these massive data is an important basis for the health
management of the power subsystem. On the one hand, because a variety of telemetry
parameters can reflect the health status of the same subsystem and equipment from
different angles; On the other hand, due to the complex relationship between the power
subsystem and the real-time telemetry parameters of the equipment, the evolution of its
operating state can be reflected by the association of multiple parameters. The mutual
penetration and cross-correlation pose new challenges for satellite operation state
feature extraction and anomaly detection.
For abnormal detection of power systems or electromechanical systems, in previous
studies, thresholds were often set for individual parameters as the state of health baseline
judgment parameters to determine the state of the system. However, more and more
scholars have begun to pay attention to the coupling relationships between internal
parameters of the system, and to propose the standard of system anomaly detection.
Wang et al. [1] introduced a spatial anomaly detection method based on attribute-related
quantitative values. The attribute correlation value is calculated by quantitative analysis,
and the spatial statistic is searched by the attribute correlation matrix and the R-tree
dynamic index structure to realize spatial anomaly detection. Reference [2] proposed a
DoS attack detection system based on multivariate correlation analysis [2]. The system
used multivariate correlation analysis (MCA) to extract geometric correlation between
network traffic characteristics, and uses anomaly-based detection principle for attack
identification. Literature 3 proposed a method called Potential Correlation-Based
Anomaly Detection (LCAD), which used a probability distribution model called a
potential correlation probability model to correlate potentials between multiple corre-
lated heterogeneous monitoring data series [3]. Sexual modeling to identify anomalies
helps to detect anomalies in the working state of the device based on the correlation
between massive data. Zhang et al. [4] built a healthy baseline by mining the correlation
of internal characteristics of vibration signals, and successfully classified the inner ring
fault, outer loop fault and roller fault of the bearing.
It can be seen that under the condition of massive data from satellite telemetry,
mining the associations and rules in the data, using the associations and rules to
construct the health baseline, it is feasible and effective for the satellite abnormal state
detection.
In this paper, the satellite power subsystem is taken as the research object, and the
satellite health baseline under different working conditions is proposed to detect the
operating states of satellite. This paper was structured as follows: The second section
elaborates the main research context and structure of this paper, and briefly introduces
the main steps of constructing a healthy baseline; the third section introduces the health
baseline construction process based on two core operating conditions of satellite power
subsystem. And through the real telemetry data of the satellite power subsystem to
analyze and evaluate its effect; the fourth section draws conclusions.
2 Methodology
This paper mainly analyzes the satellite power subsystem design data and various types
of telemetry data to find out the core operating conditions of the power subsystem.
Based on this, two types of health baselines are constructed to realize the anomaly
detection of the satellite power subsystem as shown in Fig. 1. Specific steps are as
follows:
Firstly, according to the satellite design information and data feature analysis, the
satellite operating conditions are divided and screened, that is, the sub-basin constant
current charging condition and the sub-basin discharge condition.
Research on Satellite Power Subsystem Anomaly Detection Technology 221
Baseline build
Statistical
Massive status distribution method Linear health baseline
monitoring and test Expert knowledge
verification data Nonlinear health baseline
Rule based
INPUT
…… OUTPUT
3 Case Study
In this paper, we take all telemetry of the power subsystem as filter objects, mining
their internal relationships and building a healthy baseline, and based on the statistics of
all data fluctuations, then get the fluctuation range of the constructed health baseline.
Then we combine the health baseline with the fluctuation threshold to detect the status
of the power subsystem.
Under the two core conditions of the power subsystem, two types of health baselines,
namely the linear health baseline and the nonlinear health baseline, were constructed.
Fig. 2. Health baseline between recharging current A and B under working condition 1
Table 2. Battery A: Voltage vs. recharging current health baseline parameter value list
Health baseline parameters Test parameter value Health baseline and threshold
x Recharging current \
y Battery A voltage
R
a 0.0000389 y ¼ 0:0000389 x dt þ yð1Þ
y(1) Initial charging current
b_upper 0.5 y_upper = y_pre + 0.5
b_lower –0.5 y_lower = y_pre – 0.5
Fig. 3. Health baseline between Battery A: Voltage vs. recharging Current under working
condition 1
Table 3. Health baseline between battery pack discharge current backup: A vs. B parameter
value list
Health baseline Test parameter value Health baseline and
parameters threshold
x Battery pack A discharge current \
backup
y Battery pack B discharge current
backup
a 0.952 y_pre = 0.952 x + 1.570
b 1.570
a_upper 0.967 y_upper = 0.967 x +
b_upper 1.8045 1.8045
a_lower 0.937 y_lower = 0.937 x +
b_lower 1.2025 1.2025
Fig. 4. Health baseline between battery discharge current A&B under working condition 2
Fig. 5. Health baseline between battery discharge current and voltage under working condition 2
4 Conclusion
This paper proposes a new anomaly detection method based on health baseline con-
struction for satellite power subsystem. By obtaining the different characteristics of the
telemetry data under the core conditions of the satellite power subsystem, and mining
the relationship between them, two types of health baselines based on the sub-basin
constant current charging condition and the sub-basin discharge condition are estab-
lished, that is, the linear health baseline and nonlinear health baseline. The test results
show that the method is feasible and effective for realizing the anomaly detection of the
spaceborne power subsystem.
References
1. Wang, Z., Chen, H.: Research on spatial outlier detection based on quantitative value of
attributive correlation. Comput. Eng. 32, 37–39 (2006)
2. Tan, Z., Aruna, J., He, X., et al.: A system for denial-of-service attack detection based on
multivariate correlation analysis. IEEE Trans. Parallel Distrib. Syst. 25, 447–456 (2014)
226 L. Zhang et al.
3. Ding, J., Liu, Y., Zhang, L., et al.: An anomaly detection approach for multiple monitoring
data series based on latent correlation probabilistic model. Appl. Intell. 44, 340–361 (2016)
4. Zhang, T., Lu, C., Tao, L., Li, K.: Rolling bearing fault diagnosis based on health baseline
method. In: Vibroengineering Procedia. 28th International Conference on Vibroengineering,
vol. 14, pp. 141–145 (2017)
5. Pang, J., Liu, D., Liao, H., et al.: Anomaly detection based on data stream monitoring and
prediction with improved Gaussian process regression algorithm. In: International Conference
on Prognostics & Health Management (2015)
6. Lee, H., Byington, C., Watson, M.: PHM system enhancement through noise reduction and
feature normalization. In: Aerospace Conference (2010)
7. Xiong, L., Ma, H.D., Fang, H.Z., et al.: Anomaly detection of spacecraft based on least
squares support vector machine. In: Prognostics & System Health Management Conference
(2011)
Lifetime Prediction Method of Components
Based on Failure Physics
In 1952, Shockley proposed the junction gate field-effect transistor theory, in the early
1960s when mass production of MOS transistor began, it was found that the charge
related to the thermal oxidation structure of silicon had serious influence on device
reliability [1]. People started the research on the charge in oxide layer. Since then,
people have used PoF approaches to predict device life, including high-heat treatment,
defect and corrosion, and studied its effect on device electromigration and service life
from the perspective of impurity.
In the age of transistor, people already did a raft of studies on planar state. Around
the 1970s, with the emergency of planar device (MOS), people fully studied Si/SiO2
system. In the 1970s, leakage current aroused the attention and there were oceans of
studies on soft error of packaging materials. With the rapid development and expansion
of nuclear technology in the Cold War, the studies on radiation resistance of
semiconductor devices were abundant and mature because the devices were mostly
applied for national defense and military purposes. In the early 1970s, the researchers
studied the failure mechanisms including water vapor, mechanical vibration, super-
power operation and high temperature reverse bias. In the mid-1970s, the research on
thermoelectric effect became mature and the deterioration of the interface between
metals resulted in the research on metal materials. Investigations into metallization
issues became extremely in-depth, including low ohmic contact, good bounding and
adhesion and stability of SiO2, such as Al electromigration, failure mechanism of Au-
Al bounding, aluminum film rupture at the step of the oxide layer and Al corrosion.
People began looking for new metal materials as well. In addition, the effects of failure
caused by electric stress on service life of devices were also studied around the mid-
1970s, including flashover short-circuit, transient failure caused by surge current,
solder thermal fatigue caused by intermittent action and overload loss caused by
switching power dissipation. Second breakdown was already studied at that time but it
was still a theoretically unresolved issue. Meanwhile, some progress was made in the
study on gate breakdown caused by electrostatic discharge damage of MOS integrated
circuit.
From the late 1970s to the early 1980s, Very Large Scale Integration (VLSI) was
researched and developed. As the influence of hot carrier on the device attracted
people’s attention, the research on the influence of hot carrier injection (HCI) on device
life was started. At the same time, there also emerged some studies on Drain Induction
Barrier Lower (DIBL). In the early 1990s, study of DIBL became mature with much
relevant literature published, which continued into this century. Negative bias tem-
perature instability (NBTI) was first discovered by Miura [2], Goetzberger [3] and
others in 1966. However, there was not much research on NBTI and its effect on device
reliability and life in the early stage. In the late 1980s, research on NBTI was carried
out to a certain extent, the integration degree of integrated circuits was getting higher
and people paid more attention to research on NBTI. After the 1990s, a large quantity
of research literature on NBTI was published successively, especially after entering the
period of micro- and nano- scales.
As early as the 1980s, gate oxide breakdown attracted people’s attention, and the
time dependent dielectric breakdown (TDDB) ramp voltage histogram was used to
speculate on the device’s failure rate. In 1981, Arnold Berman of IBM proposed that
the previous models for inferring reliability failures through laboratory life test were
essentially incorrect as they failed to consider the effects of temperature on TDDB and
the distribution of failure time [4]. Since then, a lot of improvement research has been
done on TDDB. In the mid and late 1970s, it was discovered that the increase of the
gate voltage VG of MOS devices could cause a rise in leakage current among the poles
[5]. In the 1980s, researchers began to study the impact of gated-induced drain leakage
(GIDL) on device life and how to reduce GIDL. Today, with the continuous
improvement of the integration degree and the decreasing thickness of the gate oxide,
the research of TDDB and GIDL has become more and more important and cannot be
ignored. On this basis, researchers often need to study its effects in depth when
predicting device life.
Lifetime Prediction Method of Components Based on Failure Physics 229
2.1 Electromigration
Electromigration is the process of transporting aluminum ions along a stream of
electrons in an energized aluminum strip at high temperature and high current density.
It is the most important failure mechanism in the electrode system of semiconductor
devices and integrated circuits (IC). The accelerated life test is affected by its elec-
tromigration. For long metal wires, the aluminum alloy strip is usually connected by the
pad without preventing metallization. The overall failure life can be controlled by
nucleation, and the current density N = 2 is generally adopted. Secondly, since the
butterfly-type NIST electromigration test structure with simple pad connection is not
suitable for the existing multi-layer metal system, very good results of electromigration
can be obtained compared with the channel feedback test structure. Thirdly, the channel
feedback test structure must be carefully designed to avoid resistance saturation and
storage effect, as they can lead to incorrect median life.
The equation [8] describing the electromigration using the Black model to deter-
mine the life is as follows:
N dðInTF Þ=dðInJÞ
Hypothesis: There is a very long aluminum-bronze metal tube with particle size
larger than line width; the mobile environment is the chip environment (80 °C) inside
the portable computer; the office environment is the chip environment (50 °C) inside
the case; the current density in mobile environment and office environment is
2.5 mA/cm2 and 2.0 mA/cm2 respectively; and J>>Jcrit, Ea = 0.8 eV, and N = 2. On
this basis, applying the above life model can get the ratio of life value in office
environment to that in mobile environment, i.e., AF:
The value of AF (about 18) is calculated after inputting the aforesaid data. It can
thus be seen that the life value in low-temperature, low-current density office envi-
ronment is 18 times larger than in high-temperature, high-current density mobile
environment. Specifically, a 1.6-times difference in current density and an 11.5-times
difference in temperature could finally translate into an 18-times increase in overall life.
2.3 TDDB
In principle, the TDDB process is divided into two stages: The first stage involves
breakdown accumulation. The characteristic of this stage is that under the influence of
electric stress, new-generated traps (charge) are accumulated inside the oxide layer and
at the Si-SiO2 interface. This results in the electric-field modulation effect inside the
oxide layer. When local electric field or current hits the critical value, the second stage,
i.e., rapid breakdown, begins. In this stage, the process of positive feedback between
electricity and heat leads to gate oxide breakdown.
Given that the oxide breakdown is due to the hole being trapped and accumulated at
local traps of the oxide layer, the trapped hole flow can be expressed as [1]:
Qp / JðEOX ÞaðEOX Þt
Where, J(EOX) refers to the F-N current density, which is directly proportional to
e−B/EOX, a is the ionizing collision induced hole generation coefficient
/ EOX / eH=EOX , B, being approximately 240 MV/cm, is a constant related to the
electron effective mass and cathode interface barrier, H is approximately 80 MV/cm,
and t is the duration. When Qp reaches the critical value, breakdown occurs:
There are two common TDDB failure models used for predicting life: E model and
1/E model.
The E model assumes that the oxide burn-in and breakdown is a thermodynamic
process during which breakdown occurs possibly because Si-O bond is destroyed due
to the mutual interaction between dipoles under thermal stress and applied electric field.
In E model, there exists a linear relationship between the electric field under Eoxcon-
dition and time, with the expression [9] as follows:
tBD ¼ t0 expðcEox Þ
In the equation, tBD refers to the life of gate oxide layerunder the TDDB stress
condition, s0 means the time of intrinsic breakdown, c is the parameter of electric field
proportion, with unit of MV/cm. All these variables are related to temperature process,
their expressions being as follows respectively:
Ea c
t0 ¼ A exp ; c ¼ bþ
KT T
Where, Eox is the electric field of oxide layer, Ea is thermal activation energy,
K refers to the Boltzmann constant, T represents the absolute temperature, and both
b and c are process-related constants. E model is usually used for evaluating the life of
gate oxide layer when the SiO2 dielectric is relatively thick and the electric field is
relatively low.
232 L. Liu et al.
1/E model (hole-induced injection model) mainly relates to the F-N tunneling
current. If the breakdown process is deemed to have resulted from gate-through current,
1/E model is generally used to predict the gate oxide life [9]:
tBD ¼ t0 expðG=Eox Þ
Where, tBD refers to the life of oxide layer under the TDDB stress condition, t0
represents the time of intrinsic breakdown, which, like G, is a constant related to
temperature process, as shown in the following expressions:
Eb 1 1 d 1 1
t0 ¼ A exp ; G ¼ G0 1 þ
Kb T 300 Kb T 300
Under the same temperature and different electric field stresses, A and G are
constants. In the logarithmic coordinate, there is a linear relationship between the
average breakdown time (TDDB life) and the reciprocal of applied electric field. 1/E
model is generally used for life evaluation when the SiO2 dielectric is relatively thin
and the electric field is relatively high.
2.4 NBTI
NBTI effect is found in PMOS devices. When the device grid is under negative bias,
the device’s saturation drain current Idsat and transconductance Gm is continually
decreased while the absolute value of threshold voltage is continually increased. An
increase in the bias action time on the grid will be accompanied by continuous
degradation of the device’s electrical parameters. There are mainly two models used to
explain the degradation mechanism in NBTI: The first is the reaction-diffusion (R-D)
model: When the vertical electric field is applied, the generated hole and Si-H bond
react to isolate the hydrogen atom and leave one charge, then the hydrogen atom keeps
away from the siliconoxide interface through diffusion. Given its very slow speed to
diffuse in the oxide layer, hydrogen thus becomes the bottleneck for the NBTI
degradation speed; The second is the charge trapping and de-trapping model. When the
electric field is applied, the charge at the Si/SiO2 interface is trapped by defects in the
oxide layer, which brings more fixed oxide charges. When the electric field is removed,
some trapped charges can be divorced from the oxide layer.
Severe recovery or passivation effect exists in NBTI effect. The longer the stress
continues, the more severe the irreversible damage to the gate oxide will be. The charge
trapping and de-trapping model can well explain the rapid recovery effect of NBTI
within a short period of time, while the R-D model can explain the phenomenon of
NBTI recovery effect existing for a long period of time after the stress condition is
removed. At present, we are inclined to believe that the two mechanisms play a joint
role in NBTI degradation.
In the testing, the two mechanisms are manifested by the exponential relationship
between electrical property degradation and time. Through a theoretical calculation of
the speed of diffusion in the oxide layer for the hydrogen atom, molecule and ion,
relationship between the device’s electrical property degradation and time is reflected
Lifetime Prediction Method of Components Based on Failure Physics 233
as a power exponent ranging from 0.165 to 0.25 in the R-D model and less than 0.165
in the charge trapping and de-trapping model. The two models are applicable to dif-
ferent scopes of time after the electric field condition is removed, that is, a time lag
between when the electric field stress is removed and when the electrical property
parameters begin to be measured, is closely related to the performance of the tested
device. With the adoption of new testing means and improvement of test device per-
formance, the time lag is greatly reduced. New researches have generally found a
relatively small power exponent ranging from 0.08 to 0.14. Researchers have put
forward testing methods like fast NBTI and on-the-fly NBTI to reduce the NBTI
recovery effect [10].
Under the old testing methods, a relatively long time delay based on the recovery
effect result showed a smaller value in electrical property degradation compared with
that of rapid testing. But this does mean that the impact of NBTI was underestimated.
The reason is, as study suggests, with a reduction in the electric field stress, the failure
time difference calculated under different testing speeds is smaller accordingly. As a
result, the electric-field acceleration factor varies by testing speed. By deducing the
device life from the result calculated under a relatively slow testing speed, the actual
service life of the device is sometimes underestimated.
ln t ¼ A þ B=Tj
Where, A and B are constants. Therefore, determining the working junction tem-
perature is an important constant in the reliability design considered for the device use.
In general, for silicon devices, the maximum allowable junction temperature is 150 °C–
175 °C in case of metallic packaging, and 125 °C–150 °C in case of plastic packaging.
Heat is transferred through three means, including radiation, convection and con-
duction. The heat generated in the active area of the semiconductor chip is transferred
to the adjacent areas through heat conduction, making temperatures in different chip
areas unevenly distributed. In other words, a temperature field is generated. The
temperature field distribution of the chip can be calculated by solving the heat con-
duction equation under certain boundary conditions, which is helpful for the reliability
of heat design. Generally speaking, heat Q transmitted from section A is directly
proportional to the temperature gradient in this direction:
In the equation, RTjc represents the thermal resistance between the collector junc-
tion and case of the transistor. Tj and Tc refer to the junction and case temperature
respectively. Pc is the dissipated power of the transistor. Based on this, thermal
resistance can be defined as the junction temperature rise caused by power consumption
per unit of transistor (unit:°C/W).
In fact, driven by switch and impulse voltage, the temperature in the active area of
the transistor will go through some relaxation time (thermal time constant s) before
gradually reaching the steady state when the temperature field and thermal resistance
will become stable. Before that, all of them constitute a function showing exponential
change over time.
RTS is called transient thermal resistance, and s is related to thermal resistance and
thermal capacity of the material.
QT / Q0 pðLÞ ¼ Q0 expðLÞ
L ¼ UB =ðqkH EM Þ
Lifetime Prediction Method of Components Based on Failure Physics 235
Where, p(L) means the probability that the hole crosses the Si-SiO2 potential barrier
uB without being hit after a long enough distance L. The hole and total number of holes
injected into the oxide layer are in direct proportion to p(L). kH is the free path of holes
and QT is the charge injected into the oxide layer because of band-band tunneling. EM
is the maximum horizontal electric field.
ID ¼ A ES expðB=ES Þ
ES ¼ ðVDG 1:2Þ=3TOX
R
∵Q0 = ID dt , where ID is the GIDL tunneling current.
R
∴QT/ Q0 p(L) = ID dtexp(UB =ðqkH EM ÞÞ
The one-dimensional GIDL tunneling current model is as follows:
ID ¼ A ES expðB=ES Þ
ES ¼ ðVDG 1:2Þ=3TOX
In the equation, A is a constant, B amounts to 21.3 MV/cm, ES refers to the electric
field at the interface of the drain-gate overlap region, 1.2 is the minimum band cur-
vature that can occur with the tunneling at the vertical interface, and 3 is the dielectric
constant ratio of SiO2 to Si.
In addition, there are also some factors which have to be taken into account for their
huge impact on predicting the device life, such as corrosion, electrostatic discharge
damage, radiation damage, soft errors and packaging.
3 Acceleration Model
In the formula, AF(T,V) is the temperature and voltage acceleration factor; VUSE is
the voltage in actual use; VSTRESS is the voltage applied during test; coefficient B,
which varies by failure mechanism, has a default value of 1. Other parameters are the
same as those of the Arrhenius model.
In the formula, AF(T, rh) is the temperature and humidity acceleration factor; rhuse is
the humidity in actual use; rhstress is the test humidity; coefficient n is generally set at 3.
AF(T, V, rh) is the temperature, voltage and humidity acceleration factor. Other
parameters are the same as above.
Lifetime Prediction Method of Components Based on Failure Physics 237
4 Conclusion
This paper takes the PoF reliability analysis (TDDB, HCI, NBTI and off-state char-
acteristics are among the current mainstream methods) of semiconductor devices as the
starting point. The paper further offers a detailed summary of PoF-based methods to
predict the device life and their development history, in an effort to provide relatively
systematic reading materials for studying the device failure analysis and life prediction.
Furthermore, the paper gives a brief introduction to several acceleration models needed
and optional when predicting the device life as a supplement to the life prediction
methods.
References
1. Shi, B., Jia, X., et al.: Reliability of microelectronic devices (1999)
2. Miura, Y., Matukura, Y.: Jpn. J. Appl. Phys. 5(180) (1966)
3. Goetzberger, A., Nigh, H.E.: Proc. IEEE 54, 1454 (1966)
4. Berman, A.: Reliability physics symposium. In: 19th Annual Digital Object Identifier,
362997 (1981)
5. Ji, X.: Photoelectric coupling – a probe into simulating the switching type with MOSfield
effect transistor. Appl. Electron. Tech. 5 (1979)
6. Yao, L.: Reliability Physics. Electronic Industry Press, Beijing (2004)
7. Campbell, S.A.: The science and engineering of microelectronic fabrication. In: Zeng, Y.,
Yan, L., et al. (2nd edn). Electronic Industry Press, Beijing (2003)
8. Zhao, X., Wu, J., et al.: A study on semiconductor device life model based on failure
mechanism. Electron. Compon. Dev. Appl. 9(12), 69–71 (2007)
9. Wang, X.: TDDB study of thin gate oxide layer. Micronanoelectron. Technol. 6, 12–20
(2002)
10. Hicks, J., Bergstrom, D., et al.: 45 nm transistor reliability. Intel Technol. J. 12(2), 131–144
(2008)
11. Jian, C., et al.: The enhancement of gate-induced-drain-leakage (GIDL) current in short-
channel SOI MOSFET and its application in measuring lateral bipolar current gain b.
IEEE EDL 13(11), 572–574 (1992)
12. Cai, L.: Calculation of semiconductor device life. J. Hangzhou Dianzi Univ. 32(5), 312–314
(2012)
Spacecraft Technology and Application
Fault Tolerant Method for Spacecraft Bus
Based on Virtual Memory
1 Introduction
In recent years, the development of spacecraft has turned to the direction of intellec-
tualization, networking and integration. Autonomous health and task planning have
become the hotspots of spacecraft development. These functions are emerging with the
wide application of computers in spacecraft, such as bus technology. As an important
means of communication between onboard computers, 1553B bus is the key part of
spacecraft function realization. If the 1553B bus chip fails, it may cause partial or
overall bus communication function failure of the spacecraft, seriously affecting the
service and safety of the entire spacecraft.
1553B bus was first used in the field of aviation in 1980’s, and gradually became
the control bus widely used in aerospace area. The most famous bus controller chip is
the BU61580 which is designed by DDC. It is advanced communication engine
between onboard computer. High level reliability is needed because of its importance.
In the process of spacecraft design, the following methods are adopted to improve
the reliability of spacecraft:
(1) Multi-level redundant backup strategy. Spacecraft equipment usually adopts
double or even three backups to avoid the failure of a mission caused by the
failure of a certain device. Such as some function can be realized by both onboard
computer and ground control computer.
(2) The selection of high-level chip. Special design in onboard chip can avoid the
damage caused by space particles.
With the application of software on spacecraft, it is possible to adjust the config-
uration of system by modifying onboard software. In this paper, a fault-tolerant method
of spacecraft bus chip based on virtual memory is introduced to avoid RAM damage in
the bus chip, which is the same principle.
2 Principle
Secondly, it writes message control words into stack area. The physical RAM address
is 0x0000*0x00FF for A area and 0x0F00*0x0FFF for B area. To send a message,
the data is written into the message block according to the address specified in the
message control word, ranging from 0x108-0xEFB. Then the message count is changed
by modifying value of message count A or message count B. Finally, after sending, it
sets START register in order to start transmission the message, and judges the message
sending status by reading the value of special address in stack.
When sending bus message periodically, repeat the above steps.
In this procedure, the chip RAM is utilized as shown in Fig. 1 below.
Message Message
sending flow receiving flow
BU61580
Initial Initial
BU61580 Physical memory BU61580
address
Write subaddr Write subaddr
control data to REG and control data to
STACK RAM(0x0000-0x107) STACK
Change Message
RAM(0x101)OR Sending
Count
(0x105)
As shown in Table 1, there are 93 message blocks which store message data.
However, due to the limitation of stack size, BU61580 is able to send 64 messages at
most in one frame.
The special hardware designed for virtual memory is called memory management
unit (MMU). It performs efficient address mapping and translation. Due to MMU,
discontinuous and different types of physical addresses appear to be continuous and
single type.
The address space model of real-time system can be divided into four kinds: (1) plat
address space, which is original physical address; (2) equivalent single address space
(SAE), which maps physical address to virtual address directly; (3) single address space
(SAS), which maps physical address to virtual address point to point but maybe
physical address is discontinuous while virtual address is not; (4) multiple address
space (MAS), which maps physical address space to different processes and uses them
separately [3].
There are two parts in the fault-tolerant method for spacecraft bus: self-detection and
memory management.
Opportunity: The central computer on orbit, as the core of satellite data processing, is a
real-time multi-task system. It acts as BC unit in the bus network and the machine-time
is rather tense. Therefore, bus RAM self-detection should be carried out during the
minimum priority task, when other data processing tasks have been finished in a cycle.
Content: As shown in Table 1, there is fixed location area in bus chip, whose mapping
can’t be transformed. Message block area is located at 0x108-0xEFB physical address.
It occupies a large amount of internal RAM space in bus chip, but in practice, a small
Fault Tolerant Method for Spacecraft Bus Based on Virtual Memory 245
amount of the area is utilized. Mapping of the message block can be adjusted when
parts are damaged. Therefore, the content of self-detection is the message block area.
4 Verification
The fault-tolerant method for spacecraft bus chip based on virtual memory is applied on
a MEO satellite. The MEO space environment is harsh and the chip is greatly affected
by space radiation.
The normal traffic of bus communication is 0.119 Mbps, which includes: important
data storage, telemetry parameters, bus terminal status polling, and instructions. Similar
data is sent in the same frame, in which the routine telemetry parameters and bus
polling messages need to be sent once every 500 ms to 4 slave computers. In addition,
the telemetry acquisition parameters need to be strong real-time, so other messages can
be interrupted. The fault-tolerant method of bus chip RAM based on virtual memory is
used to manage message blocks.
After the satellite enters orbit, all pages (physical address of message block) are in
good self-detection status at the beginning of its life. When the remote telemetry
messages apply for addresses, there may be some pages occupied. The memory
management module assigns a page table for each message, as shown in Table 2.
The satellite has a breakdown of bus chip RAM after 36 months of satellite launch.
Some address reading and writing exceptions of 0x400*0x7FF in RAM area are
found. Write 0x5555 to each 16bit and read it out, then write 0xaaaa to each 16bit and
read it out. The results are shown in Fig. 3.
It can be determined in Fig. 3 that the 0x400-0x7FF area of the bus chip has been
damaged. After the bus RAM fault tolerance strategy based on virtual memory detects
the fault, the status of pages 23 (physical address 0x3E8*0x407) *55(physical
address 0x7E8*0x807) is set to be unavailable. The memory management module
assigns a page table for each message, as shown in Table 3.
As can be seen from Table 3, the fault-tolerant method of bus chip RAM based on
virtual memory can skip the area where reading and writing exceptions occur, and
provide reliable memory pages for CPU, so as to achieve the purpose of fault tolerance
of bus chip.
5 Conclusion
In this paper, a fault-tolerant method of on-orbit satellite bus chip based on virtual
memory is proposed. Without increasing hardware devices, the local fault of RAM can
be avoided autonomously on orbit by means of self-detection and page management of
the bus chip message physical address, and the reliability of satellite can be improved.
This method can be extended to other satellite equipments and ground systems to
improve the fault tolerance of the whole system.
References
1. ACE/Mini-ACE Series BC/RT/MT Advanced Communication Engine Integrated 1553
Terminal User Guide. DDC User Guide (1999)
2. Ling, Z., Zhang, L.: An implementation of the virtual memory management mechanism of the
embedded system. Sci. Technol. Eng. 10(27) (2010)
3. Xu, R.: Virtual memory techniques for real-time systems. Comput. Eng. Sci. 28(11), 116–118
(2006)
248 N. Zhao and W. E.
4. Sun, D., Wu, W., Zheng, L., Zhao, M., Li, B.: Design of a dynamic memory manager for
embedded real-time systems. J. Chin. Comput. Syst. 35(5) (2014)
5. Wang, X.: Design of multi-platform 1553B bus simulator. Comput. Measur. Control 20(2)
(2012)
6. Li, G.: Research of the Virtual Memory Management Technology on ARM. Beijing Jiaotong
University, Mater thesis (2013)
The Design of Interactive Framework
for Space-Exploration Robotic Systems
1 Introduction
Space is fundamentally one of the most challenging domains that humans have ever
tried to explore. For exoplanet soil/rock sampling tasks, restricted by incomplete initial
environment information, uncertain execution duration and large communication delay
[1], and so on. For this reasons, space- exploration spacecraft or robot is required to be
powerfully autonomous and reliable. However, it is limited by the current computing
and storage level of computers on spacecraft, the development of artificial intelligence
technology, and the high reliability and safety requirements of spacecraft missions,
making spacecraft have powerfully autonomous is a great challenge. Lightweight, low
volume, simple and convenient application need to be considered. This study, we
propose a design of interactive framework for application of space- exploration robotic
systems.
2 Related Work
With the continuous development of space action, the field of exploration is expanding.
Space-exploration spacecraft have also evolved from simplicity to complexity, from a
single mission to a variety of tasks, from relying mainly on manual operation to
gradually autonomous development. “Surveyor” (1966–198), “Luna16/20” (1970–
1972)”, “Luna24” (1976), “Viking” (1976), “Sojourner” (1997), “Hayabusal” (2003),
“Mars express” (2003), “Rosetta” (2004), “Curiosity” (2011), “Jade Rabbit” and so on.
These successful space exploration actions represent a significant trend.
In order to realize the goal of space exploration spacecraft/robot with autonomous
and reliability, a complete system framework is necessary. Base on the design of
framework, through experiments and knowledge learning, the intelligent level of
spacecraft is gradually optimized and improved to meet the application requirements of
various missions.
The US Air Force Research Laboratory (AFRL) Space Vehicles Directorate, the
team of Orbit Logic, PnP Innovations, and Emergent Space are developing an
Autonomous Planning System (APS) framework. The core of APS is specialized
Autonomous Planning Agents (SAPAs) and Master Autonomous Planning Agent
(MAPA), SAPAs generate plans, and the plans are integrated by [2].
Many artificial intelligence techniques have been used to learn and reason in a way
that humans describe the world [4, 5]. In the described method, a set of POs is
generated using a specific use case developed by the expert system, and these operators
are then optimized using an improved PRODIGY planner [6] and the PO is optimized
using a modification of the version space method named OBSERVER. In [7], a TRIAL
plan and learning strategy is proposed. It learns and optimizes POs when generating
plans and executing plans. Until the system deadlocks, external teachers are called to
participate in controlling the completion of tasks. In [8], a system architecture design
solution named EXPO was proposed. This system uses PRODIGY as a benchmarking
program to improve knowledge in multiple domains.
The characteristics of the space exploration mission determine that the robot system
has a high rate of undefined operations in an unknown environment, and it is extremely
important to be able to react and process quickly when encountering undefined oper-
ations. Therefore, we place greater emphasis on the online learning capabilities pro-
vided by the architecture. Teacher intervention occurs during the execution of the task,
providing only the operational instructions that need to be performed, and returning the
control of the task to the system after the operation. Our strategy for action guidance is
similar to other types of tele-operation technology [6]. The robot system provides a set
of functions (actions) in advance, and the teacher (served by the ground operator)
simply indicates which functions (actions) to perform in a given situation. After
instructing actions, the robot is allowed to perform actions and generate state transitions
for learning [9]. In particular, the teacher does not supervise the behavior and does not
provide clues as to the relevant attributes of the world that should appear in the
successful execution of the task [10]. This greatly reduces the number of human-
computer interactions, and it should be noted that, depending on the complexity of the
application, indicating all relevant properties of the world in every possible situation
The Design of Interactive Framework for Space-Exploration Robotic 251
faced by the robot may be complicated. To alleviate this burden, we propose a method
of automatically learning world-related attributes that should exist to allow POs to
perform successfully [11].
The framework design described in this paper provides two different modes of oper-
ation: the ground development mode and the task execution mode. The ground
development model is mainly used for training and improving system intelligence, and
the task execution mode is used for deep space exploration of spacecraft on-orbit task
execution. The core of the two model frameworks is the planning module and the
learning module. The planning module searches, generates, and optimizes the sequence
of motion controls needed to achieve the mission objectives based on established
execution rules and constraints. The learning module is used to interact with the
operator to create, edit, and repair these execution rules and constraints.
Schematic of the Interactive framework is shown in Fig. 1. The framework includes
planning module (PM), learning module (LM), Teaching module (TM), Operation
basic set module(OBM), Control rule set module (CRM), Emergency response module
(ERM), Control execution interface module (EIM), execution module (EM), environ-
ment Perception and modeling module (EA&M). The space-exploration spacecraft
system, the application examples corresponding to two different operation modes in the
framework are shown in Fig. 2.
(a) the ground development mode (b) the ground development mode
The ground development mode is used in the development stage before the mission
is executed. The model is shown in Fig. 2(a). The system consists of Substitute
spacecraft, Substitute DSN (Deep Space Network) and GSS (Ground support system).
The EM and the environment perception part of EA&M functions are provided by
Substitute spacecraft, and the EIM function is provided by Substitute DSN. The PM,
LM, TM, OBM, CRM, ERM and the environment modeling part of EA&M functions
are all included in the GSS. Initially, the OBM consists of a limited number of meta-
operations that described in accordance with the prescribed the declarative instruction.
The manual operator interacts with the LM of the system by the EIM to form the
original CRM. The PM is based on this CRM, receive task-command and planning to
generate an initial operation sequence, through the EIM controls the EM sequentially
complete the state transition specified by the operation sequence. Meanwhile the
EA&M continuously monitors and establishes or updates the environmental state
model, and feeds back the latest environmental state model to the PM, which executes
according to the feedback status update or control operation sequence. If the planning
module finds that the feedback status is undefined or does not find a solution in the
CRM, then the ERM is activated. The operator interacts with the LM of the system to
complete the CRM and exit the ERM. The PM continues to complete the control
execution of the planning and operation sequence until the task-command requirements
are met.
The mission execution mode is shown in Fig. 2(b). Unlike the ground development
model, the system consists of spacecraft, DSN, and GSS in mission execution mode. In
addition to EM and the environment perception part of EA&M functions, the spacecraft
is also loaded with PM, LM, TM, ERM and optimized OBM and CRM. The envi-
ronment modeling part of EA&M which need large computational complexity and
large storage requirements,is retained in the GSS. The GSS also includes a complete
ground development mode framework to synchronize Substitute spacecraft with the
mission spacecraft state.
The Design of Interactive Framework for Space-Exploration Robotic 253
When the system is running, the task commands received by the PM may be task
commands such as “detect and collect samples at target A”, requiring the system to
understand these task commands and translate them into instructions that control the
EM. The declarative description of the states of the world that currently exists is need
to be provided. These descriptions need to be unambiguous and can be programmed to
generate control sequences for controlling EM without confusion. Based on this, the
planner attempts to find a sequence of actions that converts the current state to a state
that satisfies the target of the task command specification.
Where ‘sp’ is the states of precondition requirements, ‘a’ is the action, ‘e’ is the
expected results, and ‘P’ is a probability estimate, which indicates the probability of
reaching the ‘e’ when ‘sp’ is observed and action ‘a’ is executed.
P ¼ Pr esp ; a ð2Þ
A PO encodes the changes that should be observed when an action ‘a’ is executed.
The ‘sp’ contains the initial values of the state descriptors changed by the action ‘a’,
while the ‘e’ contains their final values. The action ‘a’ describes the action that needs to
be executed. For example, a ‘PO’ for an action of move the robot end actuator from ‘A’
position to ‘B’ position by ‘default’ mode and description of ‘velocity’ may have the
following parts:
actuator with ‘default’ mode and ‘velocity’. And AtpositionðBÞ is a predicate with a
value true when the robot end actuator at ‘B’ position. In this case, the state descriptor
that changes with the action is AtpositionðAÞ. However, for this change to occur, it is
also necessary that robot is hold on, which is a fact specified by the descriptor
Ishold ðrobotÞ at remains unchanged by the action.
The PO provides a concise description of the action state and execution results. Of
all the descriptors associated with completing the task, the particular PO only considers
those descriptors associated with its effect, and other descriptors may be ignored or
ignored. For instance, if the PO in (3) is used in a task that also requires other actions,
e.g. the current position is not exactly at the ‘A’ position, a private PO,
pk ¼ fsk ; ak ; ek ; Pk g, is need be completed
This is because the a private PO, ‘pk’, not relevant to ‘move the robot end actuator
from ‘A’ position to ‘B’ position’. In general, the number of descriptors used in the
state representation is much larger than the number of descriptors used in a particular
PO [16].
In the proposed framework, not all of the POs are available for planning and some
will only be handled by the learning method. We make this distinction when needed.
The training instances for learning comprise state transitions of the form:
pt ¼ ð S t ; a t ; S t þ 1 Þ ð5Þ
Where t is the time step, st is the state before executing the action, at is the executed
action, and st þ 1 is the state after executing the action. A positive instance pt;iþ for the
PO pi ¼ fsi ; ai ; ei ; Pi g is defined as that where all of the predicates in the precondition
are observed in the state before the execution of the action, the action coincides with
the action of the state transition, and the predicates in the effect are observed in the state
after the execution of the action:
p
t;i ¼ fsi St ; ai ¼ at ; ei 62 St þ 1 g ð7Þ
description of the operation that was learned has been included in the CRM is defined
as Plearned .
The main algorithm includes two key steps algorithm,
and . FindPlan Algorithm can use binary tree search
algorithm, here we focus on the Learned algorithm.
Learning and planning are the most important components of the interactive frame-
work. The following sections focus on these two aspects.
4.1 PO Evaluation
How to evaluating each PO to determine its probability in Eq. (2) is a very important
issue. In our application, the evaluation of each PO is strictly limited to the current state
and the risk in the implementation of the PO. we divide the PO in the task into two
categories.
256 W. Shi et al.
The first category, pdef ¼ sdef ; adef ; edef ; P , pdef can be verified by a large number
of tests, And when sdef 2 St ¼ et1 , Calculate P according to the formula
nþ
P¼ ð8Þ
n þ þ n
Where X describes the robot’s own configuration description and Yt describes the
configuration of the moving part on the robot in the current state. When the function is
0, the moving part of the robot is described in the current state. No contact is made with
the body structure. In addition, we also need to define the function:
(
0; 8ei 2 E; 8yj 2 Yt ; minxi 2X;yj 2Yt y2j e2i [ C
GðE; Yt Þ ¼ ð10Þ
D; other
4.2 PO Generation
The states processing rules contained in CRM are limited and cannot completely cover
or cope with complex and variable state transitions in real tasks. PM search generation
POs that reflect the best state transition is a typical optimal path planning problem.
Finding the optimal path in a large number of states space requires a lot of computational
time investment. In order to improve system reliability and reduce time investment,
ERM is started to generate new POs in two cases. (1) When planning, it is found that
there is undefined actions definition, and the corresponding operation rules need to be
added in the CRM. (2) When the execution of a PO has an unexpected effect.
sc ¼ fsi 2 St jsi 62 St þ 1 g
ð13Þ
ec ¼ fsi 2 St þ 1 jsi 62 St g
Where sc and ec are the sets of the changed descriptors, respectively, St is the state
before and St þ 1 is the state after executing the action. Then, a new PO, pinst , is created
by using sc , ainst and ec as the precondition, action, and effect parts, respectively. The
new PO becomes available immediately for planning. However, the new PO only
considers the changed descriptors in its precondition, so some unchanged causative
descriptors may be missing. In this case, the execution of the newly generated PO may
have an unexpected effect.
if such a candidate pc can be found, then add pc to the CRM. Since eexe is an unexpected
state, such a candidate OP cannot be found which is a high probability event to bring the
system control into a transition state. The transition state is generally selected from the
safe parking state, which is a limited state set Sinterim . And the safe
parking state selection
strategy is handled by the proximity principle sk ¼ minsi 2Sinterim e2exe s2i ,The problem
turns into finding a PO set :
ð14Þ
5 Conclusion
References
1. Colby, M., Yliniemi, L., Tumer, K.: Autonomous multiagent space exploration with high-
level human feedback. J. Aerosp. Inf. Syst. 13(8) (2016)
2. Ella, H., Josh, N., Marc, S., Doug, G., Tim, E., Ken, C.: Onboard Autonomous Planning
System, SpaceOps Conferences 5–9 May 2014, Pasadena, CA SpaceOps 2014 Conference
(2014)
3. Truszkowski, W.F., Hinchey, M.G., Rash, J.L., Rouff, C.A.: Autonomous and autonomic
systems: a paradigm for future space exploration missions. IEEE Trans. Syst. Man Cybern.
Part C: Appl. Rev. 36(3), 279–291 (2006). https://doi.org/10.1109/tsmcc.2006.871600
4. Argall, B.D., Chernova, S., Veloso, M., Browning, B.: A survey of robot learning from
demonstration. Robot. Auton. Syst. 57(5), 469–483 (2009)
5. Veloso, M., Bala, J., Boloedorn, E., Bratko, I., Cestnik, B., Cheng, J., De Jong, K., Dzeroski,
S., Fisher, D., Fahlman, S., et al.: The NONK’s problems: a performance comparison of
different learning algorithms, Technical report CMU-C5-91-197. Carnegie-Mellon Univer-
sity (1991)
6. Veloso, M., Carbonell, J., Perez, A., Borrajo, D., Fink, E., Blythe, J.: Integrating planning
and learning: the prodigy architecture. J. Exp. Theor. Artif. Intell. 7(1), 81–120 (1995)
7. Benson, S.: Inductive learning of reactive action models. In: Proceedings of the Twelfth
International Conference on Machine Learning, pp. 47–54. Morgan Kaufmann (1995)
8. Gil, Y.: Learing by experimentation: incremental refinement of incomplete planning
domains. In: Proceedings of the Eleventh International Conference on Machine Leaning
(1994)
9. Rybski, P., Yoon, K., Stolarz, J., Veloso, M.: Interactive robot task training through dialog
and demonstration. In: 2007 2nd ACM/IEEE International Conference on Human-Robot
Interaction (HRI), pp. 49–56. IEEE (2007)
10. Nicolescu, M., Mataric, M.: Natural methods for robot task learning: instructive
demonstrations, generalization and practice. In: Proceedings of the Second International
Joint Conference on Autonoumous Agents and Multiagent Systems, pp. 241–248. ACM
(2003)
11. Polushin, I.G., Dashkovskiy, S.N.: A small gain framework for networked cooperative
teleoperation. In: Proceedings of the 8th IFAC Symposium on Nonlinear Control Systems,
pp. 90–95. Bologna, Italy (2010)
12. Agostini, A., Torras, C., Worgotter, F.: Efficient interactive decision-making framework for
robotic applications. J. Artif. Intell. 247, 187–212 (2017)
Research and Design of Hierarchical FDIR
in Spacecraft
1 Introductions
at least keep satellite safe until it enters visible area again, so that the ground control
system can intervene and handle the failure.
Fault detection, isolation and recovery (FDIR) to ensure the safe and reliable
operation of spacecraft has become an indispensable content of spacecraft.
The concept of hierarchical FDIR was proposed by R. Gessner of EADS-Astrium
[3], which was extracted and summarized from the experience of multiple projects of
ESA. In China’s satellite design, satellite safe mode has become an important method,
but the concept of systematic and hierarchical FDIR design has not yet been formed.
Based on the project experience of remote sensing satellite of CAST2000 platform, this
paper elaborates the design concept of hierarchical FDIR of satellite. Through the
centralized management at system level, autonomous disposal at subsystem level and
redundant design of equipment level, it optimizes the parameters of software and
hardware design of satellite FDIR and realizes the FDIR architecture at satellite system
level.
The article [3] shows the design flow of ESA. The Fig. 1 shows the design flow of
CAST2000 satellite. This figure based on the ESA design flow and mixed the typical
CAST develop phase. The main flow is similar. Our design flow focuses on project
development instead of study or science experiment. Also for project, we emphasis
heritage, so the study phase is simplified to mission analysis phase. In this phase, we
should identify a basis heritage project and technical status changes. For the second
phase, instead of ESA definition phase, we define design phase including preliminary
design and critical design. The develop and test phase is for the FDIR implementation
and test in different levels. In orbit phase, we have same FDIR activity as ESA.
Mission analysis phase: according to the mission requirements, the requirements of
autonomous satellite processing and the preliminary FDIR requirements at the system
level are proposed.
Research and Design of Hierarchical FDIR in Spacecraft 263
The second step is FDIR analysis, which including system failure handling strat-
egy, subsystem/unit level failure list based on FMEA, design of system and
subsystem/unit interface, software design. The main work is close-loop the FMEA
result by FDIR analysis. The following table shows the example of unit level FMEA
and its FDIR requirement to system level (Table 3).
• Level 3 is system handling layer, that is, when subsystem fails and being judged by
the onboard computer, the fault subsystem is disposed. So that the subsystem is
restored to normal without affecting the whole satellite mission.
• Level 2 is subsystem handling layer, that is, when fault occurs under subsystem
level and the subsystem (sub-computer) independently judges, the fault processing
is completed within the subsystem, without affecting the subsystem functions.
• Level 1 (Bottom) is unit handling layer, that is, when unit (software/hardware)
autonomous judge fault occurring, and the fault processing is completed within the
unit, so that the unit can recovery without affecting the unit function.
Each FDIR layer has specific definition of I/O interface. The input interface is the
alarm symbol or specific TMs of lower layer. The output interface has two kinds: one is
sending the alarm symbol or specific TMs upper layer, the other one is sending tele-
command to lower layer (switch off payload or switch to redundant unit, etc.).
Health Status Monitor. OBC is the core unit of FDIR. In order to detect fault as soon
as possible, the OBC could monitor the system and subsystem by check the “health
status symbol” of each subsystem. The health status monitor is designed as hierarchical
design: lower level generates health status symbol to upper layer. OBC is on the top
layer and monitor the whole satellite health status (Table 5).
References
1. Rabenau, E. et al.: Mars express mission planning – expanding the flight box in flight. In:
SpaceOps2010 Conference, Huntsville, USA, 25–30 April 2010
2. Yang, W.: A control model for satellite orbit maintenance. Chin. Space Sci. Technol. 1, 11–15
(2001)
3. Gessner, R., et al.: Hierarchical FDIR Concepts in S/C Systems [EB/OL]. http://arc.aiaa.org,
https://doi.org/10.2514/6.2004-433-249
4. Lang, L.E., et al.: Research and design of FDIR techniques for satellite avionics. Comput.
Eng. Des. 35, pp. 2607–2611 (2014)
5. Jiang, L., et al.: Fault Detection, isolation and recovery design for micro-satellites. In: 2016
Sixth International Conference on Instrumentation & Measurement, Computer, Communi-
cation and Control, Harbin, China, 21–23 July 2016
6. Schmidt, M., et al.: The ECSS Standard on Space Segment Operability. SpaceOps2004,
Montreal, Canada (2004)
Research on Spacecraft Network Protocol
Based on Space Packet
1 Introduction
With the wide application of space technology and the diversification of spacecraft
functions, space communication has evolved from spacecraft-ground to inter-satellite
links. The information exchange relationship within and between the spacecraft is
increasingly complicated. The unified information network service requires the multi-
spacecraft, multi-application process in spacecraft, and multi-user task coordination. It
requires integrated network communication among devices to provide reliable opera-
tion and management for the integrated network [1]. For this reason, CCSDS has
extracted the concepts of sub-package telemetry (TM), sub-package remote control
(TC) and advanced on-orbit system package (AOS). The latest research results in this
area is standard space packet protocol (SPP), it solved the limitation that packet uti-
lization standard (PUS). The PUS service is too closely related to TM and TC, only
focused on satellite-earth link interaction [2]. The SPP can enable space missions to
efficiently transmit space application data with different types of characteristics in a
network including satellite-ground and inter-satellite links, and can realize intercon-
nection and intercommunication between spacecraft and spacecraft under CCSDS
standards.
This paper studies the application of space packet protocol, explores and proposes data
exchange strategies among satellite, earth, inter-satellite and intra-satellite information
nodes, which is helpful to promote the unification of protocol and information sharing,
promote the integration of resources in various fields, and improve the space application
capability and efficiency.
2.1 Architecture
In the traditional spacecraft on-board data network, space packet protocol is located at
the network layer. The subnet protocol only includes the on-board subnet protocol,
which uses packet service, memory access service, equipment discovery service, etc. to
realize SOIS data service, uses the convergence sublayer to realize the mapping
between data link and data service, and uses 1553B, space wire and other on-board
subnet protocols to realize the data link layer and physical layer. With the joining of
intra-satellite links, the subnet protocol is extended to UDP/IP and other internet
protocols. As the carrier of asynchronous message transmission and other services in
the application support layer, the space packet protocol can realize the subscription and
distribution of messages in the space internet and on-board equipment networks.
The interaction process of space packets in the network is shown in Fig. 1. The path
from the source to one or more destination through the subnet is called the logical data
path (LDP) [8]. The path id, including the source, the terminal and one or more subnets,
authenticates each LDP uniquely. When the application data passes through the LDP
subnet, the protocol provided by the subnet layer is used. The accessor carries the
subnet when the path is inside the spacecraft and accesses the space subnet when the
path passes through the space internet. The protocols used in the on-board subnetwork
and the space subnetwork are independent of each other, and the logical data paths
between the subnetworks can be configured and implemented through the management
system to map the relationships between the protocols.
Due to the development process, PCM remote control system and CCSDS standard
remote control space link protocol [9] are mainly used in the uplink of spacecraft space-
to-ground links, AOS [10] is mainly used in the spacecraf-ground to inter-satellite
links, and the adjacent space link protocol is being preliminarily applied. The internal
communication bus of the spacecraft also presents diversified development trends,
including 1553B bus and CAN bus at low speed, 1394 bus, SpaceWire bus and
Ethernet bus at high speed. When different spacecrafts interoperate or share informa-
tion, space packet protocol, as a protocol data unit with the smallest granularity, has the
advantages of good compatibility with various protocols, small resource overhead and
flexible customization, and can assume the responsibility of generalization of interfaces
Research on Spacecraft Network Protocol Based on Space Packet 271
between spacecraft in various fields. The following analyzes and illustrates the appli-
cation methods of space packet protocol from the perspective of compatibility between
different layers of space packet protocol and network architecture.
4 Verification
The global navigation system is composed of dozens of satellites, which often have
inter-satellite communication links, and can measure and control all satellites through
limited ground stations. For remote sensing satellites, a limited number of ground
stations means a limited observation time, because remote sensing satellites generally
do not need to form constellations. The spacecraft network protocol based on space
package proposed in this paper can enable remote sensing satellites, navigation satel-
lites and ground stations to support each other at the network layer of the spacecraft
network, thus enabling all kinds of satellites to cooperate with each other and give play
to their respective advantages (Fig. 4).
When the ground user initiates the TC operation of the spacecraft, any satellite node
can be selected to inject the TC space package according to the visibility. The earth-
visible satellite node extracts the space packet APID naming domain, selects the link
network to route to the destination node according to the space internet network layer
protocol, the destination node identifies the space packet APID, searches for corre-
sponding equipment through an addressing mechanism, transmits the equipment-borne
subnet to the terminal system, and executes the instruction unit in the space packet by
the application process in the terminal system.
When the remote sensing satellite and navigation satellite work together, the
spacecraft and gateway learn the APID naming domain of service providers and service
requestors through asynchronous message transmission services to form address
Research on Spacecraft Network Protocol Based on Space Packet 275
routing tables in configuration servers, registrants, and nodes. At the same time, service
demanders need to bind the auxiliary guide of the space package and the analysis rules
of the data domain generated by the service provider’s application process.
The network protocol based on space packet proposed in this paper will inevitably
become the basis of interactive support in future space-based networks due to its small
field overhead, strong scalability, good compatibility with link protocols at various
stages, and taking into account the characteristics of low-speed and high-speed link
performance. Due to the flexible application of the space packet protocol, there are
differences in the current use methods in various models. This paper starts with the
application methods of the space packet protocol, makes clear the characteristics and
the addressing mechanism of the space packet protocol, and studies the interaction
methods between the space packet protocol and the carrier subnet, the space subnet and
the application support layer. It is suggested that in future spacecraft design in various
fields, the space packet protocol should be adopted to construct the integration network
interoperability model of space and ground to achieve the goal of resource integration
and interactive support among different types of spacecraft.
References
1. Zhao, H.: Building an intelligent and easy way for spacecraft with integrated electronic
technology. Spacecraft Eng. 24(6), 1–6 (2015)
2. Merri, M., Cooper, S.: What has CCSDS SM&C to do with ECSS PUS? In: SpaceOps 2010
Conference (2010)
3. Zhang, Q., Guo, J., et al.: Space data system. China science and technology publishing
house, p. 118, July 2016
4. He, X.W., Zhu, J., et al.: Application method of spaceborne standard interface service in
spacecraft. Spacecraft Eng. 24(6), 52–58 (2015)
5. CCSDS713.0-B-1 Space Communication Protocol Specification (SCPS)-network protocol
(SCPS-NP). Washington DC: CCSDS (1999)
6. CCSDS 702.1-B-1. IP over CCSDS Space Links. Washington D.C.: CCSDS (2012)
7. Scott, K.,Burleigh, S.: Bundle protocol specification [EB/OL], 15 November 2015. http://
tools.ietf.org/html/rfc5050
8. CCSDS 133.0-B-1 SPACE PACKET PROTOCOL. Washington D.C.: CCSDS (2003)
9. CCSDS 232.0-B-3 TC space data link protocol. Washington D.C.: CCSDS (2015)
10. CCSDS 732.0-B-3 AOS space data link protocol. Washington D.C.: CCSDS (2015)
Research on Instrument Requirements
and Configuration for High Resolution
Infrared Observations
Abstract. High resolution infrared remote sensing can realize all-day, non-
contact, high-precision temperature measurement such as the atmosphere, water
and other objects, which is an important means of environmental monitoring and
resource detection. But High-resolution infrared data are relatively scarce data
sources, especially thermal infrared. This paper introduces the research on
instrument requirements and configuration for high resolution infrared obser-
vations. By using large aperture and integrated optical system combined with
linear push broom, high resolution multi-spectral imaging can be realized which
covers multiple detection channels from visible light, short wave infrared, mid-
wave infrared and long wave infrared spectral bands, and remote sensing data
with a panchromatic resolution better than 1 m and an infrared resolution better
than 5 m can be obtained, which can serve environmental protection, mineral
exploration, and urban remote sensing after ground processing. Through the
research in this paper, it provides a reference for the design and engineering of
high resolution infrared payload.
1 Introduction
High-resolution mid-wave and long wave infrared data are relatively scarce data
sources. In the field of terrestrial observation in China, most of the payloads are limited
in some applications such as environmental protection, mineral exploration, and urban
remote sensing due to the lack of infrared spectrum, especially the long-wave infrared
spectrum. The spatial resolution of ASTER short-wave and long-wave infrared
developed by Japan is 30 m and 90 m respectively, which is one of the sources of our
infrared data, but the payload has been invalidated. In this paper, the concept of high-
resolution from visible light to thermal infrared multi-spectral integrated camera pay-
load is proposed from the perspective of Instrument Characteristics.
2 Instrument Requirements
This chapter analyzes the basic spatial resolution, swath, spectral bands and accuracy
requirements from the application in environmental protection, mineral exploration,
urban remote sensing and other fields.
2.2 Swath
The imaging swath is mainly designed to meet the needs of fast revisiting and global
coverage. For high-resolution payloads, the width is limited by the detector, and it is
generally difficult to achieve a large swath.
Among the international mainstream of the same type of payload, the MTI which
has the same resolution has a swath of only 12 km, and the Worldview-3 with a short-
wave resolution of 3.7 m is only 13.1 km wide. Among the low and medium resolution
satellites, Landsat-8 has a swath of 185 km, but its corresponding short-wavelength and
long-wavelength infrared spectrum resolution is only 30 m/100 m. ASTER has a swath
of only 30 km when the short-wavelength infrared spectrum resolution is 30 m and the
long-wavelength infrared spectrum resolution is 90 m.
For urban black and odorous water, straw burning, the warm water discharge in
nuclear power plant, evaluation of urban heat island effect and other applications, since
the target scale is relatively small, the 30 km imaging swath capability can meet the
requirements, which can cover monitoring targets and sensitive areas of the sur-
rounding environment. However, for large inland water bodies such as Taihu Lake,
large-scale monitoring areas such as nature reserves, they require large-scale obser-
vations superior to 100 km.
278 Q. Wang et al.
2.4 Accuracy
The distribution of surface temperature is an important research object in the fields of
environmental monitoring, urban heat island, vegetation ecology, etc. It is also an
important parameter in remote sensing models such as surface flux, soil moisture
content and crop estimation. According to the research of application requirements, the
accuracy of surface temperature inversion is better than 1K.
temperature should have the highest emissivity for most natural surfaces. The emis-
sivity value between 8 and 10 lm is relatively low, and the volatility is also higher than
the value of the 10 to 13 lm interval [4].
In summary, the atmospheric window of the main infrared observation in the
thermal infrared band is mainly 8–9.4 lm and 10–12.5 lm. These two windows are the
most concentrated bands of thermal radiation energy at room temperature. The
detection information mainly reflects the emissivity and temperature characteristics of
the ground objects. Among them, the window has a high atmospheric transmittance and
a higher surface emissivity in the 10–12.5 lm window.
Long-wave infrared data needs to be inverted to obtain the surface temperature and
emissivity information required for the application. The accuracy of different inversion
methods is different. There are three commonly used surface temperature inversion
methods. One is single-channel algorithm (SC) which only contains one thermal
infrared channel, the other is dual channel algorithm or split window algorithm
(SW) which requires two thermal infrared channels, the third is temperature and
emissivity separation algorithm (TES) which requires multiple thermal infrared bands,
at least three or four thermal infrared bands.
Through the configuration of multiple thermal infrared spectral bands, the tem-
perature emissivity separation algorithm can be used to simultaneously invert the
surface temperature and emissivity. For example, the TES algorithm is proposed for
ASTER data. The ASTER sensor has five thermal infrared spectral bands, so we can
learn from ASTER’s thermal infrared spectrum configuration and add two variations to
improve the ASTER configuration [5]. First of all, The spectrum around 8.3 lm is not
considered because of the atmospheric effects. Second, The spectral segment located at
11.3 lm is placed with 12 lm, and the spectral segment of 10.6 lm is combined to use
the split window algorithm, since the 13 and 14 spectral bands of ASTER are too close
to fit the split window algorithm.
In summary, the thermal infrared is configured with four spectral bands which are
8.475–8.825 lm, 8.925–9.275 lm, 10.3–11.3 lm, and 11.5–12.5 lm.
Research on Instrument Requirements and Configuration 281
Scanning
• Smaller scale detector
• Easy to achieve large field of view
Application requirements Instrument requirements • Short cell dwell time, low signal to noise ratio
• Satellite attitude/stability is affected
• 5m IR resolution
• 30km swath • 6000~8000 pixels
• 1K temperature • 0.05K NETD Pushbroom
inversion accuracy • Higher signal to noise ratio
• Relatively simple system / high reliability
• High geometric accuracy of image data
• Small impact on satellite platforms
• Long line array infrared detector is needed
5 Conclusion
In this paper, the concept of high-resolution visible light to thermal infrared multi-
spectral integrated camera payload is proposed. From the spatial resolution, the infrared
multi-spectral has a resolution better than 5 m, and the panchromatic has a resolution
better than 1 m. From the spectral configuration, it covers multiple detection channels
from visible, short-wave, medium-wave to long-wave. Simultaneous phase observation
of different spectral segments is realized on a remote sensor, which improves the
information content and accuracy of the data.
References
1. Sobrino, J.A., Del Frate, F., Drusch, M., Jiménez-Muñoz, J.C., Manunta, P., Regan, A.:
Review of thermal infrared applications and requirements for future high-resolution sensors.
IEEE Trans. Geosci. Remote Sens. 54(5), 2963–2972 (2016)
2. Hook, S.J.: NASA 2014 The Hyperspectral Infrared Imager (HyspIRI) – Science Impact of
Deploying Instruments on Separate Platforms, California (2014)
3. Sobrino, J.A., Jiménez-Muñoz, J.C.: Minimum configuration of thermal infrared bands for
land surface temperature and emissivity estimation in the context of potential future missions.
Remote Sens. Environ. 148, 158–167 (2014)
4. Lagouarde, J.-P., Bach, M., Sobrino, J.A., Boulet, G.: The MISTIGRI ther-mal infrared
project: scientific objectives and mission specifications. Int. J. Remote Sens. 34(9–10), 3437–
3466 (2013)
5. Ramsey, M.S., Realmuto, V.J., Hulley, G.C., Hook, S.J.: HyspIRI Thermal Infrared
(TIR) Band Study Report, California (2012)
6. Johnson, W.R., Hook, S.J., Foote, M.: Infrared instrument support for HyspIRI-TIR. Proc.
SPIE 8511(23), 02 (2012)
A Method for Solving Generalized Implicit
Factorization Problem
Abstract. The problem of factoring RSA moduli with the implicit hint was
firstly proposed by May and Ritzenhofen at PKC’09 where unknown prime
factors of several RSA moduli shared some number of least significant bits
(LSBs), and was later considered by Faugère et al. where some most significant
bits (MSBs) were shared between the primes. Recently, Nitaj and Ariffin pro-
posed a generalization of the implicit factorization problem. Let N1 ¼ p1 q1 and
N2 ¼ p2 q2 be two distinct RSA moduli, Nitaj and Ariffin showed that when a1 p1
and a2 p2 share enough bits, N1 ; N2 can be factored in polynomial time, where a1
and a2 are some unknown positive integers. They also extended their work to the
case of k ð 3Þ moduli. In this paper, we revisit Nitaj-Ariffin’s work and
transform the problem into solving small roots of a modular equation. Then by
utilizing Coppersmith’s method, for the case of two moduli we improve Nitaj-
Ariffin’s result when the unknowns a1 ; a2 are relatively small, and our result is
always better than Nitaj-Ariffin’s result for the case of k ð 3Þ moduli.
1 Introduction
RSA [13] is one of the most widely deployed public-key cryptosystem. Its security
relies on the difficulty of factoring large composite integer. A brief description on the
key generation algorithm of the RSA scheme is given as follows:
Key Generation of RSA: Let N ¼ pq be an RSA modulus, where p and q are primes
of the same bitlength. Randomly choose an integer e such that gcdðe; uðNÞÞ ¼ 1,
where uðNÞ ¼ ðp 1Þðq 1Þ and calculate d such that ed 1ðmod uðNÞÞ by the
Extended Euclidean Algorithm. The public keys are N and e, and the private key are p,
q and d.
Although the problem of constructing efficient algorithm to factor N ¼ pq has been
studied for decades, there are still no polynomial time algorithms known except for
quantum algorithm. And it is believed as a mathematical hard problem in computational
number theory. However, the factorization may become feasible in polynomial time
under several special cases, such as small decryption exponent attacks [1, 15].
In 2009, May and Ritzenhofen [10] proposed the problem of factoring RSA moduli
with the implicit hint and showed that if the unknown prime factors of several RSA
moduli shared enough number of LSBs, the moduli can be factored in polynomial time.
Later, Faugère et al. [4] extended the problem to the case of unknown prime factors
shared MSBs. In 2011, Sarkar and Maitra [14] related both May-Ritzenhofen’s work and
Faugère et al.’s work to computing approximate common divisors problem and slightly
improved previous results. Since then, the methods for solving implicit factorization
problem has been well studied and deeply understood by many researches [7, 12].
Recently, Nitaj and Ariffin [11] proposed a generalization of implicit factorization
problem. Suppose that Ni ¼ pi qi are k distinct n-bit RSA moduli with an-bit qi where
a 2 ð0; 1Þ: If there exist kð 2Þ unknown positive integers ai satisfy that ai pi share tn
LSBs or MSBs where t 2 ð0; 1Þ and ai 2bn where b 2 ð0; 1Þ for i = 1, …, k. Then
Nitaj and Ariffin showed that Ni can be factored in polynomial time if
t 2a þ 2b þ 1; for k ¼ 2;
k k 2 ð1Þ
or t [ k1 a þ k1 b; for k 3:
2 Preliminaries
In 1996, Coppersmith [2, 3] successfully applied the L3 lattice basis reduction algo-
rithm to find small roots of modular equations, typically called Coppersmith’s method.
To describe the sketch of Coppersmith’s method, we first give a brief review on the
definition of lattices. Let L be a lattice which is spanned by k linearly independent
vectors v1 ; . . .; vk 2 Zn . Namely, lattice L is composed by all integer linear combina-
tions, c1 v1 þ . . . þ ck vk , of v1 ; . . .; vk , where c1 ; . . .; ck 2 Z. Then the set of vectors
v1 ; . . .; vk is called a lattice of L and K is the lattice dimension of L. Moreover, there
will be infinite lattice bases for any lattice L whose dimension is greater than 1.
In 1982, A.K., Lenstra, H.W. and Lovász, L. introduced the famous L3 lattice basis
reduction algorithm to find a lattice basis with good properties in polynomial time.
286 Z. Sun et al.
Lemma 1. (L3 , [6, 9]) Let L be a lattice of dimension k. Applying the k algorithm to
the basis of L, the output reduced basis vectors v1 ; . . .; vk satisfy that
kðkiÞ 1
kvi k 24ðk þ 1iÞ detðLÞk þ 1i ; for any 1 i k: ð2Þ
We also state a useful lemma from Howgrave-Graham [5] which gives a sufficient
condition to transform a modular equation into an integer equation. We define the norm
P
of a polynomial gðx1 ; . . .; xn Þ ¼ ði1 ;...;in Þ ai1 ;...;in xi11 . . .xinn as
X 1
kgðx1 ; . . .; xn Þk ¼ ð a2
ði1 ;...;in Þ i1 ;...;in
Þ2 : ð3Þ
Moreover, since the obtained polynomials he1 ðx1 ; . . .; xn Þ; . . .; hen ðx1 ; . . .; xn Þ are
some integer combinations of the polynomials hi ðx1 ; . . .; xn Þ which are used to con-
struct lattice, he1 ðx1 ; . . .; xn Þ; . . .; hen ðx1 ; . . .; xn Þ have the same roots ( xe1 ; . . .; xen ) modulo
pm . Then if the norm of he1 ðx1 ; . . .; xn Þ; . . .; hen ðx1 ; . . .; xn Þ satisfy the second condition
kðk1Þ
p ffiffi
1 m
of Lemma 2, namely if 24ðk þ 1nÞ detðLÞk þ 1n \ p k
we have that hold over the integers.
e e
Here we ignore h1 ðx1 ; . . .; xn Þ ¼ 0; . . .; hn ðx1 ; . . .; xn Þ ¼ 0 small terms and only
simply check whether detðLÞ\pmk does hold or not. Then based on the following
A Method for Solving Generalized Implicit Factorization Problem 287
heuristic assumption, we can solve the roots xe1 ; . . .; xen from the polynomials
h1 ð xe1 ; . . .; xen Þ ¼ 0; . . .; hn ð xe1 ; . . .; xen Þ ¼ 0.
a1 N1 ¼ q1 ðp þ 2tn ~p1 Þ;
. . .. . . ð7Þ
ak Nk ¼ qk ðp þ 2tn ~
pk Þ:
Since Eqs. (1) hold, the vector v ¼ aq1 ; aaa12q2 ; . . .; aaa1kqk 2 L1 . According to
Gaussian heuristic, the length of the shortest non-zero vector of L1 is
rffiffiffiffiffiffiffiffi
k 1 ðk1Þtn
detðL1 Þk ffi 2 k : ð9Þ
2pe
ðk1Þtn
Then with a good possibility, the vector v is the shortest vector when kvk 2 k ,
k2
namely t k1
k
a þ k1 b. By applying L3 reduction algorithm to the lattice L1 , the
vector v ¼ aq1 ; aaa12q2 ; . . .; aaa1kqk can be found out, which means one can obtain some
unknown multiples of q1 ; . . .; qk . Then one can easily factor Ni by computing greatest
k2
common divisors. However, when t\ k1 k
a þ k1 b, the reduced basis ðk1 ; . . .; kk Þ
doesn’t contain vector v. For this case, we can represent the vector v into the form with
a linear combination of reduced basis. Namely, there exist integers x1 ; . . .; xk such that
v ¼ x1 k1 þ . . . þ xk kk . Moreover, we can obtain following modular equations
where ai is an integer satisfying ai lij mod Nj for 1 j k and can be calculated from
lij and Nj .
To solve the above homogeneous linear modular equation, we utilize the following
theorem introduced by Lu et al. [8].
Theorem 2. LetN be a sufficiently large composite integer (of unknown factorization)
with a divisor p p N b . Furthermore, let f ðx1 ; . . .; xn Þ 2 Z ½x1 ; . . .; xn be a homoge-
neous linear polynomial in nðn 2Þ variables. Under Assumption 1 we can find all the
solutions ðy1 ; . . .; yn Þ of the equation f ðx1 ; . . .; xn Þ 0 (mod p) with
gcdðy1 ; . . .; yn Þ ¼ 1, and jyi j N ci for i ¼ 1; . . .; n if
Xn n
1
i¼1
c i 1 ð1 bÞ n1 nð1 bÞ 1 ð1 bÞn1 : ð12Þ
For the homogeneous linear equation in k variables modulo q1 . . .qk ffi ðN1 . . .Nk Þa ,
by Theorem 2 with the variables xi \ðN1 . . .Nk Þd ffi 2kdn , for i ¼ 1; . . .; k, we can solve
the variables when
k
1
kd 1 ð1 aÞk1 kð1 aÞ 1 ð1 aÞk1 ; ð13Þ
namely,
t ð k 1Þ k
1
a þ kb 1 ð1 aÞk1 kð1 aÞ 1 ð1 aÞk1 : ð14Þ
k
k k
1
t a þ kb 1 þ ð1 aÞk1 þ kð1 aÞ 1 ð1 aÞk1
k1
ð15Þ
1 k2
¼ k ð1 aÞ 1 ð1 aÞk1 þ b:
k1
we can find the vector v, then we can factor Ni for i ¼ 1; . . .; k. This concludes the
proof of our Theorem 1.
Our method also can be extended to the case of the multiples of primes share MSBs
and we can obtain the same bound. As it is shown, for the case of two moduli, i.e.
k ¼ 2, our result is better than Nitaj-Ariffin’s result when the unknowns a1 ; a2 are
relatively small, and our result is always better than Nitaj-Ariffin’s result for the case of
kð 3Þ moduli.
4 Conclusion
References
1. Boneh, D., Durfee, G.: Cryptanalysis of RSA with private key d less than N0.292. IEEE
Trans. Inf. Theory 46(4), 1339–1349 (2000)
2. Coppersmith, D.: Finding a small root of a univariate modular equation. In: EUROCRYPT
1996, pp. 155–165 (1996)
3. Coppersmith, D.: Finding a small root of a bivariate integer equation factoring with high bits
known. In: EUROCRYPT 1996, pp. 178–189 (1996)
4. Faugère, J.-C., Mariner, R., Renault, G.: Implicit factoring with shared most significant and
middle bits. In: PKC 2010, pp. 70–87 (2010)
290 Z. Sun et al.
5. Howgrave-Graham, N.: Finding small roots of univariate modular equations revisited. In:
Cryptography and Coding 1997, pp. 131–142 (1997)
6. Lenstra, A.K., Lenstra, H.W., Lovász, L.: Factoring polynomials with rational coefficients.
Math. Ann. 261(4), 515–534 (1982)
7. Lu, Y., Peng, L., Zhang, R., Hu, L., Lin, D.: Towards optimal bounds for implicit
factorization problem. In: SAC 2015, pp. 462–476 (2015)
8. Lu, Y., Zhang, R., Peng, L., Lin, D.: Solving linear equations modulo unknown divisors:
Revisited. In: ASIACRYPT 2015, Part I, pp. 189–213 (2015)
9. May, A.: New RSA vulnerabilities using lattice reduction methods. Ph.D. thesis, University
of Paderborn (2003). http://ubdata.uni-paderborn.de/ediss/17/2003/may/disserta.pdf
10. May, A., Ritzenhofen, M.: Implicit factoring: on polynomial time factoring given only an
implicit hint. In: PKC 2009, pp. 1–14 (2009)
11. Nitaj, A., Ariffin, M.: Implicit factorization of unbalanced RSA moduli. J. Appl. Math.
Comput. 48(1–2), 349–363 (2015)
12. Peng, L., Hu, L., Xu, J., Huang, Z., Xie, Y.: Further improvement of factoring RSA moduli
with implicit hint. In: AFRICACRYPT 2014, pp. 165–177 (2014)
13. Rivest, R.L., Shamir, A., Adleman, L.M.: A method for obtaining digital signatures and
public-key cryptosystems. Commun. ACM 21(2), 120–126 (1978)
14. Sarkar, S., Maitra, S.: Approximate integer common divisor problem relates to implicit
factorization. IEEE Trans. Inf. Theory 57(6), 4002–4013 (2011)
15. Wiener, M.J.: Cryptanalysis of short RSA secret exponents. IEEE Trans. Inf. Theory 36(3),
553–558 (1990)
OCV-Ah Integration SOC Estimation of Space
Li-Ion Battery
Dawei Fu(&), Lin Hu, Xiaojun Han, Shijie Chen, Zhong Ren,
and Hongyu Yang
Abstract. The state of charge of battery is a key, basic parameter of the battery
management, which represents the current capacity of the battery and is a health
criterion for the consistency of each cell. The accurate estimation of SOC will
provide effective technical support for extending battery life and enable the
battery to give full performance in the best state. In this paper, an optimized
OCV-Ah integration method is proposed. It can eliminate the influence of
internal resistance on the estimation error and provide an online estimation,
which is suitable for space Li-ion battery. Compare to the experimental value,
the estimation accuracy of calculated SOC is better than 4%. This method has
been applied to the analysis of a space battery, and the fault cell is identified
with the performance difference between the cells.
1 Introduction
Li-ion battery is characterized by higher energy ratio, smaller thermal effect, lower self-
discharge rate and higher cell voltage. However, it has the disadvantages of bad over-
charge and over-discharge resistance. Because of the gradually diverging of the state of
charge (SOC) of each cell, it will lead to over-charge or over-discharge. In order to
improve the life of Li-ion battery, an effective management is needed.
SOC is a key, basic parameter of the battery management, which represents the
current capacity of the battery. In addition, it is a health criterion for the consistency of
each cell, which will determine the battery management strategy. The accurate esti-
mation of SOC will provide effective technical support for extending battery life and
enable the battery to give full performance in the best state.
At present, the SOC estimation method of space Li-ion battery is generally
Ampere-hour (Ah) integration method, which is simple with poor accuracy and low
reference value. In this paper, the principle and application range of Ah integration
method, open circuit voltage (OCV) method are introduced. And an optimized OCV
method combined with Ah integration method is proposed. It can eliminate the
influence of internal resistance on the estimation error with estimation accuracy better
than 4%.
Qt
SOC ¼ ð1Þ
Q
For Li-ion battery, the relation between OCV and SOC isn’t linear with a plat
region as shown in Fig. 1 from 20%–80% SOC. But in a certain range of voltage, there
is a good one-one mapping relation [2].
4.1
3.9
3.7
OCV
3.5
3.3
3.1
0% 20% 40% 60% 80% 100%
SOC
But after enough standing, it will estimate the SOC accurately, providing initial value
or reference for error correction of other methods.
3 Battery Model
The relation between battery performances with each factor should be described in the
battery model. Good model generally can describe the above relation accurately with
parameters easy-to-get and lower order for implementation.
I
I Rp
R0
R0
UOC Cp
Uout
UOC
Uout
This model can describe the constant current mode of batteries. But the internal
polarization and self-discharging of the battery are not involved in this model.
Due to the low data volume, the static data can only provide an unsmooth curve as
shown above with bad accuracy. The charging and discharging curves are smoother
with more testing data. However, there is a large deviation to the static curve. It means
that all these three curves can’t estimate the SOC accurately.
OCV-Ah Integration SOC Estimation of Space Li-Ion Battery 295
Considering the R-int battery model in the Sect. 3, during the charging or dis-
charging of battery, the relation between output voltage and OCV is
Uout1 þ Uout2
UOC ¼ ð6Þ
2
The OCV-SOC curve is the average of charging and discharging curve with the
same current. And, it is between the two curves in Fig. 3 in accordance with Eq. (6).
Uout1 ¼ p1 x9 þ p2 x8 þ p3 x7 þ p4 x6 þ p5 x5
ð7Þ
þ p6 x4 þ p7 x3 þ p8 x2 þ p9 x þ p10
Uout2 ¼ p1 x9 þ p2 x8 þ p3 x7 þ p4 x6 þ p5 x5
ð8Þ
þ p6 x4 þ p7 x3 þ p8 x2 þ p9 x þ p10
According to Eq. (6), the OCV-SOC curve of this type space Li-ion battery is
Uout ¼ p1 x9 þ p2 x8 þ p3 x7 þ p4 x6 þ p5 x5
ð9Þ
þ p6 x4 þ p7 x3 þ p8 x2 þ p9 x þ p10
In consequence, Eq. (9) is a good expression for the OCV-SOC curve of the space
Li-ion battery. And it will provide accurate estimation of SOC.
For a spacecraft using Li-ion battery as the energy storage device, there are nine
cells in series. The cell voltage of the battery on June 11 and July 5 are listed in
Table 2. In this duration, the battery is only discharged by the measurement circuit.
The corresponding SOC of the cell voltage in Table 2 is calculated according to the
Eq. (9) and shown in Table 3. From the table, it is concluded that the decrease of SOC
of cell 7 is faster than the others after twenty-three days’ standing. That means, the self-
discharge rate of cell 7 is larger than the others with poorer performance.
The charge holding performance of each cell is listed in Table 4. The charge
holding rate of cell 7 is 90.41%, but all the others’ are larger than 95% after twenty-
three days’ standing. So the fault cell is cell 7 with lower charge holding rate.
where SOC0 is initial state of charge, I is the charging current, Q is nominal capacity
and η is the charging efficiency. For the Li-ion battery, the online estimation of
charging mode and discharging can be obtained using Eq. (10) where SOC0 is esti-
mated by the OCV-SOC equation.
For the Li-ion battery in the above section, the initial voltage and SOC before
charging, the voltage and SOC (Experimental and Calculated) after charging of 3Ah is
shown in Table 5. The results show that the largest deviation is 3.84% between the
calculated and experimental value, with accurate estimation.
5 Conclusion
At present, the SOC estimation method of space Li-ion battery is generally Ah inte-
gration method with poor accuracy. In this paper, an optimized OCV-Ah integration
method is proposed. It can eliminate the influence of internal resistance on the esti-
mation error and is suitable for space Li-ion battery. Compare to the experimental
value, the estimation accuracy of calculated SOC is better than 4%. This method has
been applied to the analysis of a space battery, and the fault cell is identified with the
performance difference between the cells.
References
1. Zhang, J., Lee, J.: A review on prognostics and health monitoring of Li-ion battery. J. Power
Sources 6007–6014 (2011)
2. Broussely, M., Perelle, M., McDowell, J., et al.: Lithium ion: the next generation of long life
batteries-characteristics, life predictions and integration into telecommunication systems. In:
Proceedings of IEEE International Telecommunications Energy, V01.2, pp. 194–201 (2000)
3. Gomez, J., Nelson, R., Kalu, E.E., et al.: Equivalent circuit model parameters of a high-
power Li-ion battery: thermal and state of charge effects. J. Power Sources 4826–4831
(2011)
4. Santhanagopalan, S., White, R.E.: State of charge estimation using an unscented filter for
high power lithium ion cells. Int. J. Energy Res. 152–163 (2010)
5. Hu, Y., Yurkovich, S.: Battery cell state-of-charge estimation using linear parameter varying
system techniques. J. Power Sources 198, 338–350 (2012)
6. Ng, K.S., Moo, C.S., Chen, Y.E.: Enhanced coulomb counting method for estimating state
of charge and state of health of lithium-ion batteries. Appl. Energy 1506–1511(2009)
7. Lee, S.J., Kim, J.H., Lee, J.M., et al.: The state and parameter estimation of a li-ion battery
using a new OCV-SOC concept. In: 2007 IEEE Power Electronics Specialists Conference,
vol. 1–6, pp. 2799–2803. IEEE, Electron Devices Soc & Reliability Group, New York
(2007)
8. Di Domenico, D., Fiengo, G., Stefanopoulou, A., et al.: Lithium-ion battery state of charge
estimation with a Kalman filter based on an electrochemical model. In: 2008 IEEE
International Conference on Control Applications, vol. 1, 2, pp. 425–430 (2008)
9. Lee, Y.S., Wang, W.Y., Kuo, T.Y.: Soft computing for battery state-of-charge (BSOC)
estimation in battery string systems. IEEE Trans. Electron. 55(1), 229–239 (2008)
10. Bakkera, S., Maat, K., van Weeb, B.: Stakeholder’s interests, expectations, and strategies
regarding the development and implementation of electric vehicles: The case of the
Netherlands. Transp. Res. Part A Policy Pract. 66, 52–64 (2014)
Analysis and Experimental Study on Influence
Factors of Spacecraft Power Cable
Temperature
Abstract. The design of the power cables is related to the safety of the satellite
power supply. High temperature of the power cables may lead to short circuit,
cable transmission performance degradation, cable damage or other risks. The
temperature of the power cables is difficult to get accurate result by theoretical
calculation and simulation analysis. In this paper, cable temperature rise tests
under different constraints are carried out, and the influence of the factors on
cable temperature changes is analyzed. Based on the effect of various factors on
cable temperature rise, the design methods and treatment measures of the power
cables in practical application are summarized. The results can provide a ref-
erence to the design of the spacecraft power cables.
1 Introduction
The spacecraft cable network is responsible for the transmission of energy and signals
between the various devices, mainly has power cables, signal cables and high fre-
quency cables. The main power cables have higher load current and higher temperature
rise. Once a short circuit occurs, the cable temperature rises more severely or even
burns, which is related to the safety of the satellite. The ambient temperature of the
spacecraft is generally good, but there are also working conditions with high ambient
temperature reaching 50–55 °C and the load power is larger. Under the severe working
conditions, the reliability of the cable is put forward higher. To ensure reliable trans-
mission, the ability to carry a certain load is required and the temperature rise also
needs to meet the derating requirements.
The main power cables bundle of the spacecraft is generally thick, and limited by
the space layout and the through-hole, it is impossible to achieve separate lashing and
laying. When laying cables, power cables, signal cables, buses are often partially
processed together. The temperature of complex cables has many influencing factors,
and it is difficult to accurately estimate the temperature of the real cable bundle of the
spacecraft by establishing a simulation model. On spacecraft, the temperature mea-
surement is usually performed by the thermistor. Due to the limited number and
location of the measuring points, the monitored cable temperature may not be the
highest temperature, and there is a hidden danger. In addition, according to the
development process of the spacecraft, the monitoring of the cable temperature is only
in the late stage of development, and if the temperature is too high, the cost of the
change is large. Therefore, it is necessary to study the cable temperature. Cable tem-
perature measurement points selection, cable temperature rise factors, the relationship
between temperature and wire selection, and cable laying design need to be solved.
For the cable temperature rise problem, there are many researches [1–4] on the
high-voltage transmission power cables in atmospheric environment, but there are few
studies on the cable temperature rise of spacecraft. Due to lack of convection in a
vacuum environment and the effective heat conduction path on spacecraft, it relies on
radiation to dissipate heat, and the cable heat dissipation problem is more severe. In this
paper, temperature rise tests of single cable and multi-cables are carried out, the actual
use on the spacecraft is simulated, the influencing factors of the cable temperature rise
are analyzed, and the influence degree of each influencing factor on the cable tem-
perature rise is given. This paper proposes constructive suggestions and guidance for
the design of the cables.
Vacuum tank
Test box
Temperature
Thermocouple
Measuring
Equipment
Bulkhead W01
electrical W02
connector W03
W04
W05
Power Source
Voltage
W06
W07
Electronic Load W09
Load
Load+
Load-
Length/m
Lashing point
Fig. 3. The multi-cables lashing diagram (W01, W02, W03, W04, W05)
The temperatures of the cables are measured by thermocouples, which are set at the
beginning (marked as -01), middle (marked as -02) and end (marked as -03) of the
single cable. There are two temperature measurement points for each temperature
measurement position, namely inner temperature named as W01N-03 and outer tem-
perature named as W01Y-03 by taking the end position of W01 cable as an example.
The inner temperature of the multi-cables at a specific location is defined as W01N-A
by taking the inner temperature of W01 at position A as an example. The bonding
points of the inner temperature and the outer temperature are required to be set at the
same section of the cable temperature measurement position. The inner temperature
measurement point is placed at the center of the cable, and the outer temperature
measurement point is placed on the outer surface of the cable, as shown in Fig. 2.
The cables are tested under vacuum environment, at ambient temperatures of 25 °C,
45 °C, and 55 °C, respectively. The current of the cable is loaded according to a certain
timing, and the temperature data is recorded every minute. The design of test cables is
shown in Table 1.
1.5
0.5
0 50 100 150 200 250
Time(min)
2.3 Test 2: The Temperature Rise Test of Single Cables with Two Tying
Methods at Different Ambient Temperatures
The single cables: The test cables are W07 and W09. The difference between the two
test cables is only the way of tying. W07 is the whole cable tied together. W09 is tied
by grouping, that is, four twisted pair wires are tied together. The loading current is as
follows (Fig. 5):
3
Current(A)
2.5
W06/W09 Current
2
0 50 100 150 200 250 300 350 400
Time(min)
0
0 50 100 150 200 250 300 350 400 450 500
Time(min)
W07 Cable Temperature (Ambient Temperature 25 ) W07N-02 Cable Temperature at different Ambient Temperatures
140 160
140
120
120
100
)
)
Temperature(
Temperature(
100
80
W07N-01 80
W07N-02
60
W07N-03 60
W07Y-01 W07N-02 @25
40 40 W07N-02 @45
W07Y-02
W07Y-03 W07N-02 @55
20 20
0 50 100 150 200 250 0 50 100 150 200 250
Time(min) Time(min)
2.32 A, that is the deducted current value of the cable, however, when the current is
2.05 A, the maximum temperature when the cable is stabilized is less than 120 °C.
80
)
60
Temperature(
W06N-02@25
40
W09N-02@25
W06N-02@45
20 W09N-02@45
W06N-02@55
W09N-02@55
0
0 50 100 150 200 250 300 350 400
Time(min)
80
)
60
Temperature(
40
W02N-A
W02N-B
20 W02N-E
W02N-F
W02N-G
0
0 100 200 300 400 500
Time(min)
The temperature rise curves of cables named W01 to W05 at the F lashing position
are shown in Figs. 11, 12 and 13 at different ambient temperatures. W01 cable is not
energized and is bundled with other cables. The temperature of W01 cable will also
change with the temperature of surrounding cables. The highest temperature position
Analysis and Experimental Study on Influence Factors of Spacecraft 307
Cable Temperature at different Ambient Temperatures Cable Temperature at different Ambient Temperatures
W02N-F @45
140 W02N-F @25 140
W03N-F @45
W03N-F @25
120 120 W04N-F @45
W04N-F @25
W05N-F @45
W05N-F @25
)
)
80 80
60 60
40 40
20 20
0 100 200 300 400 500 0 100 200 300 400 500
Time(min) Time(min)
Fig. 11. Cable temperature at ambient Temp25 °C Fig. 12. Cable temperature at ambient Temp45 °C
308 B. Zhao et al.
W02N-F @55
140
W03N-F @55
120 W04N-F @55
W05N-F @55
)
100 W01N-F @55
Temperature( 80
60
40
20
0 100 200 300 400 500
Time(min)
4 Conclusion
Through the test and analysis of this paper, the following conclusions can be drawn:
(1) For the case that power cables are concentrated and the load is large, temperature
monitoring needs to be set and the position should be specified. The location of
the temperature measurement should be selected as close as possible to the middle
position of the cable and the lashing position where the number of wires is the
highest, and paste it at the centerline of the cable.
(2) Power cables should be laid as close as possible to the deck to facilitate heat
dissipation.
(3) When signal cables are bundled with power cables, the temperature of signal
cables increase with the temperature of power cables, and finally approaches the
external temperature of power cables. Power cables should be laid separately and
separated from other cables, which is beneficial to the heat dissipation of power
cables, avoiding the temperature influence on other cables, and reducing the
influence range in the case of power cables failure.
(4) The high-power cables can be tied by grouping to facilitate heat dissipation and
avoid local overheating, which can reduce the cable temperature.
(5) Considering the temperature rise in the case of complex cables lashed together,
the high-power concentrated cable should be further derated on the basis of the
specified derating to ensure that the cable temperature rise does not exceed the
requirements.
(6) In the early stage of spacecraft design process, high-power laying path and suf-
ficient access should be reserved when planning the overall layout and cable path.
Other methods can be considered to solve local overheating and insufficient access
problems, such as high-voltage bus, buried cable or wireless transmission.
Analysis and Experimental Study on Influence Factors of Spacecraft 309
References
1. Blums, J., Budahs, M., et al.: MV Cable Temperature Characteristics in Dependence of load
Current, Ambient Environment and Temperature. Determination of the Critical Environment
(2016)
2. Du, B.: Research on Temperature Field and Ampacity of Power Cables. Zhengzhou
University, Zhengzhou (2016)
3. Lu, Z., Zhang, H., Ao, M., et al.: Research of thermal time constant and steady criterion used
in the power cables ampacity test. J. Northeast Dianli Univ. 36(5), 25–31 (2016)
4. He, J., Jiao, Y., Ye, D., et al.: Simulation and computation of temperature field and ampacity
of conduit cable laying in different ways. Electr. Meas. Instrum. 53(3), 99–104 (2016)
5. Zhang, P., Wu, L., Jiang, D., Wu, L., Zhang, Y.: The temperature rise model of spacecraft
cables in vacuum environment. Spacecr. Environ. Eng. 30(3), 235–239 (2013)
6. Geng, L., Peng, F., Wang, T., et al.: Analysis of the impact of high-power electric cable
heating on satellite thermal design. Spacecr. Environ. Eng. 31(1), 83–87 (2014)
7. Kong, W., Wang, B., Lao, S., Ai, Y.: Study on fire initiation of wire insulations on board the
satellite. Chin. J. SpaceSci. 28(1), 28–32 (2008)
Spacecraft System Autonomous
Health Management Design
Yong Lei(&), Quanyou Qu, Deyin Liang, Yilan Mao, and Xi Chen
1 Preface
With the development of aerospace technology and applications, users’ requirements for
spacecraft have been upgraded from the survival to continuous execution of missions.
The spacecraft mission has the characteristics of high cost and complex system com-
position. There is limited ground monitoring and control time, ground fault diagnosis
and disposal are not always timely. At the same time, with the increasing number of
spacecraft operating in orbit, the pressure to monitor and manage spacecraft is growing.
In order to improve the spacecraft’s continuous service capability, improve the timeli-
ness of fault handling, and reduce the pressure on ground management, it is necessary to
study and improve the spacecraft autonomous health management capabilities [1].
Spacecraft health describes the ability of spacecraft systems, subsystems, and
components to perform their design functions. The spacecraft health management can
be defined as a management activity directly related to its health status, that is, it can
independently monitor and diagnose its own health status, in the event of an anomaly,
it can be handled autonomously and return to normal work mode or minimize the risk
of safety and the impact on the mission. Health management uses monitoring and
diagnosis as the main means and it is the decision-making process and execution
process with perception as the core.
This paper carries out the research and design work of spacecraft system autono-
mous health management, and presents a hierarchical and distributed spacecraft health
management system architecture design. Based on this, the spacecraft health data
management mechanism and system health management strategies have been resear-
ched and designed.
© Springer Nature Singapore Pte Ltd. 2019
S. Sun et al. (Eds.): ICSINC 2018, LNEE 550, pp. 310–316, 2019.
https://doi.org/10.1007/978-981-13-7123-3_37
Spacecraft System Autonomous Health Management Design 311
Depending on the complexity of the spacecraft system, the autonomous health man-
agement architecture can be divided into centralized, distributed, and hierarchical and
distributed types. The core of the centralized architecture is a central management
controller or processor that collects and processes all health information. This type is
suitable for simple spacecraft. The distributed architecture consists of tasks such as
status monitoring, fault detection, and isolation processing independently. The
advantage is that system integration and testing are easy. The disadvantage is the
inability to perform data fusion between subsystems, the reliability of fault diagnosis
and detection is lower.
Considering the advantages and disadvantages of centralized and distributed
architecture, and the spacecraft system usually consists of multiple sub-systems and
products, its functions, composition, working modes, etc. are more complex, thus
providing a hierarchical and distributed autonomous health management architecture
design that combines the advantages of centralized and distributed [2], as shown in
Fig. 1. The architecture is divided into system, subsystem and component layers, it
include main functions such as health data management at all layers, fault perception,
data fusion diagnosis, and fault handling based on knowledge database.
Health data
management
Spacecraft
GRound
managem
Subsystem Sub-system data ent Center
fault fusion fault knowledg
layer level health
perception handling e database
Data diagnosis
Component Component
layer level health
Sensor 1 Sensor n BITm
Data
The system layer mainly completes the following functions: acquiring, storing, and
processing the health data, and adopting different strategies to interact with the ground
system according to the critical degree of health status; exchanging health information
with each subsystem, in the event of a fault, according to system fault-tolerant
resources and Knowledge model, perform system-layer and subsystem-layer fault
handling and reconfiguration [3]; integrate health data and information from different
subsystems to eliminate data inconsistency, identify and isolate faults, and obtain more
reliable subsystem health status.
312 Y. Lei et al.
The subsystem layer mainly completes the following functions: obtaining sub-
system, component-layer health data and information, and performing fusion pro-
cessing, performing subsystem-layer, component-layer fault handling and reconfigu-
ration according to subsystem-layer fault detection, diagnosis, and knowledge model;
exchanging the subsystem health data and fault handling information with the system
layer.
The component layer will complete the following functions: according to the self-
test information or internal sensor information of the component, the fault diagnosis of
the component is performed and the fault management strategy is executed according
to the knowledge model; exchanging the component health information and fault
handling information with the subsystem layer.
System
Health data health Spacecraft Ground
storage unit Data management center
Generation Data
scheduling
Data sharing
and storing
Health Data sharing Data bus
Table 1. System level common health data generation and processing method
Method Description Application
Health Synthesize and analyze the health status of each Generate the health status of the
status subsystem, perform fusion diagnosis spacecraft for the ground to
synthesis processing, and obtain system health status quickly and accurately
summary information, including system understand the health status of
working mode, system event report index, the spacecraft
energy/attitude/load/communication/propulsion
and resource margin evaluation
Statistical According to the specific period (one track, one Used to analyze and diagnose
analysis day, one month, load working period, etc.), long-term characteristics of key
count the maximum/minimum value and parameters
average value of key parameters affecting the
health status of the system
Point For the key parameters of continuous change, it Draw curves for key parameters
sampled can also be sampled and stored according to the in a specific period
period, and generated a curve according to the
period (one track, one day, one month, etc.)
Mutation Performs system mutation monitoring and Used to detect a sudden change
monitoring important telemetry parameters, including 0-1 or an excessive range of
mutation, excessive variability of transitions, changes
and out of range
Event Corresponding event reports are generated for Used for further analysis,
reporting faults and abnormal events monitored at the location and disposal of faults
system level, and record all relevant data on the ground
System Real-time record system latest working status, Record the critical state of the
working system configuration status, and reliable storage system and use it for system
status through multiple methods recovery in case of fault
records
314 Y. Lei et al.
Health Data Sharing. The purpose of spacecraft health data sharing is to synchronize
system state, improve the accuracy, robustness and efficiency of health management.
Spacecraft system layer health data sharing is mainly through the following ways.
Publish the whole system parameters, health status, system configuration status, etc.
to the sub-systems, and synchronize the current status of the system.
Correlate the current working parameters between the relevant subsystems for
verification diagnosis. To ensure that all kinds of data required for fault detection are
comprehensive and accurate, and improve detection accuracy.
Store the critical state of the system in different storage units or sub-systems. When
a sub-system fails, the working parameters before the failure can be restored to
maintain the continuity of the system as much as possible; the health data storage unit
can also perform fusion analysis on health data to achieve accurate fault diagnosis and
location.
Health Data Scheduling. Health data scheduling includes spacecraft internal
scheduling and scheduling between the spacecraft and ground. The principle of data
scheduling is to transmit critical health data in a timely and on-demand manner.
The spacecraft internal scheduling adopts a periodic and burst-compatible
scheduling strategy. The periodic scheduling strategy is used for periodic messages,
and the burst-compatible strategy is adopted for the changing messages, which can
ensure that information is delivered in a timely manner and save communication
bandwidth resources. In addition, all health data is stored in the health data storage unit,
so that all data can be queried when needed.
AOS-based channel and packet two-level scheduling strategy is adopted between
the spacecraft and ground. The channel is used to distinguish different types of data, the
channel scheduling uses a policy triggered by an instruction or an abnormal event. The
packet scheduling strategy adopts a two-dimensional dynamic scheduling method
based on the shortest transmission period and the importance of packet.
The spacecraft system layer health management strategy mainly includes: system
monitoring strategy, payload mission safety strategy, power supply safety strategy,
transfer to safety mode strategy, system recovery strategy, system reconfiguration
strategy, etc., and the implementation of each health management strategy depends on
lots of basic technologies, including fault detection, diagnosis, data fusion, software
development, PUS [4] (Telemetry and telecommand packet Utilization) anomaly event
handling mechanisms. The composition of the spacecraft system layer health man-
agement strategy is shown in Fig. 3.
Spacecraft System Autonomous Health Management Design 315
5 Conclusion
According to the spacecraft health management objectives and requirements, this paper
proposes a hierarchical and distributed health management design, from the aspects of
health management architecture, health data management and health management
strategy. The hierarchical and distributed architecture has the advantages of simple and
efficient implementation, sufficient data fusion and sharing, and accurate fault diag-
nosis. The health data management is based on the system management unit and the
health data storage unit. The health data flow design scheme, health data generation and
processing method, health data sharing mechanism and scheduling method are put
forward. Finally, for the health management application, the health management
strategies such as system monitoring, payload mission safety, power supply safety,
safety mode; system recovery and system reconfiguration are designed.
The system layer health management scheme proposed in the paper is reasonable in
architecture, complete in health data management, effective in strategy, and can adapt
to the health management needs of complex spacecraft. It can be used as a reference for
spacecraft system layer health management research, system design, and spacecraft
ground control system design.
References
1. Tipaldi, M., Bruenjes, B.: Spacecraft Health Monitoring and Management Systems (2014).
978-1-4799-2069-3/14 ©IEEE
2. Kolcio, K., Breger, L., Zetocha, P.: Model-Based Fault Management for Spacecraft
Autonomy (2014). 978-1-4799-1622-1114 ©IEEE
3. Figueroa, F., Walker, M.G.: Integrated System Health Management (ISHM) and Autonomy,
8–12 January 2018, Kissimmee, Florida (2018)
4. ECSS. Ground systems and operations - Telemetry and telecommand packet utilization,
ECSS-E-70-41A (2003)
5. Fayyaz, M.: Survey and future directions of fault-tolerant distributed computing on board
spacecraft. Adv. Space Res. 58, 2352–2375 (2016)
Design and Simulation Verification of Ground
Charge Equipment for Li-Ion Battery Pack
Lin Hu(&), Dawei Fu, Hongyu Yang, Shuo Feng, Jianbo Du,
Jinchen Zhao, and Chengzhi Lu
Abstract. Li-ion battery packs are aerospace energy storage components that
have been used and have good prospects. Because it is not overcharge resistant
but needs to use them in series, the complex requirements of charging man-
agement have become an important factor restricting their development. It
designed a li-ion battery pack ground charging device from the satellite interface
to meet the demand in this paper, this device used CC-CV charge method,
equalization charging for cells, and has completed simulation verification, laid a
good foundation for subsequent charging equipment.
1 Introduction
Compared with Ni-Cd and Ni-MH batteries, li-ion battery has higher operating voltage
without memory effect, smaller volume, lighter weight, higher energy ratio, higher
discharge current, longer life, lower self-discharge rate, and wider available tempera-
ture range and so on. Therefore, li-ion battery is the first choice when we need the
higher performance, reliability, security electrical energy storage device [1].
The charging process of the li-ion battery is completely different from the other’s,
as the performance of li-ions decreases, the imbalance between battery cells becomes
more serious after repeated charge and discharge cycles, the li-ion battery is not
overcharge resistant, overcharging will seriously affect the cycle life of the li-ion
battery. The charge termination voltage is generally controlled within 1% accuracy.
Thus, in order to meet the needs of routine testing, it is necessary to use a charging
equalization circuit to manage the charging process of the battery pack, it should design
special charging device for li-ion battery packs. It designed a more efficient equal-
ization charging ground equipment for li-ion battery packs in this paper, which charges
from the interface of the spacecraft with the existing charging power supply.
2 Systematic Review
2.1 Charging Mode
The development of charging mode starts with constant current charging but this
method is not easy to control. Constant voltage charging method, keeps the charging
voltage constant during the charging process, has some adaptability, and is closer to the
optimal charging curve than constant current charging method. However, the initial
charging current is too large, which has a great impact on the life of the battery, and
even causes the battery damage, so it is also hardly used. Compared with the two
methods, the stage charging is more reasonable, especially the two-stage charging
method which combines constant current and constant voltage. This method charges
constant current to the rated voltage and uses constant voltage charging to complete the
remaining charge [2]. Through the combination of the previous two charging methods,
the battery safety is better guaranteed. Although the charging control circuit is more
complex, because of its short charging time and high charging safety, it plays a
dominant role in the charging of li-ion batteries. The charging device designed in this
paper adopts this charging mode.
The design of the equipment mainly includes the design of the system (including the
whole system, constant current charging loop, constant voltage charging loop), related
modules design (including equalization module, system stability, overvoltage protec-
tion module, etc.) and the simulation of the system. The setup of system and simulation
of the circuit use Oracd Pspice.
Design and Simulation Verification of Ground Charge Equipment 319
In view of the current common and the increasing demand for battery capacity with
the development of aerospace, the 70 Ah battery pack is selected as the charging target.
In the constant current charging process, the charging current is selected to be 10 A,
which improves the charging efficiency and safety; in the constant voltage charging, the
charging current is selected to be 0.7 A. The single cell voltage range is set to 2.7 V to
4.2 V. For the common 30 V spacecraft bus voltage, 7 cells are connected in series, the
voltage of the whole pack is from 18.9 V to 29.4 V. Agilent 6670 series power supply
is used as the charging device.
Fig. 2. Constant current charging loop Fig. 3. Constant voltage charging loop
320 L. Hu et al.
As the system topology shown in Fig. 1, the feedback of the current loop is
obtained by collecting the voltage across the current sampling resistor. The feedback of
the voltage loop is the output voltage division.
DVEA R6
gain ¼ ¼
DVSENSE R5
Figure 2 shows the constant current charging loop. Through the two resistors R5
and R6 on the periphery of the amplifier, when performing the gain calculation of small
signals, it is known
The system selects 7.5 times gain to ensure the accuracy of charging current within
3%. It is proved by simulation that the system is stable and can meet the charging
current accuracy. If the sampling resistor is considered at the same time, the loop gain
is 3.
The design of the constant voltage charging loop is shown in Fig. 3, using the
output voltage divider for feedback. The gain of the loop is the same as the gain
calculation of the constant current charging. The error amplifier gain is 40 when the
sampling gain is not considered, and the gain obtained by considering the sampling
gain is 5.39.
There is a series relationship between the battery cells. During the charging process, the
positive and negative voltage of each battery cell are floating.
Each equalization module is shown in Fig. 5: each single cell is connected in
parallel with a shunt resistor R, and the output of this current is controlled by switch;
each module samples and calculates the total voltage of the entire pack and the cell, it
compares the average voltage of the entire battery with the cell, so that, during the
Design and Simulation Verification of Ground Charge Equipment 321
charging process, relatively smaller charging current is achieved for the higher voltage
battery and larger for the lower ones, so the entire pack is matched. According to
m 1
Vcomp þ Vcomp ¼ ðVcell þ Vcell Þ ðVBAT þ VBAT Þ ð1Þ
mþ1 m
1 1
2Iripple ¼ ðVin VoutÞD ð2Þ
L f
among them,
D ¼ Vout=Vin. ð3Þ
Simulate the charging current ripple at some selected part of the simulation process.
According to Fig. 8, the maximum charging current of the sample is 10.02 A, the
minimum is 9.98 A, and the ripple is less than 0.02 A, which is much less than 1%. It
meets the requirements of the indicator. Figure 9 reflects the ripple variation of the
current flowing through the inductor, it can be seen that the ripple value is less than 1%.
it can also meet the requirements of the indicator.
Fig. 10. Charging current with voltage Fig. 11. Charging current with input
voltage
During the charging process, as the battery voltage gradually increases, the
charging current decreases accordingly. However, under the regulation of the system’s
negative feedback, the charging current is relatively stable. The simulation results in
Fig. 10 show that when the battery voltage increases from 18.9 V to 29.4 V, the ripple
decreases and the charging current decreases accordingly. During the whole change,
the DC component of the charging current decreases from 10.25 A to 9.92 A. The
change is within 3%.
When charging li-ion battery through the interface of the satellite, an existing
ground power source can be used as the supply. The design of the system with a large
input range can increase the universality of the system accordingly. This design selects
40 V as the input voltage selection, and the input voltage range can be selected from
324 L. Hu et al.
35 V–55 V. Figure 11 shows the simulation result. When the input voltage changes
from 35 V to 55 V, the current is between 9.95 A and 10.10 A, and the variation range
is within ±0.5%.
Fig. 12. Output voltage ripple of CV loop Fig. 13. Inductor current of CV loop
Figure 12 shows the output voltage of the constant voltage loop. The maximum
voltage is 29.989 V, and the minimum is 29.256 V, the average is 29.622 V.
Considering that the voltage drop of the sampling resistor at the negative end of the
battery pack is 0.28 V, it sets the output voltage as 29.68 V, the ripple is 0.367 V
which is 1.2% compared with the rated voltage. Figure 13 shows the inductor current
value, the maximum is 744.473 mA, the minimum is 670.244 mA, the average
inductor current is 707.359 mA, the average inductor current design of the constant
voltage loop is 700 mA, the error is 1%, and the ripple is 37.1 mA, it is 5%. Since the
main control object of constant voltage charging is lower voltage ripple, the current
performance is generally sacrificed, so the system is not as good as the constant current
loop in terms of current characteristics.
Fig. 14. CC loop voltage at the SW point Fig. 15. CV loop waveform at the SW point
Figure 14 shows the voltage at the SW point which is between the switch and the
inductor in the constant current loop. Figure 15 shows the voltage at SW point in the
constant voltage loop. The input voltage is PWM modulated by the switch, the max-
imum 36 V and the minimum –1 V SW voltage are obtained. The loss of the 36 V
compared to the 40 V input voltage is resistance consumption of the resistance in the
switch, and the negative 1 V voltage comes from the voltage drop of the freewheeling
Design and Simulation Verification of Ground Charge Equipment 325
diode. According to this figure, it can be seen that the output current/voltage of the
charging system is stable. According to the above simulation results, it can be seen that
the current loop is stable and can reach various indexes.
Li-ion battery packs are aerospace energy storage components that have been used and
have good prospects. Because it is not overcharge resistant but needs to use them in
series, the complex requirements of charging management have become an important
factor restricting their development. How to satisfy the requirement of equalization
charge of li-ion batteries in spacecraft AIT stage has become one of the problems that
restrict the upgrading and optimization of spacecraft energy storage devices.
It designed a li-ion battery pack ground charging device from the satellite interface
to meet the demand in this paper, this device used CC-CV charge method, equalization
charging for cells, and has completed simulation verification, laid a good foundation for
subsequent charging equipment.
References
1. Slimm, M., et al.: Lithium-ion batteries for space. Space Power 502(416), 477 (2002)
2. Pan, J.: Lithium battery intelligent management system. Doctoral dissertation, Zhejiang
University (2004)
3. Bian, Yankai, et al.: Equilibrium control and design of Li-ion batteries. J. Northeast Electric
Power Univ. 26(2), 69–72 (2006)
Motion Control of Robot Mobile Platform
Based on Indoor Positioning System
Abstract. This paper studies and designs the electric omnidirectional mobile
platform which based on the Mecanum wheel for the robot automation
assembly. The host computer communicates with the indoor positioning system
and the mobile platform to remotely track and control the mobile platform. At
the same time, the visualization module can view the movement trajectory of the
mobile platform, the state of the platform and the positional relationship with the
obstacle in real time, and can perform basic motion control and path planning
and execution operations on the mobile platform. By properly arranging the
location of the base station, the global positioning of the 100-m plant can be
achieved, and the positioning accuracy can reach 10 cm, which is better applied
to the development of spacecraft and provides technical support for the real-
ization of smart factories in the future.
1 Introduction
With the maturity of robotic precision assembly technology, more and more spacecraft
models use this technology to assemble large-weight equipment, and achieved good
results. However, there are still some problems in the use process [1], in which the lack
of automation of the mobile platform is the main problem affecting its ease of use.
The existing robot mobile platform is purely mechanical and relies on manual
pulling and moving, and has the following problems:
(1) The transfer is time-consuming and labor-intensive, and the manual control has
low control accuracy for the entire system position;
(2) Insufficient steering ability, it is not convenient to adjust the direction of the robot
within a limited station;
(3) There are hidden dangers to people and products during the manual pulling
process.
The degree of automation of the manual pull-type mobile platform and the
industrial robot does not match, which affects the application efficiency and effect of
the entire system. It is necessary to introduce an automated electric mobile platform to
further improve the transport adjustment capability of the robotic precision assembly
system, making it better suited for spacecraft assembly.
The omnidirectional mobile technology based on the Mecanum wheel structure has
mature applications [2, 3], and the technology is applied to the robot precision
assembly system, which can realize the automatic movement of the system’s zero
turning radius and improve the use efficiency [4]. After adopting the electric omnidi-
rectional mobile platform, the system can be more conveniently transported at different
stations according to the application requirements, and the system orientation can be
flexibly and quickly adjusted to improve the automation level and application efficiency
of the entire system. The operator can precisely control the movement of the mobile
platform through the control terminal, and improve the position control precision of the
system. Compared with the manual pulling mode, the safety risk to the person and the
product can be greatly reduced.
In addition, after adopting the electric omnidirectional mobile platform, the robot
system will have the capability of remote automatic control, combined with the indoor
positioning system, which can realize high-precision remote closed-loop control in the
use site. Intelligent factory is an important development trend of industrial production
in the future. It realizes remote path planning and closed-loop control of mobile
platform, which can provide technical support for the realization of smart factory in the
future. This project has carried out research on remote visualization CNC software of
robot mobile platform, aiming at realizing remote path planning and closed-loop
control of mobile platform.
2 Overall Design
2.1 Functional Requirements
It can communicate with the indoor positioning system [5, 6] and the mobile platform
through the host computer, and then remotely track and control the robot mobile
platform. At the same time, the visualization module can view the movement trajectory
of the mobile platform, the state of the platform (coordinates, speed, etc.) and the
positional relationship with the obstacles in real time. It enables basic motion control
and path planning and execution of mobile platforms.
paths [7, 8]. In addition, the system can provide related operation auxiliary interfaces
such as compilation of NC program files, such as editing and saving of files, and data
feedback of programs written.
Fig. 2. Robot mobile platform remote visualization control software overall program framework
Motion Control of Robot Mobile Platform Based on Indoor Positioning System 329
The status monitoring functions carried by the system include: online display of
real-time position information of the platform (obtained by the indoor positioning
system), platform movement speed and angular velocity information, alarm informa-
tion, and obstacles (in the vehicle body safety zone or alarm zone) information [9]. In
addition, the system will also realize the feedback and monitoring of the stable support
state information, and achieve the virtual support requirement of the platform by
independent drive adjustment of four independent supports, the sub-module including
the drive adjustment module and the status display module. The control software will
realize the transmission of instruction information and status monitoring with the
platform in a wireless manner, and achieve the expected application goal of remote
real-time control and monitoring. The overall framework of the specific scheme is
shown in Fig. 2.
3 Detailed Design
Based on the above design, the basic interface of the developed control software is
shown in Fig. 3. The independent mode and the modules can be switched to each other.
The input and output of the control commands are visualized. The important state
monitoring information (such as alarms) is displayed in real time in the fixed window
position, and is not affected by the switching of each module.
Fig. 3. Preliminary framework of remote visualization CNC module for robot mobile platform
The data communication of the system module adopts the wireless mode, and the
underlying driving communication with the omnidirectional mobile platform is based
on the RS232/485 data protocol.
330 Z. Jiang and L. Zhang
with a transmission distance of less than 10 m. The global positioning of the entire
plant can be achieved by rationally arranging the location of the base station, and the
positioning accuracy can reach 10 cm [12] (Fig. 5).
4 Software Introduction
4.1 Technology Implementation
You can choose to set the mobile platform to “stable support adjustment”/”stable
support release”.
• Mobile platform
• Computer
• Indoor positioning system
• Mobile platform wireless communication equipment
5 Conclusion
The wireless communication between the upper computer and the indoor positioning
system and the robot mobile platform is realized, and the movement of the mobile
platform is controlled through the visual interface operation, and the movement is
performed according to the trajectory. By properly arranging the location of the base
station, the global positioning of the 100-meter plant can be achieved, and the posi-
tioning accuracy can reach 10 cm.
References
1. Zhou, Y., Zhang, J., Le, Y., et al.: Application of mobile robot technology in aerospace
manufacturing industry. Mech. Des. Manuf. Eng. (2018)
2. Zhu, Y.: Research on Design and Performance of Omnidirectional Vehicle Based on
Mecanum Wheel. Yanshan University, Qinhuangdao (2016)
3. Chung, J.H., Yi, B.J., Kim, W.K., et al.: The dynamic modeling and analysis for an
omnidirectional mobile robot with three caster wheels, vol. 03, pp.3091–3096 (1994)
4. Lu, W.: Structural design of omni-directional wheel moving mechanism. Mech. Electron.
(2006)
5. Liu, H., Darabi, H., Banerjee, P., et al.: Survey of wireless indoor positioning techniques and
systems. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 37, 1067–1080 (2007)
6. Lee, B.-G., Chung, W.-Y.: Multitarget three-dimensional indoor navigation on a PDA in a
wireless sensor network. IEEE Sens. J. 11, 799–807 (2011)
7. Wang, Z., Guo, G.: Current status and prospects of mobile robot navigation technology.
Robotics (2003)
8. Luo, M., Wu, H., Du, Y.: Research on mobile robot navigation based on human-computer
interaction. J. Comput. Measur. Control, 1291–1293 (2013)
9. Lu, X., Zhang, G.: Research on navigation method of indoor service robot. Robotics (2003)
10. Deng, P., Fan, P.: Principle and application of wireless positioning in cellular system. Mobile
Commun. (2000)
11. Lee, J.-Y., Scholtzr, A.: Ranging in a dense multipath environment using an UWB radio
link. IEEE J. Sel. Areas Commun. 20, 1677–1683 (2002)
Motion Control of Robot Mobile Platform Based on Indoor Positioning System 335
12. Cao, S.: Research on Key Technologies of Ultra-Wideband Indoor Positioning System.
Guilin University of Electronic Technology (2006)
13. Li, Z., Wang, H.: Design and implementation of remote human-computer interaction
software platform for mobile robots. J. Comput. Measur. Control 121–125 (2017)
14. Pfeifer, R., Iida, F., Bongard, J.: New robotics: design principles for intelligent systems.
Artif. Life 11(1/2), 99–120 (2005)
A Process Method and Simulation Analysis
of Spacecraft Wing Root Cable Fixing
Abstract. In order to avoid the situation that the cable is damaged during the
unfolding and folding process of solar wings, a spacecraft solar wing root cable
tying process is designed. Through the simulation of the root-hinge cable laying
process, it is concluded that setting the appropriate reserved length and cable tie
point, improving the installation flexibility of the wire harness, and the wire
harness can be prevented from being stressed after the connector is plugged,
thereby improving the service life of the cable. Then through the simulation of
the motion of the root cable, it is concluded that the interference of the root cable
is not affected during the deployment of the solar wing, and the correctness of
the static simulation is verified, which provides a theoretical basis for the
satellite assembly operation.
1 Introduction
The solar wing is the main energy supply device of the spacecraft. Due to the limitation
of the vehicle, the solar wing folds and the gathers in the fairing during the launching
stage. When the spacecraft is separated from the vehicle, it expands into a plane and
locks. Therefore, the reliable deployment of the solar wing is the key to ensuring its
normal on-orbit operation and accurate orientation of the sun [1–3].
Due to the uncontrollable expansion speed of the solar wing energy-free devel-
opment agency, the active deployment mechanism has increasingly become a hot field
in research of the solar wing mechanisms at home and abroad. However, at present,
there are no optimized design schemes for the solar wing root cable lashing point and
the solar wing root cable reserved length. The solar wing root cable tying and fixing
method has not been finalized yet, and the solar wing unfolding and folding process
exists the danger of squeezing, pulling, and damaging the cable. Therefore, this paper
simulates and analyzes the cable tying of the root cable, and determines the reserved
length of the cable, the way of walking, and the cable state after the solar wing is
rotated [4–7].
2.2 The Multiple Deployment of the Solar Wing Root Hinge and Cable
Design
In the course of research, the design model was first confirmed, and the distance
between the lower plane of the root hinge inserting the plate and the center of the root
hinge was confirmed. There are 4 connectors for the solar wing hinge design. They are:
X07, X08, X09, X10. Among them, X07 is a rectangular connector, X08, X09 and X10
are circular connectors. Currently, the circular connector has Two tail covers, one is a
straight tail cover, and the other is a tail cover. The X8 adopts a straight tail cover when
designing, and the X9 and X10 adopt a curved tail cover. In order to avoid the X9 tail
cover and the X10 interference, rotate the X9 tail cover, as shown in Fig. 2.
338 K. Xu et al.
Fig. 2. Schematic the diagram of the solar wing root cable routing
3 Simulation Analysis
The length of each electrical connector lead in the design: the X07 lead length is about
350 mm, the X08 lead length is about 300 mm, the X09 lead length is about 210 mm,
the X10 lead length is about 140 mm.
Due to the design of the cable, the cable is considered as a rigid body, and the true
state of the cable under the influence of its material parameters is not considered, and
whether the minimum bending radius, the maximum tension and the like are reached.
This simulation is based on the real material parameters of the cable for real-time
simulation of the one-dimensional finite element.
In order to meet the minimum bending radius requirement of the cable, the standard
length of the cable needs to be bundled. After flexible processing, the parameters of the
cable are input, the parameters are shown in Table 1.
The length of the lead is the design length after conversion. The result of the
standard length of the rectangular connector and the round connector tail, cable must be
tied, as shown in Fig. 3. The blue segment is the minimum bending radius prompt, the
red segment indicates interference with the rigid body structure.
A Process Method and Simulation Analysis 339
Fig. 4. Sets the lashing point (pink represents the maximum tension of the cable, red represents
the pull-off state)
After preliminary simulation, when the X07 lead length is 368 mm, the X08 lead
length is 318 mm, the X09 lead length is 213 mm, and the X10 lead length is 145 mm
can meet the working requirements.
The position of the ligature point to the lower section is: the X07 lead length is
68 mm, the X08 lead length is 77 mm, the X09 lead length is 73 mm, the X10 lead
length is 68 mm.
A Process Method and Simulation Analysis 341
(a) Minimum clearance between the extended limit position and the solar wing
(b) The minimum distance of the cable from the bottom hinge
It can be seen from Fig. 6 that in the whole process, there is no influence on the root
cable, and there is no interference or hooking, so there is no stress.
4 Conclusion
Through static and dynamic simulation analysis, the optimal design of the lashing point
and the cable reserved length under the overall wiring state of the solar wing root cable
is realized, avoiding the minimum bending radius, the maximum tension and the
occurrence of interference, hooking and other undesirable phenomena, guarantee the
safety of the spacecraft solar wing working state.
References
1. Fei, K.: Design and Performance Test of Repeated Folding and Locking Solar Wing
Mechanism. Zhe jiang University of Technology, Hangzhou (2013)
2. Yang, J., Huang, H., Liu, Y., et al.: The evidential network for fault tree analysis with
imprecise knowledge. Int. J. Turbo Jet Engines 29(2), 111–122 (2012)
3. Zhou, Z., Wu, Y., Wang, J., et al.: Development status and trend of the circular solar wing.
Spacecr. Eng. 24(6), 116–122 (2015)
4. Zhang, L.: Design and Research of the Spacecraft Step-and-Expanding Solar Wing.
Shanghai Jiaotong University, Shanghai (2012)
5. Li Entrusted: Tension design and analysis of the solar wing linkage device. Chin. Space Sci.
Technol. 26(2), 52–57 (2006)
6. Eacret, D.L., White, S.: ST8 validation experiment: Ultraflex-175 solar array technology
advance: deployment kinematics and deployed dynamics ground testing and model
validation. In: AIAA 2010-1497. AIAA, Washington D.C. (2010)
7. Hua, D.: Design and Analysis of the Repetitive Locking and Release Mechanism of the Solar
Wing. Harbin Institute of Technology, Harbin (2008)
8. Wang, X., Yan, H., Zhou, Z.: Development status and trend of two-dimensional multiple-
expansion solar wing technology. In: Expandable Spatial Structure (2014)
9. Liu, Z., Wu, Y., Qi, H., et al.: Evaluation method for reliability of spacecraft solar wing
expansion. Chin. Space Sci. Technol. 52–56 (2013)
10. Solar, J.: Dynamic Fault Tree Analysis of Solar Wing Drive Mechanism. The university of
Electronic Science and Technology of China (2012)
11. Ren, S., Shang, H., Shang, H., Qi, H.: Simulation analysis of deployment dynamics of a two-
dimensional unfolded solar wing. Spacecr. Eng. 21(4), 32–36 (2012)
12. Murphy, D.M.: MegaFlex-the scaling potential of UltraFlex technology. In: AIAA 2012-
1581. AIAA, Washington D.C. (2012)
13. Xinggao, Z., Fengxi, C., Tingwei, L., et al.: Research on vibration test specifications of solar
wing drive mechanism based on modal analysis. Min. Mach. 38(5), 171–173 (2017)
14. Hu, M., Li, W., Chen, W., et al.: Motion simulation and functional experiment of repeated
folding mechanism of sectoral solar wing. Chin. J. Space Sci. 36(1), 92–98 (2016)
Study on 3D Cable Network Design Method
Yi Yuan(&), Wei Yu, Xiaoyi Ru, Zhou An, Qi Miao, Xuhua Hu,
and Zhenpeng Ding
1 Introduction
standalone device model to carry out the general configuration layout design and form
the whole-spacecraft configuration layout model. The logic design of cable network
forms the table of the cable branch connection relation according to the electric con-
nector information and contact distribution relation in the IDS file. Based on the whole-
spacecraft configuration layout model, the general designer starts the 3D cable network
design according to the connection directing in the table of cable branch connection
relation to form the 3D cable network model finally [1].
standalone model
device model design activity
design
installation of electric
connector plug
electric connector
designation
3D cable design activity
3D cable
network design bobbin design network
path planning
model
harness design
branch design
table of cable
cable network branch
logic design connection
design activity
relation
(2) The design link interface is unclear. The current 3D cable network design is based
on the general configuration model. When the general configuration layout model
is adjusted, the 3D cable network mode will change automatically, which is not
helpful to clear the responsibility interface and control the technical state of
model.
(3) The key links are lower automatized. The cable network design is composed of
several links i.e. analysis of cable branch file, installation of electric connector
plug, electric connector designation, wire gauge design, routing path planning and
3D modelling etc., all of which will be manually completed by the designers. In
the early stage of the cable network design, the 3D models are constructed by IDS
and the standalone development units manually, hence lack the corresponding
tools and software.
Therefore, to realize the fast 3D cable network design, we must start from the
following aspects:
(1) Normalize the device. Realize the automatic modelling of device based on the IDS
file and the unify the model patterns. Identify the key information (i.e. electric
connector name) required by the 3D cable network design in the automatic
modelling to facilitate the subsequent design and extraction of cables.
(2) Clarify the working interface and connector. Construct the 3D design input model
of whole-spacecraft cable network according to the whole-spacecraft configura-
tion layout model. When the spacecraft configuration layout is changed, the input
models can be updated only after the designers execute the change and confir-
mation. This can help the designers to control the technical design state.
(3) Develop the specific tools according to the design platform. Realize the branch
length analysis, electric connector designation, installation of electric connector
plug, cable directing planning, wire gauge design, 3D modelling automation and
simplify the design process.
standalone device
specification
model design
model
whole-spacecraft
configuration
configuration design activity
layout design layout model
installation of electric
input interface connector plug
model for cable design activity
electric connector
network design designation 3D cable
3D cable
path planning bobbin design network
network design model
harness design
branch analysis
branch design
table of cable
cable network
branch
logic design connection design activity
relation
Figure 2 is the ideal development mode of 3D cable network. Except for the design
activities such as configuration layout, cable logic design and cable path planning that
are inseparable from the manual intervention, other works can be automatically realized
through the software tools. After making the information preparation for the 3D cable
network design, the modelling process of 3D model is a black box that can lower the
complexity of cable network design and reduce the design workload.
The fast design of 3D cable network contains four parts namely the device normal-
ization modelling, construction of input interface model, analysis cable network con-
nection relation and automatic modelling of cable network. Now the implementation
method will be expounded in four aspects:
Method of development
Input Step and implementation
Electric connector
IDS file Templet
installation
(2) Pro/E takes the +Z axis in the coordinate system as the cable inlet. The coordinate
system shall be designated under the Pro/E cable module, so that it can be
identified in modelling. The cable design coordinate system shall be preset on the
template of cable plug and make sure the +Z direction is outward. Also, the open
interface Pro/Toolkit of Pro/E has provided a large number of cable network
implementation functions. During the software implementation, all connector
plugs in the current model can be traversed first; the cable installation coordinate
system on each cable plug can be obtained through feature analysis; the batch
setting implementing the cable installation coordinate system can be designated
cyclically through software; each cable installation coordinate system is endowed
with a unique internal sign by interpreting the whole-spacecraft unique sign of the
electric connector corresponding to the cable plug; the cable installation coordi-
nate system endowed with sign can be identified by Pro/E subsequently (Fig. 8).
(3) Main cable path is realized through the routing network equipped on the Pro/E
software. Its principle is to preplan a virtual path on the spacecraft model, the
Pro/E will search the main path with the shortest distance from the directing
automatically during implementing the cable network directing design. Under the
circumstances of various cable networks and complicated directing, try the best to
make sure several cables can share the common path, so that the cables can be
bound and tied up. The main cable path design shall take the whole-spacecraft
cable network layout into full consideration; whether the main path is reasonable
has determined whether the cable directing is reasonable and whether the target of
cable weight control can be realized [5, 7].
(4) Pro/E provides a large number of automatic realization interfaces of cable net-
work; however, the realization process is rather complicated and the developers
must have deep understanding on the realization mechanism of Pro/E cable net-
work. To lower the development difficulty, Pro/E also provides the neutral file
format (nwf) to define the wire gauge and connection logic of the cables in the
neutral file. After Pro/E reads the neutral file, the wire gauge can be established
automatically to connect the automatic identification of the start point and end
point. The developers only have to make fewer encoding work to construct the 3D
cable network modelling automatically [6] (Fig. 9).
350 Y. Yuan et al.
Fig. 9. Definition example of nwf file wire gauge, component and connector, connection
relation
According to the neutral file, development of the fast design function of 3D cable
network modelling can be simplified into obtaining the wire gauge and logic con-
nection relation of cable network from the table of cable branch connection relation and
writing it into the neutral file according to the canned format of the neutral file. As each
wire coordinate system has been identified during designating the installation interface
of cable network, when Pro/E reads the neutral files, it can identify the cable network
design port on the model corresponding to the logic relation automatically. By adopting
the canned 3D cable network construction function of Pro/E, the 3D cable network
modelling can be generated as shown in the following figure [2–4].
4 Conclusion
References
1. Chen, Y.: An Introduction of Digital Design for Spacecraft, pp. 386–387. Science and
Technology Press, Beijing, China (2010)
2. Wang, W.: The Example of Pro/E wildfire 4.0 Second Development, pp. 53–55. Tsinghua
University Press, Beijing (2010)
3. Wu, L., Chen, B.: The Base of Pro/ENGINEER second development, p. 1. Publish House of
Electronics Industry, Beijing (2006)
Study on 3D Cable Network Design Method 351
4. Lin, L.: Second Development of Pro/E Wildfire 2.0 with Pro/Toolkit, pp. 529–530. Publish
House of Electronics Industry, Beijing (2005)
5. Li, Y., Li, C., Liang, D., et al.: Pro/Engineer 4.0 Self-Study Manual, pp. 156–164. Publish
House of Electronics Industry, Beijing (2010)
6. PTC Corporation. Pro/TOOLKIT Wildfire 2.0 APIWizard. PTC Corporation, Needhan (2011)
7. Heng, S.: Learn Pro/Engineer the Right Way, pp. 325–326. China Youth Publishing Group,
Beijing (2007)
A Similarity Approach for Thermal Distortion
Measurement of Large Spacecraft Structure
1 Preface
With the development of remote sensing satellite technology, image geolocation and
resolution will be promoted further, which always requires large dimension and low
thermal distortion. For example, a pointing error of 1 arcsecond in 500 km SSO orbit
will cause 2.5 m horizontal error. Analysis is the primary tool now in structure thermal
stability design [1, 2], and design result need to be validated by the thermal distortion
test. In the 1990s, both NASA JPL and ESA did research on thermal distortion test
methods [3–7]. In these methods, coordinate measurement machine [8] and theodolites
[9] have relative low precision. Though laser interferometer has precision of sub-micro,
it is vulnerable to environment disturbance. 3D-DIC (digital image correlation)
methods has developed rapidly in recent years, and measurement precision can be
submicros, but it is still difficult for large scale structure measurement because of
influence by environment vibration, thermal control difficulty, high cost and long test
time. This paper presents a method to evaluate the thermal stability of large-size
structure by means of measuring its down-scaling model.
Similarity relations of thermal distortion are based on the three similitude theorems
[10], and from that we deduce the similarity relations of thermal distortion between
similar structures.
When applied with thermal load, the thermoplastic equations of solid are [1]:
8
< C ¼ ður þ ruÞ=2
r Tþf ¼ 0 ðIn XÞ ð1Þ
:
T ¼ 2lC þ ðkJ ðCÞ 3K0 aðh h0 ÞÞI
u; C; T are displacement vector, strain tensor, and stress tensor respectively. I is unit
tensor, and J ðCÞ is the trace of C. k and l are Lame coefficients, and f is body force. a
is thermal expansion coefficient, and E is modulus of elasticity. m is Poisson’s ratio, and
K0 is equivalent modulus of compression.
The boundary conditions on boundary @X of region X are:
8
<u ¼ u On @u X
n T¼t On @t X ð3Þ
:
n T þ ku ¼ 0 On @e X
f ðx; r; l; E; q; t; u; g; x; a; hÞ ¼ 0 ð4Þ
Above parameters should be same in the downscaling model and full-size structure,
and from that, the following conditions should be met:
1=2 1=2
kr ¼ kE ; kt ¼ kl kE kq ; k u ¼ kl ;
ð7Þ
kE k1 1 1=2 1=2 1
kg ¼ l kq ; kx ¼ kE kq kl ; kh ¼ k1
a
From (9), E doesn’t affect u=l, but ah does, which means the module of elasticity
does not influence distortion if the structure contains only one single material.
Parameter scale should meet the following equation:
ku ¼ kl ; kh ¼ k1
a lim ð10Þ
x!1
And we get:
up ¼ um ku ¼ um kl ¼ um lp =lm ð11Þ
Even if kh 6¼ k1
a , ah is proportional to u from (1)–(3), which means the down-
scaling model is still similar to its full-size structure. We also get:
um l p ap hp
up =um ¼ kl kh ka ) up ¼ ð12Þ
l m am hm
A Similarity Approach for Thermal Distortion Measurement 355
Since temperature change doesn’t affect the similarity relation between models, to
keep the similarity of the downscaling model and its full-size structure, following
conditions should be met:
Equation (15) shows, only if the module of elasticity scale and the thermal
expansion coefficient scale of the downscaling model and its full-size model are kept
same for each material, the thermal distortion will be similar.
Since the material property can’t be chosen freely, the Eq. (15) is difficult to fulfill.
Some materials which has no effect on thermal distortion should be discarded as
necessary (Fig. 1).
Payload
1
6 21
4
22
5
7 3 23 8
Z
10
Y X
2 11 15
19 14
18 20 12 13
16
17
9
Platform
Figure 2a shows that the thermal distortion ratios of different positions are same,
and are all equal to the ratio of structure dimension if given same material compositon.
From Fig. 2b, thermal distortion ratios vary with positions even structures have similar
(a) (b)
(c) (d)
geometric shapes, only because of different constitutive relations. Figure 2c and d show
that E does not influence the thermal distortion ratio, but the thermal expansion
coefficient a does.
Table 2. Comparison of the thermal distortion tests for different size structures
Structure type Small-size Middle-size Large-size
Structure size *0.2 m *2 m *3 m
Camera number 2 8 16
Test duration 3 days 20 days 30 days
Thermal expansion coefficient −1.1 10−6/°C −1.3 10−6/°C −1.2 10−6/°C
358 H. Gao et al.
Relative distance
change (μm)
Relative distance is 150mm,αeff≈-1.1 10-6/
Image No.
Temperature
( )
Image No.
Fig. 3. 3D-DIC thermal distortion test and results for small-size structure
Relative deformation (%)
Image No.
Fig. 4. 3D-DIC thermal distortion test and results for middle-size structure
Relative deformation (%)
Image No.
Fig. 5. 3D-DIC thermal distortion test and results for large-size structure
A Similarity Approach for Thermal Distortion Measurement 359
4 Conclusion
Similarity relations between the downscaling model and its full-size structure were
deduced, and a thermal distortion measurement system was established to test struc-
tures with different sizes. It was shown that the measurement precisions for structures
with dimensions from 0.2 m to 3 m were very close, and it was feasible to evaluate the
thermal stability performance of large structure from the measuring results of its
downscaling model, which will reduce the cost, time and difficulty.
References
1. Marschall, C.W., Maringer, R.E.: Dimensional Instability—An Introduction. Permagon,
New York (1977)
2. Wolff, E.G.: Introduction to the Dimensional Stability of Composite Materials. DEStech,
Lancaster (2004)
3. Jacobs, S.F.: Variable invariables—dimensional stability with time and temperature. In:
Critical Reviews of Optical Science and Technology, vol. CR-43, pp. 181–204. Society of
Photo-Optical Instrumentation Engineers, Bellingham, WA (1992)
4. Giaretto, V., Miraldi, E., Ruscica, G., Cabiati, F., Dosio, D.: The Dimensional Stability of a
CFRP Structure Probed with Sub-Millimetre Accuracy in Varying External Conditions, ESA
Paper SP-386 (1996)
5. Cabeza, I., Pradier, A.: Dimensional Stability Verification of an Optical Bench for Space
Applications Under Thermal-Vacuum Environment, ESA Paper SP-386 (1996)
6. Edeson, R., Morris, N., Tatnall, A., Aglietti, G.S.: Dimensional stability testing on a space
optical bench structure. AIAA J. 47(1), 219–228 (2009)
7. Garnich, M.R., Liu, P., Akula, V.M.K., Fitch, J.F., Long, D.S.: Approach for increasing the
static dimensional stability of composite space Structures. AIAA J. 47(4), 1005–1013 (2009)
8. Greenway, P., Tosh, I., Morris, N.: Development of the TopSat Camera, ESA, Paper SP-554
(2004)
9. Figoski, J.W.: Alignment and test results of the Quickbird telescope using the ball optical
systems test facility. In: Proceedings of SPIE: the international society for optical
engineering, vol. 3785, pp. 99–108 (1999). https://doi.org/10.1117/12.367607
10. Murphy, G.: Similitude in Engineering. The Ronald Press Company, New York (1950)
11. Peters, W.H., Ranson, W.F.: Digital imaging techniques in experimental stress analysis. Opt.
Eng. 21, 427–431 (1981)
Design Technology of Accelerated Life
Test Based on the Load Spectrum
Compilation and D-efficiency
Abstract. The accelerated life test is a highly efficient method for long life
evaluation of products, but the operation profile validity and test time opti-
mization are unavoidable in the actual application. Herein, a design technology
research was conducted. First, the load spectrum of temperature and vibration on
operation profile of a rail transportation equipment was measured. Thus the load
spectrum was compiled based on 3-parameter rain-flow counting method and
operation profile was equivalently designed based on Corten-Dolan cumulative
damage model to ensure the validity. Then step-stress accelerated life test model
based on generalized Eyring model and Weibull distribution was developed.
The design method of step-stress accelerated life test with the shortest test time
was given. Finally, the design method was applied in signal output board of rail
transportation case. The residual life assessment was made for the 2-year-used
equipment taking the current usage profile as the first step of the step-stress
accelerated life test which verifying the design method.
1 Introduction
It’s an efficient way of residual life assessment to do the statistical analysis for the
accelerated life test data in the acceleration process of failure or performance degra-
dation by loading higher stress above normal without changing the failure mechanism
of the product [1]. The operation profile validity and test time optimization are the
problems to be solved in the application of the accelerated life test in recent years.
In the 1950s, M. Matsuiski and T. Endo proposed rain-flow counting method
regarding the material’s stress and strain, providing theoretical basis for the conversion
of measured load history to test load cycling. Chen introduced measured load spectrum
compilation based on the rain-flow counting method and equivalent conversion tech-
nology for accelerated verification of load spectrum [2]. Zhu derived the calculation
method of equivalent fatigue stress range based on Corten-Dolan accumulative damage
rule which can make equivalent conversion of loading order and nonlinear factors in
the load spectrum compilation [3]. Liao and Elsayed first presented the optimal design
method for test of products with logarithmic-distribution life which can equivalently
© Springer Nature Singapore Pte Ltd. 2019
S. Sun et al. (Eds.): ICSINC 2018, LNEE 550, pp. 360–370, 2019.
https://doi.org/10.1007/978-981-13-7123-3_43
Design Technology of Accelerated Life Test 361
2 Overall Framework
The residual life assessment could be made for products operated for some time by
after the measurement, compilation and equivalent conversion of temperature and
vibration load spectrum. The general idea are followed.
First, compilation and equivalent conversion of temperature and vibration load
spectrum is conducted. The temperature and vibration response date of the key part on
the signal output board in Beijing railway Yan-Fang line under unmanned operation is
measured. Thereby, the compilation of randomicity, volatility and irregularity regarding
the measured data is made based on 3-peak-valley rain-flow counting method. Mean-
while, the measured load spectrum is equivalently conversed to constant-stress test
profile based on Corten-Dolan nonlinear cumulative damage model.
The test time optimal design based on D-optimality criterion is conducted as fol-
lows. The step-stress nonlinear cumulative damage model based on generalized Eyring
model and Weibull distribution is developed and it can make the convert of fault or
censoring time in different stress step. The D-optimality optimal method for constant
stress and step stress accelerated life test is established by limiting the determinant
value of Fisher information matrix.
Finally, the residual life assessment is made for the signal output board applying in
rail transportation. The test profile of accelerated life test is designed with the load
spectrum compilation and equivalent conversion. At the same time, the signal output
board step stress test plan according to the usage profile is set by the test time optimal
design method of D-efficiency. Hence, the residual life assessment of the signal output
board already operated for some time is completed through experiment and data
analysis.
362 J. Xia et al.
No.1 No.2
1
The door keeps closed during measurement to ensure the board located in real
environment. The measurement results are shown in Figs. 3, 4 and Table 1. The power
spectral density curve by vibration sensor is regular in Fig. 3(1). After compilation
about randomicity, volatility and irregularity, the power spectral density for accelerated
life test is shown in Fig. 3(2). The follow-up research mainly covered the compilation
and equivalent conversion of FPGA temperature curve as shown in Fig. 4.
PSD(g^2/Hz)
5×10-5g2/Hz
Frequency/Hz
0 900 1000
(1)Measurement results (2) Compilation results
Dre is the constant amplitude after the equivalent conversion, ni and Mi represent
the cycles and amplitude for the i-level stress, respectively, rl and Ml refer to the
maximum value of alternating load and alternating stress amplitude, d is the empirical
constant and d = 5.8. k and m are parameters to be identified.
According to the mathematical physics relation between the Weibull distribution
and S-N curve, k and m can be calculated from Weibull cumulative distribution
parameters as below.
k ¼ ðb þ 1Þ lnðAa A0 Þ¼ 0:86
m ¼ b ¼ 0:1881
Design Technology of Accelerated Life Test 365
The equivalent profile of the FPGA temperature curve could be obtained in Fig. 6
after the calculation of equivalent conversion function based on Corten-Dolan theory.
The equivalent constant-amplitude temperature equals to 31.8882 °C, and it could be
used as the first stage for the step-stress to get the operation profile of the accelerated
life test.
Ti and Vi represent the temperature and vibration under the i level of stress. k is the
level of accelerated stress. A, B and D are function parameters to be identified.
Due to the equal value of shape parameter mi, the equivalent time sij under the stress
of Tj, Vj conversed from work time si under the stress of Ti, Vi could be calculated as,
D
gj
B B Vj
sij ¼ si ¼ eTj Ti si ; i 6¼ j ð4Þ
gi Vi
366 J. Xia et al.
rk is the failure data under the k stress. ai is the compensation from the test time
s1 ; s2 ; . . .; si1 under the stress of (T1, V1), (T2, V2), …, (Ti-1, Vi-1) to the failure time
ti1 ti2 . . . tiri si and a1= 0,
X
i1
ai ¼ s1i þ s2i þ . . . þ si1;i ¼ ðgi sj =gj Þ i ¼ 2; . . .; k:
j¼1
X
n
L¼ fIi ½ ln r þ zi expðzi Þ þ ð1 Ii Þ½ expðni Þg ð7Þ
i¼1
The Information matrix F is both the inverse matrix of the covariance matrix of R
(R = F−1) and the expectation of negative two order partial derivative matrix about
likelihood function, so it could be written as,
0 1
A11 A12 A13 A14
X
n B
A22 A23 A24 C
F¼ B C ð9Þ
@ A33 A34 A
i¼1
symmetrical A44
respectively, the two plans are considered to be equal based D-efficiency. And that’s
the definition of D-efficiency based equal Fisher information matrix determinants.
Thereby, the mathematical model of D-efficiency of the total test time for tem-
perature and vibration step-stress accelerated life test when the shortest test time is the
constraint condition could be deduced as below,
min tðA; B; D; mÞ
jM j
s:t: s
jM0 j
Ti ¼ T1 þ ði1ÞðT
k1
k T1 Þ
; i ¼ 1; 2 k
ði1ÞðVk V1 Þ
Vi ¼ V1 þ k1 i ¼ 1; 2 k
Tk and Vk are the highest level of the test stress as design variables. And the D-
efficiency s is also the design variable and it’s set to 0.95 in this paper.
The optimized test time decreased to 1080 h, 920 h shorter than the conventional
plan with the same evaluation accuracy which is more time-saving. The detail of the
optimized test plan is shown in Table 2.
17475000 1
ln tR ¼ 133:4688 þ þ 0:4851 ln V þ ln½lnð1=RÞ
Tk 0:0128
When the duty ratio was set to 2/3 which means the product is in work of 16 h each
day, the average of residual life is calculated to be 15.62 years with confidence (c) of
0.7 and reliability (R) of 0.8.
6 Conclusion
The design technology of accelerated life test based on the load spectrum compilation
and D-efficiency is proposed in this paper, which accomplished the mutual support and
verification of measured load, test optimization and residual life assessment. The
technique ensures the effectiveness and economy of the accelerated life test. The main
creative work is as blow.
(1) The temperature and vibration load spectrum are measured for the signal output
board in operation and it is compiled and equivalently conversed into test profile
which is efficient for the accelerated life test design.
(2) The nonlinear accumulate damage model of step stress is deduced and the test
optimization method based on D-efficiency is defined for the Fisher information
matrix determinant is equal. It established the theoretical foundation for the
accelerated life test optimization design.
(3) A residual life assessment case for the signal output board in rail transportation is
given. It verified the feasibility and practicability of the load spectrum compila-
tion, test optimization and residual life assessment.
References
1. Chen, X., Zhang, C., Wang, Y., et al.: The technology and application of accelerated life test.
National Defense Industry Press (2013)
2. Chen, Q.: Research on the equivalent load spectrum method in accelerated validation of the
fatigue life. Beihang University (2014)
3. Zhu, H., Yu, Z., J, L.: Calculation method of equivalent fatigue stress range based on Corten-
Dolan accumulative damage rule. J. Highw. Transp. Res. Develop. (2010)
370 J. Xia et al.
4. Liao, H., Elsayed, E.A.: Equivalent accelerated life testing plans for log-location-scale
distributions. Nav. Res. Logist. (NRL) 57, 472–488 (2010)
5. Ng, H.K.T., Balakrishnan, N., Chan, P.S.: Optimal sample size allocation for tests with
multiple levels of stress with extreme value regression. Nav. Res. Logist. 54, 237–249 (2007)
6. Cai, C.: The development of automotive fatigue life test technique based on the real road load
spectrum. Automotive Collaborative Innovation Center of Chongqing University (2015)
7. Zhao, C.: Study on preprocessing and feature extraction of high speed rail vibration data
based on cloud computing. Southwest Jiaotong University (2013)
Realization of Interconnecting
Application of Non-secret-related Network
and Secret-Related Network Based
on Unidirectional Optical Shutter
Abstract. The terminals, networks and systems involved in the daily work,
scientific research and production of military enterprises are all defined as secret-
related. In accordance with the national requirement, physical isolation from the
internet and internal non-secret-related network must be maintained. But at the
same time, each enterprise also deploys a private network for testing, experi-
menting and processing, and the data exchange between the networks must be
completed by manual transfer, which can be as far as one or two hours fast and
as slow as half a day, and has been far from meeting the requirement of military
enterprises’ scientific research tasks in the information age. In this paper, the
utilization of unidirectional optical shutter is proposed to realize the data
transmission between non-secret-related network and secret-related network.
Furthermore, the feasibility is analyzed and the technical verification is
completed.
1 Introduction
The whole process of transfer work requires manual participation, and the operation
is tedious and laborious, which leads to a large delay in data transfer, seriously affecting
work efficiency.
2 Design Plan
concerned. Therefore, this paper mainly analyzes the data exchange from the test
network to the secret-related network, which has been able to meet the transfer need of
more than 80% of scientific research personnel, and achieve the stated object.
distinguishing it from the existing IP address field; on the other hand, it adds network
control measures to the address field to strictly restrict data access. For the ACL
reference under the VLAN interface, the following principles should be complied with:
(1) Secret-related network segments in the data exchange area: Secret-related net-
works are only allowed to be accessed by the user departments and other
departments are prohibited from accessing;
(2) Test network segment in the data exchange area: only “user departments of the
unit” within the test network are allowed to access, and “externally-connected
areas (external units) of the test network” are prohibited to access.
Firewall Design. In the data exchange area, two firewalls are configured as the logical
boundaries of secret-related network and test network respectively. According to the
minimum release principle of data, the source IP address, destination IP address and
TCP/UDP port are restricted. They mainly are:
(1) Restricting the access of the data sending server to the terminal IP address of the
“user department of the unit” in the test network, being convenient for the sub-
mission of data transmission application and uploading of the data;
(2) Restricting the “data sending server” solely to send the data to the interface of the
test network of the optical shutter equipment.
(3) Restricting the interface of optical shutter devices in the secret-related network
can only send data to the data receiving server.
(4) Restricting the access to the data receiving server only by the “user department”
terminals in the secret-related network, being convenient for the examination and
approval of application procedures and downloading of data.
Anti-virus Wall Design. The anti-virus wall is connected to the link in a serial way,
aiming at checking and killing Trojans, viruses and malicious attacks on the data
entering the secret-related network. The feature library must be ensured to be updated
at least once a week.
Intrusion Detection Design. By configuring the port mirroring function on the secret-
related network interchanger, the traffic mirroring can be sent to the intrusion detection
machine to detect and analyze the data so as to ensure that no malicious attack data are
mingled. The feature library must be ensured to be updated at least once a week.
Network Architecture. The application mode of the test network has its particularity.
In the test process, the terminal of the task cooperative unit also needs to join up the
network to obtain data. The current test network is a two-tiered architecture, where all
terminals share the same address field, and there are no effective control measures for
access isolation between the terminal of the unit and the terminal of external unit. In
view of this, it is necessary to redesign the test network, by adding access control
376 Q. Miao et al.
measures, standardizing terminal security policy settings, etc., which can reduce
security risks from the source. The redesigned test network is shown in Fig. 5.
The network adopts the divisional design, which is divided into the terminal area of
the unit and the external terminal area. The access control between the regions is
completed through the firewall, restricting the server IP address which can be accessed
by the external terminal area, and setting up the anti-virus wall on the critical link to
avoid the external terminals’ carrying viruses.
Safety Control Measures. Although the safety protection standard and requirement of
the test network are not as high as those of the secret-related network, the working
ideas are the same. The safety control within the network mainly consists of the
following aspects:
(1) Firewall: Increase the deployment of firewalls, in accordance with the principle of
“traffic minimization” to control the flow of data within the network, specially
focusing on the control of the data access from external terminal;
(2) Antivirus: Antivirus system is deployed in the “terminal area of the unit”, which is
C/S architecture. The computer needs to be installed with an anti-virus software
client. Through the configuration strategy, the virus library of the system is kept
up-to-date and the virus is checked and killed at least once a week. The anti-virus
wall will be deployed in a serial way between the regions to avoid the virus carried
by external terminals which will affect the operation of the terminal of the unit.
(3) Network access control: a network admission control system will be deployed
within the network to restrict the access, except for the MAC address of legitimate
terminals to enter the network, and wipe out the illegal terminal access through the
functions such as “MAC address whitelist”;
Realization of Interconnecting Application of Non-secret-related Network 377
(4) Inspection software: network security inspection system software will be installed
in the computers and servers in the “terminal area of the unit”. The computer user
records, system access behavior, connection records of USB interface, etc. are
checked regularly by a specially-assigned person, and the problems found will be
corrected in time.
3 Technical Identification
In order to test the actual performance of the above design schemes, a company has carried
out the small-scale verification. The local controlled interconnection is performed
according to the topological link shown in Fig. 4. The key device is unidirectional optical
shutter device developed by domestic enterprises. The uploading and downloading of
data are accomplished through a dedicated transfer system. Compared with the traditional
way, the intuitive improvement can be seen in Table 1.
More detailed data records are available (running cycle is about 3 months) as
follows:
(1) Transmission rate: matching 1000 Mbps Ethernet and the data transmission of
1 GBps size needs about 12 s;
(2) Transmission delay: the delay of data transmission from “data sending server” to
“data receiving server” is second level.
(3) System stability: there are totally more than 200 times of data transmission, single
data size ranges from 10 MB to 15 GB, and one-time success rate of data
transmission is about 95%;
(4) Multi-person parallel operation: there are up to 30 people online simultaneously,
conducting data uploading, data downloading and the operation of process
approval respectively. The system has no obvious delay and functions well.
(5) Security: No virus, malicious code and Trojans are found in the secret-related
network, and no illegal data transmission event occurs.
378 Q. Miao et al.
4 Conclusion
Through the above analysis and technical verification, it can be seen that the controlled
“interconnection” between the non-secret-related network based on one-way optical
shutter and the secret-related network can ensure that the data can only enter the secret-
related network unidirectionally and without any feedback, so as to avoid the missing
of secret-related information and the risk of leakage. The construction of security
protection system in data exchange area and test network can ensure that the data
entering the secret-related network does not contain virus and malicious code, and will
not affect the stable operation of the secret-related network. In the process, supple-
mented by the formulation of management system, the management main body and
responsibilities of all parties of “interconnection” are defined, which can reduce more
than 90% of the “interconnection” risk.
The interconnection mode discussed in this paper has the value to be expanded,
which can be used as a reference for the construction of secret-related networks and
other non-secret-related networks used internally. The controlled “interconnection” of
the network, on the one hand, practices “intelligent manufacturing 2025” and makes
bold innovations and changes; on the other hand, it effectively optimizes work process
and improves the efficiency of research in military enterprises.
Realization of Interconnecting Application of Non-secret-related Network 379
References
1. Yan, M.-y., Wei, L.: The design of the one-way isolation net gap and its transmission
reliability. Value Eng. 219–220 (2014)
2. Ao, L-q, Chen, Z.: Research on information resource sharing scheme based on network
isolation technology. Softw. Guide 16(6), 163–167 (2017)
3. Zuo, Z.-y., Pu, X.-p.: Analysis and practice of non-confidential information exchange in
military enterprises. Inf. Secur. Commun. Priv. 117–122 (2016)
4. Wang, Y.-j., Yang, J.-h., Guo, G.-t., Liu, Y-t.: Analysis and prospect of physical isolation
technology for network security. Inf. Secur. Commun. Priv. 117–122 (2016)
5. Zhai, J.: Network Edge—Evolution of Thought of Safety Protection 23–24 (2013)
6. Zhang, X.-q.: Analysis of principle and function of unidirectional optical shutter. Netw.
Secur. Technol. Appl. 99–100 (2016)
7. Wei, X.-z.: Study on Dual-network Safe Access on the Basis of Double Unidirectional
Optical Shutter. Beijing University of Chemical Technology (2015)
8. Wan, Y.-l., Zhu, H.-j., Liu, H.-z., Zhang, K.-f.: Research on reliability of unidirectional
transmission system based on optical shutter. Hacking Def. Res. 25–27 (2010)
9. Liu, B.: Design and Implementation of Multiple Networks Synthetic Integrated System.
Northeastern University (2014)
10. Zhong, J., Li, Q.: Research on information security management based on network isolation
technology. Sci. Educ. J. 158 (2016)
Application of Ion Beam Etching Technology
in Spacecraft Encoder Lithography
Abstract. The encoder is a core component of encoding sun sensor, the pho-
tolithography precision and quality has the serious impact on the measurement
accuracy of sun sensor. Traditional encoder photolithography uses wet etching
technology, and line edge has sawtooth and burr phenomenon, photolithography
accuracy is low, can not meet the aerospace products high precision and high
quality requirements. In this paper, iron beam etching technique is used to verify
the feasibility of ion beam etching technology used in encoder, which solves the
problem of edge sawtooth and burr, and the technical problems of low pho-
tolithography accuracy, which laid the foundation for the further development of
products to higher precision.
1 Introduction
In recent years, Chinese spacecraft are developing to the direction of low cost, high
precision and high reliability, and spaceflight products need to be miniaturized, high
precision and high quality. The same is true for encoding sun sensors which are used as
a general product of spacecraft [1].
The encoder is the core part of the encoding sun sensor. By coating and pho-
tolithography, 14 code channel patterns arranged according to certain rules are pre-
pared, which are composed of more than 300 rectangular windows with different width
of 0.07 mm–8 mm. The lithography accuracy and quality of the encoder seriously
affect the measuring accuracy of the encoding sun sensor, and this error is random
error, which can not be corrected by the way of error compensation. Therefore, the
influence of this error must be reduced by improving the lithography accuracy [2].
The traditional lithography of the encoder adopts wet etching technology, and there
are some phenomena such as sawtooth and burr on the edge of the graph. The
lithography accuracy is low, which can not meet the requirements of high precision and
high quality of aerospace products. This paper attempts to use ion beam etching
technology to verify the feasibility of ion beam etching technology used in encoding
sun sensor lithography, in order to solve the technical problems of irregular edges and
low lithography accuracy of product lithography.
Ultraviolet lamp
Development liquid
Mask
pattern
Check
Edge burr
Edge sawtooth
Fig. 2. Unqualified products with edge sawtooth and many burrs in photolithography (X40)
In order to analyze the specific situation of the unqualified item, the unqualified
encoder produced within one year was investigated statistically. There are 91
unqualified encoders participating in the statistical investigation, from the coating,
lithography, flanging, cleaning of the encoder, surface shape and other aspects of the
collection, formed an unqualified items questionnaire, see Table 1. According to the
results of the survey, the arrangement is drawn, as shown in Fig. 3.
It can be seen from the questionnaire of unqualified items and the Pareto diagram
that: in the unqualified products produced within one year, the proportion of pho-
tolithography is 68.2%, in which the main item (42.9%) is the precision of lithography,
and the minor item (25.3%) is that the edge of lithography is not up to standard.
According to the principle of quality improvement, two main quality problems caused
by the low pass rate of encoder can be solved by solving the problems of over-precision
and unqualified edge of lithography.
Application of Ion Beam Etching Technology in Spacecraft Encoder Lithography 383
encoder lithography, it is necessary to determine the ratio of the thickness of film layer
to the rate of etching layer thickness through selective ratio test, so as to ensure that the
photoresist can protect the coating which does not need to be etched in the process of
etching. After verifying the feasibility of ion beam etching, through a series of
experiments, the key optimum technological parameters of ion beam etching are
selected, and the experimental samples are prepared. Finally, the edge quality and
lithography accuracy of the samples are analyzed and evaluated under microscope, and
the application effect is confirmed.
Technical scheme determination
Based on the above ideas, the ion beam etching test of chromic film encoder is
carried out, and the test scheme is shown in Table 1.
(2) Technical Measures
In this paper, IBE-150B standard ion beam etching machine is used to carry out ion
beam etching test of chromic film encoder. The main performance index of ion beam
etching machine is shown in Table 2.
Table 2. Main performance indexes of IBE-150B standard ion beam etching machine
Bed dimension U150 mm (maximum sheet diameter 4 in.)
Ion beam incident angle 0–90º arbitrary adjustment
Etching rate 10 nm–200 nm/min (relating to etching materials
and processes)
Etching inhomogeneity ±5%
Ar+ ion energy range 100–1000 eV, Continuous adjustable
Ion beam current density 0–1 mA/cm2, Continuous adjustable
Effective ion beam diameter U100 mm
The sample of ion beam etching of encoder is sent to the professional inspection
mechanism for lithography accuracy detection, and the number of drawing figure is 124
per piece of encoder. Testing qualification and testing devices are as follows.
• Testing qualification:
• The uncertainty of calibration results is evaluated and expressed in accordance with
the requirements of JJF1059 series standards.
• Calibration of environmental conditions:
• Temperature: (20±0.2) °C Humidity: (50±10)% RH
• Calibration of metrology base (standard) devices (including reference materi-
als)/main instruments (Table 3).
Application of Ion Beam Etching Technology in Spacecraft Encoder Lithography 385
Where x stands for the horizontal ordinate and y for the ordinate.
The experimental results show that under this condition, the etching rate of chro-
mium film is 17.5 nm/min, the etching rate of photoresist is 10 nm/min, the selection
ratio of chromium film and photoresist is 1.75:1, that is to say, under the same con-
ditions, the etching rate of chromium film is faster than that of photoresist by ion beam.
This is also a need trend. After obtaining this data, we can draw a conclusion that when
the thickness of chromium film is 350 nm under the condition of medium energy
250 V and medium plasma beam 100 mA, if the thickness of photoresist is more than
200 nm, the etching of chromium film can be guaranteed in theory. When chromium
film is etched clean, photoresist can also play a masking effect.
the etching rate, but the larger the voltage, the larger the ion energy will lead to the
substrate damage. Therefore, it is necessary to select the appropriate voltage to
maintain the appropriate etching rate. In addition, the ion beam density is also one of
the main factors directly affecting the etching rate. Ion beam energy and beam current
in ion beam etching are controlled [7]. The two key parameters of the effect, the higher
the energy and the higher the beam current, the faster the etching rate of the chromium
film is, but the greater the damage to the photoresist. In order to guarantee that the
photoresist can not only play a better protective role, but also improve the etching
efficiency, it is necessary to set different process parameters for process test [8].
According to the experience of ion beam etching parameter selection, three groups
of experiments have been carried out, including low energy 200 V, high beam current
120 mA, medium energy 250 V, medium beam current 100 mA, high energy 500 V
and low beam current 80 mA. The test parameters are shown in Table 5. The etching
results of chromium film encoder are shown in Table 6.
Table 5. Ion beam energy and beam parameters of chromium film encoder
Sample Energy and beam Parameter Etching Chromium etching
number time rate
009 Low energy, high 200 V, 25 min 14 nm/min
beam 120 mA
010 Medium energy and 250 V, 20 min 17.5 nm/min
beam 100 mA
011 High energy, low 500 V, 30 min 11.7 nm/min
beam 80 mA
Under the Leica M2500 metallographic microscope (Fig. 4, left) and 200 times
(Fig. 4, right), the black rectangular strip edge is the edge of the etched figure. As can
be seen from Fig. 4, the chromium film layer in the window is etched cleanly, and the
graphic edges are neat. No sawtooth, no burr, good etching effect.
It can be seen from the above experimental conditions that the etching parameters
of high energy 500 V and low beam current 80 mA have not been etched clean and do
not meet the requirements of lithography index. Therefore, the test conditions are not
suitable. The etching parameters of medium energy 250 V and medium beam current
100 mA are adopted, the graphic edges are neat, the sawtooth and burr are free, the
resolution is high, the lithography index is satisfied, and the etching efficiency is the
highest. So the medium energy 250 V and the medium beam 100 mA are chosen as the
best parameters of ion beam etching.
Fig. 4. 010 Chromium film encoder sample window edge etching effect (X200)
The ion beam energy and beam current used in the preparation process are the same
and the operation process is the same. The edge effect of lithography pattern was
observed under Leica M2500 metallographic microscope by random sampling of 1 Cr-
6 number disk out of 10 etched samples. It can be seen from the figure that the edge of
lithography pattern, sawtooth and burr of wet etching sample (Fig. 5) is serious. The
edge of the Ion beam etching of Cr-6 sample (Fig. 6) lithography is neat, no sawtooth,
no burr, and the edge of the ion-beam etched lithography is improved obviously.
Fig. 5. Edge etching effect of wet etch- Fig. 6. Ion-beam etching of Cr-6 code-
ing encoder (X100) disk graphics edge effect (X100)
Fig. 7. Lithography error distribution curve numbered as Cr-7 encoder (−0.6 lm–0.7 lm)
lithography index of 2.0 lm, the lithography accuracy of the Cr-7 encoder prepared
by ion beam etching technology meets the requirements. The precision distance
between the upper and lower specifications is sufficient margin.
In order to observe the whole lithography error distribution of 10 encoders, the box
diagram is drawn by using 124 lithography precision test data extracted from each
encoder, as shown in Fig. 8. In the graph, the horizontal coordinate represents the
encoder number, the vertical coordinate represents the error value, the unit is lm, the
specification limit is LSL = −2.0 lm under the lithography precision, the upper
specification limit is USL = 2.0 lm, and the black spot line in the diagram is the
median line. It can be seen from the box diagram in Fig. 8 that the lithography errors of
the 10 encoders are within the upper and lower specification limits, which meet the
requirements of the lithography accuracy 2.0 lm. Where the maximum lithography
Application of Ion Beam Etching Technology in Spacecraft Encoder Lithography 389
error appears in the Cr-5 encoder, the maximum value is 1.3 lm, the maximum value is
35% margin from the upper specification limit, the minimum lithography error appears
on the Cr-5 encoder, the minimum value is −1.4 lm, and there is a 30% margin at the
minimum distance.
5 Conclusion
In this paper, a lithography method based on ion beam etching is designed according to
the problems existing in the sun sensor encoder lithography of spacecraft. Through a
series of experiments, the feasibility of ion beam etching applied to the encoder
lithography of spacecraft solar sensor is verified. The application results show that the
ion beam etching technique can ensure that the edge of the lithography of the encoder is
neat, no sawtooth, no burr, and at the same time, the precision of encoder lithography
has been improved from 3 lm to 2 lm. Ion beam etching technology has
improved the quality and precision of the lithography pattern of the encoder of the solar
sensor of spacecraft.
References
1. Zheng, K.: The development direction and prospect of solar sensor. Control Eng. pp. 6–9
(2004)
2. Tu, S.: Satellite Attitude Dynamics and Control. Aerospace Press, Beijing (2003)
3. Zhou, H., Yang, H.: Present situation and prospect of lithography and microfabrication
technology. Micro/Nanoelectron. 49, 936–939 (2012)
4. Orloff, J., Utlaut, M., Swanson, L.W.: High Resolution Focused Ion Beam, Chapter 2, In:
Physics of Liquid Metal Ion Sources. Kluwer Academic Publisher (2003)
5. Li, J.: Progress of dry etching technology in semiconductor device process. Micromachining
3, 43–48 (1993)
390 S. Qin et al.
1 Introduction
The communication satellite power supply system is generally composed of solar cell
array, power controller, battery group and distribution system. Solar cell array gener-
ates energy, and the power is adjusted by power controller to form a primary power
supply bus. In addition, the power controller stores the excess energy generated by the
solar cell array in the battery set for use when the satellite solar light is invisible or the
output power of the solar cell array is insufficient. The distribution system is respon-
sible for distributing the primary bus energy to each load, which not only ensures the
low energy loss, but also ensures the safety of energy transmission [1].
The satellite power supply system is related to the lifeblood of the satellite, and the
failure or instability of the power supply system is directly related to the success or
failure of the satellite mission. For example, the INTERSAT19 satellite solar panel
failure caused the satellite power loss of 50%, the satellite mission was seriously
affected, the domestic satellite power supply system transmission path SADA fault,
resulting in satellite mission failure. The statistical results of the faults of satellite power
system in orbit at home and abroad in recent years show that the problems concerning
solar cell array, power supply regulation, battery group and distribution fault are rel-
atively high, and the omnidirectional design of power supply system is carried out. It is
urgent to improve the reliability and operation stability of satellite power system.
The design and verification technology of communication satellite power supply sys-
tem mainly includes digital simulation verification technology, semi-physical simula-
tion verification technology and all-physical test verification technology.
Simulation database
Simulate
Solar cell array occlusion data Device simulate data data
Solar cell array power predict data Load simulate data
Fig. 1. Full digital simulation verification platform for super power supply system
Fig. 2. Principle block diagram of super power supply system semi-physical simulation
platform
Validation Technology in Super Power Supply System 395
Twinkle Prima
load ry bus
Bus load
Interface unit for Platform of energy
Impedance test
battary group management strategy
system
• Test the correctness and matching of the interface between the devices of the power
supply system, and test the power supply path;
• A primary bus output characteristic test, including bus impedance, bus ripple and
bus transient response test, etc.
• Energy management strategy test, including battery charge and discharge man-
agement logic verification, various operating conditions switching threshold test
and so on.
396 R. Jiao et al.
The simulation results of output power of solar cell array are as shown in Fig. 4. The
bottom of the diagram shows the occlusion of the two solar arrays in the north and
south, the gray is occluded, the white is unoccluded, and the occlusion rate curve of the
upper right corner of the diagram can be seen. In the simulation period, the maximum
power loss rate of solar cell array is about 25%. The middle diagram of the right side of
the diagram shows the output curve of solar cell array current in the same simulation
time period, which shows the dynamic characteristics of the output current of solar cell
array in the case of occlusion. The bottom right corner of the diagram shows the typical
values of the solar cell array voltage output histogram in this time period.
The following figure shows the simulation results of impedance and dynamic
characteristics of power supply system.
Figure 5 shows the output impedance and load input impedance diagram simulated
in S3R domain. The maximum output impedance of the power supply system is 25.4 m
X, which does not intersect with the input impedance and the system is stable.
Figure 6 shows the BDR output impedance and the maximum load input impe-
dance diagram. The maximum output impedance of the power supply system is 38 m
X, which is not intersected with the input impedance and the system is stable.
Figure 7 shows the characteristics of output impedance and load input impedance
in BCR domain. The maximum output impedance of power supply system is 26 m X,
which is not intersected with load input impedance, and the system is stable.
Validation Technology in Super Power Supply System 397
Fig. 5. Impedance simulation results in S3R Fig. 6. Impedance simulation results in BDR
domain domain
Table 1 is the test list of power supply system test system, including power system
interface inspection [11], power system health inspection, power system function test
and power system special test. The test of power system interface is to ensure the
matching and safety of single-machine interface between power systems. Power system
health check is to do low-power level of the power system functional inspection. The
function test of power supply system is to carry out 1:1 function verification of power
supply system according to orbit condition. The special test of power supply system is
to carry out FDIR and reliability growth test of power system [12].
398 R. Jiao et al.
Table 1. List of test items for power supply system test system
NO. Project category Project name
1 Power system interface PCU interface check
2 check Battery group interface check
3 Others interface checks
4 Power system health check PCU health check
5 Battery group health check
6 Others health check
7 Power system function test Busbar overvoltage protection function test
8 S3R shunt regulation function test
9 S3R shunt regulation function test
10 BCR charging function test
11 BDR discharge function test
12 BDR discharge redundancy function test
13 Battery charge function test
14 Battery discharge function test
15 Special test of power system Incoming and outgoing shadow switching function
test
16 Load step dynamic response test
17 FDIR test of power supply system
18 Power system reliability growth test
4 Conclusion
In this paper, the design and verification system of super power supply system is
designed, the full digital simulation verification platform and the all physical test
certificate platform are built, and the practical application is carried out. The application
results show that, the full digital simulation platform effectively verifies the system
characteristics of the power supply system and ensures that the design of the power
supply system meets the requirements of the satellite system. The whole physical test
platform is used to test and verify the satellite power system products, and the relia-
bility of the system is tested to ensure that the power supply system does not have any
doubt. After that, the semi-physical simulation of the super-power supply system will
be carried out, and the influence of the power supply system on the power supply
system will be simulated under the condition of part of the device failure. To ensure the
satellite power supply system single failure service continuity, double failure to ensure
satellite safety.
Validation Technology in Super Power Supply System 399
References
1. Li, G.-x.: Introduce to Spacecraft Power System Technology. China Aerospace Publishing
House, Beijing (2008)
2. Zhang, Y., Jiang, D.-s., Liu, Z.: Research on Modeling and Simulation of Satellite Electrical
Power System
3. Ma, X., Zhang, D., Xu, D.: S3R&S4R control strategy analysis and simulation. Manag.
Control Technol. 26(6), 45–50 (2007)
4. Miao, D., Zhang, D.: Simulation of S4MPR application for spacecraft power supply system.
21(19), 6298–6302 (2009)
5. Zhao, G.-w., Han, F.-t.: MPPT technology for high orbit and high power satellite power
system, design and research of power technology, 1002-087X 08-1675-04 (2016)
6. Ma, P.-b., Wu, J.-h., Ji, J., Xu, X.-l.: Software environment and application of dSPACE real-
time simulation platform. J. Syst. Simul. 1004-731X 04-0667-04 (2004)
7. Zhnag, X., Guo, Y., Tong, L., Chen, H.: dSPACE real-time simulation and controller
parameters tuning for three-phase PWM rectifier. Trans. Chin. Electrotech. Soc. 28, 219–224
+232 (2013)
8. Liu, L.: Faults simulation and diagnosis of satellite power system. Dissertation for the Master
Degree in Engineering, Harbin Institute of Technology (2015)
9. Xiao, L.: Faults simulation and diagnosis of satellite primary electrical power system.
Dissertation for the Master Degree in Engineering, Harbin Institute of Technology (2016)
10. Wang, H.: Spacecraft electrical test technology. Beijing Institute of Technology Press (2018)
11. Jiao, R., Chen, Y.: Design and realization of automatic system in satellite low frequency
interface test. In: Proceedings of Spacecraft Extend Life and Reliability Study, pp. 397–406.
China Aerospace Science and Technology Corporation, Beijing (2011)
12. Song, S., Pan, J.: Key ability of EPS in integration test. In: Proceedings of Technology
Committee of Beijing Institute of Spacecraft System Engineering, pp. 23–26. Beijing
Institute of Spacecraft System Engineering, Beijing (2009)
Research on Application Method of 3D Digital
Simulation Technology in Spacecraft Assembly
1 Introduction
the foundation for full 3D digital design and manufacturing of spacecraft products,
which provides references for assembly of spacecraft products and can promote
implement of related industry work.
Assembly process design can prepare the process for spacecraft product assembly.
Since the product design information is received, the assembly process plan is prepared
according to the structural features of products, the assembly process model is created
by combining the manufacturing resource information, assembly process design pro-
cess of spacecraft is digitalized, the assembly task is scheduled and assembly process is
simulated under the 3D environment, and the feasible assembly process execution
documents can be obtained [7, 8].
With complicated load assembly of a spacecraft model as one example, digital
applications of its assembly process are studied by using the design method for digital
assembly process in this paper.
2. The camera body and the lens hood are installed on the spacecraft floor via the
integrated structure;
3. The camera radiation plate bracket is installed on the floor of the spacecraft;
4. The south and north radiation plate and 4 heat pipes (GFY07-4) of the camera have
no direct installation interfaces with the whole star. The radiation plate is installed
on the radiation plate bracket.
5. Two ends of the heat pipes are connected to the radiation plate via the hoop (4
hoops at each end). The middle part of the heat pipe is connected to the heat
collector plate under the camera body via 3 hoops. After all heat pipes are installed
in place, the gap in the contact area among heat pipes, radiation plate and heat
collector plate is less than 0.2 mm.
Fig. 3. An operator supports tray and hoop together with 4 heat pipes
(g) The operator outside the unwheeling keeps “U-shaped” coplane by using the
terminal tube hoop and fastener via the regulation plate on the side of the terminal
tube (tooling). The operator inside the unwheeling regulates looseness of the
fastener at the camera heat collector plate to assist the operator outside the
unwheeling to regulate the “U-shaped” coplane of the heat pipe, shown as the
Fig. 4.
(h) The operator inside the unwheeling completes all hoops and fasteners at the heat
collector plate of the camera, fasten them by the regulated torque, and disassemble
the regulation plate on the side;
(i) Detect the gap in the contact part between the heat collector plate and the heat
plate by using a feeler. The gap shall be less than 0.2 mm;
(j) Disassemble 4 heat pipes and scrape and coat GD414 silicon rubber in the contact
area of the heat pipe of the heat collector plate of the camera.
(k) Install 4 heat pipes again according to the step (4)–(10);
Wipe the overflowing GD414 silicon rubber by using the dust-free cloth dipped with
the absolute ethyl alcohol;
(l) Install the camera’s bottom cover again;
4 Conclusions
References
1. Jayaram, S., Jayaram, U., Kim, Y.J., et al.: Industry case studies in the use of immersive
virtual assembly. Virtual Reality 11(4), 217–228 (2007)
2. Christiand, J.Y.: Assembly simulations in virtual environments with optimized haptic path
and sequence. Original Res. Artic. Robot. Comput. Integr. Manuf. 27(2), 306–317 (2011)
3. Yin, Z.P., Ding, H., Xiong, Y.L.: A virtual prototyping approach to generation and evaluation
of mechanical assembly sequences. Proc. Instn. Mech. Engrs. Part B(S0954-4054), (04): 87–
102 (2004)
408 B. Zhang et al.
4. Selvaraj, P., Radhakrishnan, E., Adithan, M.: An integrated approach to design for
manufacturing and assembly based on reduction of product development time and cost. Int.
J. Adv. Manuf. Technol. 42, 13–29 (2009)
5. Curran, R., Gomis, G., Edgar, T., et al.: Integrated digital design for manufacture for reduced
life cycle cost. Int. J. Prod. Econ. 109, 27–40 (2007)
6. Ning, R., Zheng, Y.: Research progress and development trend analysis of virtual assembly
technology. Chin. Mech. Eng. (15), 1004–1398 (2015)
7. Zhang, L., Zhu, X., Liu, Z., et al.: Research on planning technologies of spacecraft product
assembly process. Chin. Mech. Eng. (02), 265–268 (2014)
8. Han, W., Ma, P.: Research on design and simulation application of digital assembly process.
Chin. Technol. Inf. (05), 127–129 (2015)
9. Bullinger, H.J., Richter, M., Seidel, K.A.: Virtual assembly planning. Hum. Factors Ergon.
Manufact. 10(3), 331–341 (2000)
Big Data Workshop
Network Traffic Prediction Based on LSTM
Networks with Genetic Algorithm
1 Introduction
Cloud computing has become a popular topic in the field of information computing
resources. There are an estimated 50 billion connected devices worldwide by 2020 [1].
Interconnections among these devices would cause generation of massive data. These
data must be stored and processed so that their profits can be explored and utilized.
Many large data centers are built to provide different type of services. Network traffic
prediction is a fundamental service that is needed to enable any traffic management
operation, such as differentiating traffic pricing and treatment (e.g., policing and
shaping) and security (e.g., firewall, filter and anomaly detection) [2].
Numerous models have been proposed to exploit the temporal and self-similar
property of network traffic classification and prediction. Existing models, such as auto
regressive integrated moving average (ARIMA) [3] and support vector machines [4],
are mainly linear models that cannot well describe high dimensional data and capture
complex nonlinear relationships [5]. To cope with nonlinear modeling, neural network-
based regression algorithms are widely used since neural networks are capable of
modeling highly nonlinear and complex structures [6]. However, it is reported that they
are prone to over-fitting problems and cannot always achieve adequate performance in
certain applications [7]. Recurrent neural network (RNN) is also advocated to predict or
class the time series data. In this paper, we want to explore the potential of deep neural
networks as feature extractors and learn long term temporal dependencies by LSTMs, a
type of RNN [8]. Reference [9] applied the bilinear recurrent neural network (BLRNN)
to the real world Ethernet network traffic data set and obtained decent results. However,
the parameter setting was based on local information, which was not globally opti-
mized. For random network traffic, structural parameters in LSTM model should be
carefully set to achieve high prediction accuracy.
Genetic algorithm (GA) is a class of evolutionary algorithms (EAs). GA is widely
used for search and optimization problems, where structures of chromosome and
evolutionary operations such as selection, crossover and mutation, are mimicked. In
recent years, significant attention has been attracting by integrating GA with deep
learning. More specifically, it has been utilized for more efficient selection of hyper-
parameters, e.g. kernel size, and network structures [10, 11].
In this paper, we apply GA-LSTMs to network traffic prediction, where GA is
developed to optimize the hyper-parameters in the LSTM model. The GA-LSTMs
model is mainly composed of two parts. One is the LSTMs model itself that serves as a
mapping rule between input and output variables and also the core in fitness function.
The other part is GA that optimizes the hyper-parameters in the LSTMs model.
Experimental results demonstrate that GA-LSTMs outperforms ARIMA and LSTMs in
terms of the prediction accuracy.
The rest of this paper is organized as follows. Section 2 introduces the LSTMs
model and briefly describes GA. Section 3 presents the GA-LSTMs method and the
implementation process. Section 4 discusses the experiment design and results. Finally,
the conclusion is given in Sect. 5
2 The Model
2.1 LSTM
In this section, we introduce deep learning-based structures used for network traffic
forecasting. LSTMs are good at memorizing information for a long time. As the
accuracy of the model may be affected by more previous information, LSTMs become
an appropriate choice for employ. LSTMs have been demonstrated to be effective in
language translation [12] and speech recognition [13].
Figure 1 shows a basic LSTMs architecture. The network model is composed of
three stages: input, feature extraction (based on LSTMs), and regression stage (flatten
layers). We use the window method in the input stage. The input window [Xt–n, Xt–n+1,
…, Xt–2, Xt–1, Xt] is consecutive samples for regression, where each of them is
Network Traffic Prediction Based on LSTM Networks 413
normalized to fit in [0, 1]. The historical samples are forwarded to the LSTMs for
temporal traffic feature extraction and then to flatten layers for single upcoming net-
work traffic prediction.
2.2 GA
GA is a meta-heuristic that mimics the biological evolution according to the rule of
nature, i.e. the survival of the fittest. We encode all hyper-parameters in the LSTMs
model as a chromosome (also called individual/solution), where each hyper-parameter
is a gene with predefined value range. A population consists of a number of chro-
mosomes. The initial population is randomly generated. Each individual is evaluated by
some performance metrics of the LSTMs model. Fitter individuals are more likely to
survive than others in the population in the selection process. Then, crossover and
mutation are performed to the selected individuals. Selection, crossover and mutation
are repeated to evolve the population for promising values for the hyper-parameters in
the LSTMs model until that a predefined number of generations are finished.
414 J. Chen et al.
3 Methodology
In order to predict the future service traffic accurately, a LSTMs model based on GA is
proposed, where the flowchart is illustrated in Fig. 2. LSTMs is for network traffic
prediction while GA is utilized to tune the hyper-parameters in the LSTMs network.
Besides, the LSTMs network is integrated into the fitness evaluation in GA.
1 Xn
MAE ¼ jyðtÞ xðtÞj ð1Þ
n t¼1
rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
1 Xn
RMSE ¼ ð yð t Þ xð t Þ Þ 2 ð2Þ
n t¼1
Where y(t) and x(t) are the predicted and actual output values at time t, respectively,
and n represents the number of predictions.
4.4 Results
The actual data points are verified by various models, including ARIMA, LSTMs and
GA-LSTMs. The network traffic prediction results are shown in Fig. 3. In Fig. 3, we
sample these data points every five minutes because the amount of data points are too
many to be shown. Figure 4 is an extracted segment from Fig. 3.
In Figs. 3 and 4, it is easily seen that GA-LSTMs has a higher prediction accuracy
than models based on LSTMs and ARIMA. Besides, data points in the next 240 min
are also forecasted in Fig. 5. The forecasting data points are almost the same with the
actual traffic data, reflecting a promising performance of GA-LSTMs. In fact, compared
with neural network, the GA-LSTMs model not only reduces the computational
complexity, but also improves the generalization ability for the model.
Fig. 5. The network traffic prediction results obtained by the GA-LSTMs model
Table 2 shows the MAE and RMSE results of ARIMA, LSTMs and GA-LSTMs.
Apparently, GA-LSTMs performs significantly better than the other two methods,
indicating the superiority of GA-LSTMs for network traffic prediction.
418 J. Chen et al.
5 Conclusion
A long short-term memory method based on genetic algorithm is presented for network
traffic prediction in this paper. The network traffic is forecasted by training the GA-
LSTMs model. The original big data can be analyzed effectively by GA-LSTMs.
Experimental results show that GA-LSTMs achieves much better prediction results
than ARIMA and LSTM. Our future work will consider user perception aware
applications and resource allocation schemes in cellular networks [14, 15].
References
1. Kumar, J., Goomer, R., Singh, A.K.: Long short term memory recurrent neural network
(LSTM-RNN) based workload forecasting model for cloud datacenters. Procedia Comput.
Sci. 125, 676–682 (2018)
2. Khater, N.A., Overill, R.E.: Network traffic classification techniques and challenges. In:
International Conference on Digital Information Management, pp. 43–48 (2015)
3. Moayedi, H.Z., Masnadi-Shirazi, M.A.: Arima model for network traffic prediction and
anomaly detection. In: International Symposium on Information Technology, pp. 1–6 (2008)
4. Nikravesh, A.Y., Ajila, S.A., Lung, C., Ding, W.: Mobile network traffic prediction using
MLP, MLPWD, and SVM. In: International Congress on Big Data, pp. 402–409 (2016)
5. Leland, W.E., Taqqu, M.S., Willinger, W., Wilson, D.V.: On the self-similar nature of
Ethernet traffic (extended version). IEEE ACM Trans. Netw. 2(1), 1–15 (1994)
6. Specht, D.F.: A general regression neural network. IEEE Trans. Neural Netw. 2(6), 568–576
(1991)
7. Vanli, N.D., Sayin, M.O., Delibalta, I., Kozat, S.S.: Sequential nonlinear learning for
distributed multiagent systems via extreme learning machines. IEEE Trans. Neural Netw.
Learn. Syst. 28(3), 546–558 (2016)
8. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780
(1997)
9. Park, D.: Structure optimization of bilinear recurrent neural networks and its application to
ethernet network traffic prediction. Inf. Sci. 237(13), 18–28 (2013)
10. Hossain, D., Capi, G.: Genetic algorithm based deep learning parameters tuning for robot
object recognition and grasping. World Acad. Sci. Eng. Technol. Int. J. Mech. Aerosp. Ind.
Mechatron. Manuf. Eng. 11(3), 629–633 (2017)
11. David, O.E., Greental, I.: Genetic algorithms for evolving deep neural networks. In: Genetic
and Evolutionary Computation Conference, pp. 1451–1452 (2014)
12. Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In:
Neural Information Processing Systems, pp. 3104–3112 (2014)
Network Traffic Prediction Based on LSTM Networks 419
13. Graves, A., Jaitly, N.: Towards end-to-end speech recognition with recurrent neural
networks. In: International Conference on Machine Learning, pp. 1764–1772 (2014)
14. Xu, L., Luan, Y., Cheng, X., Xing, H., Liu, Y., Jiang, X., Chen, W., Chao, K.: Self-
optimised joint traffic offloading in heterogeneous cellular networks. In: 16th IEEE
International Symposium on Communications and Information Technologies, pp. 263–267.
IEEE Press, Qingdao (2016)
15. Xu, L., Cheng, X., et al.: Mobility load balancing aware radio resource allocation scheme for
LTE-advanced cellular networks. In: 16th IEEE International Conference on Communica-
tion Technology, pp. 806–812. IEEE Press, Hangzhou (2015)
On Multicast-Oriented Virtual
Network Function Placement: A Modified
Genetic Algorithm
1 Introduction
Currently, the main research stream on VNF placement focuses on unicast com-
munications. For example, Refs. [3, 4] studied the establishment and placement of
VNFs. On the other hand, with the emergence of more and more multimedia appli-
cations, e.g. remote education and IPTV, multicast has become one of the key com-
munications technologies. In NFV, network service should also be provided to
multicast sessions. The multicast-oriented VNF placement (MVNFP) problem, how-
ever, has not attracted enough research attention. In [5], NFV-enabled multicast routing
based on SDN was studied, where the branch and bound method was adopted. Then,
Xu et al. [6] proposed the concept of pseudo-multicast tree. The same network func-
tionality is only deployed once, which means any two VNFs in a multicast tree stand
for two different network functions. Nevertheless, some data flows need to travel longer
distance before reaching their destinations. Longer distance is more likely to lead to
larger delay, which to a certain extent deteriorates the performance of NFV.
Genetic algorithm (GA) has been successfully applied to solve the unicast-oriented
VNFP problem, where resource allocation in cloud datacenter is considered [4].
Compared with traditional mathematical methods, GA is self-organizing, adaptive, self-
learning, etc., especially suitable for tackling the NP-hard problems. The MVNFP
problem has been proven to be NP-hard.
This paper formulates a MVNFP problem with the average transmission delay in a
multicast tree minimized and proposes a two-step approach to handle it. The first step
adopts the Dijkstra’s algorithm to build a multicast tree once there is a multicast service
request. The second one deploys the VNFs for the tree, where a modified GA
(mGA) with problem-specific chromosome encoding, crossover and mutation is
devised. Compared with population based incremental learning, ant colony optimiza-
tion and particle swarm optimization, the proposed mGA achieves better solutions and
convergence.
Given a multicast service request MR, we use Dijkstra’s algorithm to find the
multicast tree GT ¼ ðVT ; ET Þ with the average propagation delay minimized, where
GT G; VT V, and ET E. The objective is defined in Eq. (2).
Minimize:
1 XjDj
Delaybfr ðPathi Þ ð2Þ
jD j i¼1
X X X
Delayaft ðPathi Þ ¼ Path
e2Es!f i
DelayðeÞ þ Path
e2Ef !fi
DelayðeÞ þ Pathi
e2Ef
DelayðeÞ
1 k kþ1 jF j !di
ð3Þ
Pathi
where Es!f 1
; EfPath i
k !fk þ 1
; EfPath i
jF j !di
EPathi , i ¼ 1; 2; . . .; jDj and k ¼ 1; 2; . . .; jF j 1. Es!f
Pathi
1
represents the link set of the sub-path from s to the node hosting f1 on Pathi . EfPath i
k !fk þ 1
stands for the link set of the sub-path from the node hosting fk to the node hosting fk þ 1
on Pathi . EfPath i
jF j !di
is the link set of the sub-path from the node hosting fjF j to destination
di on Pathi .
In the VNF placement, this paper aims to find an appropriate SFC deployment
solution for the multicast tree. The objective and constraints are shown in Eqs. (4)–(7).
Minimize:
1 XjDj h XjF j i
Delay aft ð Path i Þ þ Delay i
ð f k Þ ð4Þ
jD j i¼1 k¼1 Path
Subject to:
[
i
v2VPath
Udpl ðvÞ ¼ f1 ; f2 ; . . .; fjF j ð5Þ
\
Udpl ðuÞ Udpl ðvÞ ¼ £; 8u; v 2 Pathi ; u 6¼ v; i ¼ 1; 2; . . .; jDj ð6Þ
X
RAvailable ðvÞ fk 2Udpl ðvÞ
RConsumed ðfk Þ; v 2 V ð7Þ
Objective (4) is to minimize the average transmission delay of the multicast tree,
where the propagation delay on links and the processing delay on nodes that host VNFs
are both considered. Constraint (5) specifies that all VNFs in F are deployed on each
source-destination path, e.g. Pathi . Constraint (6) explains that any two VNFs on the
same source-destination path are different, which helps to improve the utilization of
compute resource. Constraint (7) specifies that the compute resource consumed by
deploying VNFs cannot exceed the available compute resource a node can offer prior to
the VNF placement.
It then calculates the fitness value of each solution in the population via fitness evalu-
ation. Later on, it selects solutions into a mating pool, where better solutions are selected
with higher probability. After that, crossover and mutation are applied to the solutions in
the mating pool to generate a new population. The selection, crossover and mutation
operations are repeated to evolve the population until termination condition is met.
A typical GA has the following steps [7]:
1. Initialization: Set the generation number s ¼ 0. Set the predefined maximum
generation number to Ng and randomly generate Np individuals to form population
PðsÞ.
2. Fitness evaluation: Calculate the fitness value of each individual in PðsÞ.
3. Selection: Repeatedly select an individual from the population until a mating pool
with Np individuals is generated.
4. Crossover: The crossover operation is performed to PðsÞ with a crossover proba-
bility Pc .
5. Mutation: The mutation operation is performed to PðsÞ with a mutation
probability Pm .
The population PðsÞ goes through the selection, crossover, and mutation
operations and becomes Pðs þ 1Þ.
6. Termination condition: If some termination criteria are satisfied, the evolution of
population stops and the best individual obtained in the evolutionary search is
output as the optimal solution.
In mGA, we set the population size to Np and the maximum number of itera-
tions to Ng . We use Eq. (4) as the fitness function.
After crossover, the population undergoes mutation operation, where each position
in the individuals is chosen with a mutation probability Pm . Suppose the k-th position
of the j-th region in individual Xi , xij;k , is chosen for mutation. Then, a node is randomly
j
selected from node set VPath and used to replace xij;k .
4 Performance Evaluation
In order to verify the effectiveness of mGA, we compare it with three widely used
evolutionary algorithms, including population based incremental learning (PBIL), ant
colony optimization (ACO) and particle swarm optimization (PSO). There are six
instances for experimentation, where three real-world networks from the Internet
426 X. Wang et al.
In mGA, we set Pc ¼ 0:7 and Pm ¼ 0:05. In PBIL, the learning rate and the
mutation probability are set to 0.1 and 0.05, respectively. In ACO, the pheromone
evaporation rate is set to 0.5 and the two heuristic factors are set to 0.1 and 0.2,
respectively. In the PSO, we set the inertia weight to 0.2 and the two positive constants
c1 ¼ c2 ¼ 2, respectively. For all algorithms, we set Np ¼ 100 and Ng ¼ 300. Each
algorithm is run 20 times for each instance and the best fitness curve is provided in
Fig. 2.
In terms of the convergence, PBIL, PSO, and ACO are easily trapped into local
optima in early stage of evolution, e.g. they converge around 50th generation. mGA, on
the other hand, converges gradually because of the diversity preservation provided by
the proposed mutation scheme. It not only helps to explore wider area in the search
space but also avoid damaging feasible individuals. Regarding the solution quality, it is
no doubt that mGA always obtains the best solution in each instance. This is because
the proposed crossover and mutation operations provide efficient global exploration
and local exploitation over the search space.
5 Conclusion
References
1. ETSI: Network Functions Virtualisation; Architectural Framework. Standard no. GS NFV
002 v1.1.1. ETSI (2013)
2. Cohen, R., Lewin-Eytan, L., Naor, J.S.: Near optimal placement of virtual network
functions. In: Conference on Computer Communications, pp. 1346–1354. IEEE (2015)
3. Khebbache, S., Hadji, M., Zeghlache, D.: Scalable and cost-efficient algorithms for VNF
chaining and placement problem. In: Innovations in Clouds, Internet and Networks, pp. 92–
99. IEEE (2016)
4. Rankothge, W., Le, F., Russo, A., Lobo, J.: Optimizing resources allocation for virtualized
network functions in a cloud center using genetic algorithms. IEEE Trans. Netw. Serv.
Manage. 14(2), 343–356 (2017)
5. Zhang, S.Q., Zhang, Q., Bannazadeh, H.: Routing algorithms for network function
virtualization enabled multicast topology on SDN. IEEE Trans. Netw. Serv. Manage. 12(4),
580–594 (2015)
6. Xu, Z., Liang, W., Huang, M.: Approximation and online algorithms for NFV-enabled
multicasting in SDNs. In: International Conference on Distributed Computing Systems,
pp. 625–634. IEEE (2017)
7. Beasley, D., Bull, D., Martin, R.: An introduction to genetic algorithms. Artif. Life 3(1), 63–
65 (1999)
428 X. Wang et al.
8. Batagelj, V., Brandes, U.: Efficient generation of large random networks. Phys. Rev. E Stat.
Nonlin. Soft Matter Phys. 71(3), 036113 (2005)
9. Xu, L., Luan, Y., Cheng, X., et al.: WCDMA data based LTE site selection scheme in LTE
deployment. In: 1st International Conference on Signal and Information Processing,
Networking and Computers, pp. 249–260. CRC Press Taylor & Francis Group, Beijing
(2015)
10. Xu, L., Cheng, X., et al.: Mobility load balancing aware radio resource allocation scheme for
LTE-advanced cellular networks. In: 16th IEEE International Conference on Communica-
tion Technology, pp. 806–812. IEEE Press, Hangzhou (2015)
PM2.5 Concentration Forecast Based
on Hierarchical Sparse Representation
1 Introduction
In recent years, haze pollution in China has shown an overall trend of increase [1].
PM2.5 (particle size less than 2.5 l) as the primary pollutant in the process of heavy
haze pollution, has been proved to have significant harm to human health [2]. Since the
implementation of the State Council’s air pollution prevention and control plan in
2013, the pollution situation in Beijing-Tianjin-Hebei has been obviously alleviated
and the air quality has improved significantly. However, the pollution weather has
occurred from time to time. Especially in the “2+26” cities, the pollution phenomenon
is more prominent, which is still significantly higher than that in developed countries.
Therefore, it has great importance to search for a suitable method to predict air
pollution.
At present, the North China Regional Meteorological Center introduced the WRF-
CHEM model and established the North China Regional Meteorological Prediction
System (BREMPS) locally [3]. It can simulate the temporal and spatial distribution of
regional atmospheric pollutant concentration. However, due to the uncertainties of the
emission list and the imperfections of physical and chemical mechanisms, the pre-
diction errors of PM2.5, PM10, O3 and other key pollutants are large [4]. In the recent
ten years, machine learning has been rapidly rising and has become the cornerstone of
technology in the era of big data. Fan et al. have constructed a space-time prediction
model based on the missing value processing algorithm and the cyclic neural network
to realize the prediction of air quality [5]. Experimental comparison verified the validity
of the space-time prediction framework based on depth learning. Dai et al. combined
particle swarm optimization (PSO) and support vector machine (SVM) to establish a
rolling prediction model to predict the concentration of PM2.5 in the next 24 h [6].
Although the above methods have achieved good results to some extent, there are still
some problems. Such as, the learning speed of neural network method is slow; the
support vector machine is difficult to achieve multi-classification.
Sparse representation is a focus in the current research. It is especially widely used
in computer vision, machine learning and pattern recognition. The representation
coefficient of this method contains many information, which is helpful for classifying
[7]. Moreover, It is also used for voice signal processing [8] and visual target tracking
[9]. The final result of sparse representation is depend on a few coefficients. So, the
computational complexity is greatly reduced. Since it doesn’t need to know the exact
relationship of dependent variables and independent variables, it is perfect for pre-
dicting when the relationship between predictors and prediction objects is ambiguous.
This paper uses observational data and two numerical forecast data to predict the PM2.5
concentration. It constructs 4000 samples of historical databases based on fuzzy C
means algorithm (FCM). Then, it use meteorological elements and pollutant concen-
tration to predict. It firstly use the first-level sparse representation to divided the test
samples into six categories. Then, It use the second-level sparse representation to
regress PM2.5 concentration with classification and its corresponding sub databases. By
experiments with the data predicted by RMAPS-CHEM and ECMWF in 2018, the
method in this paper has proven to have potential in PM2.5 concentration forecast. It
can improves forecast accuracy of PM2.5 concentration and is suitable for the analysis
of other meteorological data.
2 Information
2.1 Data Sources
Meteorological Data. Air temperature, wind direction, wind speed, humidity and
other observational data of meteorological stations in Beijing-Tianjin-Hebei by 3 h
between January and March, 2013–2017.
PM2.5 Concentration Forecast Based on H-SRC 431
3 Core Algorithm
Figure 1 gives the structure chart about predicting PM2.5 concentration forecast by
hierarchical sparse representation. This method is mainly include two parts. The first
part is classification: it firstly uses observational meteorological elements to construct
historical database D. Then input meteorological data predicted by RMAPS-CHEM
and ECMWF and historical database D. Finally, it can divided test samples into six
categories by first-level sparse representation. The second part is regression: it firstly
divides historical database D into six subclass databases. Then it inputs the results of
classification and subclass databases. Finally, it can regress PM2.5 concentration by the
second-level sparse representation.
Historical database D
about 4000 samples
How to select samples to ensure diversity of samples is very important, after being
classified by the PM2.5 concentration. Because it cannot be classified more meticu-
lously from the surface. In 1974, Dunn proposed FCM algorithm. From the most
algorithms of cluster, it is a successful one and is widely used in fuzzy cluster. Thus, we
use it to classify further data. Each sample is further divided into 100 categories by
FCM algorithm. It needs to select the number of each class according to the proportion
of each class randomly, which is constantly adjusted in the experiment. Finally, it
selects about 4000 samples for constructing the historical database D,
D ¼ ½x1 ; x2 x4000 .
can use historical PM2.5 concentration to represent the future PM2.5 concentration
sparsely. Figure 2 gives the structure diagram of H-SRC.
f1 f2 f3 f4 f5 f6 D1 D2 D3 D4 D5 D6
β1 β2 β3 β4 β5 β6 z1 z2 z3 z4 z5 z6
Regress
y1 y2 y3 y4 y5 y6
ec ¼ kf Dc ac k2 ð2Þ
434 R. Zhao et al.
Equation (3) is the formula for calculating the second sparse coefficients. The
second sparse coefficients is the coefficients of subclass test sample f corresponding to
its subclass historical database Dk .
After getting the second sparse coefficients, we use these coefficients and the PM2.5
concentration of subclass historical database to regress PM2.5 concentration yk . zkj
represents the PM2.5 concentration in j sample of k subclass historical database.
j ¼ 1; 2; ; mk , In it, mk is samples’ number of k subclass historical database Dk .
X
mk
yk ¼ bkj zkj ð4Þ
j¼1
4 Experimental Results
This paper does the experiment with the meteorological data predicted by RMAPS-
CHEM and ECMWF to verify the effect of using H-SRC to predict PM2.5 concen-
tration. The data is from Jan. 1 to Mar. 31 in Beijing-Tianjin-Hebei, 2018.
4.2 The Accuracy and Mean Absolute Error of H-SRC and RMAPS-
CHEM
In order to analyse the forecasting effect of H-SRC, Table 2 gives the forecasting
accuracy rate of two forecasting methods for three hours from January to March in
2018 at Beijing, and Table 3 gives the forecasting error of two forecasting methods for
three hours from January to March in 2018 at Beijing.
From Table 2, the accuracy of H-SRC in 0–35 lg=m3 , 75–115 lg=m3 , 75–115
lg=m3 , 115–150 lg=m3 is 25.28%, 13.34%, 14.28%, 23.08% higher than BREMPS.
The accuracy in 35–75 lg=m3 is only 8.23% lower than RMAPS-CHEM. Because it is
PM2.5 Concentration Forecast Based on H-SRC 435
Table 2. The accuracy of two methods at Beijing from January to March 2018
Methods 0–35 35–75 75–115 115–150 150–250 >250
Hierarchical sparse representation 71.31% 20.34% 26.67% 28.57% 61.54% /
RMAPS-CHEM 46.03% 28.57% 13.33% 14.29% 38.46% /
Table 3. The mean absolute error of two methods at Beijing from January to March 2018
ðlg=m3 Þ
Methods 0–35 35–75 75–115 115–150 150–250 >250
Hierarchical sparse representation 17.90 44.21 55.51 46.01 69.76 /
RMAPS-CHEM 49.52 49.95 82.42 50.09 106.61 /
easily misreported in 75–115 lg=m3 . On the whole, H-SRC has a better forecasting
effect, especially when the weather is excellent.
From Table 3, the mean absolute error of the six types of pollution in this method
is 31.62 lg=m3 , 5.74 lg=m3 , 26.91 lg=m3 , 4.08 lg=m3 , 36.85 lg=m3 lower than
BREMPS. Taken together, all mean absolute errors of PM2.5 concentration by this
method are less than RMAPS-CHEM. So, the effect of H-SRC is better.
436 R. Zhao et al.
5 Conclusion
Acknowledgments. This work has been supported in part by the National Key Research Pro-
gram of China (Grant No. 2016YFC0203301), the National Natural Science Foundation of China
(Grant No. 61773220, 61502206), the Nature Science Foundation of Jiangsu Province under
Grant (No. BK20150523).
References
1. Wang, Y.H., Liu, Y.J.: The characteristics of long-term changes of smog and haze in China
in the past 50 years and its relationship with atmospheric humidity. Chin. Sci. Earth Sci. 44
(1), 37–48 (2014)
2. Fang, J.X., Zhao, C.Y., Zhu, W.H.: Concentration monitoring and forecasting model of PM_
(2.5) during the heating period of Lanzhou city. Environ. Sci. Technol. 37(4), 80–84 (2014)
3. Liu, H., Rao, X.Q., Zhang, H.D., et al.: Comparative test of the forecast effect of
environmental meteorology numerical prediction model. J. Meteorol. Environ. 33(5), 17–24
(2017)
4. Zhao, X.J., Xu, J., Zhang, Z.Y., et al.: Beijing regional environmental meteorology
prediction system and PM2.5 forecast inspection. J. Appl. Meteorol. 27(2), 160–172 (2016)
5. Fan, J.X., Li, Q., Zhu, Y.J., et al.: Research on air pollution prediction model based on RNN.
Sci. Surv. Mapp. 42(7), 76–83 (2017)
6. Dai, L.J., Zhang, C.J., Ma, L.M.: PM2.5 short-term concentration dynamic prediction model
based on machine learning. Comput. Appl. 37(11), 3057–3063 (2017)
7. Akhtar, N., Shafait, F., Mian, A.: Efficient classification with sparsity augmented
collaborative representation. Pattern Recogn. 65, 136–145 (2017)
8. Abrol, V., Sharma, P., Sao, A.K.: Greedy double sparse dictionary learning for sparse
representation of speech signals. Speech Commun. 85, 71–82 (2016)
9. Abdessamad, J., Eladel, A., Zaied, M.: A sparse representation-based approach for copy-
move image forgery detection in smooth regions. In: International Conference on Machine
Vision, vol. 10341, 1034129 (2017)
10. Chen, X.H., Diao, Z.G., Hu, J.K., et al.: Dynamic-statistical prediction of PM2.5
concentration in China based on CMAQ model and adaptive partial least square regression.
J. Environ. Sci. 35(8), 2771–2782 (2016)
Intelligent User Profile Prediction in Radio
Access Network
1 Introduction
As one of the biggest telecom operators in the world, China Mobile has the widest
wireless access networks in China, in which the total number of base stations (BSs) has
exceeded 3 million by 2016 [1]. For the 5th Generation communication system (5G),
more BSs will continue to be installed to support up to 1,000-fold gains in capacity.
With increasingly dense deployment of base stations, wireless networks have become
more and more advanced and complicated. The wireless networks are generating a
large amount of wireless data (such as measurement report, traffic load, resource
usages, etc.) all the time, and this drives the wireless communication networks into the
era of big data [2]. In practice, the traffic load and resource usages of network are
closely related to the behavior pattern of the mobile users. Analyzing users’ behavior
pattern based on their interaction with wireless networks is a key problem in network
usage mining. Also, user mobility and behavior patterns prediction can significantly
The user profile prediction algorithm is divided into two parts: user’s location pre-
diction and service pattern prediction. For user’s location prediction, the personal
trajectory modeling can be conducted according to the users’ historical trajectory. After
modeling, when the system is inputted the continuous trajectory points of a user within
a few hours, which cell the user will visit next time will be predicted as the output
Intelligent User Profile Prediction in Radio Access Network 439
results. For service pattern prediction, when the continuous service information for a
period of time (for example, a few hours) is fed into the well trained model, the service
type and the traffic used by the user at the next moment can be predicted as the output
results. After the above two predictions are effectively combined, for a certain cell we
can know how many users will enter and leave at the next moment, and how much data
traffic they will bring in or take away. The real-time prediction of the users’ mobility
and service pattern will assist in the analysis of data traffic distribution in wireless
networks at the near future moment.
The algorithm includes three parts (see Fig. 1): offline learning, online prediction
and traffic statistics. In the offline learning stage, historical data is used to build pre-
diction models. For online prediction, the real-time data in the process of user
movement will be fed into the model that has been well trained in the offline learning
stage to obtain the prediction results. Then, the data traffic at the next moment for a cell
brought by each user can be summed up to obtain the total traffic. The other Key
Performance Indicators (KPI) prediction (for example, PRB resource utilization) can be
also made to understand the load condition of the network at the near future moment.
Data
collec on
Mobile
network
data Data pre-
Historical loca on data
processing
Offline
learning
Loca on data APP type and traffic
Real- me Real- me APP type training data training
loca on data and traffic data
User me- User me-
loca on model service model
Online
Predic on
User locaton APP type and
predicton traffic predic on
Traffic Sta s cs
PðTp jd 2 nj Þ Pðd 2 nj Þ
Pðd 2 nj jTp Þ ¼ PG : ð1Þ
k¼1 PðTp jd 2 nk Þ Pðd 2 nk Þ
Where G represents the total number of destinations for a certain user. A location
where a user resides for more than a certain time can be selected as the destination. The
longer the residing time is, the more likely the location is to be a destination, but this
method may miss the places where users only stay at a short time. For Pðd 2 nj Þ, it is
the priori probability of the destination nj , it can be calculated as
Where Sd2nj is the total number of the historical trajectories that with the destination
nj . Stotal is the total number of historical trajectories. Only a priori probability of the
grid nodes that users have ever reached is not zero. That means, only the grid nodes that
users have arrived before can be predicted as the user’s next destination. It is important
to note that the prior probability and the time are relevant. So, for better location
prediction, the trajectory of the user’s working day and the weekend is analyzed and
counted, respectively. Where PðTp j d 2 nj Þ is the posteriori probability that indicates
the probability of passing through the trajectory Tp when the destination is nj . It can be
calculated as:
STp ;d2nj
PðTp j d 2 nj Þ ¼ ð3Þ
Sd2nj
Where STp ;d2nj refers to the number of trajectories Tp when the destination is nj . For
the process of user trajectory prediction, the user’s destination is predicted by real-time
trajectory data, and then the cell with the highest handoff probability from the current
Intelligent User Profile Prediction in Radio Access Network 441
location to the predicted destination is selected. The next location of the trajectory Tp is
the prediction point.
If the user is at the current location point nc , the destination predicted is nj , there
may be different tracks from the current point to the destination, the location transfer
probability from nc to nj is
The next location of the trajectory with Pðnc ! nj Þ maximum is the predicted
location point.
According to the above model, the data of the rest day and the data of the working
day are divided into two databases, which are trained and predicted respectively. The
distinction between the working and the rest day can improve the accuracy of pre-
diction model for it is in line with users’ living habits.
Where Nðqar ! qar0 & ni ! ni þ 1 Þ refers to the number of events that the APP
switches from qar ! qar0 by a user when the location changes from ni ! ni þ 1 . The next
location ni þ 1 of user can be obtained by the previous trajectory prediction, and
combine the current location ni and the current application status qar the number for
handover Nðqar ! qar0 & ni ! ni þ 1 Þ can get. Then the service type with the maximum
Pðqar ! qar0 jni ! ni þ 1 Þ will be chosen as the predicted service type.
In static state, according to the continuity of business, users’ service type is related
to the user’s current location, next time period and the probability of a service type user
will use in the next period of time. In this way, we can give the probability that the user
will use a certain service type for the next period of time:
442 Y. Qiu et al.
Ni;t;a
PAi;t;a ¼ PA ð6Þ
a¼1 Ni;t;a
Among the formula, Ni;t;a is the number of events that the service type a is used by
the user at the location ni and time t, and the A refers to the total number of user service
types.
The matrix E describes the number of users in the current cell and their state,
In this case, the distribution probability of the total traffic volume of the cell is,
0 1 0 1
S1 p1
B S2 C B p2 C
B C B C
@ . . . A ! @ . . . A ¼ pðei Þ ð13Þ
S2 n p2 n
The probability that the total traffic volume is greater than VTH in a cell is,
X
2n
pðS [ VTH Þ ¼ pðei ÞpðSi [ VTH jei Þ ð14Þ
i¼1
3 Experiments Results
11:00-11:15
00:00-00:15
01:00-01:15
01:30-01:45
02:00-02:15
03:00-03:15
04:00-04:15
02:30-02:45
03:30-03:45
04:30-04:45
05:00-05:15
05:30-05:45
06:00-06:15
06:30-06:45
07:30-07:45
08:00-08:15
08:30-08:45
14:00-14:15
15:00-15:15
15:30-15:45
23:30-23:45
07:00-07:15
09:00-09:15
09:30-09:45
10:00-10:15
10:30-10:45
11:30-11:45
12:00-12:15
12:30-12:45
13:00-13:15
13:30-13:45
14:30-14:45
16:00-16:15
16:30-16:45
17:00-17:15
19:00-19:15
17:30-17:45
18:00-18:15
18:30-18:45
19:30-19:45
20:00-20:15
20:30-20:45
21:00-21:15
21:30-21:45
22:00-22:15
22:30-22:45
23:00-23:15
Ratio of Fully Correct Prediction Ratio of Next Cell Correct Prediction
03:00-03:15
03:45-04:00
04:30-04:45
06:00-06:15
10:30-10:45
12:00-12:15
15:45-16:00
16:30-16:45
01:30-01:45
02:15-02:30
05:15-05:30
06:45-07:00
07:30-07:45
08:15-08:30
09:00-09:15
21:45-22:00
09:45-10:00
11:15-11:30
12:45-13:00
13:30-13:45
14:15-14:30
15:00-15:15
17:15-17:30
18:00-18:15
18:45-19:00
19:30-19:45
20:15-20:30
21:00-21:15
22:30-22:45
23:15-23:30
Correct App Prediction Ratio
4 Conclusions
In this study, we have proposed a new algorithm model for user profile prediction. The
proposed scheme includes two sub-algorithms: Bayesian theory is used to model the
user mobility pattern and an innovative method is designed to make service pattern
prediction by combining the time and location features. The proposed strategy has low
computational complexity. Field test results show that the accuracy and timeliness of
the algorithm is of great significance to understand the temporal and spatial data traffic
distribution in the wireless network. Also, the user profile prediction can help operators
perceive the personalized service needs of mobile users so that more refined type of
operation and maintenance can be carried out.
References
1. China Mobile Communications Corporation: Sustainability Report: Big Connectivity, New
Future (2016)
2. Xu, L., Zhao, X., Luan, Y., et al.: User perception aware telecom data mining and network
management for LTE/LTE-advanced networks. In: 4th International Conference on Signal
and Information Processing, Networking and Computers, pp. 237–245. Springer, Qingdao
(2018)
Intelligent User Profile Prediction in Radio Access Network 445
3. Xu, L., Luan, Y., Cheng, X., et al.: Telecom Big Data based user offloading self-optimisation
in heterogeneous relay cellular systems. Int. J. Distrib. Syst. Technol. 8(2), 27–46 (2017)
4. Song, L., Kotz, D., Jain, R., et al.: Evaluating next-cell predictors with extensive Wi-Fi
mobility data. IEEE Trans. Mob. Comput. 5(12), 1633–1649 (2006)
5. Tomar, R.S., Verma, S.: Trajectory prediction of lane changing vehicles using SVM. Int.
J. Veh. Saf. 5(4), 345–355 (2011)
6. Wang, J., Tang, J., Xu, Z., et al.: Spatiotemporal modeling and prediction in cellular
networks: a Big Data enabled deep learning approach. In: INFOCOM IEEE Conference on
Computer Communications, pp. 1–9. IEEE (2017)
Realization of the National QPF Master
Blender: A Big Data Approach
Jian Tang(&), Kan Dai(&), Zhiping Zong, Yong Cao, Couhua Liu,
Song Gao, and Chao Yu
1 Introduction
Precipitation is one of the most important weather forecast products and plays a key
role in many forecast-related scenarios [12, 16]. In response to the increasing demands,
national prediction centers all over the world have established quantitative precipitation
forecast (QPF) operation. [3, 8, 14, 20, 25] comprehensively reviewed the progress of
QPF technology, and pointed out that the continuous development of numerical pre-
diction models and the application of statistical post-processing techniques have led to
continuous improvement of QPF accuracy, but the added value which can be provided
by forecasters is increasingly limited. In recent years, with the development of
© Springer Nature Singapore Pte Ltd. 2019
S. Sun et al. (Eds.): ICSINC 2018, LNEE 550, pp. 446–454, 2019.
https://doi.org/10.1007/978-981-13-7123-3_52
Realization of the National QPF Master Blender 447
Table 1. List of data sources of operational numerical weather prediction models available at
CMA. See more detail in Tang et al. 2018.
Model type Model name Resolution
Global deterministic CMA GRAPES global model 0.25°/up to 240 h
ECMWF high-resolution model 0.125°/up to 240 h
NCEP global model 0.5°/up to 240 h
Global ensemble ECMWF global ensemble model 0.5°/up to 360 h
CMA T639 global ensemble model 0.5°/up to 360 h
NCEP global ensemble mode 1.0°/up to 384 h
Regional mesoscale CMA GRAPES-MESO model 0.1°/up to 84 h
CMA GRAPES-3 km model 3 km/up to 36 h
CMA-Shanghai-MESO 9 km model 9 km/up to 72 h
448 J. Tang et al.
Secondly, the demand for front-ends of QPF products is increasing fast. For
example, in hydrological applications, only the QPF with spatial/temporal resolution
smaller than 10 km/1 h could meet the requirements of flash flood forecasting [1, 24],
and the accuracy is critical [13]. In order to meet the demand, Weather Prediction
Center (WPC) of the US has established a complete product system, including the
cumulative QPF and probability QPF grid products in the future 3/6/24 h, with a spatial
resolution of 2.5 km. The Central Meteorological Observatory (also known as National
Meteorological of China, NMC) has also carried out QPF operation for many years,
and established a grid-based QPF forecast routine in 2015, using objective technologies
to convert the field forecasts into QPF products with a resolution of 5 km [4]. The
“Modern Meteorological Forecast Operation Development Plan (2016–2020)” released
in 2016 put forward a higher goal, that is requiring the release of higher resolution
gridded forecasts with spatial and temporal resolutions of 1–2.5 km and 1 h respec-
tively by 2020, and the forecast accuracy rate of heavy rain should maintain an
improvement rate of 10–20% compared with the international advanced models. Tra-
ditional QPF routine has been unable to support forecasters to achieve the above
objectives.
In addition, the widespread use of big data mining techniques in forecasting
operations challenges the value of traditional forecasters. For example, [17] evaluated
the WPC QPF and indicated that the forecaster did not show an advantage over the
objective QPF after the bias correction and downscaling procession. The same situation
exists in the QPF service of NMC. For example, the “ensemble optimal quantile”
method has approached or slightly exceeded the forecasters’ skill on the Threat Score
(TS) [3]. There are three main reasons why the objective forecasting techniques show
better scores than forecasters: (1) The traditional value of the forecaster is reflected in
the correction of the systematic deviation of the model precipitation level forecast, but
the objective QPF method can better replace this work. (2) In addition to the correction
of the magnitude of model output, the forecaster also adjusts the location and distri-
bution of the rain area through the analysis of the evolution of the weather systems,
which is very difficult, but the TS score is difficult to express or evaluate this part of
work. (3) When the heavy rain warning is issued, the forecaster would sacrifice false
alarming ratio (FAR) to secure public safeness.
Facing with the above challenges, forecasters need to find new added value based
on massive information and data mining technology to play a more important role in
NWPs. [3] pointed out that the forecasters need to be freed from the traditional manual
production, based on the ability to understand the model and the information mining
ability from well-designed artificial intelligence, and turn to be scientific decision-
makers. [7] also pointed out that there is an urgent need to develop convenient and
efficient forecast editing tools to help forecasters achieve added value. To this end, this
study proposes to design a subjective and objective QPF platform to help forecasters
cope with multiple challenges. The remainder of this paper is organized as follows.
Section 1 introduces the overall design of the platform. Section 2 explains the opera-
tional routine based on the platform design. Finally, in Sect. 3 we show the forecasting
ideas through case application and verification, and present a summary and future
development in Sect. 4.
Realization of the National QPF Master Blender 449
How to integrate massive model forecast data information to obtain accurate QPF
products is the primary problem that needs to be solved in the subjective and objective
integration of QPF platform. Using different dynamic frameworks and physical
parameterizations, model systems from different forecasting centers have assimilated
different observations, and therefore they have different advantages and disadvantages
for different regions, seasons, steps and weather types. In addition, different data can
provide complementary QPF forecast information. For example, global NWPs can
provide precipitation information for large-scale weather systems. Ensemble model
systems can provide forecast uncertainty and low probability of extreme weather
information. High resolution meso-scale model can provide characteristics such as the
shape and evolution of the convective-scale precipitation system.
At present, leading international weather centers are developing tools that can help
forecasters quickly mine massive forecast information. For example, in WPC of the
US, the QPF process [19] is done as follows. The forecasters need to verify and
evaluate the forecast results of more than 100 numerical models, and use the “WPC
MASTER BLENDER” software to rapidly select the model forecasts and give corre-
sponding weights according to verifications. Then the products with relatively coarse
resolution are generated. After that, automated post-processing techniques are used to
form high-resolution products which are then distributed to regional weather offices.
Next, local forecasters could perform grid editing with graphical forecast editors
(GFE), and finally return, merge and publish the QPF.
In the traditional QPF operation of NMC (see Fig. 1), forecasters usually make
QPF products bases on a single model QPF (usually ECMWF as the deterministic
model), with other models or objective forecast methods as a reference. Then they use
MICAPS to plot contours of different QPF levels and finally convert the contour-QPF
into grid products through automated post-processing operations. The disadvantages of
this QPF routine including the inability to use massive model data information effec-
tively, the waste of time manually drawing the contour-QPF, the out of control of QPF
post process, and the lack of proper tools to make specific products according to users’
requirements. To this end, according to the operational technology framework in short-
medium-range weather forecast proposed by [6, 7], the design of subjective and
objective QPF routine is divided into the following five steps. (1) Based on real-time
multi-model QPF, appropriate QPF source is selected based on different forecasting
steps, seasons and needs. (2) Based on history, recent forecast verification and forecast
uncertainty assessment, best guess QPF is obtained by using multi-model QPF inte-
gration technology. (3) Based on various weather conceptual models and understanding
of model bias, QPF adjustment techniques and contour correction techniques are used
to further correct the best guess field. (4) A high-resolution grid QPF is obtained by
using the gridded post-processing techniques to downscale the best guess QPF field.
(5) Finally, service products are produced according to various service needs, such as
heavy-rain warning, typhoon rainfall, process precipitation, station forecast, etc. The
process uses multiple intelligence techniques [21] to help forecasters control the entire
grid QPF process and rapidly implement the subjective corrections.
450 J. Tang et al.
Fig. 1. The comparison between traditional QPF processes and the subjective and objective
integrated QPF processes.
QPF Master Blender V1.0 was mounted on MICAPS4 [9]. MICAPS4 provides
efficient display and analysis for high-resolution model forecast, and adopts an open
framework, which provides very convenient expansion capability and a basis for QPF
platform construction. The platform adopts a layered architecture, consists of a multi-
data source integrated correction module and a grid point post-processing module. The
multi-data integration correction module includes five interactive Graphic User Inter-
faces (GUI): data source selection, multi-source QPF integration, QPF field adjustment,
contour generation and modification, and grid QPF post-processing.
Fig. 2. Forecasters’ different work modes of QPF Master Blender under different forecast
scenarios.
theoretical weight integration maximum TS of 0.369 can be achieved, and the fore-
caster’s TS score reached 0.36. This is a L3 case.
From May 16th to 17th, 2018, there were scattered heavy precipitation cases in
Huanghuai region. It can be found that almost all global models have missed the
process, but the mesoscale models have some reflection on heavy rain. The real-time
verification found that the constancy of recent model is poor, and the global models and
the mesoscale models are quite different. Then, forecasters need to fully intervene. And
this could be a L4 stage case. Forecasters need to complete the operations by com-
pletely intervening the forecast through the understanding of the precipitation process.
Facing with the challenges brought by the explosive growth of meteorological data, as
well as the increasing demand from front-ends and the wide application of objective
forecasting techniques, the traditional QPF operational routine can no longer embody
the core value of forecasters. To this end, QPF Master Blender V1.0 was designed and
implemented to help forecasters control the entire digitization QPF process from five
aspects: forecast big data selection, QPF integration, QPF adjustment and correction,
grid post-processing and service product production. Then the value of forecasters
could be redefined as “big data mining based on weather understanding”. QPF Master
Blender integrates several key technologies: the construction technology of multi-
model QPF dataset, including the design of standard data format and storage specifi-
cation, as well as the conversion technology of the data with different resolutions;
multi-model QPF weighting and probability matching technique; various adjustment
Realization of the National QPF Master Blender 453
Acknowledgment. This paper was supported by the National Science and Technology Major
Project of the Ministry of Science and Technology of China under grant No. 2017YFC1502004
and the Key Project in the National Science & Technology Pillar Program of China under grant
No. 2015BAC03B01.
References
1. Arnaud, P., Bouvier, C., Cisneros, L., et al.: Influence of rainfall spatial variability on flood
prediction. J. Hydrol. 260(1–4), 216–230 (2002)
2. Bougeault, P., Toth, Z., Bishop, C., et al.: The thorpex interactive grand global ensemble.
Bull. Am. Meteorol. Soc. 91(8), 1059–1072 (2010)
3. Bi, B.G., Dai, K., Wang, Y., et al.: Advances in techniques of quantitative precipitation
forecast. J. Appl. Meterol. Sci. 27(05), 534–549 (2016)
4. Cao, Y., Liu, C.H., Zong, Z.P., et al.: State-level gridded quantitative precipitation
forecasting system. Meteor. Mon. 42(12), 1476–1482 (2016)
5. Chen, D.H., Shen, X.S.: Recent progress on GRAPES research and application. J. Appl.
Meterol. Sci. 17(06), 773–777 (2006)
6. Chen, L.Q., Zhou, X.S., Yang, S.: A quantitative precipitation forecasts method for short-
range ensemble forecasting. J. Nanjing Inst. Meteorol. 28(04), 543–548 (2005)
454 J. Tang et al.
7. Dai, K., Cao, Y., Qian, Q.F., et al.: Situation and tendency of operational technologies in
short- and medium-range weather forecast. Meteor. Mon. 42(12), 1445–1455 (2016)
8. Ebert, E.E.: Ability of a poor man’s ensemble to predict the probability and distribution of
precipitation. Mon. Weather Rev. 129(10), 2461–2480 (2001)
9. Gao, S., Bi, B.G., Li, Y.A., et al.: Implementation and development plan of MICAPS4.
J. Appl. Meterol. Sci. 28(05), 513–531 (2017)
10. Gebhardt, C., Theis, S., Krahe, P., et al.: Experimental ensemble forecasts of precipitation
based on a convection-resolving model. Atmos. Sci. Lett. 9(2), 67–72 (2008)
11. Golding, B.W., Ballard, S.P., Mylne, K., et al.: Forecasting capabilities for the London 2012
olympics. Bull. Am. Meteorol. Soc. 95(6), 883 (2014)
12. Kim, G., Barros, A.P.: Quantitative flood forecasting using multisensor data and neural
networks. J. Hydrol. 246(1–4), 45–62 (2001)
13. Knebl, M.R., Yang, Z.L., Hutchison, K., et al.: Regional scale flood modeling using
NEXRAD rainfall, GIS, and HEC-HMS/RAS: a case study for the San Antonio River Basin
Summer 2002 storm event. J. Environ. Manage. 75(4), 325–336 (2005)
14. Li, J., Du, J., Chen, C.J.: Introduction and analysis to frequency or area matching method
applied to precipitation forecast bias correction. Meteor. Mon. 5, 580–588 (2014)
15. Malardel, S., Wedi, N., Deconinck, W., et al.: A new grid for the IFS. ECMWF Newslett.
146, 23–28 (2016)
16. Messner, J.W., Mayr, G.J., Zeileis, A., et al.: Heteroscedastic extended logistic regression for
postprocessing of ensemble guidance. Mon. Weather Rev. 142(1), 448–456 (2014)
17. Novak, D.R., Bailey, C., Brill, K.F., et al.: Precipitation and temperature forecast
performance at the weather prediction center. Weather Forecast. 29(3), 489–504 (2014)
18. Palmer, T.N., Molteni, F., Mureau, R., et al.: Ensemble prediction (1993)
19. Petersen, D., Brill, K.F., Bailey, C., et al.: The evolving role of the forecaster at the weather
prediction center. In: World Weather Open Science Conference (2014)
20. Qian, Q.F., Zhang, C.A., Gao, S.Z., et al.: Real-time correction method for ensemble
forecasting of typhoon tracks. J. Trop. Meteorol. 30(05), 905–910 (2014)
21. Tang, J., Dai, K., Zong, Z.P., et al.: Methods and platform realization of the national QPF
master blender. Meteor. Mon. 44(8), 1020–1032 (2018)
22. Toth, Z., Kalnay, E.: Ensemble forecasting at NMC: the generation of perturbations. Bull.
Am. Meteorol. Soc. 74(12), 2317–2330 (1993)
23. Toth, Z., Kalnay, E.: Ensemble forecasting at NCEP and the breeding method. Mon.
Weather Rev. 125(12), 3297–3319 (1997)
24. Zehe, E., Becker, R., Bardossy. A.: The influence of spatial variability of soil moisture and
precipitation on runoff production (2001)
25. Zhang, F.H., Cao, Y., Xu, J., et al.: Application of the logistic discriminant model in heavy
rain forecasting. Meteor. Mon. 42(4), 398–405 (2016)
26. Zhang, H.B., Chen, J., Zhi, X.F., et al.: Design and comparison of perturbation schemes for
GRAPES_meso based ensemble forecast. Trans. Atmos. Sci. 37(03), 276–284 (2014)
Artificial Intelligence Research on Visibility
Forecast
Abstract. The meteorological data in 2000 to 2017 from the China observation
meteorological stations were collected for research. The multiple time scales
variation characteristics and relations between visibility and meteorological
elements were studied to summarize the weather conditions of low visibility
weathers. The selecting factors which related to visibility and its change were
input into an artificial neural network model for training. The long-term and
meticulous visibility forecast of observation stations in China were calculated
through the European Centre for Medium-Range Weather Forecasts (ECMWF)
data. The error and TS score detection showed that the model had better ref-
erence than the China Meteorological Administration Unified Atmospheric
Chemistry Environment model (CUACE) in the first half year of 2018.
1 Introduction
Along with the increase of aerosol emissions in various urban areas in China, low
visibility weather occurs frequently [1], which not only affects the normal operation of
highways and civil aviation, but also causes continuous heavy pollution incidents and
endangers public health and safety. Visibility changes have similar regularity and
periodicity on a larger time scale, which is mainly affected by climate, topography and
other factors, while the frequent fluctuations of visibility in the same time-space region
are caused by meteorological and environmental factors [2]. Precipitation [3], boundary
layer height, wind direction [4], wind speed [5], temperature [6], relative humidity [7],
pollutant concentration [8, 9] and other factors affect the atmospheric extinction
characteristics in different ways, and then change the visibility. The optical properties
of aerosols are determined by the chemical composition, the particle size, shape and
mixed state of the aerosol, and these properties are related to meteorological factors
[10]. In recent years, with the increasing scale of industrial production, the hygroscopic
growth of pollutants in aerosols was gradually responsible for the frequent occurrence
of long-term low visibility weather [11].
Visibility has strong nonlinear characteristics. The Neural Network prediction
method directly output the prediction data by inputting various types of related data
[12], and uses its better approximation ability for nonlinear data, and can better predict
the air quality of a specific city group or a larger area. Fully connected neural network
using back propagation algorithm is a mature artificial intelligence method. It can
© Springer Nature Singapore Pte Ltd. 2019
S. Sun et al. (Eds.): ICSINC 2018, LNEE 550, pp. 455–461, 2019.
https://doi.org/10.1007/978-981-13-7123-3_53
456 C. Xie and X. Ma
approximate any nonlinear function through the hidden layer and linear output layer by
using signal forward propagation and error back propagation [13]. However, fully
connected neural network is subject to more conditions in practical applications due to
its shortcomings in convergence, difficulty in training, easy to fall into local minimum,
and no feedback in one-way propagation [14].
In order to ensure the physical significance of the input factors of the neural network
model, the correlation between the meteorological observation factors and visibility of
the national meteorological stations from 2000 to 2016 was analyzed. We analyzed
nineteen surface and near-surface meteorological observation factors such as pressure,
temperature, humidity, wind speed and wind direction and their daily changes by
person correlation analysis method. The result showed that the p values were all less
than 0.05 and was statistically significant.
The stable pressure field and the sparse pressure gradient are beneficial to the
formation of low visibility weather. When the pressure is low, the stable pressure field
will cover the boundary layer and enrich particulate matter and water vapor in the
boundary layer. In this way the possibility of low visibility weather significant increase.
The relationship between visibility and temperature change is more complicated. On
the one hand, when the pressure change is not obvious, the negative temperature
change is mainly caused by radiation cooling, which is conducive to the formation of
radiation fog, resulting in a reduction of visibility. On the other hand, when the pressure
change is a large positive value, the negative temperature change at this time is mainly
caused by the invasion of cold air, resulting in an increasing of visibility. When the
temperature in the boundary layer rises, the vertical exchange of the atmosphere is
strengthened, and the stability of the atmospheric stratification is reduced, which is
conducive to the dilution of the pollutants and the dissipation of the mist. Meanwhile,
the relative humidity in the air is lowered resulting in a significant increasing of
visibility. Relative humidity has a significant effect on visibility. Water vapor changes
visibility by changing the optical properties of aerosols in the boundary layer. On the
one hand, the extinction of water vapor itself will reduce visibility, on the other hand,
the moisture absorption of particulate matter in the aerosol increases, and the extinction
increases significantly, resulting in a reduction of visibility.
Wind has a dilution effect on pollutants and water vapor at the horizontal level. It is
difficult to appear fog and haze weather in gale weather. However, the blowing sand or
floating weather in Northwest China may occur with the increase of wind speed, which
significant reducing the visibility. Meanwhile, the topography and the urban underlying
surface affected the ground wind field. The effect of diffusion dilution cannot be fully
reflected. The wind direction also has an effect on visibility. The warm and humid air
current blown from the sea surface to the land area is rich in high water vapor, increasing
the possibility of low visibility weather. the bias from the north Although the northerly
wind is relatively dry, it is not conducive to low-visibility weather, but the polluted air
mass from the north may also increase the aerosol concentration in China and reduce
visibility. When the vertical wind shear (WS) near the formation is low, turbulent
Artificial Intelligence Research on Visibility Forecast 457
mixing almost does not exist. It’s easy to appear low visibility weather with the vertical
stratification in a stable state. In the opposite case, when the vertical WS in the near-
surface layer is large, the vertical layer is in an unstable state. The elements of the
boundary layer are evenly mixed and are not prone to low visibility weather. Mean-
while, short-term heavy rainfall, thunderstorm gale and hail weather, which reduces
visibility, are prone to occur in the case of high vertical wind shear, so the correlation
between wind shear and visibility is affected.
The inversion phenomenon in the boundary layer will hinder the vertical convec-
tion of the air meanwhile reduce the pollutant capacity of the boundary layer. In this
case the pollutants, the combined air and gas bubbles cannot be exchanged vertically
and gradually accumulate under the low environmental capacity result in a significantly
decreasing of visibility. When the humidity above 850 hPa is high, it indicates that the
unstable weather system is developing and then the cloud amount in the sky will
increase obviously or precipitation will occur. At this time, there will be no fog. But
when the surface humidity is close to saturation and the height above 850 hPa is
relatively dry, namely an “upper dry and lower wet” atmosphere structure, which is
more prone to foggy weather [15].
After the stability experiment the neural network forecasting model was deter-
mined. Ground monitoring meteorological data such as temperature, pressure,
humidity, wind speed and the ECMWF high-altitude data such as humidity, wind speed
were input to the model so as to obtain the visibility forecast of 10 days with a 3 h
interval.
4 Results
After the establishment of the model, the forecast effect of the visibility of the national
meteorological stations from January 1, 2008 to June 30, 2008 was tested.
The daily average errors of forecast data and corresponding observation data were
compared. The Mean absolute error (MAE), Root mean square error (RMSE) and
System error (SE) were utilized to test the accuracy and generality of the forecasting
model. Meanwhile, the CUACE visibility forecast model of 3 days were introduced to
this test by the same method. The test results were shown in Table 2.
As shown in Table 2, the neural network model had a better performance than the
CUACE model in Absolute error test and Root mean square error test. The accuracy of
the neural network model was maintained at a high level during the 10-day forecast
process, and gradually weakened by the extension of the forecast time. This is due to
the model belongs to off-line model, the accuracy of ECMWF data itself decreases with
the extension of prediction time and the time series iteration caused by the prolongation
of forecast date in the process of neural network modeling leads to the decrease of
forecast precision. The neural network model also had a better performance than
Artificial Intelligence Research on Visibility Forecast 459
Table 2. Error analysis of ANN visibility (outside the brackets) and CUACE forecast visibility
(outside the brackets) to monitoring data
Error024 h 048 h 072 h 096 h 120 h 144 h 168 h 192 h 216 h 240 h
MAE 3.90 4.03 4.07 4.13 4.18 4.25 4.34 4.52 4.52 4.73
(5.51) (5.55) (5.49)
RMSE 5.06 5.25 5.31 5.37 5.41 5.50 5.58 5.78 5.78 6.01
(6.86) (6.95) (6.93)
SE –1.41 –1.74 –1.76 –1.83 –1.86 –1.88 –1.89 –1.90 –1.97 –2.00
(2.94) (2.90) (3.08)
CUACE model in System error test. It is shown that the model maintains good stability
in multi-station and multi-time prediction. The visibility prediction of neural network is
less than the actual value, mainly due to the systematic error of ECMWF data.
In order to highlight the forecasting effect of the neural network model on the
difference visibility level, the average visibility of meteorological stations during the
period from January 1, 2018 to June 30, 2018 was classified according to less than
2 km, 2–5 km and 5–10 km, respectively. The daily average TS score of 1–10 days is
the correct number of predictions divided by the number of correct predictions, the
number of empty reports, and the number of false negatives. Meanwhile, the CUACE
visibility forecast model of 3 days were introduced to this test by the same method. The
test results were shown in the Table 3.
Table 3. Forecast TS score of ANN visibility (outside the brackets) and CUACE forecast
visibility (outside the brackets)
TS 024 h 048 h 072 h 096 h 120 h 144 h 168 h 192 h 216 h 240 h
score
<2 km 0.14 0.14 0.13 0.12 0.11 0.09 0.09 0.08 0.08 0.08
(0.03) (0.02) (0.01)
2–5 km 0.24 0.24 0.24 0.23 0.23 0.22 0.21 0.21 0.21 0.19
(0.10) (0.07) (0.08)
5– 0.22 0.22 0.22 0.22 0.22 0.22 0.21 0.21 0.21 0.20
10 km (0.11) (0.11) (0.13)
mainly due to the insufficient training caused by fewer samples of low visibility
weather data. It also showed that with the extension of prediction time, the prediction
effect of visibility the neural network is gradually weakening.
5 Conclusions
Low visibility weather occurs frequently, causing the serious loss of life and property,
and it’s difficult to forecast. Based on the principle of meteorology, the meteorological
elements and vertical stratification elements were analyzed to reveal their relationship
with low visibility weather. The results showed that the stable pressure field, near-
surface relative humidity, “upper dry and below wet” structure contributed significantly
to low visibility weather. By means of selecting reasonable diagnostic factors, an
artificial neural network model was established with dynamic learning rate BP algo-
rithm to simulate the low visibility weather of China. Based on the balance between
generalization ability and prediction accuracy, the tests conducted repeatedly to
determine the parameters of the artificial neural network. And then, the ECMWF data
was used as the input data in the model and obtain a visibility forecast. The error and
TS score detection showed that the model had better reference than the CUACE model.
It provided guidance for improving visibility objective forecasting technology.
References
1. Fu, H., Chen, J.: Formation, features and controlling strategies of severe haze-fog pollutions
in China. Sci. Total Environ. 578, 121 (2016)
2. Gao, Z.K., Cai, Q., Yang, Y.X., et al.: Time-dependent limited penetrable visibility graph
analysis of nonstationary time series. Physica A 476, 43–48 (2017)
3. Cao, W.H., Liang, X.D., Li, Q.C.: A study of the stageful characteristics and influencing
factors of a long-lasting fog/haze event in Beijing. Acta Meteorologica Sinica 71(5), 940–951
(2013)
4. Basahel, A., Rafiqul, I.M., Suriza, A.Z., et al.: Availability analysis of free-space-optical
links based on rain rate and visibility statistics from tropical a climate. Optik-Int. J. Light
Electron Opt. 127(22), 10316–10321 (2016)
5. Pierini, J.O., Lovallo, M., Telesca, L.: Visibility graph analysis of wind speed records
measured in central Argentina. Physica A 391(20), 5041–5048 (2012)
6. Kartha, M.J.: Surface morphology of ballistic deposition with patchy particles and visibility
graph. Phys. Lett. A 381(5), 556–560 (2017)
7. Deng, H., Tan, H., Li, F., et al.: Impact of relative humidity on visibility degradation during a
haze event: a case study. Sci. Total Environ. 569, 1149–1158 (2016)
8. Cheung, H.C., Tao, W., Baumann, K., et al.: Influence of regional pollution outflow on the
concentrations of fine particulate matter and visibility in the coastal area of southern China.
Atmos. Environ. 39(34), 6463–6474 (2005)
9. Li, Y., Huang, H.X.H., Griffith, S.M., et al.: Quantifying the relationship between visibility
degradation and PM 2.5 constituents at a suburban site in Hong Kong: differentiating
contributions from hydrophilic and hydrophobic organic compounds. Sci. Total Environ.
575, 1571–1581 (2017)
Artificial Intelligence Research on Visibility Forecast 461
10. Yu, X., Ma, J., An, J., et al.: Impacts of meteorological condition and aerosol chemical
compositions on visibility impairment in Nanjing, China. J. Cleaner Prod. 131, 112–120
(2016)
11. Lin, J., Li, J.: Spatio-temporal variability of aerosols over East China inferred by merged
visibility-GEOS-Chem aerosol optical depth. Atmos. Environ. 132, 111–122 (2016)
12. Kukkonen, J., Partanen, L., Karppinen, A., et al.: Extensive evaluation of neural network
models for the prediction of NO2 and PM10 concentrations, compared with a deterministic
modelling system and measurements in central Helsinki. Atmos. Environ. 37(32), 4539–4550
(2003)
13. Reich, S.L., Gomez, D.R., Dawidowski, L.E.: Artificial neural network for the identification
of unknown air pollution sources. Atmos. Environ. 33(18), 3045–3052 (1999)
14. Lou, W.G., Ting-Ting, Q.I., Lou, Y.Y., et al.: General regression neural network based on
tax payment assessment and its empirical research. Syst. Eng. (2015)
15. Xue-Kuan, M.A., Cai, X.N., Yang, G.M.: Study on fog synoptic characteristics and fog
forecast method in Chongqing. Clim. Environ. Res. 12(6), 795–803 (2007)
Multi-model Ensemble Forecast
System for Surface-Layer PM2.5
Concentration in China
1 Introduction
Particulate matter with diameter of 2.5 µm or less (PM2.5) is the most important
component of haze that frequently occurs in China in recent years. High PM2.5 con-
centration can reduce the atmospheric visibility [1], affect human health [2], and also
influence regional and global climate [3]. Therefore, more accurate forecast of surface-
layer PM2.5 concentrations can enhance the early warning ability of haze process and
reduce its negative impact on human lives.
Numerical prediction of PM2.5 concentrations is widely implemented in many
countries’ national operational air quality forecast systems [4, 5]. Recently years, China
Meteorological Administration (CMA) has developed some numerical PM2.5 forecast
models for specific regions of China, such as China Meteorological Administration
Unified Atmospheric Chemistry Environment for aerosols (CUACE) model for entire
China, Beijing Regional Environmental Meteorology Prediction System (BREMPS)
for northern China [6], Regional Atmospheric Environmental Modeling System
(RAEMS) for eastern China [7], and an air quality forecast system based on Com-
munity Multiscale Air Quality Model (CMAQ) for southern China [8]. However, there
are large biases between forecasted and observed PM2.5 concentrations in these
numerical models due to the uncertainties in initial conditions, meteorological fields,
emission inventories, and physical and chemical processes in aerosols’ life time.
Multi-model ensemble method is an effective way to improve forecast accuracy,
and some previous studies have applied it to forecast PM2.5 concentration. Huang
improved the air quality forecast system for Beijing by multiple linear regression
ensemble based on Nested Air Quality Prediction Modeling System (NAQPMS),
CMAQ, and Comprehensive Air Quality Modeling System with Extensions (CAMx)
[9]. Results showed that after ensemble, the biases and RMSE between forecasted and
observed daily PM10 concentrations over 28 sites of Beijing during January−April
2010 were reduced by 5.8 µg m−3 and 32−43%, respectively. However, most previous
multi-model ensemble studies in China focused on specific regions, a study for entire
China is needed. Furthermore, few studies used combination of several multi-model
ensemble methods to forecast PM2.5 concentration, which may reduce the uncertainty
of single multi-model ensemble method [10]. Different with traditional statistical
methods, Artificial Neural Network (ANN) has strong ability in dealing with non-linear
questions and is widely used in PM2.5 concentration forecast. However, most input data
of ANN in previous studies were from single numerical prediction model, few studies
used ANN in multi-model ensemble forecast. Only Zhang used Back Propagation ANN
(BP-ANN) ensemble to forecast the PM2.5 concentrations in Beijing during Olympic
Games based on NAQPMS, CAMx and CMAQ models [11]. The bias between
observed and forecasted daily PM2.5 concentrations was lower than that of mean
ensemble, with value of 22.5 µg m−3.
Therefore, we developed a multi-model ensemble system to enhance the forecast
accuracy of PM2.5 concentrations in China, which is operated and maintained by
Environmental Meteorological Center of CMA. Sections 2 and 3 describe the data and
ensemble methods, respectively. Sections 4 and 5 evaluate the forecast results during
2015−2016 and a heavy haze process, respectively.
2 Data Description
Observed hourly surface-layer PM2.5 concentrations of 1492 sites in China were from
China National Environmental Monitoring Centre. Forecasted hourly surface-layer
PM2.5 concentrations for next 72 h were from four air quality forecast models which
were CUACE, BREMPS, RAEMS, and a CMAQ model with localized improvements
operated by National Meteorological Center, Beijing Meteorological Bureau, Shanghai
Meteorological Bureau, and Guangdong Meteorological Bureau, respectively. And in
this paper, the four models were abbreviated as CUACE, NNC, ENC, and SNC,
respectively, according to their forecast region. To fit the observations, forecast data
were interpolated to 1492 sites. The study period was from January 1, 2015 to
December 31, 2016.
464 T. Zhang et al.
3 Ensemble Methods
3.1 Mean Ensemble
In mean ensemble, all models were thought to have same forecast accuracy. For each
site and forecast time, the mean ensemble was calculated as
1 Xn
Fmean ¼ F
i¼1 i
ð1Þ
n
in which n was the number of models with value of 4 in this paper, Fmean was
forecasted PM2.5 concentration by mean ensemble, and i and Fi were the model i and
its forecast value, respectively.
Xn Xn 1
jAvgi Avgobs j
Fwighted ¼ W Fi ¼ Pn Fi ð2Þ
i¼1 i i¼1 1
i¼1 jAvgi Avgobs j
in which FMLR was forecasted PM2.5 concentration by MLR ensemble. ai and a0 were
the weight coefficients solved by training stage.
The learning rule of network was defined as the minimum mean square error criterion.
Network would repeat the processes mentioned above until the error between fore-
casted and expected values smaller than preset value or maximum training times.
According to sensitivity experiments, we found out a reasonable setting of BPANN
parameters for our data, which were trainbr training function, 10 hidden layers, and a
sample length of 50 days [12].
rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
1 Xn
RMSE ¼ ðFi Obsi Þ2 ð8Þ
n i¼1
Pn
i¼1 ðFi F Þ Obsi Obs
R ¼ qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
Pn 2 ð9Þ
2
i¼1 ðFi F Þ Obsi Obs
in which n was the number of days used in evaluation and equaled to 50 in this paper.
F was the forecasted PM2.5 concentration of a specific single model or ensemble
method. Obs was the observed PM2.5 concentration.
For each of the three evaluation indexes, all single models (CUACE, NNC, ENC,
and SNC) and ensemble methods (mean, weighted, MRL, and ANN) were ranked
according to the value of index, with points from 0 to 7 corresponding to evaluation
result from worst to the best, respectively. For example, for a specific single model or
ensemble method, the smaller |NMB| value meant smaller forecast biases and higher
evaluation points. Therefore, the maximum and minimum |NMB| value of all models
and ensemble methods were given 0 and 7 points, respectively. And the total score was
the sum of the points of all three evaluation indexes. For each site, the single model or
ensemble method with highest total score was considered to be the best ensemble for
next 72 h. For a site, the points were calculated as follows
PointjNMBj;i ¼ Sort jNMBj18 i ð10Þ
in which i was a specific single model and ensemble method. Pointi was the point of i.
Sort was sorting function. Sorti was the rank of i for a certain evaluation index. The
forecasted PM2.5 concentration by best ensemble was calculated as follows
FBest ¼ Max Sort PointjNMBj;18 þ PointRMSE;18 þ PointR;18 ð13Þ
in which FBest was forecasted PM2.5 concentration by best ensemble. Max was max-
imum function.
Figure 1 presents the statistical variables of observed and forecasted PM2.5 concentra-
tions by all single models and ensemble methods during 2015–2016, including mean
value and NMB, RMSE, and R values between them. All single models and ensemble
methods could forecast high PM2.5 concentrations in China which located in Jing-Jin-Ji,
Yangtze river delta and Sichuan basin regions. However, CUACE model overestimated
concentrations in central China and Sichuan basin with NMB values of 20–80%, and
underestimated concentrations in southern and western China with NMB varied from –
80% to –5%. ENC and NNC model underestimated concentrations in northern China with
NMB varied from –40% to –10%. SNC model underestimated concentrations in southern
China with NMB values from –20% to –5%, and largely overestimated concentrations in
Sichuan basin, central China, and Yangtze river delta region with NMB values larger than
40% in most sites.
Compared with CUACE and SNC models, mean ensemble reduced the magnitude
of overestimation in central China but still with high NMB values varied from –20% to
40%. And the improvements in western and northeastern China were limited. This
indicate that mean ensemble cannot effectively improve the forecast results, and its
forecast ability is highly dependent on those of single models. After taking the biases
between single models and observations as weight coefficients, weighted ensemble
slightly reduced the overestimation of mean ensemble in central China and Yangtze
river delta region in which NMB values were in the range of ±20%. In other regions of
China, the forecast results of weighted ensemble were similar to those of mean
ensemble. And weighted ensemble presented lower R values in northern China.
Compared with mean and weighted ensemble, forecasted PM2.5 concentrations by
MLR and ANN ensembles were closer to observations, especially in western and
northeastern China. In most sites of China, the NMB values between observations and
MLR and ANN ensembles were in the range of ±5% and ±10%, respectively. But
there were more sites with R values in the range of 0.6-0.8 in northern China in ANN
ensemble compared with MLR ensemble. MLR and ANN ensembles has similar
RMSE values of 20–70, 20–40, and 10–30 lg m−3 in northern, central, and southern
China, respectively, which were smaller than those of single models, mean and
weighted ensembles. On the basis of real-time evaluations of all single models
and ensemble methods, PM2.5 concentrations forecasted by best ensemble was higher
and closer to observations in northern China compared with MLR and ANN
Multi-model Ensemble Forecast System 467
ensembles. The NMB values between observations and best ensembles were in the
range of ±5% in most regions of China. The RMSE values were similar to those of
MLR and ANN ensembles. Best ensemble presented strong positive correlation with
observations, with more sites in China had R values higher than 0.6 compared with any
single model and other ensemble method. This suggests that best ensemble combined
advantages of all single models and ensemble methods in different sites, and presented
similar average PM2.5 concentrations but better trend compared with MLR and ANN
ensembles.
Fig. 1. Observed and forecasted mean PM2.5 concentrations (lg m−3) by single models
(CUACE, ENC, NNC, and SNC) and ensemble methods (mean, weighted, MLR, ANN, and
best) during 2015–2016 and the NMB, RMSE, R values between them.
468 T. Zhang et al.
5 Case Evaluation
From February 25 to March 4, 2018, a heavy haze process occurred in the south of
Jing-Jin-Ji region, with maximum observed PM2.5 concentrations of 251, 323, and
311 lg m−3 in Beijing, Shijiazhuang, and Zhengzhou, respectively. Figure 2 presents
the time series of observed and forecasted PM2.5 concentrations by best ensemble. Best
ensemble could forecast the bimodal distribution of observed PM2.5 concentrations.
A clean and cold air swept across Jing-Jin-Ji region from north to south in the middle of
this process, leading to the transmission of pollutants from north to south. The PM2.5
concentrations rapidly decreased to good level, but the time when concentrations began
to decrease in southern site was later than that of the northern site. With the extension
of forecast time, biases between forecasted and observed PM2.5 concentrations became
larger. From 24 to 72 h’ forecast, the NMB values decreased from –22% to –25%,
from –4% to –26%, and from –12% to –15% for Beijing, Shijiazhuang, and
Zhengzhou, respectively. And R values also decreased with the extension of forecast
time, but they maintained at a high level with values in the range of 0.49–0.77. The
results above illustrate that best ensemble has the ability to forecast the magnitude and
trend of PM2.5 concentrations in heavy pollution process.
Fig. 2. Time series of observed and forecasted PM2.5 concentrations by best ensemble in
different forecast time from February 25 to March 4, 2018 in Beijing, Shijiazhuang, and
Zhengzhou.
6 Conclusion
were closer to observations, especially in western and northeastern China, with NMB
values of ±5% and ±10% in most sites of China, respectively. Best ensemble showed
stronger comprehensive forecast capability with similar NMB and RMSE values but
higher R values compared with MLR and ANN ensembles.
In the heavy haze process occurred in south of Jing-Jin-Ji region from February 25
to March 4, 2018, the NMB and R values between best ensemble forecasted and
observed PM2.5 concentrations in 3 representative cities varied from –4% to –26% and
from 0.49 to 0.77, respectively. These indicate that as the final output of multi-model
ensemble system, best ensemble performs better than single models and can provide a
strong objective reference to forecaster.
References
1. Wang, K., Dickinson, R., Liang, S.: Clear sky visibility has decreased over land globally
from 1973 to 2007. Science 323(5920), 1468–1470 (2009). https://doi.org/10.1126/science.
1167549
2. Lelieveld, J., Evans, J., Fnais, M., et al.: The contribution of outdoor air pollution sources to
premature mortality on a global scale. Nature 525(7569), 367–371 (2015)
3. Intergovernmental Panel on Climate Change (IPCC), Climate Change, The Physical Science
Basis. Contribution of Working Group I to the Fifth Assessment Report of the
Intergovernmental Panel on Climate Change. Cambridge University Press, Cambridge
(2017)
4. Savage, N., Agnew, P., Davis, L., et al.: Air quality modelling using the Met Office Unified
Model (AQUM OS24-26): model description and initial evaluation. Geosci. Model Dev. 6
(2), 353–372 (2013). https://doi.org/10.5194/gmd-6-353-2013
5. Lee, P., McQueen, J., Stajner, I., et al.: NAQFC developmental forecast guidance for fine
particulate matter (PM2.5). Weather Forecast. 32(1), 343–360 (2017). https://doi.org/10.
1175/waf-d-15-0163.1
6. Zhao, X., Xu, J., Zhang, Z., et al.: Beijing regional environmental meteorology prediction
system and its performance test of PM2.5 concentration. Q. J. Appl. Meteorol. 27(2), 160–172
(2016). https://doi.org/10.11898/1001-7313. (In Chinese)
7. Zhou, G., Xu, J., Xie, Y., et al.: Numerical air quality forecasting over eastern China: an
operational application of WRF-Chem. Atmos. Environ. 153, 94–108 (2017). https://doi.org/
10.1016/j.atmosenv.2017.01.020
8. Deng, T., Wu, D., Deng, X., et al.: Numerical forecast system of air quality photochemical
smog over pearl river delta region. Environ. Sci. Technol. 36(4), 62–68 (2014). (In Chinese)
9. Huang, S., Tang, X., Xu, W., et al.: Application of ensemble forecast and linear regression
method in improving PM10 forecast over Beijing area. Acta Sci. Circum. 35(1), 56–64
(2015). (In Chinese)
470 T. Zhang et al.
10. Liu, D., Li, L.: Application study of comprehensive forecasting model based on entropy
weighting method on trend of PM2.5 concentration in Guangzhou, China. Int. J. Environ.
Res. Public Health 12, 7085–7099 (2015). https://doi.org/10.3390/ijerph120607085
11. Zhang, W., Wang, Z., An, J., et al.: Update the ensemble air quality modeling system with
BP model during Beijing olympics. Climatic Environ. Res. 15(5), 595–601 (2010). (In
Chinese)
12. Zhang, H., Zhang, T., Li, T., et al.: Forecast of air quality pollutants’ concentrations based on
BP neural network multi-model ensemble method. China Environ. Sci. 38(4), 1242–1256
(2018). (In Chinese)
Research on Natural Language Processing
and Aspose Technology in the Automatic
Generation of Ocean Weather Public Report
1 Introduction
Ocean meteorological service requires timeliness, precision and diversity. There are
many kinds of public report service products, involving a variety of meteorological
elements and content. It also needs a lot of analysis and production time for forecasters
to meet the needs of public services. Taking the ocean weather public report as an
example, issued three times a day, it mainly focuses on the forecast and observation
description of the ocean gale and ocean fog, including the wind force, wind direction,
fog level, visibility and so on. The forecast text also needs its corresponding forecast
graphics. On the one hand, text descriptions not only need to transform various
meteorological elements and their observation or forecast values, but also need to
merge the sea areas with similar wind or fog phenomena, which can highlight the focus
and regularity of the report. On the other hand, forecasters usually revise the contents of
the report on the basis of the previous report, and make changes to various factors such
as reporting time, forecasting time and observing time. Because of the persistence of
weather phenomena, some contents of the previous report may be reused. So, the
forecaster will just make necessary changes according to the new changes. Therefore,
as far as the process of producing the ocean weather public report is concerned, the
quality of the report is highly demanded. It is necessary not only to extract and interpret
various meteorological elements efficiently and accurately, but also to analyze and
refine the characteristics and the merger rules of the sea areas based on meteorological
elements, and to express them in natural language, supplemented by graphics. Finally,
a meteorological service product that is easily understood and accepted by the public
comes into being. At the same time, there are some heavy and simple repetitive labor.
Abroad research began in the 1970s on the automatic generation method of weather
forecast text, it has gone through the earliest word substitution method (Computer
Worded Forecasting, CWF) and the natural language processing technology (NLP) in
the early 1990s. Some text generation systems for specific weather forecasting fields
have also been developed and applied. Domestic research is relatively late, mainly
focusing on the simple conversion from data codes to their corresponding text
descriptions. In 2000, Shanghai Jiao Tong University launched a preliminary study on
multi-language weather forecast text automatic generation system (MLWFA) based on
natural language processing technology [1–3]. About the automatic generation of office
documents, including data, text, graphics, and so on, there are COM components which
refer to the Microsoft Office to generate and control the operation of office documents
and document generation technology based on XML technology, but the implemen-
tation is too difficult. Aspose technology using a series of APIs for document
manipulation and management in the application program can achieve the creation and
editing functions of Microsoft Office Word documents, and then complete the auto-
matic generation of the report document [4, 5]. Therefore, combined with the
requirements of producing the ocean weather public reports, this paper analyzes a large
number of historical reports published daily by National Meteorological Center of
China, and proposes an automatic generation method of ocean weather public reports
by using Natural Language Processing technology and Aspose technology.
2 Principle
The computer automatic generation method of ocean weather public reports is to solve
the problem of generating the report document based on the ocean weather data. This
paper proposes the following basic ideas: Firstly, read the ocean weather forecast data,
and match the codes or values of various meteorological elements with the code of the
sea area one by one. Then, select the meteorological elements according to the forecast
focus, and Merge eligible sea areas and their associated meteorological elements into
different sets. Secondly, these meteorological elements in each set are filtered and
merged after their corresponding sea areas according to certain rules, such as the wind
force, the wind direction, the increasing and decreasing trend of wind force. Then,
natural language generation technology is applied to reasonably organize the meteo-
rological Information such as the forecast period, the corresponding sea area, the wind
force, the wind direction, the wind trend, and etc. Furthermore, microplanner, sentence
Research on Natural Language Processing and Aspose Technology 473
optimization and post-processing are used to form the text paragraph in natural lan-
guage expression. Finally, Aspose document management technology is used to
automatically generate the report document based on templates and historical docu-
ments, accompanied with automatic reference and update of related content and
standardization of file formats. The basic process is shown in Fig. 1:
3 Key Technology
3.1 Text Content Analysis and Feature Extraction
Text analysis and feature extraction are to analyze the content and structure features of
a certain type of text, and provide extraction rules for the automatically generated text.
For the text of the ocean weather public report, it is generally necessary to perform
feature analysis and extraction on time variables, regional variables, meteorological
element variables, and short sentences and paragraphs involved.
The statistical analysis of more than 600 historical ocean weather public reports
written by forecasters, focusing on those expressions with the natural language, gives
the following results. Firstly, meteorological element variables include weather terms
describing the wind or gust, fog, and visibility at sea and their corresponding intensity
(magnitude), for example, wind forces of 7, 8–9, 4–6 level, wind directions in the
north, in the southeast, from the north to the northwest, fog levels includes “light fog”,
“fog”, “dense fog”, visibility levels include “visibility less than 500 m”, “visibility less
than 1 km”, “visibility less than 10 km”. Secondly, regional variables include 18 areas
of China’s inshore sea, which forecasters usually describe jointly with geographical
names that represent geographical locations such as the Bohai Sea, Taiwan Strait, and
the terms that represent the azimuth and area ratio, such as the Northeastern part of sea,
the Southeastern part of sea, the Central part of sea, and the majority of sea. The short
sentence structure feature refers to the general sentence expression form describing the
above-mentioned variables involved in meteorological elements. The multiple short
474 X. Bai et al.
sentence structures are extracted and merged into a sentence template. Figure 2 shows a
simple example of a sentence template. The brackets (【】) and information in them
represent a variable which can be time, the location, wind force, wind direction, fog
level, visibility distance, illustration number, etc.
The Description Rule of Wind Force. The ocean weather texts are prioritized
according to the level of wind force. Areas with strong gust (usually referred to 7–8
level and above) are first classified, and areas with the same wind force and similar
wind direction are grouped together. In principle, 4–5 level wind and below are not
required to be written in the report, but if all winds are 4–5 level and below in summer,
4–5 level and its areas are still written. The gust level is always one level than the
average wind, for example, if the wind is 6–7 level, the gust is 8 level.
The Description Order of Meteorological Elements. The general sequence is wind
force, wind direction, fog, visibility.
The Merger Rule of Areas. Areas with the same wind increase and the same wind
decrease trend, the same or adjacent wind direction (no more than 45° deflection), or
the same fog level, can be combined to describe, so that the text description is more
compact and the overall weather change trend is more clear.
The Fuzzy Description Rule of Areas. The final output text needs to be merged and
other subsequent optimization processing, mainly involving the language consolidation
rules of adjacent areas. By analyzing massive historical texts, words like the majority of
sea and inshore sea are introduced to describe multiple adjacent areas to make the text
expression more compact. For example, when three areas of Yellow Sea exist at the
same time, they will be summarized as the majority of the Yellow Sea. The north-
western and northeastern part of the South China Sea will be summarized as the
northern part of the South China Sea. Regarding nine areas from Bashi Channel to the
Southern part of the South China Sea, if more than 7 areas exist, they can be sum-
marized as the southern China Sea. If involved areas are more than 70%, they can be
summarized as China’s inshore seas.
The ocean weather public report is a Microsoft Word document. Figure 3 shows a
custom template word document. The brackets (【】) and information in it represent
the variable fields. The other parts are constant fields. The data in constant and variable
fields can be either texts or images. Positions of variables can be determined by
“bookmark” or defining variables and can be initialized. Next, methods based on
Aspose. Words, such as Document, DocumentBuilder, Bookmark, InsertImage can be
used to search, locate, fill and modify the text, image and other elements in the variable
field and modify the format. Finally, a new report Word document by Save method is
output.
4 Application
Based on the above method automatically generate the public report. Firstly, input data
mainly includes the ocean weather data for the next 1–3 days issued three times a day
by NMC and historical documents. Secondly, establish the paragraph template and
customize the document template by using natural language processing technology and
word frequency statistics technology. Finally, Aspose. Words dynamic link library is
called in the application program, developed in C# language, to generate the final
document of the ocean weather public report. The result of automatically generated text
and manually written text is shown in Fig. 4. The automatically generated report
document is shown in Fig. 5. It can be seen that variable fields in the custom template
have been filled with valid data and the font, paragraph, and the layout of the document
are consistent with the custom template.
Research on Natural Language Processing and Aspose Technology 477
5 Conclusion
This paper focuses on main issues of the automatic generation of the ocean weather
public report, including making texts closer to forecaster’s intention, ensuring the
correctness of text description in terms of grammar and semantics, and ensuring the
efficiency and correctness of the report document. The natural language processing
technology and Aspose technology are used to explore the automatic generation
method of the report document. The results are shown as following:
Automatically generated text content is correct, which is consistent with the orig-
inal data without missing important meteorological elements and any sea area. It also
coincides well with manually written texts by forecasters, which is convenient for them
to just make some small amount of modification on this basis and output the final
release document more quickly.
The automatically generated text is fluent and correct in semantics and grammar,
and the text style conforms to forecasters’ writing habits.
The automatically generated document guarantees the correctness, consistency and
standardization which effectively avoids common errors of weather elements and time
interval selection, illustration sorting, and text styles that occur during manual editing
of documents.
Forecasters can reduce much complicated manual labor and improve the efficiency
of forecasting services.
This method has strong practicability.
References
1. Wu, H., Lv, Z., Zhang, H., et al.: Text generation on weather falling area description. Comput.
Eng. Appl. 50(13), 247–251 (2014)
2. Zhang, J., Tan, X.: A method for automatically generating meteorological text products.
Meteorol. Sci. Technol. 32, 72–93 (2004)
3. Yao, T., Huang, X., Zhang, D., et al.: The research on the model for Chinese text generation
of weather forecast. J. Jiangxi Normal Univ. 22, 102–111 (1998)
4. Zhou, Q., Zhu, X., Hu, X.: Document generation method based on Aspose technology and
custom template. Comput. Syst. Appl. 24(6), 235–238 (2015)
5. Wang, X., Qing, G., Zhang, M.: Application of Aspose. Words Public Meteorol. Serv.
Comput. Program. Skills Maintenance 18, 43–45 (2015)
4G Video Service Stalling Perception
Evaluation and Optimization
Based on Big Data
1 Introduction
In recent years, mobile Internet services have developed rapidly. The proportion of
netizens using mobile phones reaches 98.3%, mainly for 4G networks [1]. The telecom
operators face a series of challenges, especially the improved requirements of users’
quality of service (QoS) [2–4]. The traffic proportion of mobile Internet video services
has exceeded 60%, which is the main driving force for network traffic growth. Cur-
rently, Tencent video service traffic accounts for more than 56% of video service traffic
[5]. Therefore, the analysis of Tencent video service stalling rate can help improve the
4G network video service experience. However, mobile video service perception
evaluation involves a wide range of processes and a large number of network elements
[6, 7]. Thus, a reasonable video perception evaluation method has always been nec-
essary and precious.
This paper proposes a 4G mobile video service perception evaluation model, based
on the network operation and user service big data analysis. This model includes the
correlation coefficient method, the interval score method, and the interval increment
method to select and analyze KQIs. Through the establishment and application of the
model, we find a problem diagnosis method of video service poor-quality cells and
content servers. Therefore, accurate evaluation and quick optimization can be per-
formed to improve service perception.
2 Evaluation Model
2.1 Model Establishment
Video service stalling is an important basis for determining the performance of 4G
video services. It can directly reflect the experience of video services. The big data
analysis method is used to mine xDRs (x detail record) on the 4G network. Based on
factors that affect video service experience, the following four analysis methods are
used to map KQIs from multiple dimensions to CEI (customer experience indicator)
[8]: Correlation coefficient calculation, segmentation statistics, normal distribution
statistics, and comprehensive evaluation. Thus, we can create an evaluation model for
user video stalling perception to accurately evaluate video service stalling perception.
The video perception model is showed as the Fig. 1.
downlink packet loss rate) affecting video service perception. Therefore, major KQI
identification, indicator weight calculation, and evaluation threshold selection are key
to the final perception scoring model [9].
A. KQI Identification & Indicator Weight Calculation
The correlation coefficient method is used to identify the KQIs with high correlation to
evaluate the video service stalling perception.
Select n items for which correlation evaluation is performed, and calculate the
stalling rate based on the 0.5% step. Then calculate the average value of KQIs in each
range to obtain the stalling rate column S = (0%, 0.5%, …, 12.5%), in which the
number of variables is 26 and 95% cells are included.
Suppose that Si is for KQI of item i, Ri is for the correlation coefficient between
KQI and Si . The calculation method of Ri is as Formula (1):
Pn Pn Pn
n Sj Sij Sj Sij
j¼1 j¼1 j¼1
Ri ¼ vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
!2ffi v ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
!2ffi ði ¼ 1; 2; . . .; n; j ¼ 1; 2; . . .; nÞ
u u
u P u
tn S 2 P S P P
n n n n
j j tn S2ij Sij
j¼1 j¼1 j¼1 j¼1
ð1Þ
i is for the sequence number corresponding to the KQI, j is for the variable numbers
corresponding to Si and S.
The correlation depends on the absolute value of Ri . Thus we identify the KQI with
the absolute value 0.8 as the evaluation basis to evaluate the mapping between KQIs
and CEIs.
The KQI weight is proportional to the correlation coefficient of the stalling rate. The
weight of each KQI can be calculated based on related coefficients. Then the calcu-
lation method of the percentile weight of the KQI indicator di is as Formula (2):
Ri
di ¼ P
n 100 ði ¼ 1; 2; . . .; nÞ ð2Þ
Ri
i¼1
6000
4000
2000
0
0 20 40 60 80 100 120 140 160 180 200 220 240 260 280 300 320 340 360
(The downlink RTT delay indicator)
C. Scoring Calculation
Referencing the video perception model, we obtain major KQI identification, indicator
weight and evaluation threshold. Then we can calculate the video perception score by
the comprehensive evaluation method.
Suppose that gi is for the evaluation coefficient of KQI index, Kih is for the eval-
uation threshold of the good value, Kiz is for the evaluation threshold of the median
value, Kic is for the evaluation threshold of the poor value, Kix is for the KQI. The
evaluation coefficients gi can be calculated by using the interval scoring method and the
interval increment method. The calculation method is as Formula (3):
8
>
> If kix is better than Kih ; then gi ¼ 1
> If k is
< worse than Kic ; then gi ¼ 0
ix
gi ¼ If kix is jKix Kiz j
better than Kiz ; and worse than Kih ; then gi ¼ jK 0:6 þ 0:4
>
> ih Kiz j
>
: If k is
ix better than Kic ; and worse than Kiz ; then gi ¼ jK ix Kic j
jKiz Kic j 0:4
ð3Þ
Based on evaluation weights di, putting the evaluation ideas of interval scoring
method and interval increment method into the evaluation process, we obtain the
service perception scoring results.
X
n
Video Perception Score ¼ di gi ði ¼ 1; 2; . . .; nÞ ð4Þ
i¼1
gi is for the evaluation coefficient of KQI index, di is for the evaluation weights.
4G Video Service Stalling Perception Evaluation and Optimization 483
3 Model Verification
3.1 Analysis Object
We select 38,441 cells with the poor stalling rate as objects for model verification.
3.2 Weight
Firstly, the stalling rate is calculated by the segment, and then the correlation coefficient
is obtained by means of calculation according to Formula (1). Finally, five KQIs whose
absolute values of correlation coefficient are >0:8 are obtained (Tables 1 and 2).
3.3 Threshold
We collect statistics on selected KQIs based on the number of cells, and then we obtain
the distribution curve. Based on the evaluation threshold, the grading threshold of each
KQI is as Table 3.
Fig. 3. The fitting curve of the perception score and stalling rate
The mapping between the score model and the actual stalling rate is high. The
stalling rate corresponding to the curve 60 points is 1.90%, and the average stalling rate
1.74% corresponds to 70 points. There is an obvious correlation between two curves.
For 38441 cells, the precision rate is 73% and the recall rate is 80.96%. Based on
the traditional analysis method, this is a good result. The scope and accuracy of the
check are considered, which proves that the evaluation model is effective and feasible.
4 Optimization Effect
During network optimization, the video stalling rate evaluation model is used to
evaluate video services based on the service IP address, Gateway, transmission
IP RAN, and wireless network cells. We can identify network elements with poor
quality based on scores.
The score of IPRAN-22 on the transmission ring is significantly lower than that on
others. The cause is that the network delay increases due to high transmission load.
After the bandwidth is expanded, the indicator becomes normal (Table 4).
Based on the evaluation of wireless network cells, TOP 1,000 cells with poor scores
are filtered out. We associate 4G network performance counters to locate the problem,
and implement optimization by means of carrier capacity expansion, parameter opti-
mization, and radio frequency adjustment. Then, the stalling rate of Tencent video
services is improved by 0.27PP, greatly improving work efficiency (Fig. 4).
2.40%
2.30%
2.18%
2.20%
2.10%
2.00%
Before optimization After optimization
5 Conclusion
This paper presents an effective and feasible evaluation model of 4G mobile video
service perception, based on the network operation and user service big data analysis.
With the application of the model, the video service stalling problem evaluation and
locating efficiency is greatly improved. In addition, it supports efficient implementation
of optimization and improves user experience in 4G network video services. At the
same time, it provides a valuable reference for E2E (end to end) perception evaluation
and optimization.
References
1. China Internet Network Information Center: The 42nd China Statistical Report on Internet
Development, July 2018
2. Xu, L., Zhao, X., Luan, Y., et al.: User perception aware telecom data mining and network
management for LTE/LTE-advanced networks. In: 4th International Conference on Signal
and Information Processing, Networking and Computers, pp. 237–245. Springer, Qingdao
(2018)
3. Xu, L., Zhao, X., Yu, Y., et al.: Data mining for base station evaluation in LTE cellular
systems. In: 3rd International Conference on Signal and Information Processing, Networking
and Computers, pp. 356–364. Springer, Chongqing (2017)
4. Xu, L., Luan, Y., Cheng, X., et al.: WCDMA data based LTE site selection scheme in LTE
deployment. In: 1st International Conference on Signal and Information Processing,
Networking and Computers, pp. 249–260. CRC Press Taylor & Francis Group, Beijing
(2015)
486 J. Yin et al.
5. Chen, C., Ke, J., Qin, D.: Discussion on Video Service Experience Evaluation and
Optimization. Designing Techniques of Posts & Telecommunications (2017)
6. Xu, L., Luan, Y., Cheng, X., et al.: Telecom big data based user offloading self-optimisation
in heterogeneous relay cellular systems. Int. J. Distrib. Syst. Technol. 8(2), 27–46 (2017)
7. Xu, L., Cheng, X., et al.: Mobility load balancing aware radio resource allocation scheme for
LTE-advanced cellular networks. In: 16th IEEE International Conference on Communication
Technology, pp. 806–812. IEEE Press, Hangzhou (2015)
8. Abdellah, S., Sara, M., El-Houda, MN., Samir, T.: QoS and QoE for mobile video service
over 4G LTE network. In: Computing Conference, pp. 1263–1269 (2017)
9. Liu, L., Zhou, WA., Song, J.: Quantitative customer perception evaluation for telecommu-
nication service. In: International Conference on Pervasive Computing & Applications, vol. 2,
pp. 912–915 (2009)
Application of the Bayesian Processor
of Ensemble to the Combination
and Calibration of Ensemble Forecasts
1 Introduction
The observational temperature data set that is used in this study was provided by MDL
of NWS. It contains T2M observations from 24 stations over the conterminous United
States (CONUS) (Fig. 1) during 1977–2010.
The NAEFS combines the National Centers for Environmental Prediction’s
(NCEP) Global Ensemble Forecast System (GEFS) with the Canadian Meteorological
Center (CMC)’s ensemble prediction system. Both GEFS and CMC systems contain a
control forecast and 20 perturbed members. All data were taken from the 0000 UTC
runs with lead time from 24 to 240 h. The gridded ensemble forecasts were interpolated
to the 24 station sites. In this study we focus on the warm half year season of April–
September. Similar results were found for the cold half year of October–March (not
shown). Two seasons were used for development of BPE (2008, 2009) and the third
season (2010) was used for cross-validation.
Application of the Bayesian Processor of Ensemble 489
3 Methodology
The idea of the Bayesian Processor of Ensemble (BPE) is to build a likelihood function
from the joint sample of ensemble forecasts and observations, and to combine it with a
prior distribution to obtain the posterior distribution of the predictand. According to
Bayes theorem:
f ðXjwÞgðwÞ
£ðwÞ ¼ ð1Þ
kðXÞ
Here gðwÞ is the prior density function of the predictand (I.e., climatological dis-
tribution); f ðXjwÞ is called the likelihood function of the predictand and describes the
conditional probability density of the forecast given an observed value; kðXÞ is the
expected density function of the predictor vector X and £ðwÞ is the posterior density
function.
Thus, the BPE includes three components: (i) estimation of the prior distribution,
(ii) modeling of the likelihood function, and (iii) determination of the posterior dis-
tribution. Details of the procedures are found in Krzysztofowicz [13]. Prior distribution
for the predictand Prior distribution is based on a long climatic sample. In this study we
use 30 years from 1977 to 2006. To increase the sample size, data are pooled from
consecutive 5 days. The daily time series are standardized to guarantee the stationarity
of the first two moments. Seasonality effect is also removed after standardization.
Instead of estimating a different distribution function for every day of the season, it is
sufficient to estimate just one distribution function. From a set of candidates, the
Weibull distribution is chosen as it fits the sample best (not shown).
xk mk 0 wk mk
x0k ¼ ; wk ¼ ð2Þ
Sk Sk
Next step is to model the stochastic dependence between the standardized forecast
and predictand. Here we employ the meta-Gaussian likelihood model of Krzyszto-
fowicz and Kelly [13] using the normal quantile transform (NQT):
where G0 is the prior distribution, K 0 denotes the marginal distribution function of the
standardized forecast x0 ; Q is the standard normal distribution function, and Q1 is its
inverse. The NQT guarantees that each marginal distribution function is Gaussian in
the sample space of the transformed variates (Z, V).
The likelihood parameters a, b and r2 can be estimated through:
a ab r2
A¼ ;B ¼ 2 and T 2 ¼ 2 ð5Þ
a2 þr 2 a þr 2 a þ r2
The posterior distribution is obtained by simply combining the prior with the
likelihood function, where the posterior distribution function £ of w, is specified by
the equation
1 1 1
£ðwÞ ¼ Q Q ðGðwÞÞ AQ ðKðxÞÞ B ð6Þ
T
The reliability component Rel measures the ability of the EPS to be statistically
consistent by comparing the observed frequency of events with their forecast proba-
bility. When perfectly calibrated (i.e., zero Rel error), a given EPS system would reach
its lowest, Potential CRPS value, which is also a function of the observational data.
4 Results
The BPE coefficients are developed and then applied for each station and each lead
time independently. Two seasons of a joint sample (2008, 2009) is used as training
data, and a third year (2010) is used as the cross-validation data. BPE outputs calibrated
distribution functions, quantile functions, density functions, and ensemble members.
Here we use 11 posterior quantiles (0.05, 0.1, 0.2, … 0.9, 0.95) for verification. The
verification is carried out for all sites and CRPS and its reliability component is then
averaged over the 24 stations and the entire verification period.
To assess the effect of merging different ensemble forecast systems, we designed
and tested two configurations of BPE with different potential predictors for the multi-
linear regression for central tendency of the likelihood function. Table 1 gives the
configuration name and the potential predictors. For simplicity, both configurations use
the GEFS ensemble as a predictor for both the central tendency and uncertainty.
The GEFS control is also offered as a predictor of central tendency in both configu-
rations, while BPE-GEFS_CMC also adds the CMC ensemble mean as a potential
predictor.
Table 1. Configuration of BPE with different potential predictors for the multi-linear regression
for central tendency of likelihood function
Model name Potential predictors of T2M
BPE-GEFS GEFS Control Member, GEFS Ensemble Mean
BPE-GEFS_CMC GEFS Control Member, GEFS Ensemble Mean, CMC Ensemble Mean
The reliability component of CRPS is displayed in Fig. 2b. Short range raw 2mT
GEFS forecasts exhibit a large (up to 3.5 F) calibration error over both the dependent
(2008–2009) and independent (2010) verification periods. This is due to the lack of
initial surface perturbations in the GEFS system, highlighting the need for the statistical
calibration of raw ensemble forecasts. Due to their miscalibration, CRPS (Fig. 2a) for
the raw short range GEFS is also high (above 4.5 F). This leads to the unusual, mostly
flat raw GEFS CRPS curves.
In contrast, the calibration component of the BPE-GEFS system is significantly
reduced to around 0.15 F at all lead times, both for the dependent and independent
samples. Therefore, post-processed ensemble forecasts accurately represent the uncer-
tainty and are very reliable. This is due to the self-calibration property of the Bayesian
492 Y. Wang et al.
Fig. 2. (a) CRPS and (b) its decomposition in reliability of raw GEFS and BPE-GEFS calibrated
ensemble forecast averaged over 24 stations during summer half year for training sample
(2008/09) and independent sample (2010).
formulation used in BPE [16]. Correspondingly, CRPS for the BPE-GEFS system is
significantly lowered, starting from above 2 F and reaching just below 4 F at longer lead
times. As with the reliability component, BPE-GEFS exhibits similarly low scores for
CRPS over the dependent and independent verification periods (cf. red vs blue dotted
curves in Fig. 2a), indicating a robust performance with no apparent overfitting.
BPE is set up to combine forecasts from any number of systems. When the CMC
ensemble mean is included as an additional potential predictor for central tendency
(BPE-GEFS_CMC) beyond the GEFS ensemble and control forecasts (BPE-GEFS),
CRPS is generally lowered (except at Day 9 lead time for 2010), thus extending pre-
dictability by 12–24 h (Fig. 3a). The combined BPE-GEFS_CMC configuration displays
calibration scores even better than the already low values found for BPE-GEFS (Fig. 3b).
Fig. 3. (a) CRPS and (b) its decomposition in reliability of BPE-GEFS, BPE-GEFS_CMC,
calibrated ensemble forecast averaged over 24 stations during summer half year for training
sample (2008/09) and independent sample (2010).
Application of the Bayesian Processor of Ensemble 493
The above verification results indicate the value of post-processing raw ensemble
forecasts with BPE in terms of enhanced skill and reliability. Encouragingly, the results
on dependent data carry over to an independent verification period. Since BPE can
combine independent input information from multiple forecast systems, it can be used
to merge output from multiple ensemble forecast systems.
References
1. Hamill, T.M., Colucci, S.J.: Verification of Eta-RSM short-range ensemble forecasts. Mon.
Weather Rev. 125, 1312–1327 (1997)
2. Wilks, D., Hamill, T.: Comparison of Ensemble-MOS Methods using GFS Reforecasts.
Mon. Weather Rev. 135, 2379–2390 (2007)
3. Gneiting, T., Balabdaoui, F., Raftery, A.E.: Probabilistic forecasts, calibration and sharpness.
Roy. Stat. Soc. 69, 243–268 (2007)
494 Y. Wang et al.
4. Glahn, H.R., Peroutka, M.R., Wiedenfeld, J., Wagner, J.L., Zylstra, G., Schuknecht, B.,
Jackson, B.: MOS uncertainty estimates in an ensemble framework. Mon. Weather Rev. 137,
246–268 (2009)
5. Gneiting, T., Raftery, A.E.: Weather forecasting with ensemble methods. Science 310, 248–
249 (2005)
6. Wilks, D.S.: Extending logistic regression to provide full-probability-distribution MOS
forecasts. Meteorol. Appl. 16, 361–368 (2009)
7. Cui, B., Toth, Z., Zhu, Y., Hou, D.: Bias correction for global ensemble forecast. Weather
Forecast. 27, 396–410 (2012)
8. Raftery, A.E., Gneiting, T., Balbdaoui, F., Polakowskil, M.: Using Bayesian model
averaging to calibrate forecasts. Mon. Weather Rev. 133, 1155–1174 (2005)
9. Hamill, T.M.: Comments on ‘‘Calibrated surface temperature forecasts from the Canadian
ensemble prediction system using Bayesian Model Averaging’’. Mon. Weather Rev. 135,
4226–4236 (2007)
10. Hodyss, D., Satterfield, E., McLay, J., Hamill, T.M., Scheuerer, M.: Inaccuracies with
multimodel postprocessing methods involving weighted, regression-corrected forecasts.
Mon. Weather Rev. 144, 1649–1668 (2016)
11. Krzysztofowicz, R., Evans, W.B.: Probabilistic forecasts from the national digital forecast
database. Weather Forecast. 23, 270–289 (2008)
12. Candille, G.: The multi-ensemble approach: the NAEFS example. Mon. Weather Rev. 137,
1655–1665 (2009)
13. Krzysztofowicz, R.: Algorithms Implementing Bayesian Processor of Ensemble (BPE), p. 75
(2016). Available upon request (2016)
14. Matheson, J.E., Winkler, R.L.: Scoring rules for continuous probability distributions.
Manage. Sci. 22, 1087–1096 (1976)
15. Hersbach, H.: Decomposition of the Continuous Ranked Probability Score for ensemble
prediction systems. Weather Forecast. 15, 559–570 (2000)
16. Krzysztofowicz, R.: Decision criteria, data fusion, and prediction calibration: a Bayesian
approach. Hydrol. Sci. J. 55, 1033–1050 (2010)
Initial Analysis of the Cell Selection Progress
in SA of 5G NR
1 Introduction
With the acceleration of the commercial pace of the 5th generation mobile networks
worldwide, ultra-high speed, ultra-high spectral efficiency, ultra-low latency and
massive connectivity based on massive IoT have become the consensus of the next
generation broadband mobile communication networks. 5G NR technology adopts
application-oriented design ideas in the formulation of standards, and adapts to the
differentiated networking requirements through flexible and variable system-level
parameter adjustment. The 5G NR has attracted wide attention from domestic and
foreign operators due to its targeted design for new spectrum (millimeter wave). In the
standard setting phase, 5G NR has formulated two networking modes according to the
actual networking development requirements. One is non-standalone networking mode
that needs to be combined with the existing 4G network, and the other is standalone
networking mode that can be independently deployed and operated. As the develop-
ment of standardization, the related networking technology details are still being
improved. No matter which networking mode is adopted, the UE cell selection
mechanism is the basic process for implementing end-to-end communication in any
communication system, and it is also an important entry point for subsequent network
In addition to the two NSA candidate schemes illustrated in Fig. 1, there are some
other candidates. However, it is generally believed that the candidate of the operator’s
Initial Analysis of the Cell Selection Progress in SA of 5G NR 497
initial most feasible system is the one on the left of Fig. 1 [1]. This paper does not
cover the other candidates. We can choose between co-site and non-co-site for NSA
networking deployment.
The 5G NR still follows the design concept of control plane and user plane sep-
aration in LTE. The SDAP sublayer is added to the user plane to provide the QoS
mapping relationship between the core network and the data radio bearer (DRB).
498 Z. Xu et al.
full band (FR1: 450 MHz–6000 MHz; FR2: 24250 MHz–52600 MHz), the UE can
synchronize downward frequency according to the pre-built operating frequency band
of the country’s operator, thereby shortening the initial boot cell search time. The 5G
NR terminal is step-locked in accordance with the accuracy of the synchronization
raster in frequency synchronization. The 5G NR sync grid is not fixed with 100 kHz as
the synchronization precision in LTE/NB-IoT, but combined with the calculation
formula of GSCN (global synchronization channel number) and center frequency point
of SSB [3, 4]. It is worth mentioning that the UE needs to further clarify the center
frequency by determining the location of the 20 PRBs in the frequency domain
occupied by the SSB. Since the OFDM subcarrier spacing of the 5G NR is flexible, the
UE needs to make a traversal search attempt according to the subcarrier spacing cor-
responding to the different working frequency bands to determine the actual subcarrier
spacing of the SSB in the initial downward synchronization. After locking SSB to
achieve frequency domain synchronization, UE can achieve time domain synchro-
nization of symbol level precision by combining carrier frequency and acquired sub-
carrier spacing [5].
other SI system messages for the EN-DC data transmission mode of the non-
independent networking. Of course, the system design can also perform SIB1 design
for compatibility with the independent networking mode, but the UE cannot listen to
the decoding.
239 P
B
C
192 H
182
P P
P S
Subcarrier B B
S S
Number C C
S S
H H
56
47 P
B
C
H
0
0 1 2 3
The UE first searches for the PSS primary synchronization signal, and the PSS
length of the NR is 127 pseudo-random sequence, using the frequency domain
BPSK M sequence. The PSS maps to consecutive 127 subcarriers in the middle of 12
PRBs, occupies 144 subcarriers, and performs guard intervals on both sides without
transmitting power. After the UE searches for the PSS, the subcarrier spacing of the
SSB can be obtained.
The SSS frequency domain is similar to the PSS. It maps to 127 subcarriers in the
middle of 12 PRBs and occupies 144 subcarriers. After the UE searches for SSS, it can
Initial Analysis of the Cell Selection Progress in SA of 5G NR 501
ð1Þ
obtain NID . The only physical layer cell ID in NR is determined as the following
formula:
The different symbol positions occupied by PSS, SSS, PBCH, and DM-RS are
given in Table 2, where v ¼ NID
cell
mod 4.
-- ASN1START
-- TAG-MIB-START
-- TAG-MIB-STOP
LTE, mainly transmitted through the downlink PDSCH channel, and the PDSCH
channel needs the DCI of the PDCCH to be scheduled. The UE needs to obtain the
PDCCH channel information of the scheduled RMSI in MIB, perform blind detection
on the PDCCH to obtain the RMSI. This information in the MIB is the pdcch-
ConfigSIB1 field [11].
The parameter ssb-SubcarrierOffset in the MIB message can be used to confirm
whether the current SSB has a common CORSET (control resource set) in the fre-
quency domain, that is, the Type0-PDCCH common channel is configured. Therefore,
it can be determined whether the frequency domain of the current SSB is configured
with SIB1. When the UE detects that the current SSB in the frequency domain is not
configured with SIB1 under certain conditions, the parameter pdcch-ConfigSIB1
contained in the MIB message can be used to detect whether the Type0-PDCCH public
search space provides CORSET in the frequency domain of the next SSB and the
biased SSB. If the UE still does not detect the public CORSET, it can be considered
that the cell is not configured with SIB1, and then the cell search procedure for the
corresponding frequency point obtained by the previous frequency-locked SSB will be
abandoned [12].
The PDCCH channel in the NR corresponds to multiple search spaces, including
common search space and UE-specific search space. The Common Search space of
Type 0 is only used for RMSI scheduling. In NR, CORESET (Control Resource SET)
is introduced for the physical Resource SET for the PDCCH channel. One cell PDCCH
channel corresponds to multiple CORESET sets, and the CORESET set has an ID
number, where CORESET 0 represents a physical resource set corresponding to the
search space.
the 5G NR cell user plane can be independently scheduled by listening to the physical
control channel. The UE may acquire the 5G NR system configuration parameter by
using the camped 4G LTE cell RRC reconfiguration message. The LTE cell configures
the system configuration parameter of the 5G NR as the second cell group related
parameter of the EN-DC, and encapsulates it into an 8-byte string group for trans-
mission. The UE can quickly obtain the SSB message by decoding these auxiliary
configuration information. These auxiliary information mainly involves the following
four aspects:
Subcarrier interval (numerology) information of SSB: 5G is flexible and config-
urable for OFDM system level parameter subcarrier spacing, and subcarrier spacing of
SSB can only be configured for 15 kHz or 30 kHz (carrier frequency <6 GHz),
120 kHz or 240 kHz (carrier frequency >6 GHz);
Frequency domain location information of the SSB: the starting node of the fre-
quency domain (absoluteFrequencyPointA), the frequency domain location of the SSB
(absoluteFrequencySSB, defined as the no. 0 subcarrier of the no. 10 index RB in
SSB), the frequency band (frequencyBandList) and actual carrier start position (off-
setToCarrier) or bandwidth (carrierBandwidth) based on different subcarrier spacing;
Time domain location information of SSB: The SSB can repeatedly transmit at a
specific candidate position within a half-frame (5 ms), and the number of repetitions
and whether the candidate time domain position is actually configured by the SSB can
be known by the parameter ssb-PositionsInBurst. If the candidate location is not
configured with SSB, the frequency domain location of the corresponding SSB does
not need to be used for other channels or signal transmission. At the same time, the
auxiliary time domain information also includes the repetition period (ssb-
periodicityServingCell) for configuring the SSB half-frames, and the default value is
repeated every half frame (5 ms).
5 Conclusion
antennas; after receiving the PBCH, the UE also receives system messages transmitted
on the PDSCH. Finally get the complete system message.
References
1. Zhang, J., Xie, W., Yang, F.: Architecture and solutions of 5G ultra dense network.
Telecommun. Sci. 44–51 (2016)
2. Liu, Y., Li, X., Ji, H.: Key technologies of network self-organization in 5G ultra-dense
scenarios. Telecommun. Sci. 44–51 (2016)
3. Chen, M.: New network planning technology for 5G ultra-dense networking. Mob.
Commun. 28–29 (2016)
4. Lei, Q., Zhang, Z., Cheng, F.: 5G radio access network architecture based on C-RAN.
Telecommun. Sci. 106–115 (2015)
5. 3GPP TS 38.331 V15.2.0 NR; Radio Resource Control (RRC) protocol specification
(Release 15)
6. Zhou, H.: 5G wireless access network architecture design. Electron. Sci. Technol. 102–105
(2017)
7. 3GPP TS 38.101-1 V15.2.0 NR; User Equipment (UE) radio transmission and reception;
Part 1: Range 1 Standalone (Release 15)
8. 3GPP TS 38.213 V15.2.0 NR; Physical layer procedures for control (Release 15)
9. Zhang, J., Xie, W., Yang, F.: Mobile edge computing and application in traffic offloading.
Telecommun. Sci. 32, 132–139 (2016)
10. Bai, L., Liu, T., Yang, C.: Interference coordination method and performance analysis in
ultra-dense network. Sig. Process. 1263–1271 (2015)
11. Xu, L., Cheng, X., et al.: Mobility load balancing aware radio resource allocation scheme for
LTE-advanced cellular networks. In: 16th IEEE International Conference on Communica-
tion Technology, pp. 806–812. IEEE Press, Hangzhou (2015)
12. Xu, L., Luan, Y., Cheng, X., Cao, X., Chao, K., Gao, J., Jia, Y., Wang, S.: WCDMA data
based LTE site selection scheme in LTE deployment. In: 1st International Conference on
Signal and Information Processing, Networking and Computers, pp. 249–260. CRC Press
Taylor & Francis Group, Beijing (2015)
NB-IoT Network and Service Collaborative
Optimization
1 Introduction
Mobile Monitoring. Such services mainly include services such as personnel and item
location tracking, and are sensitive to network connectivity, power consumption, and
mobility metrics. Involved indicators: Attach success rate, RRC connection success
rate, uplink rate, and downlink rate.
Mobile Control Class. Such services mainly include services such as shared bicycle,
smart wear, and mobile payment, and are sensitive to network connectivity, service
connectivity, power consumption, and mobility metrics. Involved indicators: Attach
success rate, RRC connection success rate, RRC connection delay, uplink rate,
downlink rate.
Pcmax is the maximum transmit power of the terminal, P0-Npusch is the initial target
power of the UE, PL is the path loss, and a is the road loss compensation weight, which
is usually 0.7–0.8.
According to the calculated uplink path loss, it considers the difference between the
uplink and downlink frequencies, and estimates the downlink path loss, and further
approximate the RSRP of the location of the UE. The formula is as follows:
Among them, RSTXPwr is the reference signal power, and D is the correction value
for the uplink path loss.
In the NB-IoT network planning and design, GSM (900 MHz) is planned according
to 1:4 site, and the depth coverage requirement scenario can adopt 1:2, 1:1 planning
scheme, and on-demand planning to ensure good coverage of the NB-IoT network; The
antenna uses the 2T4R deployment principle to replace the site with 2T4R antenna
feeds to improve the uplink performance of the NB-IoT network. NB and GSM
independent antenna feeder systems should be deployed in key areas and overlapping
coverage areas to ensure that NB optimization does not affect GSM network coverage.
The outdoor road edge coverage RSRP is not lower than –84dBm, and the SINR is not
lower than –3 dB. It is expected that no less than 30 dB penetration loss can be
reserved to meet the general scene depth coverage requirements [8].
Tests show that NB-IoT is 1:1 compared with GSM networking mode 1:2, RSRP is
about 4–7 dB, and 1:4 is about 10–12 dB lower than 1:1. However, in the 1:2 net-
working mode, the SINR is 1.5–2 dB higher than the 1:1, and the SINR is 0.6–2.4 dB
higher than the 1:4. The pairs under different load loads are shown in Table 1.
Tests show that NB-IoT in 1:4 network can reach the same level of GSM coverage
and has certain wear and tear redundancy. If 1:1 networking is adopted, not only the
investment is too large, but also the problem that the network noise rise is obvious and
the interference control is difficult. In the initial stage of network construction, NB-
IoT’s coverage enhancement technology can be fully utilized to achieve basic coverage
of NB-IoT compared to the GSM900 macro station’s 1:4 mode, which can support a
510 P. Liu et al.
wide range of business development needs and effectively control investment risks. It
should be noted that when 1:4 networking, NB-IoT can basically meet indoor demand,
95% coverage can reach –107 dBm, and receiving sensitivity has 25 dB surplus, which
can meet basic indoor coverage requirements, and some deep coverage can be Deepen
coverage based on actual conditions.
Fig. 2. Comparison of coverage performance between NB-IoT total antenna feed and
independent antenna feed.
NB-IoT Network and Service Collaborative Optimization 511
The test results show that the outdoor road coverage, the NB-IoT independent
antenna feeder has a RSRP increase of 0.86 dB and a SINR increase of 1.67 dB. The
test pairs in the outdoor coverage room are shown in Table 2.
The test results show that the outdoor coverage room, the NB-IoT independent
antenna feed has a 2.5 dB increase in RSRP and a 3.5 dB increase in SINR.
Summary: Under the NB-IoT independent antenna feed scenario, the RF opti-
mization adjustment scheme will not be affected by the different coverage targets of
GSM and FDD networks, and no joint optimization is needed. The azimuth can be
independently combined with the requirements of NB-IoT network coverage opti-
mization. The adjustment of the mechanical downtilt angle, the electronic downtilt
angle and even the antenna hanging height optimizes the RF optimization effect.
According to field test research, the independent antenna feeder network has a sig-
nificant effect on the optimization of depth coverage.
The maximum number of users in the 24 h of a single cell in the test area is
collected. It can be seen from that the statistical data is consistent with the on-line time
of the street light, as shown in Fig. 4.
NB-IoT network defaults that the PRACH channel occupies 12 subcarriers, and
each user access occupies one subcarrier. Under the better conditions of the wireless
environment, 12 terminals initiate random access at the same time, and the RRC access
time to all terminals is 0.93 s. The Msg2 to Msg5 resource allocation process takes
0.92 s, and the 1 s supports up to 12 concurrent users. From this point of view,
concurrent access is mainly limited by the resource allocation in the process of Msg2 to
Msg5.
The NB-IoT access capability mainly considers two aspects: one is the maximum
number of RRC connected users, and the other is the number of valid RRC connected
users. At present, the maximum number of RRC connected users of Huawei and ZTE is
600.
Through calculation, the maximum number of valid RRC connected users in
100 ms is 150, calculated as follows:
The maximum effective number of users in the uplink theory. In the Single-tone
15 kHz subcarrier mode, each user occupies one RU, and the resources occupied by the
PRACH are not considered.
In the simulation calculation, the maximum number of valid RRC connected users
is 50, which is calculated as follows:
Simulate the maximum number of effective users. Under the better conditions of the
wireless environment, the Single-tone 15 kHz subcarrier mode is used for uplink
scheduling simulation. Each uplink user occupies an average of 2.5 RUs. Considering
the PRACH resource consumption and scheduling efficiency, the 8 ms average
scheduling is calculated. The number of uplink users is four.
Number of valid RRC connected users ð100 msÞ ¼ 100 ms=8 ms * 4 ¼ 50 ð5Þ
In theory, each uplink user occupies 1 RU. In fact, the row user occupies an average
of 2.5 RUs because the number of valid users decreases to 50. One RU frequency
domain is 15 k and the time domain is 8 ms.
Summary: NB-IoT access capability is mainly limited by the capability of con-
current access processing. Constrained concurrent access will generate a large amount
of retransmission signaling, causing signaling avalanche. The research and verification
of the concurrent access processing capability of NB-IoT base stations under different
coverage levels is carried out. By verifying the concurrent access test of 60 terminals,
the access can be concurrently transmitted under coverage level 0 and coverage level 1,
and only 36 concurrent accesses can be accessed under coverage level 2.
514 P. Liu et al.
5 Conclusion
NB-IoT network construction is initially completed, but the overlapping coverage and
interference problems are prominent; the overall coverage quality is up to standard, but
there are still coverage holes and continuous weak coverage sections; at this stage, NB-
IoT is mainly covered by macro stations, which can solve the problem. 90% indoor
general scene coverage, however, some indoor complex and difficult scenarios need to
further study the low-cost, easy-to-deploy NB indoor scene coverage scheme. In the
optimization analysis of the problems that have occurred in NB-IoT, it can be found
that the terminal management and control capabilities are lacking. Each terminal
manufacturer can define the frequency and time of the terminal access network. If a
large number of terminals simultaneously access the network, it will lead to network
congestion and even network congestion. Signaling storm; when the market is required
to promote the business, the traction customer is aligned with the NB-IoT platform, and
the technical department communicates with the industry manufacturers before the
business application commercial. In addition, end-to-end analysis methods are lacking:
the enabling of industry terminals, in addition to the ability of operators to end-to-end
pipelines, depends on the ability of the manufacturer to manage the platform. For
example, the lamp management platform fails to control the lights, and the motorcycles
cannot be reported. Cycling status, etc.; end-to-end analysis means should be imple-
mented and improved as soon as possible to clarify the problem demarcation. For non-
network reasons, urge customers to check.
References
1. He, X., Song, L.: NB-IoT uplink coverage performance based on rate requirements.
Telecommun. Sci. 149–156 (2016)
2. Hao, H., Ye, L., Noya, Z.: Capacity performance analysis of NB-IoT independent
deployment. Mobile Commun. 78–84 (2017)
3. Guo, B., Liu, Y., Zhang, Y.: Discussion on NB-IoT wireless throughput and low power
consumption technology. Mobile Commun. 84–89 (2017)
4. Xu, L., Chen, Y, Schormans, J., et al.: User-vote assisted self-organizing load balancing for
OFDMA cellular systems. In: 22nd IEEE International Symposium on Personal Indoor and
Mobile Radio Communications, pp. 217–221. IEEE Press, Toronto (2011)
5. Chao, Z., Gao, Y., Ding, H.: Analysis of NB-IoT performance. Mobile Commun. 47–52
(2017)
6. Feng, C.: Discussion on the strategy of narrowband internet of things deployment. Mobile
Commun. 64–68 (2017)
7. Yan, S., Huang, J.: NB-IOT coverage analysis based on theory and measurement. Mobile
Inf. 72–74 (2016)
8. Wang, Y.: NB-IoT technology and network deployment. Electron. Test. 60–61 (2017)
9. Ritao, C., Anda, D., Fanli, M.: Preliminary study on planning objectives and planning of
NB-IoT. Telecommun. Sci. 137–143 (2016)
10. Liu, Y., Dong, J., Liu, N.: Key technology and planning simulation method of NB-IoT.
Telecommun. Sci. 144–148 (2016)
NB-IoT Network and Service Collaborative Optimization 515
11. Zhang, J.: Research on NB-IoT deployment strategy of China mobile. Mobile Commun.
25–30 (2017)
12. Xu, L., Zhao, X., Luan, Y., et al.: User perception aware telecom data mining and network
management for LTE/LTE-advanced networks. In: 4th International Conference on Signal
and Information Processing, Networking and Computers, pp. 237–245. Springer, Qingdao
(2018)
A Gender and Age Prediction Algorithm Using
Big Data Analytic Based on Mobile APPs
Information
Jie Gao(&), Tao Zhang, Jian Guan, Lexi Xu, and Xinzhou Cheng
1 Introduction
2 Relevant Work
To the best of our known, the feasibility of demographic inference through various
tracing data of the customers has been proposed many times in the past. For example,
the authors of [4] proposed a solution of predicting gender, age and religion tendency
of the mobile users based on the search queries from SNS, such as Facebook, i.e.. In
[5], the authors developed a scheme to predict demographics such as relationship, age
and gender. The scheme is not only based on the behavioral features of application
usage, voice call usage, and SMS usage, but also refers to the environment features,
such as Bluetooth and WiFi devices detected per day on mobile phones. Also in [6],
Suranga gave a warning that there will be multiple privacy and security issues with the
data collection through over-permission platform and share with other companies. In
order to verity the affection, the authors presented a framework to predict mobile users’
gender based on installed APPs simply and the accuracy could reach around 70% in the
numerical results.
Mobile phones are widely used worldwide [7–9]. The proposed gender and age
prediction algorithm is based on the LightGBM method by collecting the tracing data
such as installed APPs, records of APPs usage, type and price of the mobile phone. In
order to enhance the accuracy of prediction, cross verification scheme is applied in the
proposed GAPA. Through repeated iterative calculations based on the training data set,
we use the most accurate set of the features’ importance to train the forecast data set.
Then we check the accuracy of this calculation model by the log-loss function.
518 J. Gao et al.
In order to enhance the accuracy of prediction, LightGBM algorithm and cross veri-
fication scheme is applied in the proposed GAPA. Figure 1 presents the process of
GAPA and the whole process can be divided into five steps: data collection, feature
engineering, model training, cross verification and results evaluation.
N
Step 2 Step 3 Step 4
Step 1
Training Set
Feature
Begin Data Collection Model Training Meet the target?
Engineering
Forecast Set
Data Processing Model Training
Step 5
Results Evaluation
End
We represent the male and female as 1 and 2, we divide the age as 10 years for each
group. For example, the users who are in the 25–30 group will be represented by 3 in
the age feature.
What is more, we select game and physical health as the representative APP series
for different gender to analyze the rule more deeply. In Fig. 5, the physical health series
is a very popular series amount the female consumers during 20 to 50 years old. The
penetration of game series demonstrate a young tendency in male group, as the users
below 40 years will be more likely to install and play for fun.
A Gender and Age Prediction Algorithm Using Big Data Analytic 521
where ni;j is the number of times that term t occurs in document d, the denominator is
the total number of words in d.
The inverse document frequency is a measure of how much information the word
provides. It can be described in (2):
jDj
ITFi = log ð2Þ
jfj : ti 2dj gj
where |D| is the total number of documents in the corpus, {j: ti 2 dj} is the number of
documents where the term t appears.
Then TF-IDF is calculated as (3):
The TF-IDF value will increase as the number of times that a word appears in the
corpus proportionally.
core algorithm of GAPA. The parameters for the lightGBM in Python is set as follow
(Table 4):
1X N X 22
Loss ¼ yij lnðpij Þ ð4Þ
N i¼1 j¼1
where i is the number of users in forecast set, j is the different group of users divided by
gender and age, yij is whether a user is in the group of j or not, and pij is the probability
calculated by the GAPA for each user.
In a solution, the Loss-log is the cumulative sum of the error between the reality and
the prediction result. Apparently, 0 is the ideal target of the proposed algorithm.
4 Numerical Result
In this paper, the case analysis of machine learning is based on the user’s mobile
information collected by tracing platform. The data sets is processed and analyzed by
Sklearn in the Python environment (Fig. 6).
From the above, we can observe a disciplinary rule that the performance of GAPA
will be improved as the accumulation of features. Because all the basic features are
gains for the Sklearn algorithms. The inflection point happens when we add the
‘Top30–40 apps is installed’ features in the algorithms. Because this ten features
contain some interferential features that is an interference to classify the forecast data
set. The best result in our experiment is 2.67 by using the LightGBM algorithm.
A Gender and Age Prediction Algorithm Using Big Data Analytic 523
4.5
4.35
4.26
4
3.90 3.88
3.5
3.46
3.26
3.20 3.19
3.15
3.06 3.05
3 3.02
2.80
2.71
2.67
2.5
+BRAND +PRICE +DURATION +TFIDF_SUM +SOCIAL +PHYSICAL +TOOLS +TOP20APP IS +TOP40APP IS
INSTALLED INSTALLED
5 Conclusions
In this paper, a gender and age prediction algorithm based on big data analytic and
machine learning is proposed for study. The proposed framework can predict the user’s
information by machine learning algorithm with accuracy of 2.67. The proposed
algorithm considers more aspects and features than the algorithms in [4–6]. Also, the
precision can be improved unceasingly. In the end, we give the performance of GAPA
is significant by analyzing the collected data in the last part. And the results show that
the GAPA scheme can be generalized in the area of targeted advertising.
References
1. Jakir, K., Fenil, A., Mithila, S.: Different approaches and methods for targeted advertisements
by predicting user’s behavioral data and next location. In: Conference 2018, ICISC, pp. 1345–
1350. IEEE (2018)
2. Chen, T., Carlos, G.: Xgboost: a scalable tree boosting system. In: Conference 2016, ACM,
pp. 785–794. IEEE (2016)
3. Ke, G., Meng, Q., Finley, T.: LightGBM: a highly efficient gradient boosting decision tree. In:
Conference 2017, NIPS, pp. 342–353. NIPS (2017)
4. Bi, B., Shokouhi, M., Kosinki, M.: Inferring the demographics of search users: Social data
meets search queries. In: Conference 2013, World Wide Web, pp. 131–140. IEEE (2013)
5. Aarthi, S., Bharanidharan, S., Saravanan, M.: Predicting customer demographics in a mobile
social network. In: Conference 2011, International Conference on Advances in Social
Networks Analysis and Mining, pp. 553–554. IEEE (2011)
6. Chen, J., Wang, C., He, K.: Semantics-aware privacy risk assessment using self-learning
weight assignment for mobile apps. IEEE Trans. Dependable Secur. Comput. pp, 1 (2018)
7. Xu, L., Luan, Y., Cheng, X.: Telecom big data based user offloading self-optimisation in
heterogeneous relay cellular systems. Int. J. Distrib. Syst. Technol. 8, 27–46 (2017)
524 J. Gao et al.
8. Xu, L., Cheng, X., Chen, Y., Chao, K., Liu, D., Xing, H.: Self-optimised coordinated traffic
shifting scheme for LTE cellular systems. In: 1st EAI International Conference on Self-
Organizing Networks, pp. 67–75. Springer press, Beijing (2015)
9. Xu, L., Zhao, X., Luan, Y.: User Perception aware telecom data mining and network
management for LTE/LTE-advanced networks. In: 4th International Conference on Signal
and Information Processing, Networking and Computers, pp. 237–245. Springer press,
Qingdao (2018)
Medium-Extended-Range Weather Forecast
Based on Big Data Application
1 Introduction
With the development of modern weather forecast services, the amount of data in all
fields of meteorological services, including meteorological observations, numerical
model simulations, objective derivatives and diversified services, has grown geomet-
rically. The massive data is used in MERWF services, which constitutes a big data
environment. At present, the national forecast operation system is the Meteorological
Information Combine Analysis and Process System (MICAPS) developed by the China
Meteorological Administration (CMA). The system has been developed to the fourth
edition and has begun to support the storage and access of massive real-time data. But
the fusion of real-time data and historical data, as well as the big data analysis technical
integration, still cannot satisfy the modern MERWF. Its existing client/server infras-
tructure is also not conducive to effectively integrating the multitudinous supporting
technologies and products developed in recent years. There is very little time for
forecasters to use big data analysis and fine-tuning. At the same time, the current
MERWF is still challenged by the limitations of basic theory, model level,
Supported by the National Science and Technology Support Program of China (2015BAC03B07).
predictability, etc., which has become a hindrance for forecasters to efficiently correct
and integrate multi-source forecast information. It makes forecasters unable to get more
forecast information and conduct sufficient data mining in a limited time and has
seriously affected the efficiency of work. Therefore, according to the needs of business
development, it is urgent to build a new generation operational system which could
support the integrated business processes and big data analysis, and this is exactly what
this paper will introduce.
The MERWF operational system is a specialized big data application system built on
the MICAPS4 network platform [3]. It adopts the browser/server system architecture
and chiefly includes three parts: basic environment of big data, big data processing
system and front-end interactive analysis display (Fig. 1).
database management system (MySQL) is used to satisfy this requirement. The system
builds comprehensively distributed database architecture by introducing a variety of
databases, which basically meets the storage and access requirements of massively
concurrent system.
Compared
Filter Filter Sort
Forecast
Eigenvalue1
Eigenvalue1 Eigenvalue1 Eigenvalue1
Index Eigenvalue2 Index
Eigenvalue2
Eigenvalue3
Index 3 Eigenvalue2 ... Index n Eigenvalue2
Eigenvalue3 Eigenvalue3 ...
Eigenvalue4
(a) (b)
(c) (d)
Fig. 3. (a) Observational precipitation on June 23th, 2012, and the corresponding
(b) SVM_MEF forecast, (c) EC_C forecast, and (d) EC_M forecast [19].
Medium-Extended-Range Weather Forecast 531
compared with control forecast (EC_C) and ensemble average forecast (EC_M) in the
European Centre for Medium-Range Weather Forecasts, forecast effect of SVM_MEF
is significantly better.
The browser of MERWF operational system includes multi-source big data, subjective
and objective comprehensive analysis and forecast, service and evaluation, etc. It is
divided into 8 functional modules: historical background, monitoring, numerical model
interpretation, predictability analysis, objective forecasting, business evaluate, product
production, and customer service (Fig. 4). Each module has a unified application logic
relationship, which is consistent and reflects the professional technical process of
MERWF services.
product production
customer service
Multi-source Comprehensive
big data Big data analysis analysis and
historical and application forecast
background predictable analysis
monitoring objective forecast
numerical model business evaluate
interpretation
The visual user interface contains three parts: the functional area, the attribute area,
and the display area. The function area is located at the top of the page, including the
system function main menu and sub menu. The attribute area is situated on the left side of
the page, and is flat-designed for different modules, including data attribute character-
istics such as mode type, meteorological element, geographic range, etc. The display area
532 Y. Li et al.
is located in the right side of the page, which displays product information such as
graphics, text, and corresponding attribute operations.
The MERWF usually need to display multi-dimensional data comprehensively. The
visual products with high business value should be produced through analysis,
reconstruction, and information extraction. Products often contain a large amount of
data such as historical data, real-time data, forecast data and inspection data, which can
reflect the changing trend, abnormal signals and predictability of the weather system.
The MERWF operational system uses distributed data storage and distributed com-
puting technology, and significantly improves the application capability and access
efficiency on multi-type, high-dimensional, and complex-related meteorological data.
The system has realized efficient integration of massive real-time data and historical
data, as well as the integration of big data with forecasting techniques. The specialized
big data application system of MERWF is initially constructed. Furthermore, the
MERWF business process which integrates monitoring, forecast and service is also
established.
The system has been applied in the national forecast business department and some
provincial meteorological business departments, which has shown good versatility and
stability. However, big data applications are still focused on big data mining within the
meteorological industry. In the future, it will gradually expand to cross-industry data
and form a broader data application platform. We will focus on such intelligent
development directions, including forecast technology integration, efficient function
expansion, improving on system performance, personalized service and convenient
management. With the continuous development and improvement of the system, it
could efficiently support the development of national MERWF services.
References
1. Wang, R., Huang, X., Zhang, B., et al.: Design and implementation of a real time parsing
and storage system for massive meteorological data. Comput. Eng. Sci. 37(11), 58–64
(2015)
2. Wang, R., Wang, J., Huang, X., et al.: The architecture design of MICAPS4 server system.
J. Appl. Meteorol. Sci. 29(1), 1–12 (2018)
3. Hu, Z., Gao, S., Xue, F., et al.: Design and implementation of MICAPS4 web platform.
J. Appl. Meteorol. Sci. 29(1), 45–56 (2018)
4. STORM Documentation 1.1.0. http://storm.apache.org. Accessed 30 June 2016
5. Sun, D.W., Zhang, G.Y., Zheng, W.M.: Big data stream computing: technologies and
instances. J. Softw. 25(4), 839–862 (2014)
6. Lang, H., Ren, Y.: a fast search algorithm based on redis emory database. Comput. Appl.
Softw. 33(5), 40–43 (2016)
7. Wu, L., Zhang, F.: A study on WebGIS client based on HTML canvas. Geomat. World 7(3),
78–82 (2009)
Medium-Extended-Range Weather Forecast 533
8. Li, Y., Jin, R., Zhou, N., et al.: An analysis on characteristics of heavy rainfall processes
during the Meiyu season in Jianghuai region. Acta Meteorologica Sinica 75(5), 717–728
(2017)
9. Li, Y., Jin, R., Lin, Y.: The forecast of dekad temperature abnormal of china in winter.
Meteorol. Mon. 25(7), 41–45 (1999)
10. Huang, W., Zhang, H., Bao, Y., et al.: Objective forecast of medium-term average
temperature anomalies based on EOF technique. J. Meteorol. Sci. 37(4), 461–566 (2017)
11. Zhang, F., Cao, Y., Xu, J., et al.: Application of the logistic discriminant model in heavy rain
forecasting. Meteorol. Mon. 42(4), 398–405 (2016)
12. Dai, K., Cao, Y., Qian, Q., et al.: Situation and tendency of operational technologies in short-
and medium-range weather forecast. Meteorol. Mon. 42(12), 1445–1455 (2016)
13. Chen, B., Dai, K., Guo, Y., et al.: Precipitation verification and analysis of ECMWF
ensemble statistic products in 2013 flooding season. Torrential Rain Disasters 34(1), 64–73
(2015)
14. Du, J., Chen, J.: The corner stone in facilitating the transition from deterministic to
probabilistic forecasts-ensemble forecasting and its impact on numerical weather prediction.
Meteorol. Mon. 36(11), 1–11 (2010)
15. Yang, Q.: Prospects and progresses in the research of the methods for 10–30 days extended-
range weather forecast. Adv. Earth Sci. 30(9), 970–984 (2015)
16. He, W., Kong, M., Zhao, H.: Research on meteorological prediction with Bayesian classifier.
Comput. Eng. Des. 28(15), 3780–3782 (2007)
17. Ma, X.-K., Cai, X.-N., Yang, G.-M., et al.: Study on fog synoptic characteristics and fog
forecast method in Chongqing. Clim. Environ. Res. 12(6), 795–803 (2007)
18. Huang, W., Niu, R.: The medium-term multi-model integration forecast experimentation for
heavy rain based on support vector machine. Meteorol. Mon. 43(9), 1130–1137 (2017)
19. Roushangar, K., Alizadeh, F.: A multiscale spatio-temporal framework to regionalize annual
precipitation using k-means and self-organizing map technique. J. Mt. Sci. 15(7), 1481–1497
(2018)
Application Research of Big Data in Heavy
Rainfall Forecast Model in Meiyu Season
1 Introduction
Meiyu rainfall generally occurs over the Yangtze-Huaihe River Valley (YHRV) from
the beginning of June to July each year. Therefore, it is also known as the Jianghuai
Meiyu which accomplished by frequent heavy rain, large precipitation and concentrated
rainy days. According to statistics, the precipitation of the Jianghuai Meiyu accounts for
about 45% of total summer rainfall. Studies have shown that the Meiyu rain belt usually
shows a quasi-zonal direction around 30°N in eastern China, presenting a long strip
shape [1, 2], while it shows an east-northeast direction in the Korean Peninsula and
Japan [3, 4]. Therefore, the location and residence time of the Meiyu rain belt often
determine the spatial distribution variability of precipitation over the YHRV. In addi-
tion, due to the narrow shape of the YHRV, the location of the rain belt is often the key
and also the difficulty in the weather and climate forecast operation. Previous studies
have paid great attention to the Meiyu rain belt. Some discussed the causes of the rain
belt from the perspective of Meiyu front structural characteristics [3]. Some studied the
variation characteristics of rain belt location, range and intensity on inter-annual and
inter-decadal time scales [5]. Besides, some other researchers discussed the factors
affecting the location of the rain belt, such as the winter sea temperature and the various
subsystems of the summer monsoon circulation [6, 7]. The relationship between the
westerly jet and Meiyu rainfall shows that when the westerly jet is strong (weak), the
rain belt locates further south (north) [8]. In the short-range and medium-range (1–10
days) forecasts, forecasters often refer to the 500 hPa 5840 geopotential meter isoline
when considering the Meiyu rain belt location over YHRV, because the isoline usually
corresponds to the Meiyu front location [9].
From June 30th to July 4th, 2016, the strongest precipitation process occurred since
the beginning of the Meiyu season this year. The middle and lower reaches of Yangtze
River MLYR suffered heavy rainfall with an extreme precipitation of 800–948 mm
observed in some areas of Huanggang, Hubei Province. It is shown by post analyses
that compared with observations, the rain belt intensity in short-range and medium-
range subjective and objective forecast is obvious weaker, and the east part of the rain
belt is forecasted to be in further north area. The updating forecasts all indicated that the
main rain belt would center in MLYR, the Huaihe Valley and the Huanghuai area. In
observations, the heavy rainfall located along the Yangtze River and presented a typical
quasi-east-west Meiyu rain belt. However, both the numerical models and the sub-
jective forecast indicated that the heavy rainfall center would be near the Huaihe River
(figure omitted), where was about one latitude further north than the observations.
According to the forecast, the flood control in the Huaihe Basin would be quite
challenging because of high population density and lack of the estuary. Therefore, the
impact of the error in the rain belt location forecast is great, although it is only one
latitude further north. In view of this, it is necessary to clarify the reasons for the rain
belt location forecast error. In the following, the big data approach will be used for the
study of the aforementioned issue.
The grid precipitation data provided by the National Meteorological Information Center
(from early June to the middle of July during 1980–2009), is used to analyze the distri-
bution of pentad rainfall in Meiyu seasons over the past 30 years. 33 precipitation pro-
cesses with quasi-east-west rain belt are selected. At the same time, the NCEP/NCAR
reanalysis data [10] is used to analyze the atmospheric circulation during the same period.
To explore the mechanism of the precipitation process from June 30th to July 4th, 2016,
both ECMWF ensemble forecast data from June 21st to July 4th, 2016 and the daily
rainfall data observed at 20 o’clock every day are used. Here we defined the position of the
zonal-mean precipitation maximum in the middle-lower Yangtze River reaches from
110° to 120°E as the location of the rain belt for the period from June 30th to July 4th.
536 S. Yin et al.
Fig. 1. (a) The average precipitation distribution of 33 precipitation processes; (b) the zonal
average of the precipitation in Figure (a).
Fig. 2. The variation characteristics of rain belt locations of 33 precipitation processes. The
histogram shows the rain belt location corresponding to the right coordinate, with its average
value denoted by the red dotted line; the short-dotted line with + is the normalized series
corresponding to the left coordinate.
Application Research of Big Data in Heavy Rainfall Forecast Model 537
Based on previous studies and forecasters’ experience, it is considered that the location
of 200 hPa westerly jet, South Asia High, 500 hPa 5840 geopotential meter isoline,
850 hPa and 700 hPa shear line, as well as the intensity of West Pacific subtropical
high (WPSH) are all related to the rain belt location. So, is there still a close rela-
tionship between the precipitation process in quasi-zonal rain belt and these factors?
The correlation coefficients between the rain belt location and 200 hPa maximum wind
speed center, the location of its corresponding zero wind speed line, and the 5840
geopotential meter isoline are calculated. The corresponding correlation coefficients are
0.49, 0.63 and 0.67 respectively, and reach the 99%, 99.9% and 99.99% significance
levels respectively. Therefore, the above high-level circulation systems have significant
influences on the location of the typical quasi-east-west Meiyu rain belt.
Fig. 3. The comparisons between rain belt locations and meridional wind zero isotaches at
(a) 500 hPa, (b) 600 hPa, (c) 700 hPa, and (d) 850 hPa during 33 precipitation processes.
The rain belt locations and shear lines at different levels during 33 precipitation
processes are further compared and analyzed (Fig. 3). Here we use the location of the
zero isotach of the meridional wind as the shear line. The results show that the 850 hPa
shear line was often absent, indicating that the rainfall was mainly warm area pre-
cipitation without cold air in low level. Thus the distinct wind shear was hard to find.
Although similar situations could be detected in the middle and upper levels, shear
lines become more and more obvious with the rise of height. For example, distinct
shear lines can be found easily at 700 hPa in the first 26 cases, and they are found in
almost all the cases at 500 hPa. On one hand, it is revealed that the system in high level
is often more stable than the lower one. On the other, it shows that the cold air is weak
in most cases and not obvious in low levels. In general, the shear line does determine
the location of the rain belt in some cases. At 700 hPa to 500 hPa, the correlation
coefficients between shear lines and the rain belt locations all exceed 0.52. However,
538 S. Yin et al.
this relationship is not stable, especially for 850 hPa – the reference level often used in
the operational forecast of central and eastern China.
In the past forecasts, many front-line forecasters often take the location of low level
(850 hPa and 700 hPa) shear line and middle level (500 hPa) 5840 geopotential meter
isoline into serious consideration when forecasting the rain belt location during heavy
rainfall. When it comes to the medium-range forecast, most of them only focus on the
location of the 5840 geopotential meter isoline. It is revealed that with the lengthening
of the forecasting time, the middle and high level systems and their relationships with
the rain belt locations are generally more stable than that of the low-level ones.
5 Case Analysis
The Meiyu season in 2016 was from June 19th to July 17th. Six heavy precipitation
processes occurred during this period, of which the one from June 30th to July 4th was
the strongest and most influential. However, verification shows that serious errors
appeared in the forecasting rain belt location and heavy rain center by both numerical
models and subjective forecasts.
The average circulation situation during the precipitation from June 30th to July 4th
reflects that cooperating with WPSH, the summer monsoon extended from the north-
west side of the WPSH to the Jianghuai and Huanghuai areas (figure omitted). A wide
low trough controlled the mid-high latitudes of Asia. Blocking highs were observed in
both of the Ural Mountains on the west side of the low trough and the Okhotsk Sea on
the east side. This circulation pattern is often referred to as “two ridges and one trough”
in daily forecast operation. It is one of the typical circulation patterns in the Meiyu
season, indicating that the circulation situation is relatively continuous and stable. This
precipitation process lasted for 5 days with the cumulative rainfall of 100–250 mm in
MLYR, some areas reaching 300–400 mm, and parts of areas exceeding 600 mm. It
increased the severity of flood control in the Yangtze River Basin.
The medium-range subjective forecast (5 days earlier before the end of the process)
showed that the heavy rainfall center of the precipitation process would be along the
Huaihe River, which is obviously one latitude further north than the observation.
Referring to the experiences using the characteristic line (5840geopotential meter
isoline) commonly in the forecast, the sliding correlation between the rain belt location
and 5840 geopotential meter isoline during the Meiyu season is analyzed (Fig. 4). It
can be seen from Fig. 4a and b that there was a certain correspondence between the rain
belt location variation and the correlation coefficient trend before June 30th. The
correlation coefficient increased (decreased) when the rain belt moved to the north
(south). When it comes to July 1st, the coefficient between them decreased rapidly,
revealing that the 5840 geopotential meter isoline could no longer indicate the rain belt
location, and the 500 hPa circulation system could not dominate the rain belt location
any more, like what it had done in several previous precipitation processes.
To find out why the rain belt location could no longer be forecasted by 5840
geopotential meter isoline, the average rainfall in MLYR (26–32°N, 115–120°E) is
defined as the rainfall intensity index in the key area, and the correlation between the
Application Research of Big Data in Heavy Rainfall Forecast Model 539
Fig. 4. (a) The sliding correlation coefficient of the rain belt location and 5840 geopotential
meter isoline from June 18th to July 4th, 2016; (b) 110°–120°E zonal average precipitation.
Fig. 5. The correlation coefficients between the rainfall intensity forecasted on the 30th and
height field at (a) 500 hPa, (b) 700 hPa, (c) 850 hPa and (d) 925 hPa. The shadow denotes the
areas passing 90% confidence level, and the isoline is the average 500 hPa geopotential height
during the precipitation process.
rainfall intensity index based on the ensemble forecast and the 500–850 hPa geopo-
tential height is analyzed (Fig. 5). Figure 5a shows that the correlation coefficient in
MLYR does not pass the 90% confidence level test, indicating that there is no obvious
relationship between the rainfall intensity and 500 hPa geopotential height, or the 5840
geopotential meter isoline. However, in the lower levels (700–925 hPa), the correlation
is significant with the coefficient passing the 90% confidence level test, revealing that
the heavy rainfall or location variation during the precipitation process is only closely
related to the circulation below 700 hPa, while the influence of the higher layer
(500 hPa) circulation is relatively weak.
Further analysis also shows that the relevant significant regions from 700 hPa to
925 hPa in Fig. 5a–c are located on the northwest side of WPSH, i.e. the active
southwest monsoon region in the low level, indicating that the southwest monsoon has
540 S. Yin et al.
played a significant role in the intensity variation of this precipitation process. Figure 6
shows a vertical section along 115–120°E, from which it can be seen that the northerly
wind (cold air from the north) generally influences the area to the north of 35°N and
slopes northward with the rise of height during this rainfall. Therefore, the cold air was
far from the heavy rainfall center, and its contribution was insignificant because it was a
warm area precipitation in the monsoon region. It is quite different from the average
circulation pattern of 33 processes in Fig. 6b. In 33 classic Meiyu precipitation pro-
cesses, around the strong rainfall center (30–31°N), there were both low-level shear
lines (the interface of east and west wind) and middle-low-level (850–500 hPa) cold
and warm air confluence (the interface of south and north wind) (Fig. 6b). From June
30th to July 4th, no northerly cold air participated in the precipitation process, resulting
in deviations between the actual rain belt location and the classic forecast model results
(or 33 process averages). In the classic forecast model, the rain belt central location
(5840 geopotential meter isoline) is generally in the cold and warm air convergence
zone accompanied by a strong convergence rise, which is conducive to heavy rain
(Sugimoto et al. 2003).
Fig. 6. (a) The vertical section of 115–120°E averaged vectorial wind field, the zonal wind (red
contour), meridional wind (shaded) in the middle and low levels from June 30th to July 4th,
2016; (b) the same as (a) except for 33 historical cases; (c) the heavy rainfall center variation of
33 historical cases.
6 Conclusion
In this paper, the big data analysis method is used to explore the variation character-
istics of the rain belt location during the typical Meiyu season in the past 30 years. The
results show that the rain belt location is closely related to the 500hpa 5840 geopo-
tential meter isoline, which is in line with the forecast operation experience. However,
errors appear in the forecast of the precipitation process from June 30th to July 4th,
2016 by the classic forecast model, indicating that there are still some limitations in it.
Application Research of Big Data in Heavy Rainfall Forecast Model 541
The results of this paper show that the northerly cold air only appeared in the middle
and upper troposphere during the heavy precipitation process from June 30th to July
4th, 2016. There was a certain distance between the cold air and the rain belt center,
that is, the heavy precipitation process was not caused by the confluence of cold and
warm air. Considering the location of WPSH, it can be seen that this precipitation
process occurred in the warm area of the monsoon region, which was caused by the
convergence of warm air. Under this formation mechanism, the 5840 geopotential
meter isoline of the classic forecast model loses its forecasting ability. Instead, cold air
participation in the precipitation process is the key factor determining the accuracy of
the forecast model. On one hand, this conclusion is conducive to enriching and
deepening the forecaster’s understanding of the rain belt location variation during the
Meiyu season. On the other, it provides important reference for forecast correction.
There are still many problems in the paper that need further study. For example, why
did the numerical model fail in the rain process forecast? What are the reasons for the
difference in the performance of different numerical models? These will be discussed in
the future work.
Acknowledgement. This work was supported by grants of the National Natural Science
Foundation of China (41575066) and the National Key Technology Research and Development
Program of the Ministry of Science and Technology of China (2015BAC03B04,
2015BAC03B06 & 2015BAC03B07).
References
1. Si, D., Ding, Y.H., Liu, Y.J.: Decadal northward shift of the Meiyu belt and the possible
cause. Chin. Sci. Bull. 55(1), 68–73 (2010)
2. Xu, W.G., Jiang, J.: Characteristics of the rain belt of Meiyu between inter-annual and inter-
decadal climate variations. J. Nanjing Univ. 40(3), 292–303 (2004)
3. Ding, Y.H., Liu, J.J., Sun, Y., et al.: A Study of the Synoptic-Climatology of the Meiyu
System in East Asia. Chin. J. Atmos. Sci. 31(6), 1082–1101 (2007)
4. Oh, J.H., Kwon, W.T., Ryoo, S.B., et al.: Review of the research on Changma and future
observational study (KORMEX). Adv. Atmos. Sci. 14(2), 207–222 (1997)
5. Wei, F.Y., Xie, Y.: Interannual and interdecadal oscillations of meiyu over the middle-lower
reaches of the changjiang river for 1885–2000. Q. J. Appl. Meteorol. 16(4), 492–499 (2005)
6. Zong, H.F., Zhang, Q.Y., Chen, L.T.: Temporal and spatial variations of precipitation in
Eastern China during the Meiyu period and their relationships with circulation and sea
surface temperature. Chin. J. Atmos. Sci. 30(6), 1189–1197 (2006)
7. Su, T.H., Xue, F.: The intraseasonal variation of summer monsoon circulation and rainfall in
East Asia. Chin. J. Atmos. Sci. 34(3), 611–628 (2010)
8. Sampe, T., Xie, S.P.: Large-scale dynamics of the Meiyu-baiu rainband: environmental
forcing by the westerly jet. J. Clim. 23, 113–134 (2010)
9. Sugimoto, S., Hirakuchi H.: Simulation of precipitation caused by a Baiu front: an
evaluation study with radar data. In: Weather Radar Information and Distributed
Hydrological Modelling, vol. 282, pp. 51–58. IAHS Publication (2003)
10. Kalnay, E., Kanamitsu, M., Kistler, R., et al.: The NCEP/NCAR 40-year reanalysis project.
Bull. Am. Meteorol. Soc. 77, 437–471 (1996)
Predictability of an Extreme Rainfall Event
in North China
1 Introduction
From July 18th to 20th, 2016, the Sichuan Basin, the area from Jianghan to North
China, and the Huanghuai area suffered from the strongest rainfall process since the
flood season in 2016. The rainfall in 22 counties and cities such as Daxing of Beijing
City and Jingjing of Hebei Province broke the historical records. The heavy rainfall
also caused disasters such as mountain torrents, landslides, debris flow and urban
waterlogging in provinces like Hubei, Hunan, Hebei, Shanxi and Guizhou, resulting in
many casualties. The heavy rainfall event is characterized by two remarkable features.
First, the rain area jumped rapidly from the Yangtze River Basin to North China, which
connected two rainy seasons – the Jianghuai Meiyu season and North China rainy
season. It can be seen that the rain belt moved drastically on the 18th and the 19th,
moving from 28°N (the Yangtze River Basin) to the vicinity of 38°N (North China),
crossing 10 latitudes northward within two days. The rainfall intensity also increased
significantly on the 19th, reaching the magnitude of rainstorm. In comparison, the
change of the rainfall zone showed a slow v-shaped variation in the east-west direction,
extending from east to west from the 16th to 19th. After reaching the largest magnitude
on the 19th, the rain zone began to move back from west to east (Fig. 1c). Another
feature is that on the initial stage, large errors appeared in the forecasts from both the
ECMWF numerical model (Fig. 1d), which generally performed perfectly, and the
subjective forecast (figure omitted). Figure 1d shows the forecast of the rain belt
location from 1200 UTC July 19th to 1200 UTC July 20th (hereinafter referred to as
the 20th) from different initial forecast times. It can be seen that before the 16th, all the
forecasts indicated that the rain belt was within the area from the middle and lower
reaches of the Yangtze River to the Huanghuai area (30–35°N), and the average rainfall
intensity was weak. After that, the forecasted rain belt moved to around 40°N,
revealing that the ECMWF deterministic model gave the information 96 h in advance
that the rain belt would jump northward, while it failed to grasp this feature before that
time point.
Fig. 1. (a) Distribution of cumulative precipitation in North China and other places from July
18th to 20th, 2016. (b) Time series of zonal averaged precipitation in 110–120°E from July 16th
to 22th. (c) Time series of meridional averaged precipitation in 26–42°N from July 16th to 22th.
(d) Rain belt location forecast on July 20th by ECMWF deterministic model initialized at 1200
UTC from July 10th to 19th.
Previous studies have pointed out that the rain belt jumps northward two times and
stays in three quasi-stationary phases every year in China. And it moves northward
periodically [1–4]. Among them, the third stagnation period from mid-July to mid-
August represents the beginning of the rainy season in North China and Northeast
China. During this heavy rainfall process, the rain belt jumped from the Yangtze River
Basin to North China, indicating the end of the Jianghuai Meiyu and the beginning of
the rainy season in North China. However, in most medium-range forecasts, subjective
544 J. Ma et al.
or objective, there are obvious errors, especially in the forecast of the rain belt location.
So, what causes these forecast errors? How long is the predictability of this extreme
rainfall process by numerical model? In this paper, to solve the above problems, the
causes of the dramatic rain belt location variation in a short period of time and its
mechanisms are analyzed. Besides, the big datasets of deterministic forecast and
ensemble forecast are used to discuss the predictability of the rainfall process in the
medium-range forecast.
It is generally believed that the difference in the initial fields is an important reason for
forecast errors of the numerical model. Ensemble numerical forecast provides an
effective way to correct initial bias and model errors. Compared with the deterministic
forecast, ensemble numerical forecast can improve the effectiveness and reliability of
the forecast by providing probabilistic prediction involving multiple members or
possible scenarios through characterizing and describing the uncertainty of initial filed
[5–7]. Some members of the ensemble forecast may capture some “missing” infor-
mation to help forecasters, decision makers and the public make better decisions. Big
data set of ensemble forecast has also been widely used in recent years due to its great
application value in operational weather forecast and forecast theory research. There-
fore, forecasting and analysis field in ECMWF deterministic model (spatial resolution:
2.5° 2.5°, forecasting time: 240 h), and the big data set from ECMWF ensemble
forecast model (spatial resolution: 0.5° 0.5°, forecasting time: 360 h with 51
members) are used to analyze the atmospheric circulation and the predictability of this
heavy rainfall process by statistical methods such as empirical orthogonal function
(EOF), correlation analysis and composite analysis.
In addition, the NCEP/NCAR rainfall reanalysis data with a spatial resolution of
0.5° 0.5° is selected to verify the forecast results. The location of blocking high is
quantified as a blocking high north boundary position index, defined as the averaged
latitude of the 500 hPa 5880 gpm isoline in 120–130°E. The rainstorm north boundary
position index is defined as the zonal averaged 50 mm precipitation north boundary
position in 110–120°E.
Through the analyses of circulation background, it can be seen that the low vortex
system of the lower layer (850 hPa), the monsoon and West Pacific Subtropical High
(WPSH) are the main circulation systems affecting the rainfall process (figure omitted).
The low vortex system gradually moved northward when it moved eastward along the
edge of WPSH, so the main rainfall area was obviously related to the WPSH north
boundary position. Figure 2a shows the changes of WPSH from July 18th to 20th. It
can be seen that with the eastward movement of the low vortex system in the middle
and lower reaches of the Yangtze River, WPSH began to extend westward and move
Predictability of an Extreme Rainfall Event in North China 545
northward. Its western ridge point extended from 125°E to 110°E, and the 5880 gpm
isoline above the 120–130°E region (often used to define the north boundary of WPSH)
also jumped from 30°N to 40°N. Near the Bohai Sea and the East China Sea, a high
pressure ridge was formed, which slowed down the low vortex moving eastward and
northward. Therefore, the rainfall was prolonged and the accumulated rainfall increased
obviously in North China, the Huanghuai area and the nearby areas.
The heaviest rainfall in North China during this rainfall process is on July 20th, 2016,
and North China is the region with the greatest error in precipitation forecasting. Thus, the
region of 38–42°N and 115–120°E is selected as the key area for the following analyses.
The correlation between the average rainfall intensity in the key area and 500 hPa
geopotential height field is calculated (Fig. 2b). Two systems affecting the variation of the
rainfall intensity in the key area are found: the blocking high in 120–130°E and the low
vortex in 110–120°E, both of which have passed the 99.9% confidence test. It also
confirms that the strong blocking high is favorable for heavy rainfall in North China, and
the effect of the low vortex is reflected in the inverted trough on the north side.
Fig. 2. (a) The variations of the WPSH position from July 18th to 20th. The black, red, and
green lines represent the 18th, 19th and 20th respectively. (b) Distribution of the correlation
coefficients between the average rainfall intensity in the key area (38–42°N, 115–120°E) and
500 hPa geopotential height field (shaded means the passing of 99.9% confidence test) from
ECMWF ensemble forecast initialized at 1200 UTC July 16th. The contour is 500 hPa
geopotential height on the 20th.
In addition, from the composite analysis of 850 hPa wind field based on the
blocking high north boundary position index, it can be seen that when the high located
to the north of its normal position, a cyclonic anomalous wind field was detected in
North China (Fig. 3a), increasing the low vortex intensity (Fig. 3d). It was favorable
for the enhancement of the southwest monsoon and pushed the low-level jet northward
(Fig. 3a and c), and vice versa (Fig. 3b). The comparison of the north boundary
position index series of blocking high and the rainstorm shows that their variation
trends were consistent, and their correlation coefficient reached 0.61 (passed the 99.9%
confidence test). It is revealed that in 110–120°E, the average rain area location and
intensity variation were obviously affected by the geopotential height field of the region
(30–45°N, 120–130°E).
546 J. Ma et al.
Fig. 3. Composite analyses of 850 hPa wind field based on blocking high north boundary
position index on July 20th from ECMWF ensemble forecast initialized at 1200 UTC July 16th
(shaded represents the passing of 99.9% confidence test), for (a) high index members, (b) low
index members, (c) difference between high and low index members, and (d) ensemble mean
wind anomalies.
As the position variation of the east section of WPSH has a significant impact on the
rain belt location during the heavy rainfall, it is necessary to analyze the causes for the
deviation of the forecasted rain belt location from the geopotential height field of 120–
130°E. Is the failure of predicting the variation characteristics of WPSH in 120–130°E
the reason for the omission of the sudden northward jump of the rain belt?
First, the ECMWF deterministic forecasts are verified. Figure 4 shows the ECMWF
deterministic forecasts of WPSH on July 20th, which were initialized on July 13th to
19th. It can be seen that the variation of WPSH north boundary in 110–120°E was
obviously smaller than that in 120–130°E, and it was very close to the observations.
The results show that the subjective and objective forecasts of the western part of the
rain belt (110–115°E) were closer to the observations. However, the subjective and
objective prediction errors in the eastern part of the rain belt (115–120°E) were
obvious, which might be caused by the great impact of the WPSH in 120–130°E on the
east part of the rain belt. From July 13th to 19th, with the approaching of time, the
range of WPSH near 120–130°E predicted by the model continuously adjusted,
jumping from 32°N to 40°N. The model shows great uncertainty for the prediction of
WPSH in this region (Fig. 4a). In Fig. 4a, the WPSH on 20th predicted by the model
initialized on July 13th to 15th was in the south thus the moving path and influence
range of the low vortex were both in the south, and the overall rainfall area and the
strong rainfall center were in the south of North China. The forecasted WPSH on 20th
Predictability of an Extreme Rainfall Event in North China 547
initialized on July 16th to 18th began to adjust toward the “blocking situation”, so the
predicted rainfall area gradually moved northward with an enhanced intensity.
Therefore, from the perspective of forecasting time, only the ECMWF deterministic
forecast initialized on July 16th has reference values. It can be seen that the rainfall in
key areas on 20th (North China and nearby areas) initialized on July 16th was sig-
nificantly enhanced when compared with that initialized on July 12th–15th. With the
previous analyses of the rain belt’s northward jumping, it can be seen that the maxi-
mum forecast leading time of ECMWF deterministic model is only 96 h for the rain
belt location.
Fig. 4. (a) The ECMWF deterministic forecast of WPSH on the 20th initialized at 1200 UTC
every day from July 13th to 19th, and the observations on the 20th. The contour number is the
corresponding starting date; (b) the rainfall intensity forecast for key areas on July 20th by
ECMWF deterministic model initialized at 1200 UTC every day from July 12th to 17th.
Based on the empirical orthogonal function (EOF) analysis of rainfall field on July
20th initialized at 1200 UTC on July 12th, the main distribution mode presented
northeast-southwest direction (the first principal component with an interpretation
variance of 21.9%) (Fig. 5a), and the main rainfall centers were located in the
Huanghuai and Jianghuai areas. It represented the forecast trend of most members, that
is, the rain belt moved northward while moving eastward, but most of its north
boundary was within the Huanghuai area (represented by member 1 and 4). The
correlation analysis between the time coefficient of the first principal component and
the 500 hPa height field show that there is a significant correlation between the main
rainfall mode and the 500 hPa geopotential height field in the region (120–130°E, 35–
45°N), indicating that the variation of the 500 hPa height field has a significant
influence on the rain belt distribution. The results are consistent with our analysis of the
circulation background field. Figure 5b shows another rain distribution type, in which
the rain belt was carried northward to the north of Huanghuai-Jianghuai area, while the
middle and lower reaches of the Yangtze River were negative, indicating that the
rainfall was less and the Meiyu season tended to end. It corresponds to members of
2/6/14/24. The third mode explanation variance is small and it is unnecessary to be
analyzed here.
We also analyze the distribution characteristics of 500 hPa geopotential height
fields and the amount of precipitation on July 20th when the time coefficients of the
548 J. Ma et al.
Fig. 5. EOF analysis of the precipitation on 20th from the ECMWF ensemble forecast
initialized at 1200 UTC July 12th, for (a) the first principal component and (b) the second
principal component. (c) The distribution of correlation coefficients between the time coefficients
of the first principal component and the synchronous 500 hPa height field.
above modes are greater than 1.5 standard deviations (figure omitted). It is found that
when the first and second modes are strong, the distribution of rain areas is similar
except for the north boundary of rainstorm area. The second mode rain area is located
in the further northeast area. It indicates that the heavy rainfall center moves faster, and
its main body would move eastward into the sea. But the heavy rainfall center in the
first mode would continue to move northward. It also explains why the first mode
presents an east-west dipole distribution. The difference between the two modes might
be attributed to the development of blocking high over 120–130°E in the east side of
the rain area. Although the blocking high situations are presented in both the first and
second modes, the intensities of these two are different. It shows that the strong rainfall
area is more sensitive to the change of blocking high intensity. In comparison, the rain
belt forecasted by ECMWF deterministic model is still located in the Yangtze River
valley on July 20th, indicating that the forecast error from deterministic model is
greater.
Figure 6 shows the first two modes of the forecasted rainfall on the 20th initialized
every day at 1200 UTC from July 13th to 16th, 2016. With the approaching of leading
time, the rain belt of the first mode (Fig. 6a, c, e, g) and its north boundary continu-
ously move to the northeast, indicating the northward movement of the forecasted rain
belt, which is closer to the observations. The explained variance of the first mode varies
little with time, and the significant correlation region between the first mode time
coefficient and the height field is similar to that of Fig. 5c, indicating that the ensemble
forecast is stable for the first mode. Compared with the first mode, the horizontal
distributions of the second mode change obviously with the approaching of leading
time (Fig. 6 b, d, f, h). In the forecasts initialized on the 13th and 14th, the second
mode is mainly located in the north part of the Jianghuai and the Huanghuai areas. In
the forecasts initialized on the 15th and 16th, the main rain areas in the second mode
jumped northward to North China. Therefore, the ensemble forecast initialized on the
Predictability of an Extreme Rainfall Event in North China 549
15th can be used as a reference as it reveals that the rain belt would jump northward to
North China and nearby areas during this rainfall process.
Fig. 6. The first mode (a, c, e, g) and the second mode (b, d, f, h) of EOF based on the 20th
rainfall forecasted by ECMWF ensemble model initialized at 1200 UTC July 13th (a and b), 14th
(c and d), 15th (e and f) and 16th (g and h). The explained variance is given in the lower right
corner.
In conclusion, from the point of view of forecast leading time, compared with the
deterministic forecasts (started from the 16th), the ensemble forecasts (started from the
15th) could give a more reasonable forecast 24 h in advance, so it has a significant
advantage in forecasting. The analyses of the first two modes of EOF reveal the
uncertainties of rain belt location prediction and the causes, which is of important
reference value for improving the medium-range forecast.
5 Conclusions
In this paper, the predictability of medium-range forecast for an extreme rainfall event
from July 18th to 20th, 2016 is discussed. In this event, both subjective and objective
forecasts show great errors on the forecast of the north boundary position of the rain
belt. To find out the reason, the atmospheric circulation factors influencing the rain belt
location variation are firstly analyzed. The results show that the intensity variation of
blocking high in 120–130°E greatly affects the rain belt location in 110–120°E. When
the blocking high is stronger, the average 5880 gpm isoline in 120–130°E is in a further
north area, so is the rain belt in 110–120°E. Then, the rainfall intensity in North China
is also higher.
The further study of atmospheric circulation shows that when the blocking high is
stronger, the low vortex system in North China is stronger, and the southwest airflow
between them is obviously stronger, which is favorable for the northward transport of
water vapor. Therefore, it leads to the more north location of rain belt in the East China
and the strong rainfall in North China. Based on the characteristics of heavy rainfall
550 J. Ma et al.
and the forecast of the above-mentioned atmospheric circulation factors, the pre-
dictability of extreme rainfall in North China is discussed by big data set of ECMWF
ensemble and deterministic model forecasts. The results show that the maximum
forecast leading time for the rain belt location by deterministic model is 96 h, while it
can reach 120 h by ensemble model.
Acknowledgement. This work was supported by grants of the National Natural Science
Foundation of China (41575066) and the National Key Technology Research and Development
Program of the Ministry of Science and Technology of China (2015BAC03B04,
2015BAC03B06 & 2015BAC03B07).
References
1. Tao, S.Y.: Torrential Rain in China. Science Press, Beijing (1980)
2. Guo, Q.Y., Wang, J.Q.: Interannual variations of rain spell during predominant summer
monsoon over China for recent thirty years. Acta Geogr. Sin. 36(2), 187–195 (1981)
3. Ding, Y.H.: Summer monsoon rainfalls in China. Q. J. R. Meteorol. Soc. 70, 373–396 (1992)
4. Ding, Y.H., Li, C.Y., He, J.H., et al.: South China Sea Monsoon Experiment (SCSMEX) and
the East Asian monsoon. Acta Meteorologica Sin. 62(5), 561–586 (2004)
5. Matsueda, M., Tanaka, H.L.: Can MCGE outperform the ECMWF ensemble? SOLA 4,
77–80 (2008)
6. Park, Y.Y., Buizza, R., Leutbecher, M.: TIGGE: preliminary results on comparing and
combining ensembles. Q. J. R. Meteorol. Soc. 134(637), 2029–2050 (2008)
7. Johnson, C., Swinbank, R.: Medium-range multi-model ensemble combination and calibration.
Q. J. R. Meteorol. Soc. 135(640), 777–794 (2009)FloatPlacement>
Research on Visibility Forecast Based
on LSTM Neural Network
1 Introduction
Atmospheric visibility refers to the maximum horizontal distance of a target object that
can be seen by a person with normal vision under the current weather conditions [7].
Visibility as an important physical parameter for characterizing atmospheric trans-
parency reflects air pollution and atmospheric environmental quality in the region.
When continuous and wide-ranging low-visibility weather conditions occur, it often
causes frequent occurrences of traffic accidents and delays in aircraft movements,
causing huge losses to people’s lives and property. Therefore, accurate forecasting of
low visibility weather is a difficult problem in meteorological work and plays a positive
role in preventing the occurrence of various traffic accidents.
2 Model Data
2.1 Data Source
Meteorological stations in Beijing, Tianjin and Hebei region are located in the North
China Plain. They are affected by the East Asian monsoon climate, with similar cli-
matic background and the same range of relative humidity, so similar weather
Research on Visibility Forecast Based on LSTM Neural Network 553
phenomena are also found in the same weather conditions. The data used in this paper
is the conventional ground observation data and high-altitude data from October to
February of 2015–2016, including wind speed, relative humidity, temperature, and so
on. The model is trained by constructing a historical database of the conventional
ground and high-altitude data, and the data derived from the EC model with the same
elements as above from January to February and October to December in 2017 is used
to forecast.
The value interval is [–1, 1], a complete negative correlation is denoted by –1, a
complete positive correlation is denoted by +1, and a non-linear relationship is denoted
by 0. Table 1 gives the factors selected based on the calculation results.
3 Model Structure
Neural network is a deep learning technique which simulates the structure of biological
neural network and is used to model nonlinear statistical data [4]. The neural network is
connected by a large number of artificial neural units in a prescribed manner, including
an input layer, a hidden layer, and an output layer, and external input information is
dynamically processed by adjusting the structural parameters of the internal unit [6].
The LSTM model is a new network structure, which solves the lack of long-term
memory in the traditional recurrent neural network. It contains a memory unit and three
control gates, namely the forget gate, the input gate and the output gate, to realize
storage and control of information (Fig. 1).
ht-1 ht ht+1
tanh
A σ σ tanh σ
A
Xt-1 Xt Xt+1
The forget gate determines some of the information that is discarded from the cell
state. The gate reads the output of the previous layer and the input of the current time
point, and outputs a value between 0 and 1 to the current cell state. Completely
reserved with the value 1 and completely discarded with the value 0. The calculation
formula is:
Input gate determines the amount of new information added to the cell state. The
implementation of the gate requires two steps: First, the sigmoid activation function
determines the information that needs to be updated, and the new candidate vectors will
be generated by the tanh activation function and added to the state ct . The calculation
formula for this process is:
Then, the cell’s state needs to be renewed. Multiply the forget gate ft with the state
ct1 of the old cell, discard the information that needs to be discarded, and add a new
candidate vector. The calculation formula is:
Research on Visibility Forecast Based on LSTM Neural Network 555
Output gate controls the influence of long-term memory on current output. This
output is not only based on cell status, but also a filtered version. A sigmoid activation
function is used as a valve to selectively output part information of the cell’s state, and
the tanh activation function is used to scale the value of the cell’s state between –1 and
1, and multiply the output of the sigmoid activation function with this value. It will
ultimately determine which part of the cell’s information will be output. The calcula-
tion formula of this gate is:
ht ¼ ot tanhðct Þ ð7Þ
4 Experimental Analysis
After the model was established, the data of meteorological stations in Beijing, Tianjin
and Hebei region from October to December of 2017 was predicted by the LSTM
model. Effect of the model is verified, and compared with the results of the BP neural
network. In order to evaluate the prediction effect at different levels of visibility,
classification is performed according to the intervals of 0–1 km, 1–5 km, 5–10 km and
above 10 km. The model effect is analyzed by the Mean Absolute Error (MAE) and the
Root Mean Square Error (RMSE) of the predicted value and visibility.
From Table 2, the mean absolute error (MAE) and the root mean square error
(RMSE) between the forecast results of LSTM model and the observed results of
visibility are smaller than those of BP neural network model. Among them, the root
mean square error (RMSE) of the LSTM model is 4.81 km, which is 0.25 km lower
than the BP neural network. It shows that predicted values of LSTM model is highly
fitted to the observed values, and the stability is good. At the visibility levels of 0–
1 km, 1–5 km and 5–10 km, the mean absolute errors of LSTM model are 0.37 km,
1.25 km and 1.57 km, respectively, which are 1.56 km, 1.58 km and 2.33 km lower
than those of BP neural network method. From the perspective of the mean absolute
error, when the visibility level is below 10 km, the deviation between predicted value
and observed value is less than forecast result of the level above 10 km, which indi-
cates that the prediction effect of the two models below 10 km is stable and higher than
the forecast result of the visibility above 10 km.
Define the TS score, TS = accuracy/(accuracy + empty + missing). Take visibility
less than 1 km as an example. Accuracy means that the observed value is less than
1 km and the forecast result is less than 1 km. Empty means that the observed value is
greater than 1 km and the forecast result is less than 1 km. Missing means that the
observed value is less than 1 km and the forecast result is larger than 1 km. At the same
time, the predicted value of the BP neural network model are compared, as shown in
Table 3.
556 Y. Dai et al.
From the data of Table 3, the TS scores of this method in 0–1 km and above 10 km
are 0.22 and 0.38, and the forecasting effect is obviously better than BP neural network
method, which are 0.06 and 0.17 greater than that of BP neural network method.
Experimental results in 1–5 km and 5–10 km show that the difference between the two
is only 0.02 and 0.01. In this interval, the prediction accuracy of the two methods is
relatively close. On the whole, the LSTM model has a better application effect in the
low visibility of 0–1 km and the high visibility above 10 km.
In order to visually value the forecasting effect of the LSTM model, the heavy
pollution weather data of Beijing, Tianjin and Hebei region in November 6, 2017 is
selected for testing. The missing data is removed, and a total of 159 test samples are
tested. As shown in Fig. 2, the predicted results are visually compared with the
observed results.
25
visibility(km)
20
15
10
5
0
54406
54431
54594
54528
53392
53593
53692
53773
53792
53886
53896
54308
54408
54434
54503
54519
54534
54604
54612
54621
54632
54705
54717
Station
Fig. 2. The comparison of predicted values of LSTM model and observed values
Research on Visibility Forecast Based on LSTM Neural Network 557
The red and blue in the line chart represent the true and predicted values of
visibility respectively. As shown in the figure, the forecast effect of the model in the 1–
10 km is slightly better than the forecast effect of the model in the 0–1 km, mainly due
to the lack of heavy-pollution, low-visibility weather data samples, resulting in
insufficient training of the model. From the whole point of view, the predicted results of
the model fit well with the observed values of visibility, and even when the visibility is
extreme, the relevant information can also be captured.
5 Conclusion
The model of visibility prediction based on LSTM recurrent neural network proposed
in this paper is an effective model to deal with time series. Based on the historical
meteorological data of Beijing, Tianjin and Hebei region from October to February of
2015–2016, meteorological factors that highly correlated with visibility are selected
with the Pearson Correlation Coefficient method to build the data set. And weighted
parameters of the model are trained to predict the visibility. With the comparison and
analysis of the experiment, the following conclusions are obtained: (1) The LSTM
model shows good stability in the prediction of multiple sites under the same climate
background. (2) Due to few weather data samples with heavy pollution and extremely
low visibility, the TS score of the LSTM model in 0–1 km is lower than 1–10 km. It
can be seen that increasing the training samples with low visibility in the later period is
beneficial to improve the effect of model in 0–1 km. (3) The visibility forecast skills
based on LSTM neural network are better than that of BP neural network. The TS score
of visibility in 0–1 km is 0.22, and the mean absolute error is 0.34 km. The TS score of
visibility in 1–5 km is 0.27, and the mean absolute error is 1.25 km. The TS score of
visibility in 5–10 km is 0.22, and the mean absolute error is 1.57 km. The TS score of
visibility above 10 km is 0.38, and the mean absolute error is 6.07 km. It can be seen
that the LSTM model has higher application value in visibility forecasting and can
obtain more accurate results.
Acknowledgments. This work has been supported in part by the National Natural Science
Foundation of China (Grant No. 61773220), the National Key Research Program of China (Grant
No. 2016YFC0203301), the Nature Science Foundation of Jiangsu Province under Grant
(No. BK20150523).
References
1. Wang, Z., Li, J., Wang, Z., et al.: Numerical simulation and control countermeasures of
strong haze pollution in central and eastern China in January 2013. Chin. Sci. Earth Sci. 1,
3–14 (2014)
2. Yin, S., He, L.: Analysis of atmospheric circulation and weather in January 2015.
Meteorological 4, 514–520 (2015)
3. Zhai, X., Long, Y., Xiao, Z.: Haze weather feature analysis and visibility forecast based on
support vector machine in Wuhan. Resour. Environ. Yangtze River Basin 12, 1754–1761
(2014)
558 Y. Dai et al.
4. Zhang, W., Wang, Z., An, J., et al.: Using BP neural network to improve the prediction effect
of real-time forecasting system for air quality of Olympic games. Climatic Environ. Res. 5,
595–601 (2010)
5. Li, Y., Liu, D., Jin, L., Gao, Y.: Application of BP neural network model in Chongqing
drought forecast. Meteorology 12, 14–18 (2003)
6. Cai, Z., Han, S., Yao, Q., Zhang, M.: Study on weather visibility prediction of Tianjin based
on BP neural network, pp. 2848–2854. Chinese Academy of Environmental Sciences (2016)
7. Liu, D., Hu, B., Yuan, Y., Zhang, H.: Application of least squares support vector machine in
visibility prepackage, pp. 1386–1391. Meteorological Society of China (2009)
8. Han, W., Wu, Y., Ren, F.: Air pollution prediction based on full connection and LSTM
neural network. Geogr. Inf. World 3, 34–40 (2018)
9. Jiang, J., Zhang, G., Gao, J.: Main influencing factors of atmospheric visibility in Beijing.
J. Appl. Meteorol. 2, 188–199 (2018)
10. Fan, J., Li, Q., Zhu, Y., Hou, J., Feng, W.: Research on space-time prediction model of air
pollution based on RNN. Sci. Mapp. Sci. 7, 76–83 (2017)
Application of Artificial Intelligence
on the Image Identification of Icing
Weather Phenomena
1 Introduction
Surface observation is the basis of weather monitoring, weather forecast [1–3], analysis
and prediction on climate [4, 5], and other works. With the advancement of observation
automation and structural reform of vocational work in China, most of the surface
observational projects have realized automation. It greatly reduces the workload of
observers and makes it possible for the observers to focus on the operation system
guarantee and data processing, and to realize the unattended operation of surface
observation. However, at present, many weather phenomena, such as cloud cover,
cloud height, ground condensation and obstruction to vision, still need manual
observation. The disadvantages of strong subjectivity, low observation frequency,
sparse distribution of stations, high input and maintenance costs are becoming
increasingly prominent, which seriously affects the effectiveness of observation
automation [6–8]. In recent years, the image recognition technology based on deep
learning has developed greatly. Research on deep learning by Hinton et al. [9, 10] leads
many scholars to find new research directions and targets, significantly improving the
timeliness and accuracy of deep learning [11]. Convolutional neural network (CNN) is
an algorithm of deep learning, which can realize local connection and parameter
sharing, thus reducing the number of parameters, calculation difficulty and time [12]. It
is a multi-layer feedforward neural network specially designed for processing two-
dimensional input data [13], with strong fault tolerance, self-learning ability and par-
allel processing ability. Krizhevsky et al. [14]. proposed a deep learning algorithm
based on the core of CNN and won the champion of the 2012 ImageNet Competition
with a much lower error rate than second place. After that, the research and application
of CNN have increased exponentially and they are widely used in classifying problems
of various fields. However, the relative researches on the identification and application
of the weather phenomena are few. According to the characteristics of icing, 47 layers
of neural network are designed to analyze the existence of icing and its key features in
different stages of development, so as to automatically identify the phenomena of icing.
The weather phenomena image recognition method in the basis of deep learning is
similar to the observers’ visual observation in principle, avoiding the problems such as
small sample numbers, complex relationships between meteorological elements and the
inconspicuous convergence of model in the process of establishing the statistical
identification model. It is expected to obtain more continuous and quantitative mete-
orological observation information.
Ice accretions, such as surface glaze, rime and mixing rime, are the major weather
phenomena causing freezing disaster in China, which have great impact on industrial
and agricultural production and socioeconomic development. Especially during Jan-
uary 10–February 2 in 2008, southern China suffered from a rare synoptic process of
continuous low temperature, rain, snow and frost [15–17]. According to statistics, more
than 100 million people were affected by this disastrous synoptic process, and the direct
economic losses exceeded 110 billion yuan. Therefore, the observation of glaze and
rime plays an important role in meteorological disaster prevention and mitigation. At
present, the observation of glaze and rime is an artificial observation task. Since Jan-
uary 1 2014, the domestic meteorological stations had stopped the night observation of
Application of Artificial Intelligence on the Image Identification 561
glaze and rime, and planned to cancel the artificial observation task from 2019. In the
absence of automatic observation equipments for glaze and rime, it is necessary to carry
out the study of glaze and rime image recognition algorithm based on deep learning, in
order to satisfy the needs from weather forecast and service.
Based on the field experimental data of icing weather phenomena, including glaze,
rime and mixing rime at Nanyue Mountain Meteorological Station (Nanyue station)
and Huaihua National Reference Climatological Station (Huaihua station) in Hunan
Province, the layout environment requirements of icing weather phenomenon image
identification based on deep learning, are discussed in this paper. By using the
experimental pictures collected at Nanyue station, the CNN deep learning is used to
model and train, and it is tested with the data at the same station. This method provides
technical support for the automatic observation of icing weather phenomena, and also
plays a positive role in effectively promoting the automation of surface observation and
realizing the unattended operation of surface observation.
The Nanyue Mountain is situated in the Hengshao Basin of Hunan Province, located in
the Nanyue District of Hengyang City. Nanyue station (27° 18′N, 112° 42′E) is located
on Wangritai mountaintop in Nanyue Mountain Scenic Area. The altitude of the
observation site is 1265.9 m, and the horizontal distance to Zhurong peak (1300 m) is
only 400 m. Nanyue station is a high-altitude station, and in winter there are many
types of icing condensation weather, such as glaze, rime and mixing rime. HuaiHua
station (27° 34′N, 110° 00′E) with higher elevation and much more glaze in winter is
situated in the eastern part of Yunnan-Guizhou Plateau. The two stations can provide
abundant sample data for the research on artificial intelligence identification technology
of image recognition based on deep learning.
The field experiment used hikvision DS-2CD5028FWD/E-IRA web cameras with 2
million black-light. The cameras have functions of network shield integration, ultra-
wide dynamic infrared camera shooting, human eye bionic technology, multi-spectral
imaging technology and so on. The image size is 1920 * 1080 pixels. Due to the
infrared function of the camera, light was not filled at night during the experiments.
In the local observation field at Huaihua station, the camera was fixed only once so
its stability was poor, and the observation scene experienced many changes during the
period of the experimental observation (Fig. 1). Except the glaze observed in Fig. 1B,
the glaze and rime cannot be clearly shot in Fig. 1A, C and D because of their long
focal length, and even in Fig. 1B the observation time is very short. Therefore, the
observation environment at Huaihua station is not conducive to automatic camera
observation of ice accretion.
Nanyue station is a high-altitude station, and the observation environment is harsh
with high winds and thunderstorms. Therefore, the camera was reinforced twice and its
stability was improved significantly. The scene was fixed with a dense forest behind the
ice shelf, and there were icing wires, stainless steel, porcelain, bamboo poles, and screw
steel, etc. on the shelf. The stainless steel was painted red to test the ice accretion on
different media surfaces (Fig. 2). Comparing and analyzing the field experiment
562 X. Huang et al.
pictures and artificial observations, it can be concluded that metal has small heat
capacity, poor hydrophilicity, fast cooling and heating speed. So it is easy to form
icing, but fast for melting. While vegetation has large heat capacity, good
hydrophilicity, slow cooling and heating speed, and the melting speed of icing is
obviously lower than that on metal. The impact of different colors on icing is not
obvious. For example, the beginning of ice accretion on the metal icing shelf at 0:43 on
January 25 (Fig. 2A) is basically the same as the artificial observation. The end time of
ice accretion on the metal ice shelf is 12:30 on February 7, and the ending of vegetation
ice accretion is 11:43 on February 9. The former is 40 h earlier than the latter (Fig. 2B,
C), and the latter is basically consistent with the artificial observation. Consequently,
the icing shelf and the tall dense vegetation background are the necessary conditions for
the observation of ice accretion, and the observation at Nanyue station is consistent
with the artificial observation. During the field experimental observation period (from
January 22 to March 1 2008), few icing pictures were observed at Huaihua station, but
three icing synoptic processes occurred at Nanyue station. The specific times of arti-
ficial observation were: the mixing rime process from 0:40 on January 25 to 12:30 on
February 9 (hereinafter referred to as process 1); the rime process from 8:46 on
February 10 to 17:45 on February 11 (process 2); the mixing rime process from 19:30
on February 20 to 10:17 on February 23 (process 3). During the observation period, the
equipment was damaged due to a lightning strike. The data of February 16–21 and
February 29–30 were missed and icing data in the beginning period of Process 3 was
absent. In brief, because of good observational environment and large sample size, the
data at Nanyue station were selected for modeling.
Fig. 2. Formation and melting process of icing (A: 0:43 on January 25; B: 12:03 on February 7;
C: 11:43 on February 9)
Application of Artificial Intelligence on the Image Identification 563
The sample data for modeling and training are camera pictures photographed at Nanyue
station every 20 min from 07:00 on January 23 2018 to 23:00 on March 1 2018.
1789 pictures are selected as training samples after decontamination. Ice accretion
during the daytime (07:00a.m.–18:59) and the night (19:00–06:59a.m.) at Nanyue
station are respectively shown in Fig. 3A and B. The pictures of the daytime are
significantly better than those of the night. The blurry pictures are very unclear
(Fig. 3C) mainly due to the changes of night camera mode after February 8, thus there
are only the daytime pictures after February 8. The pictures at Nanyue station during
January 23–March 1 2018 are selected as the testing samples, which are not duplicated
and independent with the training samples.
Fig. 3. Contrast of mixing rime between day and night (A: 14:43 on February 6; B: 23:03 on
February 6; C: 04:00 on February 25)
Because of light, focus, weather and other reasons, the sizes of the photographs are
different. In order to calculate conveniently, we uniformly use the numpy, which is a
math library of python, to convert the picture into 299 * 299 pixels (not shown).
In this paper, after pre-processing and simply tagging the start and end time of ice
accretion, the training model automatically captures the shape, color, texture and other
key information of the ground icing through neural network, and directly extracts
information features from the whole picture instead of manual marking.
The training process of CNN is divided into two stages. The first stage is dis-
seminating data from the low level to the high level, or the forward propagation stage.
Another stage is when the results of the forward propagation do not correspond with
expectations, the error is propagated from the high level to the low level, or the back
propagation stage.
The main parameters of the model are shown in Table 2. Training steps denote the
steps must be run before the end of training. Learning_rate denotes the speed at which
the parameters could reach optimal values. Train_batch_size is the number of pictures
trained at a time. Validation_batch_size is the number of pictures used in the evaluation
batch. Eval_step_interval denotes the time interval of estimating the training results.
5 Result Analysis
100
75
50
25
0
1-23 1-26 1-29 2-1 2-4 2-7 2-10 2-13 2-16 2-19 2-22 2-25 2-28
Fig. 4. Time series of icing identification probability at Nanyue Station from 7:23 January 23 to
12:20 March 1, 2018
(The blue line expresses the probability of icing identified by the model of each
time level; and the red line indicates the start and end time of the three icing processes.)
Figure 5 shows the pictures of the three processes in the icing stable period
respectively, and all the probabilities are greater than 99.3%. There are many con-
densates in the stable periods of process 1 and process 3, the probability is stably more
than 99%, just like quasi-straight line. Process 2 is a rime process with shorter icing
duration and less condensate, thus the probability fluctuates (Fig. 4).
During the process of icing melting, the probability gradually decreases. The
amount of condensate in the process 1 is maximum, and it began to melt on Feb 7; the
melting time is longer and the probability curve decreases slowly. Process 3 is similar
to process 1, but the condensate is less than process 1, and the melting speed is
relatively faster, the probability curve decreases more rapidly. Process 2 is a rime
566 X. Huang et al.
Fig. 5. Pictures of icing stationary phase in the three processes (A: 13:23 on January 31
(99.9%); B: 7:03 on February 11 (99.3%); C: 13:43 on February 22 (99.9%))
process with short duration, the melting speed is the fastest and the probability curve
drops rapidly (Fig. 4).
There are 9 omissions in the testing pictures, and the probability is less than 50%.
The omission usually occurs at the beginning or near the end of the ice accretion due to
the small amount of condensate and the indistinct characteristics of the ice accretion
(Fig. 6A and B). In addition, when the condensate is little or during the transition of
camera modes between daytime and night, omissions could occur. Under these two
situations, it is also difficult to identify even by manual observation, so, such cases are
few and generally have little impact on the observations.
Fig. 6. Omissions of ice accretion (A: 7:03 on February 9 (41.84%); B: 15:43 on February 11
(36.29%); C: 18:43 on February 22 (23.63%))
There are 5 false accepted pictures in the testing pictures (the probability is greater
than 50%). The false accept mainly appears in the beginning or at the end of the
process. In Fig. 7A, the icing has almost melted, but condensation still exists on the
ground, and the probability of false acception is 50.3%. Because of the decrease of
visibility before the rime formation in Fig. 7B, the image is blurred and the FAR
reaches 85.65%. Due to the discontinuity of false accept pictures and the low FAR, it
has little effect on the observations.
In summary, the PC of identification is more than 98%. The training model can
extract the characteristics of ice accretion very well and identify the icing weather
phenomena effectively. Although the probabilities of false and missing identification
are very low, PC can be further improved by the following measures. On one hand, the
meteorological conditions can be included to eliminate the false identification. On the
other, because the ice accretion is a continuous weather phenomenon, the false
acception and omission can be further eliminated according to the consistency of the
identification criteria.
Application of Artificial Intelligence on the Image Identification 567
Fig. 7. False identification of ice accretion (A: 15:23 on February 9 (50.3%); B: 8:03 on
February 10 (85.65%))
In this paper, the camera images of the icing field experiment at Nanyue station and
Huaihua station are used to model the testing images at Nanyue station by CNN
technology, and the testing data are used to verify the model. On this basis, the
environment conditions for collecting images used for identification of icing weather
phenomena by artificial intelligence are discussed. The main conclusions are as
follows.
(1) Comparing and analyzing the experimental pictures at Nanyue station and
Huaihua station, the environmental conditions for the image acquisition equip-
ment of Nanyue station are similar to manual observation environment and are
better than that of Huaihua station. The Nanyue station is suitable for the auto-
matic observation of icing weather phenomena, including glaze, rime, mixing
rime and so on.
(2) Independent samples are used to verify the effect of the icing training model at
Nanyue station. The PC is 99.21%, the FAR is 0.28%, and the PO is 0.51%. The
probability of ice accretion during the formation period ascends, and the proba-
bility in the stable period is greater than 99.3%. But, it decreases in the melting
period. It shows that the training model can extract the characteristics of icing
weather phenomena very well, and the identification effect of icing is quite good.
However, in the initial stage of icing formation, the end stage, the transition time
of camera modes between daytime and night, or when the pictures are unclear at
night, the false acceptance or omission could occasionally occur. It is recom-
mended to set supplementary lighting at night for increasing the clarity of the
pictures.
(3) Artificial intelligence image recognition method based on deep learning can
identify the icing weather phenomena very well, and can eliminate the false
acceptance and omission by including the meteorological conditions criteria and
judging the consistency of identification. Glaze, rime and mixing rime can be
distinguished by precipitation, temperature, relative humidity, visibility and other
meteorological parameters.
568 X. Huang et al.
(4) Image identification based on deep learning needs to be trained by a lot of image
samples from different fields, backgrounds and illumination conditions. The
quantity of the image samples and the precision of the annotations probably affect
the image recognition effect. In this study, the field testing duration is relatively
short. The corresponding image acquisition standards and specifications will be
studied according to the needs of deep-learning image identification technology
and conditions of meteorological operation. The artificial intelligence model
should be further optimized and improved.
Acknowledgement. This research was funded by the Small Business Construction Project of
China Meteorological Administration (2018) “Comprehensive Meteorological Observation
Intelligent Analysis and Identification System Construction” (QXPG20174022) and Special
Project for Capacity Building of Meteorological Forecasting of Hunan Meteorological Bureau
(2016-2017) “Meteorological element product improvement based on multi-source data fusion
(YBNL16-04)”.
References
1. China Meterological Administration. The criterion of surface meteorological observation,
pp. 21–27. China Meteorological Press, Beijing (2003)
2. Huang, X.Y., Li, Z.X., Li, C., et al.: Analysis on extreme freeze catastrophic weather of
Hunan in 2008. Meteorol. Mon. 34(11), 47–53 (2008)
3. Hu, W.D., Yang, K., Huang, X.Y., et al.: Analysis on a severe convection triggered by gust
front in Yinchuan with radar data. Plateau Meteorol. 34(5), 1452–1464 (2015)
4. Wang, Z.Y., Ding, Y.H., He, J.H., et al.: An updating analysis of the climate change in
China in recent 50 years. ACTA Meteorol. Sin. 62(2), 228–236 (2004)
5. Liang, S.J., Ding, Y.H., Zhao, N., et al.: Analysis of the interdecadal changes of the
wintertime surface air temperature over mainland China and regional atmospheric circulation
characteristics during 1960–2013. Chin. J. Atmos. Sci. 38(5), 974–992 (2014)
6. Xing, H.Y., Zhang, J.Y., Xu, W., et al.: Development and prospect of automatic
meteorological observation technology on the ground. J. Electron. Measur. Instrum. 31
(10), 1534–1542 (2017)
7. Ma, S.J., Wu, K.J., Chen, D.D., et al.: Automated present weather observing system and
experiment. Meteorol. Mon. 37(9), 1166–1172 (2011)
8. Liu, L.Y., Lan, M.C., Zhu, X.W., et al.: The Comparative analysis of two cloud products of
FY2G satellite in Hunan Province. Torrential Rain Disasters 36(2), 164–170 (2017)
9. Hinton, G.E., Osindero, S., TeH, Y.W.: A fast learning algorithm for deep belief nets. Neural
Comput. 18(7), 1527 (2006)
10. Hinton, G.E., Salakhutdinov, R.R.: Reducing the dimensionality of data with neural
networks. Science 313(28), 504–507 (2006)
11. Yu, B.,Li, S., Xu, S.X., et al.: Deep learning: the key to open big data era. J. Eng. Stud. no. 3,
233–243 (2014)
12. Bengio, Y.: Learning deep architectures for AI. Found. Trends Mach. Learn. 2(1), 1–127
(2009)
13. Chan, C.H., Pang, G.K.H.: Fabric defect detection by Fourier analysis. IEEE Trans. Ind.
Appl. 36(5), 1267–1276 (2000)
Application of Artificial Intelligence on the Image Identification 569
14. Krizhevsky, A., Sutskever, I., Hinton, G.E.: Image net classification with deep convolutional
neural networks. In: Proceedings of International Conference on Neural Information
Processing System, pp. 1097–1105 (2012)
15. Wang, Z.Y., Zhang, Q., Chen, Y., et al.: Characters of meteorological disasters caused by the
extreme synoptic process in early 2008 over China. Clim. Chang. Res. 4(2), 63–67 (2008)
16. Gang, H., Chen, L.J., Jia, X.L., et al.: Analysis of the severe cold surge, ice-snow and frozen
disasters in South China during january 2008: II possible climatic causes. Meteorol. Mon. 34
(4), 101–106 (2008)
17. Ye, C.Z., Wu, X.Y., Huang, X.Y.: A synoptic analysis of the unprecedented severe event of
the consecutive cryogenic freezing rain in Hunan Province. Acta Meteorologica Sin. 67(3),
488–500 (2009)
The Realization Path of Network Security
Technology Under Big Data and Cloud
Computing
Abstract. This paper studies the cloud and big data technology based on the
characters of network security, including virus invasion, data storage, system
vulnerabilities, network management etc. It analyzes some key network security
problems in the current cloud and big data network. Above all, this paper puts
forward technical ways of achieving network security. Cloud computing is a
service that based on the increased usage and delivery of the internet related
services, it promotes the rapidly development of the big data information pro-
cessing technology, improves the processing and management abilities of big
data information. With tie rapid development of computer technology, big data
technology brings not only huge economic benefits, but the evolution of social
productivity. However, serials of safety problems appeared. How to increase
network security has been become the key point. This paper analyzes and dis-
cusses the technical ways of achieving network security.
1 Introduction
and usage model, it means get resource (Hardware, software) via network. The network
of providing resource is called ‘Cloud’. The hardware resource in the ‘Cloud’ seems
scalable infinitely and can be used whenever [9–11].
Cloud computing is the product of the rapid development of computer science and
technology. However, the problem of computer network security in the background of
cloud computing brings a lot of trouble to people’s life, work and study [12–14].
Therefore, scientific and effective management measures should be taken in combi-
nation with the characteristics of cloud computing technology to minimize the risk of
computer network security and improve the stability and security of computer network.
This paper briefly introduces cloud computing, analyzes the network security problem
of computer under cloud computing, and expounds the network security protection
measures under cloud computing.
Processing data by cloud computing can save the energy expenditure and reduce the
dealing cost of big data, so that it can improve the healthy development of cloud
computing technology. Analyzing big data by cloud computing technology can be
shown by a directed acyclic data flow graph G ¼ ðV; EÞ, and the cloud service module
in the Parallel selection mechanism is made up by a serial group V ¼ fiji ¼ 1; 2; . . .; vg
and a serial of remote data transfer hidden channels E ¼ fði; jÞji; j 2 Vg. Assuming the
date transmission distance of the data flow model in C=S framework is
T0 þ B þ is þ Td þ ji þ 1\T0 þ B þ is þ Td.
The directed graph model GP ¼ ðVP; EP; SCAPÞ explanation, EP represent
LKSET, the VP cross channel bearing the physical node set, the SCAP explains the
quantity of data unit of physical node. Besides, assuming undirected graph GS ¼
ðVS; ES; SARSÞ expresses data packet markers input by application. The process of link
mapping between cloud computing components and overall architecture can be
explained by:
Fig. 1. Built-in network link structure for big data information processing
In Fig. 1, the ith transmission package in the cloud computer is ith. Let Ti represent
the transmission time of ith. The interval of Component is mapped to thread or process
is showed by ji ¼ Ti Td , when ji ¼ Ti Td in the range of (−∞, ∞), the weight of
node I is Wi which computing time, the detail application model of big data infor-
mation processing is shown in Fig. 2
In the mobile cloud system model, the grid architecture that relies on local computing
resources and the wireless network to build cloud computing, which will select the
components of data flow graph to migrate to the cloud, Computer data processing cloud
The Realization Path of Network Security Technology 573
computing formula modeling, fGðV; EÞ; si; di; jg is the given data flow applications,
assuming that the channel capacity is infinite, the problem of using cloud computing
technology to optimize big data information processing is described as follows
1
maxmax TP ¼ ; i; j 2 f0; 1; ; v þ 1g ð2Þ
xi;yi;jxi;yi;j tp
Among them:
si X di;j ðxi xj Þ2
tp ¼ maxfmax maxðxi : Xi Þ; max max ð Þg ð3Þ
i2v i2v gp i2v ði;jÞ2E ði;jÞ2E yi;j
The energy overhead of data flow migrating between groups in mobile cloud
computing is described as:
nQ
k ¼ Intð Þþ1 ð4Þ
1Q
4.2 Convenience
Big data stored in the cloud usually affects network data. Most enterprises will connect
multiple servers so as to build computing terminals with strong performance. Cloud
computing itself has the convenience. Customers of its hardware facilities do not need to
purchase additional services. They only need to purchase storage and computing ser-
vices. Due to its particularity, cloud computing can effectively reduce resource con-
sumption and is also a new form of energy conservation and environmental protection.
4.3 Participatory
When local computers encounter risks, data stored in the cloud will not be affected, nor
will it be lost, and at the same time these data will be shared. The sharing and transfer
of raw data is generally based on physical connections, and then data transfer is
implemented. Compared with the original data research, data sharing in big data cloud
computing can be realized by using the cloud. Users can collect data with the help of
various terminals, so as to have a strong data sharing function.
574 N. Kang et al.
5 Security Issues
5.1 System Vulnerabilities
Most computer networks have risks from system vulnerabilities. Criminals use illegal
means to make use of system vulnerabilities to invade other systems. System vulner-
abilities not only include the vulnerabilities of the computer network system itself, but
also can easily affect the computer system due to the user’s downloading of unknown
plug-ins, thus causing system vulnerability problems.
8 Conclusion
This paper studies the cloud and big data technology. In the context of large data cloud
computing, the computer network security problem is gradually a highlight, and in this
case, the computer network operation condition should be combined with the modern
network frame safety technology, so as to ensure the security of the network infor-
mation, thus creating a safe network operation environment for users.
References
1. Zhou, X., Lu, L.: Application and operation of computer network security prevention under
the background of big data era. Netw. Secur. Technol. Appl. 05, 24–30 (2017)
2. Sun, H., Jia, R.: Research on enterprise network information security technology system in
the context of big data. Commun. Technol. 50(02), 334–339 (2007)
3. Xu, L., Cheng, X., Chen, Y., Chao, K., Liu, D., Xing, H.: Self-optimised coordinated traffic
shifting scheme for LTE cellular systems. In: 1st EAI International Conference on Self-
Organizing Networks, pp. 67–75. Springer Press, Beijing (2015)
The Realization Path of Network Security Technology 577
4. Gao, M.: Network security technology in big data environment. Inf. Commun. 01, 158–159
(2017)
5. Xu, L., Zhao, X., Yu, Y., et al.: Data mining for base station evaluation in LTE cellular
systems. In: 3rd International Conference on Signal and Information Processing, Networking
and Computers, pp. 356–364. Springer Press, Chongqing (2017)
6. Xu, L., Chen, Y., Schormans, J., et al.: User-vote assisted self-organizing load balancing for
OFDMA cellular systems. In: 22nd IEEE International Symposium on Personal Indoor and
Mobile Radio Communications, pp. 217–221. IEEE Press, Toronto (2011)
7. Wang, F.: Discussion on network information security in the context of big data. Digital
Technol. Appl. 05, 210 (2016)
8. Xu, L., Luan, Y., Cheng, X., et al.: Telecom big data based user offloading self-optimisation
in heterogeneous relay cellular systems. Int. Jo. Distrib. Syst. Technol. 8(2), 27–46 (2017)
9. Zhou, H.: Application of cloud computing technology in computer secure storage. Netw.
Secur. Technol. Appl. (10):78–79 (2017)
10. Xu, L., Zhao, X., Luan, Y., et al.: User perception aware telecom data mining and network
management for LTE/LTE-advanced networks. In: 4rd International Conference on Signal
and Information Processing, Networking and Computers, pp. 237–245. Springer Press,
Qingdao (2018)
11. Xu, L., Luan, Y., Cheng, X., Xing, H., Liu, Y., Jiang, X., Chen, W., Chao, K.: Self-
optimised joint traffic offloading in heterogeneous cellular networks. In: 16th IEEE
International Symposium on Communications and Information Technologies, pp. 263–267.
IEEE Press, Qingdao (2016)
12. Huang, Y.: Network information security control mechanism and evaluation system in the
context of big data. Inf. Comput. (Theor. Ed.) 20, 201–202 (2016)
13. Xu, L., Cheng, X., et al.: Mobility load balancing aware radio resource allocation scheme for
LTE-advanced cellular networks. In: 16th IEEE International Conference on Communica-
tion Technology, pp. 806–812. IEEE Press, Hangzhou (2015)
14. Xu, L., Luan, Y., Cheng, X., et al.: WCDMA data based LTE site selection scheme in LTE
deployment. In: 1st International Conference on Signal and Information Processing,
Networking and Computers, pp. 249–260. CRC Press Taylor & Francis Group, Beijing
(2015)
Rate Control in HEVC: A Survey
Abstract. Rate control is an important tool in video coding. So far, R-k based
rate control algorithm for rate control is widely used in High Efficiency Video
Coding (HEVC). In this paper, we present a survey on the algorithm of R-k
based rate control process, including bit allocation and rate control coding by
parameters calculation. Firstly, we provide bit allocation method at three levels:
group of pictures (GOP), frame, and basic unit level. Secondly, the method-
ologies of coding parameters calculation are presented. Then, we summarize the
recent research achievements and future directions and finally open problems are
discussed.
1 Introduction
2 Bit Allocation
In most rate control models, bit allocation is adapted in three levels: GOP level, frame
level, and basic unit level. For HEVC, one basic unit can be divided into one or more
coding units (CU). Except three levels above, bit allocation for the first picture is one
special level as encoder is hardly to know the first picture’s information and allocate a
proper number of bits and in most cases, the encoding parameters are appointed by
users.
Fig. 1. Figure gives a demonstration of the intra-frame period of four GOPs with a size of eight
frames. The I frame is represented by a black bar (the last one in GOP0). In this picture, only
GOP0 consists of an I frame.
2.3 BU Level
In HEVC, one BU can consist of one or more CUs. In BU level, one important topic is
how to divide CUs and code in different CU. [13] proposes an algorithm which is able
to decide the size of CU in a fast way that terminates the homogeneous CUs. After that,
for the rest CUs, it used two linear support vector machines (SVMs) to decide CU
division and termination. [7] categorizes the depth of CUs into three levels including
low-texture, medium-texture and high-texture level and take the distinguish statistical
characteristics into consideration and use Laplacian probability density function to
calculate the transform ratio residues. Besides, in order to find the best coding unit, [14]
proposes a convolutional neural network (CNN) aimed to classify units. The algorithm
can learn the region-wise images’ characteristics.
After allocating appropriate bits for each level, the encoder is required to achieve
allocated bits by adjusting coding parameters. Coding parameters calculation is usually
adapted by the bit allocation strategy. Coding strategies based on R-k model which
approximate the rate of the RD curve while taking the difference between intra-frame
and inter-frame into account, which have been proved a good performance. Figure 2
gives a typical demonstration of R-D curve in videos.
In R-k model, it’s of vital importance to accurately calculate the value of k, because
k largely depends determines the accuracy of algorithm. For this reason, the R-k based
rate control algorithm designs a special methodology to calculate the coding parameters
and update the value of k in encoding procedure by using trained data [15].
[13] proposed a new algorithm to calculate k value, which does not need trained
models. Instead, it uses piecewise linear approximations. In each frame, it linearly
approximate k by the actual rate and distortion of coded blocks.
Rate Control in HEVC: A Survey 581
4 Future Topics
As we mentioned above, a lot of works based on R-k model aim to allocate appropriate
bit rate and achieve the target bit by updating parameters, and they are proved to
achieve good performance.
In general, recent works on rate control concentrate on two subjects: one is to
consider special rate control strategies for special situation, such as [9] concentrating on
high dynamic range imaging (HDR), [18–20] aimed to screen content coding (SCC),
and the proposal in [21] applying for 3-D HEVC; the other is to apply distinct R-k
models for different parts and compute coding parameters separately, such as proposals
in [8–11]. We believe in the future, research on special situation will go deeper, with
more analyses and better development on SCC, 3-D, HDR and so on. Multi-R-k model
may also result in wider application.
Moreover, although current rate control algorithms in HEVC have a good capa-
bility, they have not been accurately designed for coding intra frames. Thus their
accuracy in achieving a target bit rate cannot reach a high level. More algorithms of
intra-coding may be proposed.
Besides, with the development of deep-learning, some algorithms based of neural
network have been applied in video coding. The proposed algorithms perform good
accuracy and low complexity in simulation. We believe deep-learning in HEVC may
be another hot topic in the next few years.
5 Conclusion
In this paper, we present a survey for rate control in HEVC from the perspective of bit
allocation and coding parameters calculation. There are kinds of bit allocation strategies
at each bit allocation level, and methodologies of coding parameters calculation are
adapted accordingly. With tractable theoretical analysis, more insight might be pro-
vided for the design of video coding, especially for special situation video coding and
multiple coding models building.
582 J. Zou and B. Li
References
1. Li, B., Li, H., Li, L.: Lambda-domain rate control algorithm for high efficiency video coding.
IEEE Trans. Image Process. 23(9), 3841–3854 (2014)
2. Choi, H., Nam, J., Yoo, J., Sim, D., Bajic, I.: Rate control based on unified RQ model for
HEVC. Document Rec. JCTVC-H0213, San Jose, February 2012
3. Wang, S., Ma, S., Wang, S., Zhao, D., Gao, W.: Quadratic q-domain based rate control
algorithm for HEVC. In: IEEE International Conference on Acoustics, Speech and Signal
Processing, Vancouver, pp. 1695–1699 (2013)
4. Wang, S., Ma, S., Wang, S., Zhao, D., Gao, W.: Rate-GOP based rate control for high
efficiency video coding. IEEE J. Sel. Top. Signal Process. 7(6), 1101–1111 (2013)
5. Wiegand, T., Sullivan, G.J., Bjontegaard, G., Luthra, A.: Overview of the H.264/AVC video
coding standard. IEEE Trans. Circuits Syst. Video Technol. 13(7), 560–576 (2003)
6. Song, F., Zhu, C., Liu, Y., Zhou, Y., Liu, Y.: A new GOP level bit allocation method for
HEVC rate control. In: IEEE International Symposium on Broadband Multimedia Systems
and Broadcasting (BMSB), Cagliari, pp. 1–4 (2017)
7. Lee, B., Kim, M., Nguyen, T.: A frame-level rate control scheme based on texture and non-
texture rate models for high-efficiency video coding. IEEE Trans. Circuits Syst. Video
Technol. 1–14 (2013)
8. Meddeb, M., Cagnazzo, M., Pesquet-Popescu, B.: Region-of-interest based rate control
scheme for high-efficiency video coding. In: Proceedings of the IEEE International
Conference on Acoustics, Speech Signal Processing (ICASSP), pp. 7338–7342, May 2014
9. Perez-Daniel, K.R., Sanchez, V.: Luma-aware multi-model rate-control for HDR content in
HEVC. In: IEEE International Conference on Image Processing (ICIP), Beijing, pp. 1022–
1026 (2017)
10. Sanchez, V., Aulí-Llinàs, F., Vanam, R., Bartrina-Rapesta, J.: Rate control for lossless
region of interest coding in HEVC intra-coding with applications to digital pathology
images. In: IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), Brisbane, pp. 1250–1254 (2015)
11. Sanchez, v: Rate control for HEVC intra-coding based on piecewise linear approximations.
In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP),
Calgary, pp. 1782–1786 (2018)
12. Erabadda, B., Mallikarachchi, T., Kulupana, G., Fernando, A.: Machine learning approaches
for intra-prediction in HEVC. In: IEEE 7th Global Conference on Consumer Electronics
(GCCE), Nara, pp. 206–209 (2018)
13. Zhang, T., Sun, M., Zhao, D., Gao, W.: Fast intra-mode and CU size decision for HEVC.
IEEE Trans. Circuits Syst. Video Technol. 27(8), 1714–1726 (2017)
14. Kuanar, S., Rao, K.R., Conly, C.: Fast Mode decision in HEVC intra prediction, using
region wise CNN feature classification. In: IEEE International Conference on Multimedia &
Expo Workshops (ICMEW), San Diego, pp. 1–4 (2018)
15. Li, B., Li, H., Li, L.: Adaptive bit allocation for R-lambda model rate control in HM, in JCT-
VC M0036. In: 13th Meeting of Joint Collaborative Team on Video Coding of ITU-T SG 16
WP 3 (2013)
Rate Control in HEVC: A Survey 583
16. Katayama, T., Song, T., Shimamoto, T.: QP adaptation algorithm for low complexity HEVC
based on a CNN-generated header bits map. In: IEEE 8th International Conference on
Consumer Electronics - Berlin (ICCE-Berlin), Berlin, pp. 1–5 (2018)
17. Maung, H., Aramvith, S., Miyanaga, Y.: Improved region-of-interest based rate control for
error resilient HEVC framework. In: IEEE International Conference on Digital Signal
Processing (DSP), Beijing, pp. 286–290 (2016)
18. Guo, Y., Li, B., Sun, S., Xu, J.: Rate control for screen content coding in HEVC. In: IEEE
International Symposium on Circuits and Systems (ISCAS), pp. 1118–1121, May 2015
19. Guo, Y., Li, B., Sun, S., Xu, J.: Rate control for screen content coding based on picture
classification. In: Visual Communications and Image Processing (VCIP), pp. 1–4, December
2015
20. Xiao, J., Li, B., Sun, S., Xu, J.: Rate control with delay constraint for screen content coding.
In: IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, pp. 1–4
(2017)
21. Song, Y., Jia, K., Wei, Z.: Improved LCU level rate control for 3D-HEVC. In: Visual
Communications and Image Processing (VCIP), Chengdu, pp. 1–4 (2016). https://doi.org/
10.1109/vcip.2016.78054783
Screen Content Coding: A Survey
1 Introduction
The early development of the well-known High Efficiency Video Coding (HEVC) and
H.264/AVC [1] is primarily concentrated on camera-captured content sequences.
However, the phenomenon that not just camera-captured content is displayed on video
devices has become popular in recent years. This video sequence containing a large
amount of moving or stationary pictures, text, together with camera-captured content is
mostly used in specific situations such as remote computer desktop access, wireless
displays, and video conferencing screen sharing [2]. This kind of content is called
screen content, which shows very different characteristics in comparison with that
captured by camera. The sequence captured by the camera has relatively good conti-
nuity in the spatiotemporal domain. While in the sequence of screen content, timely
continuous pictures may show significant movement or sudden change, and spatially, a
single image may have large flat areas with high contrast and sharp edges. Due to these
2 HEVC-SCC Extension
this concept. When this concept is applied to the video captured by the camera, it turns
out not very successful because text patterns in adjacent areas of space may share
similarity with ongoing coding blocks, while often change as space steadily changes.
Therefore, accurate match for blocks is hard to find in the same image, thus limiting the
improvement of coding performance. However, for screen content, it is difficult to find
spatial correlation between pixels in the same image. For a typical screen content
sequence, repeated patterns frequently occur in the same image. Therefore, block
copying in a single image is possible and turns out very efficient.
Therefore, Intra-Block Copy mode is proposed as a new prediction mode to take
advantage of the particular feature.
In this mode, the blocks which have previously been reconstructed are used for
predicting the prediction unit (PU) in the same image. The relative displacement from
the position of the present PU to the reference block position is shown by the dis-
placement vectors (block vectors or BVs). Then, the compensated prediction error is
coded in the same way as the residual in HEVC version 1 [3]. An illustration of IBC is
shown in Fig. 1.
In short, palette coding is a prediction method based on colors [19, 20]. All the
pixels of the encoding block (the values of three components, R/G/B or Y/Cb/Cr) are
categorized as a list of colors. The main color is the representative color, which occurs
frequently in blocks. For each coding unit (CU) in palette mode, an index table of color
is created. Every index entry in the table is combined with three sample values (R/G/B
or Y/Cb/Cr). Except barely used pixels, almost all pixels are transformed into matching
indexes in CU. These rare pixels are called escaped pixels. They are set apart in the
index table and unable to be quantized to dominant colors with ESCAPE symbols
marked. Actual values of these escaped pixels are explicitly signaled. These indexes,
ESCAPE included, are run-length coded, predicting from the upper or left neighbor. An
illustration of Palette Mode is shown in Fig. 2.
[21] proposed a pre-decision approach for fast palette mode on the basis of ana-
lyzing color complexity in video data. The approach saved approximately 74.24%
computation in game video experiments compared with HEVC-SCC palette mode.
Meanwhile, the bjøntegaard delta bit rate (BDBR) only increases about 0.36% and the
bjøntegaard delta peak signal-to-noise rate (BD-PSNR) averagely decreases 0.03 dB.
under the premise of bitrate control. In order to better allocate resources and avoid
bitrate runaway, [25] also introduces sliding window bit allocation instead of traditional
GOP bit allocation, and achieves good results.
Fig. 3. The flow chart of the improved SCC rate control scheme.
4 Conclusion
Coding tools added in HEVC-SCC extension which were primarily designed for screen
content coding have made great improvements in SCC. With new coding tools under
development, screen content coding is expected to achieve higher efficiency. In rate
control of screen content coding, aiming at the discontinuity of screen content
sequence, an abrupt frame prediction mechanism is added. Meanwhile, the bit allo-
cation and parameter updating in bitrate control are improved. The rate runaway is
prevented and the coding quality is guaranteed, thus improving the coding efficiency in
SCC.
References
1. Advanced Video Coding for Generic Audiovisual Services, ITU-T and ISO/IEC JTC1,
document ITU-T Rec. H.264 and ISO/IEC 14496-10, May 2003
2. Yu, H., McCann, K., Cohen, R., Amon, P.: Requirements for an Extension of HEVC for
Coding of Screen Content, ISO/IEC JTC 1/SC 29/WG 11, document MPEG2014/N14174,
San Jose, CA, USA, January 2014
3. Sullivan, G.J., Ohm, J., Han, W.-J., Wiegand, T.: Overview of the high efficiency video
coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 22(12), 1649–1668
(2012)
4. Boyce, J., et al.: Edition 2 Draft Text of High Efficiency Video Coding (HEVC), Including
Format Range (RExt), Scalability (SHVC), and Multi-View (MV-HEVC) Extensions,
document JCTVC-R1013, Sapporo, Japan, July 2014
5. Narroschke, M.: Extending H.264/AVC by an adaptive coding of the prediction error. In:
Proceedings of 25th Picture Coding Symposium (PCS), O5–3. Beijing, China, April 2006
6. Lan, C., Xu, J., Sullivan, G.J., Wu, F.: Intra Transform Skipping, document JCTVC-I0408,
Geneva, Switzerland, April 2012
7. Peng, X., Lan, C., Xu, J., Sullivan, G. J.: Inter Transform Skipping, document JCTVC-
I0408, Stockholm, Sweden, July 2012
8. Cohen, R., Vetro, A.: AHG8: Impact of Transform Skip on New Screen Content Material,
document JCTVC-L0428, Geneva, Switzerland, January 2013
9. Peng, X., Xu, J., Guo, L., Sole, J., Karczewicz, M.: Non-RCE2: Transform Skip on Large
TUs, document JCTVC-N0288, Vienna, Austria, July 2013
10. An, J., Zhao, L., Huang, Y.-W., Lei, S.: Residue Scan for Intra Transform Skip Mode,
document JCTVC-J0053, Stockholm, Sweden, June 2012
11. He, D., Wang, J., Martin-Cocher, G.: Rotation of Residual Block for Transform Skipping,
document JCTVC-J0093, Stockholm, Sweden, June 2012
12. Peng, X., Li, B., Xu, J.: On Residual Rotation for Inter and Intra BC Modes, document
JCTVC-O0186, Geneva, Switzerland, October 2013
13. Lee, S., Kim, I.-K., Kim, C.: AHG7: Residual DPCM for HEVC Lossless coding, document
JCTVC-L0117, Geneva, Switzerland, January 2013
14. Joshi, R., Sole, J., Karczewicz, M.: AHG8: Residual DPCM for Visually Lossless Coding,
document JCTVC-M0351, Incheon, Korea, April 2013
15. Naccari, M., Mrak, M., Gabriellini, A., Blasi, S., Izquierdo, E.: Inter-Prediction
Residual DPCM, Incheon, Korea, document JCTVC-M0442, April 2013
16. Nguyen, T., Khairat, A., Marpe, D.: Non-RCE1/Non- RCE2/AHG5/AHG8: Adaptive Inter-
Plane Prediction for RGB Content, document JCTVC-M0230, Incheon, Korea, April 2013
17. Joshi, R., Xu, J.: HEVC Screen Content Coding Draft Text 1, document JCTVC-R1005,
Sapporo, Japan, July 2014
18. Xu, X., Li, X., Liu, S.: Intra block copy for next generation video coding. In: ICMEW
(2018)
19. Guo, L., Pu, W., Zou, F., Sole, J., Karczewicz, M., Joshi, R.: Color palette for screen content
coding. In: 2014 IEEE International Conference on Image Processing (ICIP), 5556–5560
October 2014
20. Onno, P., Xiu, X., Huang, Y.-W., Joshi, R.: Suggested Combined Software and Text for
Run-Based Palette Mode. JCTVC-R0348, July 2014
21. Liu, Y., Fang, C., Sun, J., Huang, X.: Fast Palette Mode Decision Methods for Coding Game
Videos with HEVC-SCC. IEEE Trans. Circuits Syst. Video Technol. (2018)
592 B. Li and J. Zou
22. Zhang, L., et al.: SCCE5 Test 3.2.1: In-Loop Color-Space Transform. JCTVC-R0147, July
2014
23. Henrique Malvar, S., Sullivan, G.J., Srinivasan, S.: Lifting-based reversible color
transformations for image compression. In: SPIE Proceedings, 7073 (2008), https://doi.
org/10.1117/12.797091
24. Li, B., Xu, J., Sullivan, G., Zhou, Y., Lin, B.: Adaptive Motion Vector Resolution for Screen
Content. JCTVC-S0085, October 2014
25. Guo, Y., Li, B., Sun, S., Xu, J.: Rate control for screen content coding in HEVC. In: IEEE
International Symposium on Circuits & Systems, September 2015
26. Wang, S., Li, J., Wang, S., Ma, S., Gao, W.: A frame level rate control algorithm for screen
content coding. In: ISCAS, May 2018
Intelligent Fitness Trainer System Based
on Human Pose Estimation
1 Introduction
Convenient deployment. The system can be deployed conveniently in the gym and
other application scenarios, which satisfies the freedom of fitness time and space for
users.
The personalized needs of users. The system can intelligently analyze the different
characteristics of multiple fitness practitioners, give personalized guidance suggestions.
The needs of gym venues. Gym managers are eager to improve the scientific and
technological level of gym, and intelligent functions can attract a larger number of
users.
Hardware Layer. It consists of camera, display screen, server and user terminal. The
camera is responsible for obtaining user’s facial information and user’s action resource
data. The display screen is used to show the user’s action in real time, and hide the
Intelligent Fitness Trainer System Based on Human Pose Estimation 595
camera position. The user can practice the action in front of the display screen without
looking for the camera position. Server is our data processing center. We build user
database, implement alphapose platform for human pose estimation and design
movements evaluation and correction algorithm. User terminals display various ser-
vices to users and display information processing. User’s historical information and
operation behavior will be displayed and output in the terminal.
Data Layer. Data layer is divided into three parts: image data, course data and user
data. User data includes user’s face data, physical fitness data, history course data and
score data. Users can freely access personal data after matching face information
correctly (in the future, they can support sharing or backup operations). Image
resources are user posture information acquired by camera, which is uploaded to the
server during user fitness process. After image data processing, the system will give a
score and update user data.
Application Layer. The application layer provides various services, including training
courses selection and allow users to choose and match freely. Physical fitness analysis
and recommendation of personalized courses or course packages are provided; real-
time analysis and correction of user posture will be carried out during training. After
training, the training will be scored and guidance suggestions will be put forward.
Users can also view their history courses.
Presentation Layer. The presentation layer is divided into three parts, web end, APP
and desktop client.
2.2 Introduction
Intelligent fitness trainer system is a self-learning and evaluation system for fitness. It
has the functions of motion capture, correction reminder, face recognition, behavior
recording and so on. It is suitable for gyms. The system achieves the self-learning by
motion capture, correcting and reminding the fitness worker according to the standard
movement coursed, evaluating the fitness worker after the completion of fitness by
intelligent learning, perfecting personal fitness files through face recognition and
behavior records, and giving fitness suggestions by using recommendation algorithm.
The system aims to provide users with personalized fitness guidance, develop good
sports habits, while increasing their interest. And the physical map of our system is
shown in Fig. 2.
2.3 Principle
Figure 3 shows the principle and working process.
Firstly, we use the camera to recognize the users by face recognition. The server
can provide the user database information, understand its usage habits, and recommend
courses from the course resource database, and give a pre-warning of the habitual errors
in the user’s history.
596 J. Zou et al.
After the motion starts, the system will call the system motion module. First, the
server reads the images captured by the camera and process them frame by frame. By
deep learning alpha pose algorithm, human pose estimation is applied, human body is
recognized, key points of human joints are marked, and position information of user
joints in each frame is obtained. Then, it compares with the standard action to realize
the action correction. The results are shown in Fig. 4.
At the end of the whole exercise process, we will generate feedback reports,
including details of movement errors, corrective guidance and motion accuracy scores,
maximum error points, best progress points, calorie consumption, etc., to enhance the
users’ sense of fitness achievement and interest in fitness.
Intelligent Fitness Trainer System Based on Human Pose Estimation 597
3 Key Technology
In our system, the key technology is human pose estimation, which determines the
accuracy of pose recognition and has vital effect on movement correction.
Human pose estimation is a popular topic of computer vision. Compared with
traditional algorithms, algorithms based on deep learning train the networks with a
large amount of images, learn features from global space. It needn’t detect or recognize
the local feature of objects and has good robustness [1]. In practice, it is more chal-
lenging to recognize multiple people’s postures than to recognize individual people’s
postures in images. [2, 3]. In general, two frameworks for this problem have been
proposed: one is a two-step framework [4], which first detects human bounding boxes
and then the posture in each box is estimated independently; the other is a part-based
framework [5–7], which detects body parts firstly and then assemble the detected body
parts into multiple human postures.
As for video scene reconstruction framework, [9] proposes a framework for col-
laborative tracking of multiple human objects and estimating their three-dimensional
posture. [10] eliminates the re-projection error, avoids the optimization of noise
observation, and introduces geometric constraints on prior knowledge composed of
human joints at reference points.
We apply the two-step framework. We apply a regional multi-person pose esti-
mation (RMPE) framework proposed in [8]. And the framework is shown in Fig. 5.
598 J. Zou et al.
4 Conclusion
In this paper, we introduce a kind of intelligent fitness trainer system based on human
pose estimation. We use the optical camera to capture users’ movement, and achieve
pose estimation by regional multi-person pose estimation framework. The field trail
shows that our system has a good influence on fitness training, especially on accuracy
of movements. Besides, although our method makes some efforts in improvement of
RMPE algorithm, more work still need to be done in deeper research.
References
1. Liu, Y., Xu, Y., Li, S.: 2-D human pose estimation from images based on deep learning: a
review. In: 2018 2nd IEEE Advanced Information Management, Communicates, Electronic
and Automation Control Conference (IMCEC), Xi’an, pp. 462–465 (2018)
2. Sapp, Toshev, A., Taskar, B.: Cascaded models for articulated pose estimation. In: European
Conference on Computer Vision (ECCV), pp. 406–420. Springer (2010)
3. Sun, M., Kohli, P., Shotton, J.: Conditional regression forests for human pose estimation. In:
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3394–3401.
IEEE (2012)
4. Gkioxari, G., Hariharan, B., Girshick, R., Malik, J.: Using k-poselets for detecting people
and localizing their keypoints. In: IEEE Conference on Computer Vision and Pattern
Recognition (CVPR), pp. 3582–3589 (2014)
5. Chen, X., Yuille, A. L.: Parsing occluded people by flexible compositions. In: IEEE
Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3945–3954 (2015)
6. Pishchulin, L., Insafutdinov, E., Tang, S., Andres, B., Andriluka, M., Gehler, P., Schiele, B.:
Deepcut: joint subset partition and labeling for multi person pose estimation. In: IEEE
Conference on Computer Vision and Pattern Recognition (CVPR), June 2016
7. Insafutdinov, E., Pishchulin, L., Andres, B., Andriluka, M., Schiele, B.: DeeperCut: a
deeper, stronger, and faster multi-person pose estimation model. In: European Conference on
Computer Vision (ECCV), May 2016
8. Fang, H., Xie, S., Tai, Y., Lu, C.: RMPE: regional multi-person pose estimation. In: 2017
IEEE International Conference on Computer Vision (ICCV), Venice, pp. 2353–2362 (2017)
Intelligent Fitness Trainer System Based on Human Pose Estimation 599
9. Tang, Z., Gu, R., Hwang, J.: Joint multi-view people tracking and pose estimation for 3D
scene reconstruction. In: 2018 IEEE International Conference on Multimedia and Expo
(ICME), San Diego, CA, pp. 1–6 (2018)
10. Takahashi, K., Mikami, D., Isogawa, M., Kimata, H.: Human pose as calibration pattern: 3D
human pose estimation with multiple unsynchronized and uncalibrated cameras. In: 2018
IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW),
Salt Lake City, UT, USA, pp. 1856–18567 (2018)
The Overview of Multi-person Pose
Estimation Method
1 Introduction
Common top-down methods for multi-person pose estimation [4–6] are to use
person detectors, within which the estimation for single-person pose [7–9] is adopted.
The major problem of the top-down method is that if a person detector goes wrong,
there is no means of recovery. Due to the principle of this method, the running time is
in proportion to the number of people. Bottom-up methods, by contrast, are more
robust in theory and tend to weaken the relationship between the number of people in
the image and the runtime complexity. However, the bottom-up method does not utilize
global context directly from other individuals. Bottom-up approaches [10, 11] in
previous practice did not keep efficiency advantages because of the costly global
inference in eventual parsing. For example, [10] proposed a bottom-up approach to
detect candidates and associate them with individuals. [11] developed on the basis of
[10] with more solid partial detectors featured by Residual Neural Network (ResNet)
[12] and pairwise scoring algorithm dependent on images, which greatly improves the
runtime but needs a separate logistic regression due to the difficulty for pairwise
representation to regress accurately.
The rest of the paper is organized as follows: Sect. 2 introduces two top-down
methods, DeepCut and DeeperCut. Section 3 briefly describes a bottom-up method
OpenPose. Section 4 presents the method AlphaPose used in our Intelligent Fitness
Trainer System. Finally, Sect. 5 concludes the whole paper.
2.1 DeepCut
In [10], the top-down method is used to estimate the pose of many people. The so-
called top-down method is, first, to use Convolutional Neural Networks (CNN) to
detect human body, that is, body part candidates, and second, to determine whom these
joint points belong to. Finally, the Integer Linear Programming (ILP) optimization
model is used to estimate attitude. There are overlapping parts in the execution order of
the two steps.
The DeepCut model is illustrated in Fig. 1. Firstly, CNN is used to extract body
parts candidates. Each candidate region corresponds to a joint point. Each joint point
acts as a node in the graph. All the nodes represented by these candidate joints con-
stitute a complete graph, as shown in dense graph in Fig. 1. The correlation between
nodes is used as the weight between nodes in the graph. At this time, it can be regarded
as an optimization problem. The joints belonging to the same person (the nodes in the
figure) can be grouped into a group with each person as a separate class. Meanwhile,
another branch needs to mark the detected nodes to determine which part of the human
body they belong to. Finally, each person’s posture estimation is made up of classified
people and tagged parts.
The advantages of the DeepCut model can be summarized in three points. The first
is that the problem of multi-person pose estimation can be solved in the case of number
position. The distribution of each person’s nodes can be obtained by classification.
Secondly, by clustering graph-theoretic nodes, non-maximum suppression is effectively
used. Thirdly, the optimization problem is expressed as an ILP problem, which can be
effectively solved by mathematical methods.
602 B. Li et al.
However, the computational complexity of DeepCut is very costly due to the use of
adaptive fast R-CNN for body detection and ILP for pose estimation, so the following
method DeeperCut is an accelerated implementation of DeepCut.
2.2 DeeperCut
[11] proposed an improved method, DeeperCut, based on DeepCut. The improvements
are mainly on the following two aspects: (1) Using the latest residual net to extract
body part, which makes the outcome more accurate and the precision higher. (2) Using
image-Conditioned Pairwise Terms method, which can compress the nodes of many
candidate regions to fewer nodes, and that is the major reason why DeeperCut can be
stronger and faster than DeepCut. The principle of the image-Conditioned Pairwise
Terms method is to judge whether the candidate nodes are the same important node by
the distance between them.
context and allows greedy bottom-up parsing steps, which achieves high accuracy as
well as real-time performance regardless of the number of people in the image. This
structure is designed to learn part positions and association by two branches of the
prediction process in the same sequence. The approach proposed in [13] ranks first in the
first COCO 2016 Key points Challenge. The model of real-time multi-person 2D pose
estimation using part affinity fields is shown in Fig. 2.
Fig. 2. The model of real-time multi-person 2D pose estimation using PAFs [13].
The main ideas applied in this method are as follows. The first idea is using
confidence maps for joint detection. Each joint corresponds to a confidence map. Each
pixel has a confidence level. Values of each point in the confidence map is related to
the distance of ground truth. For multi-person detection, K-person confidence maps are
merged to get the maximum of each person at that point. The maximum rather than the
average is used here because the accuracy is not affected even if the peak value is very
close. In the test phase, non-maximum suppression is used to obtain candidates for
body parts.
The second idea is using Part Affinity Fields (PAFs) for body composition. For
multi-person problem, the parts of different people are detected, but each person’s body
needs to be separately combined to form a full-body, where the method PAF is used.
The advantage of this method is to include both location and direction information.
Each limb has an affinity region between two related body parts, and each pixel has a
description direction of a 2D vector. If there are more than one person overlapping at a
certain point, the vector sum of these individuals is divided by the number of people.
After obtaining confidence maps and PAFs, how to use such information to find the
optimal connection between two body-parts should be taken into consideration, which
is transformed into a graph theory problem. Hungarian algorithm used here is the third
idea. The nodes in the graph are the detection candidates in the body part, and the edges
are the optimal connections of these candidates. The weight on each edge is the
aggregation of affinity regions. So the next problem is to find a group of connections so
that no two edges share a node, that is, to look for a way of connection for edges with
largest weight.
604 B. Li et al.
The methods DeepCut [10], DeeperCut [11] and OpenPose [13] introduced in Sect. 2
and Sect. 3 all have the problem that if two people are very close, they are prone to get
ambiguity, whether they are top-down methods or bottom-up methods. In addition,
because they depend on the relationship between the two components, they lose access
to global information. Due to this problem, [17] proposed an improved method based
on top-down methods.
[17] aimed to detect the correct pose of human body, even if the inaccurate area
frame is detected in the first step. To illustrate the problems of previous algorithms,
Faster-RCNN and Single Person Pose Estimation (SPPE) Stacked Hourglass are used
for experiments. The main problems are location recognition errors and redundancy.
Actually, SPPE is easily affected by area frame errors that redundant area frames may
create redundant poses. Therefore, a Regional Multi-Person Pose Estimation (RMPE)
framework is introduced to improve SPPE-based performance. Adding symmetric
spatial transformer network (SSTN) to SPPE structure can pick out human body parts
in inaccurate region frames with high quality. Parallel SPPE branches (SSTNs) are used
to optimize their networks. Parametric pose non-maximum suppression (NMS) is used
as a solution to redundancy detection. In this structure, an ordinary pose distance
measurement scheme is applied to compare the similarity between poses. Data-driven
method is used for the optimization of distance parameters. Finally, Pose-Guided
Proposals Generator (PGPG) is introduced to enhance training data, and learn the
descriptive information of different poses in output results to imitate the generation
process of the human body region frame and further produce a larger training dataset.
In details, the method is made up of three parts. The first part is Symmetric STN
and Parallel SPPE. The human body region frame obtained by target detection algo-
rithm is not very suitable for SPPE in that SPPE algorithm is trained on a single image
and is rather sensitive to location errors. The effect of SPPE can be effectively
improved by micro-transformation and pruning. SSTN + Parallel SPPE can effectively
enhance the effect of SPPE under imperfect human body area detection results. The
structure is shown in Fig. 3.
After STN + SPPE + SDTN, the inaccurate detection frame first conducts pose
estimation and maps the estimation results to the original map, so as to adjust the
original frame and make the frame become accurate.
The Overview of Multi-person Pose Estimation Method 605
The second part is Parametric Pose NMS. Human location inevitably results in
redundant detection frames and redundant attitude detection. Due to this phenomenon,
[17] proposed a non-maximum suppression for pose to eliminate redundancy. First, the
pose of maximum confidence is selected as the reference, and the region frame close to
the reference is eliminated according to the elimination criterion. This process is
repeated several times until redundant recognition boxes are eliminated and each
recognition box is unique (no overlap beyond the threshold).
The third part is Data Augmentation. For Two-Stage pose estimation (first locating
the region, then locating the pose points), proper data enhancement can help
SSTN + SPPE to adapt to imperfect human body region location results. Otherwise,
the model may not be well adapted to strange human body location results when it is
running in the test phase. An intuitive method is to use the detected area frame in the
training phase. However, target detection only produces a localization area for a person.
By using the generated human body localization, a certain effect can be achieved.
Because the real location of each person and the detected location box have already
been generated, a large training set of samples can be created by using the same
samples as the human body test results. Through this technology, system performance
can be further improved.
The results of the four methods – DeepCut [10], DeeperCut [11], OpenPose [13],
AlphaPose [17] on the MPII Human Pose dataset is presented in Table 1. The statistics
in the table illustrate that AlphaPose is 17 mean average precision (mAP) higher than
DeepCut and DeeperCut. Quantitative results on full testing set and subset of 288
testing images used in [10]. It is noteworthy that the average accuracy of AlphaPose in
recognizing complex joints such as wrist, elbow, ankle and knee is 72 mAP, which is
20 mAP higher than the previous latest results. The final wrist accuracy of AlphaPose is
70.4 mAP and knee accuracy is 73 mAP. Some results of AlphaPose is shown in
Fig. 4. These results show that AlphaPose can accurately predict the attitude of multi-
person images.
606 B. Li et al.
5 Conclusion
In test phases, DeepCut, DeeperCut and OpenPose frequently have errors in pose
location and joints connection, while AlphaPose perform better with less errors and
faster runtime speed. AlphaPose is a relatively accurate multi-person pose estimation
tool compared with previous methods from the robustness of its algorithm. AlphaPose
can estimate poses in pictures, videos or multiple pictures, and track motion in pictures.
It has different output forms, such as PNG, JPG and AVI, which have key image forms,
as well as JSON format, which makes it a popular tool for many applications.
The Overview of Multi-person Pose Estimation Method 607
References
1. Felzenszwalb, P.F., Huttenlocher, D.P.: Pictorial structures for object recognition. In: IJCV
(2005)
2. Andriluka, M., Roth, S., Schiele, B.: Monocular 3D pose estimation and tracking by
detection. In: CVPR (2010)
3. Andriluka, M., Roth, S., Schiele, B.: Pictorial structures revisited: people detection and
articulated pose estimation. In: CVPR (2009)
4. Pishchulin, L., Jain, A., Andriluka, M., Thormahlen, T., Schiele, B.: Articulated people
detection and pose estimation: reshaping the future. In: CVPR (2012)
5. Gkioxari, G., Hariharan, B., Girshick, R., Malik, J.: Using kposelets for detecting people and
localizing their keypoints. In: CVPR (2014)
6. Sun, M., Savarese, S.: Articulated part-based model for joint object detection and pose
estimation. In: ICCV (2011)
7. Newell, A., Yang, K., Deng, J.: Stacked hourglass networks for human pose estimation. In:
ECCV (2016)
8. Wei, S.-E., Ramakrishna, V., Kanade, T., Sheikh, Y.: Convolutional pose machines. In:
CVPR (2016)
9. Ouyang, W., Chu, X., Wang, X.: Multi-source deep learning for human pose estimation. In:
CVPR (2014)
10. Pishchulin, L., Insafutdinov, E., Tang, S., Andres, B., Andriluka, M., Gehler, P., Schiele, B.:
Deepcut: joint subset partition and labeling for multi person pose estimation. In: CVPR
(2016)
11. Insafutdinov, E., Pishchulin, L., Andres, B., Andriluka, M., Schiele, B.: Deepercut: a deeper,
stronger, and faster multi-person pose estimation model. In: ECCV (2016)
12. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR
(2016)
13. Cao, Z., Simon, T., Wei, S., Sheikh, Y.: Real-time multi-person 2D pose estimation using
part affinity fields. In: CVPR (2017)
14. Wang, H., An, W.P., Wang, X., Fang, L., Yuan, J.: Magnify-net for multi-person 2D pose
estimation. In: ICME (2018)
15. Chen, X., Yang, G.: Multi-person pose estimation with LIMB detection heatmaps. In: ICIP
(2018)
16. Mehta, D., Sotnychenko, O., Mueller, F., Xu, W., Sridhar, S., Pons-Moll, G., Theobalt, C.:
Single-shot multi-person 3D pose estimation from monocular RGB. In: 3DV (2018)
17. Fang, H., Xie, S., Tai, Y., Lu, C.: RMPE: regional multi-person pose estimation. In: ICCV
(2017)
A Network Information Data Protection
Scheme Based on Superposition Encryption
Liu Zhe1,2(&)
1
School of Information and Electronics,
Beijing Institute of Technology, Beijing, China
stayhere1974@163.com
2
General Office of Supreme People’s Procuratorate, Beijing, China
Abstract. The wide application of virtualization, big data, cloud storage and
other technologies has made the data centralized storage increase security risks
and leakage risks. For the problem of internal personnel using legal identity and
role to steal internal information, this paper proposes a network information data
protection architecture based on superposition encryption from the process of
data generation, transmission, storage and processing. Based on the framework,
including information storage protection based on symmetric cryptography and
information transmission protection based on hybrid cryptography, a network
information data protection scheme based on superimposed encryption is
designed. Finally, the security of the scheme is proved.
1 Introduction
With the development of big data, cloud storage, virtualization and other technologies,
great changes have taken place for the information application model, in which a large
number of data are centralized on the server, and the end users retrieve data from the
server through application systems or virtualization technologies. This application
mode enables information to be shared more fully, greatly improving work efficiency,
but also bringing greater risks and hidden dangers to security and confidentiality, which
is partly due to information attacks through external attacks on the network, on the one
hand, due to internal personnel using work convenience to steal information.
In this context, many industries and enterprises in many fields at home and abroad
are facing increasingly severe problems of data information security, and many
scholars are constantly exploring solutions for data security protection in the era of big
data. Paper [2] analyzes the security requirements of data services and studies the
existed data protection scheme and mechanism. Combining the cloud service charac-
teristics and data services process, a secure solutions for data service in cloud needs to
meet the requirement: re-encryption method under trusted cloud service providers.
Paper [3] proposes the security framework based on the integration of cloud and client,
the cloud cooperate with the client each other to achieve the protection for user data in
the cloud, that is, the cloud will be responsible for protecting the security of the user
data in the cloud, meanwhile the client will be responsible for verifying and proving the
data safety. Yan et al. [4] discusses various aspects of cryptography and data security in
cloud computing, including secure cloud data storage, cloud data privacy protection,
trusted cloud data management and cryptography related to cloud data security. Paper
[5] brings a critical comparative analysis of cryptographic defense mechanisms, and
explores research directions and technology trends to address the protection of out-
sourced data in cloud infrastructures. Paper [6] proposes advanced algorithm for
cryptography which is totally depend on hashing function technique to generate a
secret key which is further used to encrypt and decrypt the important information.
Rajesh et al. [7] proposes a well-organized privacy preservation data-mining scheme
with data-mining perturbation merged approach, and uses the association rules with
cryptography techniques. Paper [8] presents AES4SeC, a security scheme fully con-
structed over cryptographic pairings. Paper [9] introduces an algorithm-agnostic
solution that provides both data integrity and confidentiality protection at the disk
sector layer. Yi et al. [10] proposes an efficient hardware architecture based on mul-
tivariate scheme for storage devices.
For the problem of internal personnel using legal identity and role to steal internal
information, this paper proposes a network information data protection system archi-
tecture based on superposition encryption, which includes information storage pro-
tection based on symmetric cryptography and information transmission protection
based on hybrid cryptography, then a network information data protection scheme
based on superimposed encryption is designed. Finally, the security of the scheme is
proved. The scheme achieves effective protection on the basis of satisfying the normal
application of information technology, which can ensure the security of information
data, especially the centralized storage data in the network.
and flexibility of symmetric cryptography, but also fully utilizes the security of hybrid
cryptography. In the case of ensuring the security of information storage, the appli-
cation is more flexible and meets actual needs.
The information is stored centrally on the server and protected by symmetric cryp-
tography, where the information plaintext is converted into cipher text when stored, and
then converted into plaintext after being read from the storage device.
perform finer-grained area division, and use more intensive encryption algorithms or
keys. This fine-grained use of symmetric cryptography encrypts the stored files, which
can achieve “one thing (text) and one key” under the highest security.
Storage
Hybrid cryptography is only used during the information transmission protection, and
the symmetric cryptographic key is used once and randomly generated for each time.
Thus, the symmetric cryptographic keys are different for each transmission even for the
same file, thus achieving “one person one key” and “one thing one key” and “one time
one key”.
C1 ¼ E0K0 ðPÞ
Where y2Bn, Bn is the user set, k°yʹ is the public key (encryption key) for user y, E°
is a asymmetric cryptographic encryption operation; M is the ciphertext of the sym-
metric cryptographic key encrypted by the asymmetric cryptographic operation.
Finally, the information ciphertext encrypted by the symmetric cryptography and
the symmetric key encrypted by the asymmetric cryptography are superimposed.
f:C ¼ C1 þ M
f 1 :C1 ¼ C M, M ¼ C C1
Where y 2 Bn, Bn is the user set; k°y is the private key (decryption key) of the user
y; Kʹ is the symmetric cryptographic key; M is the ciphertext of the symmetric cryp-
tographic key encrypted by the asymmetric cryptography.
Finally, use the asymmetric cryptography decrypt the obtained symmetric crypto-
graphic key, and decrypt the information ciphertext to obtain the information plaintext,
that is:
P ¼ D0K0 ðC1 Þ
5 Conclusion
The superposition of the symmetric cryptography and the hybrid cryptography is used
to protect information, which solves the problem of security and confidentiality in
information centralization to a certain extent, but also affects the normal implemen-
tation of some functions in application, such as the information retrieved by the
A Network Information Data Protection Scheme 615
application system is ciphertext data, which cannot be counted and queried. This
cryptography protection mode is mainly based on the protection of information
including core sensitive content, such as daily office files, which is more suitable for the
protection of unstructured data such as files, audio and video.
For the information security and confidentiality requirements in information
application, this paper proposes a cryptographic protection scheme based on the
superposition encryption of the symmetric cryptography and the hybrid cryptography,
which combines “one person one key” and “one text one key” to ensure the security of
information transmission due to the use of the hybrid cryptography; storage and
transmission encryption are separated from each other due to the use of superposition
encryption. This paper demonstrates the feasibility of the scheme from system con-
struction, scheme design, technology implementation, security protection, etc., which
can meet the information security protection in the information application of gov-
ernment agencies and other units, and can effectively guarantee information security of
the application modes of virtualization, big data, cloud storage.
References
1. Sibi Chakkaravarthy, S., Sangeetha, D., Venkata Rathnam, M., Srinithi, K., Vaidehi, V.:
Futuristic cyber-attacks. Int. J. Knowl.-Based Intell. Eng. Syst. 22(3), 195–204 (2018)
2. GuXin.: Studies on trusted secure data services under cloud environment. Wuhan University
(2013)
3. MaWenqi.: Research on the cloud and client fusion based user’s data security technology in
cloud storage. National University of Defence Technology (2016)
4. Yan, Z., Deng, R.H., Varadharajan, V.: Cryptography and data security in cloud computing.
Inf. Sci. 387, 53–55 (2017)
5. Kaaniche, N., Laurent, M.: Data security and privacy preservation in cloud storage
environments based on cryptographic mechanisms. Comput. Commun. (2017).
S014036641730796X
6. Kumar, M.: Advanced RSA cryptographic algorithm for improving data security. In: Golden
Jubilee Convention CSI- on “digital Life.” December 2018
7. Rajesh, N., Selvakumar, A.A.L.: Association rules and deep learning for cryptographic
algorithm in privacy preserving data mining. Cluster Comput. 2, 1–13 (2018)
8. Morales-Sandoval, M., Gonzalez-Compean, J.L., Diaz-Perez, A., et al.: A pairing-based
cryptographic approach for data security in the cloud. Int. J. Inf. Secur. 17(4), 441–461 (2018)
9. Broz, M., Patocka, M., Matyas, V.: Practical cryptographic data integrity protection with full
disk encryption extended version (2018)
10. Yi, H., Nie, Z.: Towards data storage in cryptographic systems: an efficient hardware
architecture based on multivariate scheme for secure storage applications. Cluster Comput.
(2018)
Author Index
D H
Dai, Kan, 446 Han, Dong, 124
Dai, Yuliang, 551 Han, Jindong, 32
Deng, Xiangjin, 249 Han, Xiaojun, 291
Ding, Zhenpeng, 343 Han, Yuehua, 495, 505
Ding, Zhimin, 227 Hao, Jianing, 154
Dong, Jiangbo, 78 He, Yongcong, 188
Dong, Mingtao, 32 He, Yuan, 32
K M
Kang, Nan, 570 Ma, Jie, 534, 542
Kong, Peng, 131 Ma, Xiaomei, 69
Ma, Xuekuan, 455
L Ma, Yan, 41
Lei, Rongjia, 593, 600 Mao, Yilan, 310
Lei, Yong, 146, 310 Meng, Weiqiang, 3
Li, Bingyi, 578, 584, 593, 600 Miao, Qi, 343, 371
Li, Fangyong, 360 Mo, Fan, 276
Li, Haibo, 139
Li, Hao, 336
Li, Luming, 91 N
Li, Nan, 227 Ni, Kaili, 69
Li, Xiangyuan, 593, 600 Niu, Jiaxiang, 91
Li, Xiao, 54
Li, Xiaojuan, 115 P
Li, Yan, 139 Pan, Huifang, 212
Li, Yifan, 164 Pan, Li, 146
Li, Yong, 525 Peng, Liqiang, 284
Li, Yongchang, 154
Li, Yue, 593, 600 Q
Li, Zhidong, 219 Qiao, Lei, 13
Liang, Deyin, 310 Qin, Huafeng, 525
Liang, Yin, 400 Qin, Suran, 380, 391
Liu, Bo, 13 Qiu, Yaxing, 437
Liu, Couhua, 446 Qu, Quanyou, 310
Author Index 619
R X
Ran, Qianhan, 22 Xia, Jixia, 360
Rao, Xiaoqin, 462 Xie, Chao, 455
Ren, Liang, 188 Xie, Hongrui, 61
Ren, Lu, 352 Xing, Huanlai, 411, 420
Ren, Yebing, 78 Xu, Jian, 13
Ren, Zhenyue, 173 Xu, Kai, 336
Ren, Zhong, 291 Xu, Lexi, 411, 516
Ru, Xiaoyi, 343, 371 Xu, Menglei, 525
Xu, Ran, 462
S Xu, Zetao, 495
Shen, Ao, 495 Xue, Feng, 173
Shi, Haitao, 352
Shi, Jian, 100 Y
Shi, Jianmin, 380 Yan, Hongcheng, 91
Shi, Wei, 249 Yang, Chunsheng, 173
Song, Ding, 391 Yang, Feng, 188
Song, Lei, 124 Yang, Hai, 411, 420
Song, Wenlong, 212 Yang, Hongyu, 291, 317
Sun, Bo, 219 Yang, Hua, 13
Sun, Gang, 360 Yang, Liuqing, 115, 284
Sun, Songlin, 22, 69, 593, 600 Yang, Yi, 300
Sun, Tengfei, 164 Yao, Jian, 3
Sun, Zhelei, 115, 284 Yao, Meng, 249
Ye, Chengzhi, 559
T Yin, Jinsong, 479
Tan, Tian, 100 Yin, Shan, 534, 542
Tang, Jian, 446 You, Yanan, 131
Tang, Yang, 41 Yu, Chao, 446
Tian, Hexiang, 131 Yu, Lei, 164
Tong, Yelong, 164 Yu, Wei, 343
Toth, Zoltan, 487 Yuan, Lili, 391
Yuan, Yi, 212, 343
W Yun, Lei, 391
Wan, Yuting, 105
Wang, Fengjun, 437 Z
Wang, Guozhi, 505 Zeng, Chunping, 100, 260
Wang, Haobin, 78 Zhan, Tianming, 429, 551
Wang, Hui, 471 Zhang, Bihui, 462
Wang, Luyao, 593, 600 Zhang, Boyin, 400
Wang, Qiang, 180, 204, 400 Zhang, Chenhua, 300
Wang, Qianying, 276 Zhang, Hengde, 429, 462, 551
Wang, Teng, 195 Zhang, Jingjing, 105
Wang, Wenping, 146 Zhang, Lei, 219
Wang, Xiaonan, 131 Zhang, Lijian, 326, 336
Wang, Xidong, 437 Zhang, Miao, 227
Wang, Xingyan, 400 Zhang, Qiang, 131
Wang, Xinhan, 420 Zhang, Rui, 91
Wang, Yi, 487 Zhang, Shao-po, 124
Wei, Heng, 479 Zhang, Shuai, 219
Wei, Huangfu, 78 Zhang, Tao, 115, 516
Wu, Feixia, 3 Zhang, Tianhang, 462
Wu, Lequn, 300 Zhang, Tianwei, 284
620 Author Index