Professional Documents
Culture Documents
BI and TI
BI and TI
BI and TI
Business
Intelligence and
Information
Technology
Proceedings of the International
Conference on Business Intelligence
and Information Technology BIIT 2021
Lecture Notes on Data Engineering
and Communications Technologies
Volume 107
Series Editor
Fatos Xhafa, Technical University of Catalonia, Barcelona, Spain
The aim of the book series is to present cutting edge engineering approaches to data
technologies and communications. It will publish latest advances on the engineering
task of building and deploying distributed, scalable and reliable data infrastructures
and communication systems.
The series will have a prominent applied focus on data technologies and
communications with aim to promote the bridging from fundamental research on
data science and networking to data engineering and communications that lead to
industry products, business knowledge and standardisation.
Indexed by SCOPUS, INSPEC, EI Compendex.
All books published in the series are submitted for consideration in Web of Science.
Zhipeng Fan
Editors
Business Intelligence
and Information Technology
Proceedings of the International Conference
on Business Intelligence and Information
Technology BIIT 2021
123
Editors
Aboul Ella Hassanien Yaoqun Xu
Information Technology Department School of Computer and Information
Cairo University Engineering
Giza, Egypt Harbin University of Commerce
Harbin, Heilongjiang, China
Zhijie Zhao
School of Computer and Information Sabah Mohammed
Engineering Department of Computer Science
Harbin University of Commerce Lakehead University
Harbin, Heilongjiang, China Thunder Bay, ON, Canada
Zhipeng Fan
School of Computer and Information
Engineering
Harbin University of Commerce
Harbin, Heilongjiang, China
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
v
vi Preface
General Chair
General Co-chairs
Sabah Mohammed Lakehead University, Canada
Ping Han Harbin University of Commerce, China
Wei Wang Harbin University of Commerce, China
Shizhen Bai Harbin University of Commerce, China
Fengge Yao Harbin University of Commerce, China
Zeguo Qiu Harbin University of Commerce, China
Publication Chair
Aboul Ella Hassanien SRGE, Egypt
Publication Co-chair
Zhipeng Fan Harbin University of Commerce, China
vii
viii Organization
xi
xii Contents
1 Introduction
China’s “14th five-year plan” emphasizes the need to promote the deep integration of
the Digital Economy and the real economy, enabling the transformation and promoting
of traditional manufacturing industries, engendering new industries, new modes of busi-
ness, and new engines of economic development. Manufacturing is the most important
part of the real economy. Promoting the high-quality development of the manufacturing
industry is an important strategic measure to strengthen the country through manufactur-
ing and speed up the construction of a new development pattern. This paper discusses the
internal mechanism of the influence of the digital economy on the manufacturing industry
and applies a panel data model to analyze the heterogeneity of regional influence.
The measurement system of the digital economy is developed around three dimen-
sions: The digital infrastructure economy, the scale of the digital economy, and the appli-
cation of mobile digitalization (Wei Zhuangyu 2020) [1] Xu Xianchun(2020) believe that
we can be defined and calculated the scope of the Digital Economy by using the structure
coefficient of industry value-added, the adjustment coefficient of the digital economy and
the increase rate of industry [2] Yu Shan and Fan Xiufeng (2021) considered the selection
of secondary indexes can select from the development of communication industry, such
as the Internet penetration rate, the total amount of communication service, the length
of mobile optical cable and so on [3].
The impact of the transformation and upgrading of the digital economy enabling
the manufacturing industry is studied mainly from two aspects. One is how the digital
economy improves the quality and efficiency of manufacturing; the other is how the
digital economy promotes the value creation of manufacturing. Li Chunfa(2020) the
digital economy can expand the boundaries of a manufacturing division of labor, reduce
transaction costs, transfer value distribution, and improve manufacturing productivity.
It can also cause changes in demand and stimulate consumption potential, forcing the
manufacturing sector to upgrade [4]. Jiao Yong (2020) holds that the impact of the digital
economy on the manufacturing industry can be divided into short-term and long-term.
The short-term impact is value remolding, and the long-term impact is value creation.
In the form of factor-driven to innovation-driven, product-driven to demand-driven shift
[5]. Li Yingjie and Han Ping (2021) believe that the mechanism of the digital economy
promoting the high-quality development of the manufacturing industry is mainly through
quality change, efficiency change, and power change [6].
The digital economy can reduce the transaction cost of the manufacturing industry: First,
it can reduce the Information asymmetry between enterprises and enterprises, enterprises
and consumers, enterprises, and the allocation of production factors. In the era of the
Digital Economy, the rapid spread of technologies such as big data and the Internet has
greatly improved the transparency and openness of the market while reducing the prob-
lems of “moral hazard” and “adverse selection” arising from Information asymmetry
between the consumer and the producer, between the producer and the production factor
Digital Economy and High-Quality Development of Manufacturing Industry 5
correspondence degree enhances greatly, the entire society’s factor allocation efficiency
enhances obviously. Second, the birth of the giant platform economy itself is the embod-
iment of reducing transaction costs. Essentially, the platform business content is not the
product itself but “reduces transaction costs. “. Whether it is “sharing economic model”,
or online shopping platform, takeaway platform. The core idea is to collect idle social
resources and establish a communication platform between producers and consumers
to reduce the transaction costs of both sides and improve the matching degree between
supply and demand.
The digital economy has stimulated demand creation in the manufacturing sector, which
is mainly reflected in: First, the rise of big data analytics, Internet platforms that can accu-
rately depict consumer preferences, as a special production factor, consumer demand
preference flows backward along the industrial chain, so the producer can catch the
change of consumption level more quickly and adjust the production plan better. Sec-
ond, with the rapid integration of traditional industry and the Internet, the manufacturing
industry has undergone great changes, gradually transforming from large-scale assembly
line production to Mass customization production. Through digital information tech-
nology, the producer collects and analyzes the product’s complete life cycle [8]. The
producer’s demand is communicated immediately, and the production line realizes the
flexible adjustment by the data flow. In conclusion, the research hypothesis of this paper
is put forward:
3 Status Analysis
3.1 Index System Design
Determination of the Weight of the Indicator System. Based on the five develop-
ment concepts of “innovation, coordination, green, open and sharing”, a comprehen-
sive index system for high-quality development of manufacturing industry has been
established, as well as from the digital economy infrastructure development level, the
digitization popularization dynamics, the digital economy terminal application depth to
the digital economy development level measurement, the concrete steps are as follows:
n
Yij ln(Yij )
i=1
Step 3 : Calculate the information entropy of index J: ej = − (3)
ln(n)
dj
Step 5 : Calculate the weights : wj = (5)
m
dj
j=1
Using the above method to calculate the index’s weight in the evaluation system, the
concrete results are shown in Table 3.
2019, the score rose to 5.95. From the perspective of regional economic development, the
development level of China’s digital economy presents obvious regional heterogeneity.
The development level of the Eastern Region is higher than that of the rest of the country.
The growth rate of the northeast region is relatively slow, the growth rate of the digital
economy in central China is relatively fast, and the development level of other regions
is relatively balanced except Shanxi. The western region is the most backward region
in the development of the digital economy in China. Still, it has a good momentum of
development and shows a rising trend year after year.
4 Empirical Analysis
4.1 Variable Design
1. Explanatory variable: High Level of quality development in manufacturing. Indicator
selection concerning the above measurement results.
2. The core explanatory variable: The level of development of the digital economy. The
above measurements inform the selection of indicators.
3. Control Variables: To reduce the lack of robustness of the measurement model caused
by missing variables, this paper selects the following control variables according to
the current academic research results: (1) human capital (Edu). The paper takes
primary, secondary, high, and higher education into consideration. (2) foreign direct
investment: This paper uses the ratio of provincial-level FDI to provincial-level
GDP as the proxy variable. (3) Government support (Gov). The government’s role
in transforming and upgrading the manufacturing sector is expressed in the share of
annual regional fiscal expenditure in GDP. (4) Urbanization (Urb). The proportion
of the urban population to the region’s total population will be taken as the basis
for measuring regional Urbanization. (5) Financial Development (Fin). This paper
measures the balance of deposits and loans to the regional GDP at the end of the
year. (6) the level of Economic Development (Dev). This paper uses GDP per capita
to measure the level of economic development.
In the model (6), yit represents the manufacturing high-quality development level of
area i at time t, lndigitalit represents the digital economy development level of area i at
time t, C is a constant term, J is the number of control variables, and X is the control
variable, μi , νt , εit they are regional heterogeneity, time heterogeneity, and random error.
Considering the availability and completeness of data, adopts the data from 2011
to 2019 of 30 provinces, except Tibet. Table 4 is descriptive statistics of explanatory
variables, explanatory variables, and control variables.
Digital Economy and High-Quality Development of Manufacturing Industry 9
According to the test results, each variable rejects the original hypothesis that there
is no unit root under 1% significance level. All variables are stationary sequences, and
there is no pseudo-regression in panel data regression.
First, the model is validated by a mixed regression model and a fixed-effect model,
and the selection of the model is tested using F statistics. The results show that the inter-
individual intercept terms are different, and the mixed effect model is unsuitable. Using
the random-effects model and the Hausman test, the original hypothesis is rejected at
the 1% significance level, which shows that the random disturbance is related to the
explanatory variable, so the fixed effects model should be chosen.
The Regression Coefficient of the development level of the digital economy is 0.293
and passes the test under the significance level of 0.05. After removing the influence
of Control Scalars, the regression coefficient of the influence of the digital economy on
manufacturing quality is 0.056. It can be concluded that the development level of the
digital economy promotes the high-quality development of the manufacturing industry
in China, and the higher the level of the digital economy is, the better the development
of the manufacturing industry. Increasing the investment of digital infrastructure con-
struction, increasing the popularization of the digital economy, and the application of
digital economy terminals will help promote the high-quality development of China’s
manufacturing industry. So hypothesis 1 is validated.
10 X. Xu and P. Han
Variable OLS FE RE
Lndigital 0.239*** 0.007 0.056* 0.293** 0.121*** 0.160**
(5.74) (0.11) (1.84) (2.28) (4.56) (2.55)
lnedu −0.988** 0.881 0.189
(−2.15) (1.21) (0.32)
FDI 1.667** −3.378*** −0.259
(2.72) (−4.13) (−0.37)
gov −0.667** 1.749** −0.217
(−2.51) (2.30) (−0.59)
urb 0.979** −1.905** 1.088**
(2.22) (−2.25) (2.10)
fin 0.101* 0.174*** 0.112***
(4.04) (3.12) (2.95)
lndev 0.165 0.416* 0.391***
(0.94) (1.70) (2.66)
cons 1.042*** (18.92) 0.821*** 1.179*** 3.806 1.131*** 3.976**
(0.48) (44.37) (1.31) (22.28) (2.26)
R−squared 0.3695 0.5333 0.5454 0.4050 0.5454 0.5420
F/Wald 32.99 45.21 3.37 7.44 20.83 58.18
[0.0000] [0.0000] [ 0.0675] [0.0000] [0.0000] [0.0000]
Hausman 58.14
[0.0000]
N 270 270 270 270 270 270
Note: *, * *, * * * are expressed at 10%, 5% and 1% significance levels respectively. () under FE
model is t statistic, and () under RE is Z statistic
On the one hand, the digital economy supported by 5G technology and the Internet
of things can reduce information asymmetry, accelerate the flow of production factors,
improve resource allocation, and reduce transaction costs. On the other hand, with the
development of the digital economy, the manufacturing industry has gradually changed
from a traditional product orientation to demand customization, which is the embodiment
of value creation in the manufacturing industry. Hypothesis 1 is verified. The control
variable:
Human Capital (Edu). The Regression Coefficient of human capital is 0.881. The
accumulation of human capital supports the manufacturing industry and improves inno-
vation efficiency and management level. (2) Foreign direct investment. In theory, foreign
direct investment can promote the manufacturing industry to learn advanced manage-
ment experience. However, in the fixed model, the Foreign Direct Investment Coefficient
was −3.378 and passed the hypothesis test at a 1% significance level. The reasons are:
since 2009, the growth rate of the real use of FDI in China has slowed down obviously,
and the secondary sector of the economy of the real use of FDI and its proportion have
Digital Economy and High-Quality Development of Manufacturing Industry 11
declined in recent years. (3) Government support(Gov). In the fixed-effect model, the
number of words in which the government played a role was 1.749, and it passed the
test at the significant line level of 5%. This shows that the government plays a role
in promoting the upgrading of the manufacturing industry, and fiscal expenditure as a
capital element promotes the high-quality development of the manufacturing industry.
(4) Urbanization. In the fixed-effect model, the influence Coefficient of Urbanization
by country is −1.905. It is tested at the level of 1% significance, which shows that
the degree of Urbanization significantly inhibits the high-quality development of the
manufacturing industry. Because Urbanization makes manufacturing employment costs
rise and will bring about negative effects such as the hollowing out of industries. (5).
Financial Services (Fin). In the fixed-effect model, the influence Coefficient of financial
development level is 0.174, and it passes the test under the 5% significance level. It
shows that financial service plays an important role in promoting the development of the
manufacturing industry. (6) Level of economic development (lndev). The Regression
Coefficient of economic development level was 0.416, and it passed the test under the
10% significance level. This indicates that economic growth is largely consistent with
high-quality development in the manufacturing sector.
In order to ensure the reliability of the empirical test results, this paper tests the robustness
of the model. It divides the sample of 30 provinces, cities, and autonomous regions
into the eastern, central, western, and northeastern regions according to the principle
of dividing the four economic regions, to examine the heterogeneity between different
regions. Table 6 shows that the development level of the digital economy has positive
effects on the eastern, northeastern, central, and western regions. Still, the positive effects
on the western regions are the least obvious. It shows that the digital economy in the
western region is relatively low. In order to promote the high-quality development of
the manufacturing industry, the northeast region plays the most important role. This
phenomenon may be because the sample size of the three provinces in the northeast region
is relatively small, making the regression coefficient of the variable abnormal. In general,
China’s digital economy has a positive effect on the development of manufacturing
quality, but there is obvious heterogeneity among regions. Hypothesis 2 has been verified.
12 X. Xu and P. Han
Findings:
Based on the research conclusion, this paper puts forward the following countermea-
sures and suggestions, expecting the digital economy to play a better role in promoting
the high-quality development of the manufacturing industry.
Enterprise-level: Manufacturing enterprises should vigorously promote the construc-
tion of digital platforms, make full use of big data, artificial intelligence, and other dig-
ital information technology, to promote the manufacturing industry in an intelligent,
efficient direction. At the same time, the innovative digital technology integration appli-
cation scene promotes the diffusion of advanced technology. It promotes the industrial
chain innovation upgrade, thus derived more new industries and new forms of busi-
ness. Strengthen the cooperation and sharing with scientific research institutions and
universities to improve the efficiency of technology transformation.
On the Industrial Organization level: We should dig deep into the value of the digital
economy, cultivate and develop new industries, upgrade traditional industries, push the
industrial chain, value chain to the high-end extension. In terms of industrial layout, we
should follow the law of regional heterogeneity. The distribution of manufacture under
the digital economy should be based on the difference of regional factor endowment and
location advantage, adapt to local conditions, and prevent the malignant competition
caused by the homogeneous development of the industry.
Macro-policy level: To develop the digital economy to do a top-level design, high-
quality manufacturing development to create a good business environment. Accelerate
the deployment of 5G, Internet of things, artificial intelligence, cloud computing and
other new generation of digital information technology. At the same time, the infor-
mation sharing platform between government and enterprise, enterprise and enterprise,
enterprise and consumer is constructed.
14 X. Xu and P. Han
References
1. Wei, Z.Y., Li, Y.T., Wu, K.D.: Can the digital economy promote the high-quality development
of the manufacturing industry? an empirical analysis on provincial panel data. J. Wuhan Finan.,
37–45 (2021)
2. Xu, X.C., Zhang, M.H.: A study on the scale of chinese digital economy - from the perspective
of international comparison. China’s Ind. Econ. (05), 23–41 (2020)
3. Yu, S., Fan, X.F., Jiang, H.W.: the impact of digital economy on china’s manufacturing qual-
ity going-out — from the perspective of export technological complexity enhancement. J.
Guangdong Univ. Finan. Econ. 36(02), 16–27 (2021)
4. Li, C.F, Chou, C.: The mechanism of digital economy driving the transformation and upgrading
of manufacturing industry—an analysis from the perspective of industrial chain. Bus. Res. (02),
73–82 (2020)
5. Jiao, Y.: Transformation of economy enabling manufacturing: from value remolding to value
creation. Economist (06), 87–94 (2020)
6. Li, Y.J., Han, P.: Mechanism and path of high-quality development of manufacturing industry
in digital economy. Macroecon. Manag. (05), 36–45 (2021)
7. Mohsen, A.: The impact of 5g on the evolution of intelligent automation and industry
digitization. J. Ambient Intell. Hum. Comput. (2021)
8. Gaffley, G., Pelser, T.G.: Developing a digital transformation model to enhance the strategy
development process for leadership in the south african manufacturing sector. South Afr. J.
Bus. Manag. 52(1), 12 (2021)
9. Shen, Y.H., Huang, H.: The impact of digital economy on the optimization and upgrading of
manufacturing industry structure 40(03), 147–154 (2021)
Research on the Mechanism of Promoting
the Development of High-Efficiency
and Featured Tropical Agriculture in Hainan
Relying on the Digital Economy
Abstract. From the perspective of current economic development trends, the dig-
ital economy has become a new economic growth point. As the only Hainan Free
Trade Port (FTP) in China, the FTP provides new opportunities for Hainan to
develop tropical and efficient modern agriculture with characteristics. However,
Hainan has a relatively low starting point for developing High-Efficiency and Char-
acteristic Tropical Agriculture (HECTA) in Hainan. It is necessary to explore an
effective mechanism for promoting the development of HECTA in Hainan relying
on the digital economy. It is necessary to establish and improve the cooperation
mechanism, data sharing mechanism, and talents that realize the full coverage of
digital infrastructure construction: the team training mechanism and the guarantee
mechanism at the legal and normative level.
1 Introduction
The outbreak of the new crown pneumonia epidemic has directly led to people’s social
distancing. To reduce contact and complete normal production and life, people need
the help of numbers. Therefore, online offices, online education, online medical care,
and online games are needed. Fresh food e-commerce, as the representative of the new
format of the digital economy, quickly fills up positions. Reality forces the digital econ-
omy to become a new growth point for economic development, giving the economy new
opportunities in the crisis. Based on the different duration of the epidemic, the growth
rate of the digital economy will be about 2.7–3 times the GDP growth rate, which has
a significant effect on stabilizing economic growth [1]. The Chinese government work
report pointed out: “E-commerce online shopping, online services, etc. Business for-
mats have played an important role in the fight against the epidemic. We must continue
to introduce supporting policies, comprehensively promote ‘Internet+,’ and create new
advantages in the digital economy.“ FTP is China’s only free trade port. From its devel-
opment perspective, the digital economy will also become an important starting point
for the high-quality development of FTP.
With the deepening of the understanding of the digital economy, the digital economy
has begun to be further integrated with various fields. With the emergence of the new
crown epidemic, the development of the digital economy has increased. Organic inte-
gration with “agriculture, rural areas, and farmers” has become the focus of attention
in the economic field in the past two years. Digital agriculture is an advanced stage of
agricultural modernization, and it is the only way for my country to move from an agri-
cultural country to an agricultural power. According to statistics, in 2020, the scale of
China’s rural e-commerce reached 28801.57 billion yuan, maintaining a very fast growth
rate. The important position of e-commerce in agricultural product sales channels also
highlights the integration of the digital economy and agricultural development in China.
Development.
Chen X. M. [5] analyzed the positive effects of the emergence of the digital economy
on the development of the agricultural and rural economy and put forward suggestions
and countermeasures for the problems faced in the integration of the digital economy and
the agricultural and rural economy, aiming to promote further the in-depth integration
of the digital economy and the agricultural and rural economy. Li analyzed the income
growth and income distribution effects of the rural digital economy and pointed out,
“the results of system integration show that the digital economy has a positive effect
on farmers’ income, but it also exacerbates the income gap. Obvious income-increasing
effect and negative effect on fair income. They believe that the digital economy has a more
significant impact on farmers’ agricultural income than non-agricultural income; groups
with higher socioeconomic status and education will benefit more from the development
of the digital economy. The digital economy has strengthened the widening income
gap under the conditions of promoting the overall growth of farmers’ income, posing
new challenges to public policies committed to promoting inclusive development [5].
Cai et al. [7] analyzed the opportunities, obstacles, and strategies of rural economic
development in the field of the digital economy. They pointed out that “in the field of
the digital economy, rural economic development should make full use of the significant
effects of policies and the effectiveness of e-commerce in helping farmers.
Digital technology empowers production, technological innovation releases vitality
and other important opportunities, and vigorously promotes the agricultural and digital
18 F. Jiang et al.
Current research mainly focuses on the connotation, and theoretical basis of the digital
economy itself, and its integration and development with specific industries are still
being explored. Especially in the construction of the FTP, the research on the integration
of the digital economy and tropical HECTA needs to be further deepened.
How HECTA can truly achieve leapfrog development with the help of the digital
economy and obtain more practical benefits for the guidance of practice remains further
explored.
3.1 Hainan Has a Relatively Low Starting Point for the Development of HECTA
Hainan is my country’s tropical treasure island, with unique tropical agricultural resource
endowments. The land area is 35,400 km2 , accounting for 42.5% of the country’s tropical
land area. The annual average temperature is 23–26 °C, the annual sunshine hours are
1832–2558 h, and the average rainfall is 1600 mm. Rich in biological species, annual
growth of crops, surrounded by the sea to form a natural animal disease barrier, and
unique conditions for the development of HECTA. Hainan’s tropical characteristic and
efficient agriculture not only includes tropical crops such as natural rubber, tropical fruits
such as bananas and mangos, dragon fruit, lychees, winter melons such as peppers and
beans, sweet corn, and tropical flowers such as chrysanthemums, green radish pears and
so on. However, from a national and even global perspective, Hainan Province has a late
start and poor foundation for the development of HECTA.
Research on the Mechanism of Promoting the Development of High-Efficiency 19
In 2018, Xi Jinping [2] specifically pointed out in his speech at 413, “We must strengthen
the construction of the National Southern Propagation Research and Breeding Base
(Hainan), build a national tropical agricultural science center, and support Hainan to build
a transfer base for the introduction of global animal and plant germplasm resources”.
Under the circumstances, Hainan Province has a unique advantage. It has gradually
formed southern propagation, winter melons and vegetables, tropical fruits, tropical
flowers, animal husbandry, marine fisheries, leisure agriculture, and tropical cash crops,
sold to the whole country. Even globally, a new path of HECTA has been formed. The
“Overall Plan for the Construction of FTP” proposes that the construction of a free
trade port must conform to Hainan’s positioning. “Giving full play to Hainan’s rich
natural resources, unique geographical location, and backing on the ultra-large-scale
domestic market and hinterland economy, we will seize important opportunities for a
new round of global technological revolution and industrial transformation, and focus
on the development of tourism, modern service industries, and high-tech industries. To
accelerate the cultivation of new advantages in cooperation and competition with Hainan
characteristics.”
3.3 The Digital Economy Has Inserted “Digital Wings” for Hainan’s
Development of HECTA
Necessity analysis
Free trade port construction provides new opportunities
and upgrade traditional industries, and improve industrial efficiency. Promote the digi-
tization, networking, and intelligence of the entire agricultural chain, and promote the
formation of “digital wings” in the industry, which can improve the efficiency of agricul-
tural production and ensure the quality of agricultural products. The “Overall Plan for
the Construction of FTP” also puts forward, “Data flows in a safe and orderly manner.
Under the premise of ensuring the safe and controllable data flow, expand the opening
of the data field, innovate the design of security systems, realize the full convergence of
data, and cultivate the development of a digital economy.“
Many studies have shown that the global supply chain structure is accelerating adjust-
ment, and the global supply chain, industrial chain localization, regionalization, and
decentralization are accelerating the evolution. China proposes a new development pat-
tern and pays more attention to coordination, integration, and reconstruction of the supply
chain. The international community also attaches great importance to the diversified lay-
out of the supply chain. The FTP deepens the structural reform of the agricultural supply
side, integrates and utilizes the advantages of tropical agricultural resources, vigorously
adjusts the agricultural structure, and works hard on new supplies to create the “trump
card” of HECTA. Therefore, creating new potentials for the digital economy, “great
changes in the digitally driven industrial chain,” and “new economic growth points in
the digital age” have become hot topics in FTP. Hainan’s HECTA should focus on intel-
ligent agricultural production, networked management, and Managing digitalization and
online services. Strengthen the foundation of agricultural and rural digitalization, inte-
grate data resources, enhance data acquisition capabilities, increase data development
and application, deepen integration and sharing, strengthen management services, etc.,
to form a HECTA industry layout relying on digital economy A long-term mechanism
to blaze a new path for agricultural and rural development in the new era, enlarge and
strengthen the agricultural industry, and help the rural revitalization strategy.
Research on the Mechanism of Promoting the Development of High-Efficiency 21
resources and related standards, deepen the construction of special agricultural data,
and promote information sharing and joint construction.“ At the same time, improving
information transparency, regulating hidden rules, introducing incentives and safeguards,
promoting information sharing and resource exchange among regions and departments
with policy-based support methods, and achieving mutual benefit and win-win results
are also effective means.
4.4 Establish and Improve the Guarantee Mechanism for the Legal Information
Flow Security of Data and Information Collection
As the digital economy uses big data, collecting information is the most basic link,
and the legitimacy of information collection and the security of information flow are
particularly important. Therefore, risk prevention and control must be strengthened to
ensure data security. Based on the original laws and regulations, supplement, revise
and improve the legal systems related to the digital economy and digital finance, must
highlight their forward-looking, systematic, and operability, and establish a dynamic
tracking mechanism, build a standardized measurement index system, and ensure big
data Legalize information collection, and formulate strict norms and standards for data
sources, ownership of responsibility and rights, protection of intellectual property rights,
and profit distribution.
Research on the Mechanism of Promoting the Development of High-Efficiency 23
Figure 2 shows an effective mechanism for building THECA industrial layout driven
by the digital economy.
Cooperation
Mechanism
Effective
Training Mechanism Security Mechanism
Mechanisms
Sharing Mechanism
Fig. 2. An effective mechanism for the layout of the A industry driven by the digital economy
5 Conclusion
This article analyzes the need to develop Hainan’s HECTA industries to rely on the digital
economy. To promote the development of the digital economy under the construction of
FTP, build an effective mechanism for the layout of Hainan’s HECTA industries driven
by the digital economy. The construction of a full coverage cooperation mechanism for
digital infrastructure construction, a data-sharing mechanism, a training mechanism for
agricultural digital pioneers, and a security mechanism for information flow has laid a
theoretical foundation for developing FTP’s digital economy in the HECTA.
In the construction of the FTP, the research on the impact of the development of the
digital economy on the HECTA industry is still in its infancy, and related applications are
still being improved. There are still shortcomings in the mechanism we have constructed,
and further research is needed in the future.
Acknowledgments. This work was supported by the Hainan University 2020–2021 Construction
of Basic-level Party Organizations Special Project Research(Hddj07, Research on the Construc-
tion of Basic-level Party Organizations in Colleges and Universities of FTP), Hainan University
2021 Education and Teaching Reform Research Project (HDJY2159, Hainan University applied
undergraduate top-notch innovative talent training model and practical research driven by the free
trade port construction) and Philosophy and Social Science Planning Project of Hainan Province
(HNSK(YB)21–41, Research on the Centennial Course and Experience of the Political Ecological
Construction of the Communist Party of China). This work was supported by the Humanities and
Social Sciences Research Innovation Team of Hainan University, Hainan Free Trade Port Cross-
border E-commerce Service Innovation Research Team (HDSKTD202025). This work was also
supported by Hainan Provincial Natural Science Foundation of China (720RC567, Research on
24 F. Jiang et al.
Information Compensation Mechanism of Free Trade Zone to Global Supply Chain Innovation
Coordination under BlockChain Technology, and 720RC569, Tourism Value Chain Distribution
and Ecological Optimization Mechanism of Hainan International Tourism Consumption Center
Based on System Dynamics).
Authors’ Contributions. Fan Jiang was responsible for proposing the overall idea and frame-
work of the manuscript. Shaoqing Tian was responsible for the writing of the first draft of the
manuscript and translation. Chongli Huang was responsible for the revision of the manuscript and
proofreading. Fan Jiang, Shaoqing Tian, and Chongli Huang contributed equally to this paper and
are regarded as the first authors.
References
1. Tang, D.D., Liu, X.L., Ni, H.F., Yang, Y.W., Huang, Q.H., Zhang, X.J.: Global economic
changes, china’s potential growth rate and high-quality development post-epidemic period.
Econ. Res. 55(08), 4–23 (2020)
2. Pei, C.H., Liu, H.G.: An economic analysis of Xi Jinping’s new era of opening to the outside
world. Econ. Res. 53(02), 4–19 (2018)
3. Li, C.F., Li, D.D., Zhou, C.: The mechanism of the digital economy driving the transformation
and upgrading of the manufacturing industry: an analysis based on the perspective of the
industrial chain. Bus. Res. (02), 73–82 (2020)
4. Wei, J., Liu, J.L., Liu, Y.: Digital economics: connotation, theoretical foundation, and important
research issues. Sci. Technol. Progr. Countermeas. (7), 1–7 (2021)
5. Chen, X.M.: Research on the integrated development of digital economy and agricultural and
rural economy. Econ. Manag. Abst. (12), 5–6 (2021)
6. Li, Y., Ke, J.S. Three-level digital divide: income growth and income distribution effects of
rural digital economy. Agric. Technol. Econ. (08), 119–132 (2021)
7. Cai, Z.Z., Su, X.D.: Opportunities, obstacles, and strategies of rural economy in the field of
digital economy. Agric. Econ. 07, 35–37 (2021)
8. Fu, R.Y.: The digital economy will become an important starting point for the high-quality
development of hainan free trade port. Hainan Daily, 04–22 (2021)
The Impact of Digital Economy on China’s
Low-Carbon Development
Abstract. As a new resource type, the digital economy permeates all economies
and societies, becoming an essential engine for future development. Based on
Chinese provincial panel data from 2015–2018, the article investigates the impact
of the digital economy on the development of China’s low-carbon economy using
a mediating effects model. The conclusion of the article is as follows. First, the
digital economy has an essential role in promoting the development of China’s
low-carbon economy. Second, the digital economy can encourage the development
of China’s low-carbon economy through technological innovation. Third, there is
heterogeneity in the impact of the digital economy on the development of China’s
low-carbon economy, and its impact is in the order of strong to weak: east, west,
and central. The article provides empirical evidence at the provincial level to
develop the digital economy in China’s low-carbon economy. It offers policy
recommendations for the digital economy to contribute more effectively to China’s
low-carbon economy.
1 Introduction
Global warming and the development of a low-carbon economy (LC) are increasingly
becoming a consensus [1]. Global governance has entered the stage of full implementa-
tion of the Paris Agreement, and China has entered a critical period of the low-carbon
transition.
As the future of the world economy, the digital economy (DE) is playing an increas-
ingly important role in global economic growth [2], and its higher penetration could
bring dividends to China at different levels and broaden the innovation field [3].
Currently, the process of DE development has both opportunities and challenges [4].
The 14th Five-Year Plan and the 2035 Vision Outline propose to embrace the digital era
and drive overall changes in production, life, and governance with digital transformation.
As shown in Fig. 1, the size of China’s DE in terms of value-added has expanded from
2.6 trillion yuan in 2005 to 35.9 trillion yuan in 2020, and the DE is growing rapidly
and impacting our economic life.
Driven by the concept of sustainable development, the relationship between the DE
and the environment is becoming increasingly close [5]. Li Y. et al. [6] showed that
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022
A. E. Hassanien et al. (Eds.): BIIT 2021, LNDECT 107, pp. 25–34, 2022.
https://doi.org/10.1007/978-3-030-92632-8_3
26 F. Yao and L. Li
400 41.36 45
thousand
350 34.8 36.2 40
32.9
300 35
26.1 30 Overall size of the digital
250 econom(hundred million
20.3 25
200 RMB)
20
14.2 15.2 Digital economy accounts
150
15 for the GDP ratio value of
100 (%)
10
50 5
0 0
2005 2008 2011 2014 2017 2018 2019 2020
2 Research Hypothesis
According to Porter’s hypothesis [8], environmental regulation policies may have long-
run or short-run effects on the competitiveness of firms. In the short run, environmental
regulation can increase firms’ costs and reduce their competitiveness; in the long run,
environmental regulation can force firms to choose cheaper and more environmentally
friendly technologies, improving their competitiveness. Unlike the traditional economy,
the DE brings efficiency and polarization in the area of technology integration. The
high penetration rate of the DE can effectively break the limits of time and space, blur
the boundaries between industries, direct capital flows, and improve resource utilization
efficiency. In such an era where information and data are gradually becoming the core
capital of enterprises, companies will actively take advantage of the DE to achieve
technological innovation and actively cater to the concept of LC.
ρ is the low carbon economic efficiency value to be calculated for China; si_ , sγα , slb
are the slack variables for inputs, desired outputs, and non − desired outputs;λ is the
decision unit weight, and the production unit is efficient when ρ = 1; when ρ < 1, there
is efficiency loss in the production unit.
The results calculated using the SBM model are static and do not directly reflect
the productivity changes. According to the idea of Oh [11], this paper adopts the SBM
model as the basis and uses the GML index to measure the value of low carbon economic
efficiency in China, which can be defined as follows.
1+DG (
T x t ,y t ,bt
)
GMLt,t+1 xt+1 , yt+1 , bt+1 ; xt , yt , bt = 1+DG (xt+1 ,yt+1 ,bt+1 )
T (6)
1) Benchmark model:
2) Technological innovation:
The results of the baseline regression are shown in Table 3. Table 3-(3) shows the
final results. The results from the baseline regressions show that DE has a significant
contribution to the development of LC in China for a fixed number of provinces and
times. This confirms that hypothesis H1 holds that the DE has a significant positive
effect on low carbon development in China.
30 F. Yao and L. Li
Variables LC Tec LC
DE 5.0171*** 27.5557*** 3.8280***
(0.8546) (10.62332) (0.7459)
Tec 0.0432***
(0.0064)
GDP 0.3581*** 1.4740*** 0.2945***
(0.0312) (0.3881) (0.0281)
Urban 0.0034 −0.1673*** 0.0106***
(0.0039) (0.0482) (0.0035)
Income −0.0434 −2.8867* 0.0812
(0.1371) (1.7041) (0.1177)
Trade 0.0085*** 0.1695*** 0.0012
(0.0017) (0.0215) (0.0018)
Constant −126.4274* −2596.556*** −14.37508
(74.71983) (928.8478) 65.51425
Year Yes Yes Yes
Province Yes Yes Yes
Observations 120 120 44
R-squared 0.8499 0.6154 0.4492
Standard errors in parentheses
*** p < 0.01, ** p < 0.05, * p < 0.1
5 Conclusions
Based on the sample data of 30 provinces in China (except Tibet) from 2015–2018,
this paper uses the natural logarithm of the number of new product R&D projects as
a measure of product innovation to examine whether the DE can effectively promote
the development of China’s LC through the level of product innovation and then. The
empirical results lead to the following conclusion. First, the development of the DE
has a significant role in promoting the LC in China. Second, the DE can promote the
LC in China through the intermediary path of technological innovation. Third, there is
heterogeneity in the impact of the DE on the LC in China, and its impact effect is in the
order of strong to weak: east, west, and central.
Based on the above findings, this paper further puts forward targeted policy
recommendations.
First, importance should be attached to the intermediary role of technological inno-
vation in developing the DE for the LC. Modern digital information technology is used
to channel capital flows to enterprises using low-carbon technologies, further encourage
enterprises to actively realize low-carbon technological innovation, improve the effi-
ciency of traditional industries, establish the Internet in various industries, promote the
integration of DE technologies with the real economy, and create a new digital industrial
ecology, thereby promoting LC.
Second, strengthen the infrastructure development, especially the DE infrastructure
development in the central and western regions. The eastern region will fully play its
unique regional advantages, actively promote the LC process, and then drive the central
region with its advantages. The central region should actively introduce many advanced
technologies and new industries in the DE and guide traditional industries to change
their production models to promote the rationalization and low-carbon development of
industrial structures. At the same time, the eastern and central regions will benefit from
the accumulated talents, experiences, and technologies to the western region to establish
a solid foundation of DE at the national level to realize the LC nationwide.
References
1. Muntasir, M., Zahoor, A., Shabbir, A., Haider, M., Abdul, R., Vishal, D.: Reinvigorating
the role of clean energy transition for achieving a low-carbon economy: evidence from
Bangladesh. Environ. Sci. Pollut. Res. Int. (2021)
2. Hu, L., Yao, S., Zhou, Z., Miao, N.: The importance of the construction of digital economy
major under the background of new liberal arts. World Sci. Res. J. 7(8), 104–111 (2021)
3. Sorescu, A., Schreier, M.: Innovation in the digital economy: a broader view of its scope,
antecedents, and consequences. J. Acad. Mark. Sci. 49(4), 627–631 (2021)
4. Cheng, C., Huang, H.: Big data and industrial innovation progress in jiangxi province incre-
mental effect highlights enabling digital economy cultivation. J. Phys. Conf. Ser. 1852(2),
022005 (2021)
5. Li, Z., Li, N., Wen, H.: Digital economy and environmental quality: evidence from 217 cities
in China. Sustainability. 13(14), 8058 (2021)
6. Li, Y., Yang, X., Ran, Q., Wu, H., Muhammad, I., Munir, A.: energy structure, digital economy,
and carbon emissions: evidence from China. Environ. Sci. Pollut. Res. Int. 28, 1–24 (2021)
34 F. Yao and L. Li
7. Li, X., Liu, J., Ni, P.: The Impact of the digital economy on CO2 emissions: a theoretical and
empirical analysis. Sustainability 13(13), 7267 (2021)
8. Porter, M.E., van der Linde, C.: Toward a new conception of the environment-competitiveness
relationship. J. Econ. Perspect. 9(4), 97–118 (1995)
9. Rong, H.: The research on fiscal and financial policy to coordinate support development of
low carbon economy- a case study of Heilongjiang Province of China. Int. J. u - e - Serv. Sci.
Technol. 9(12), 39–52 (2016)
10. Tone, K.: A slacks-based measure of efficiency in data envelopment analysis. Eur. J. Oper.
Res. 130(3), 498–509 (2001)
11. Oh, D.H.: A global malmquist-luenberger productivity index. J. Prod. Anal. 34(3), 183–197
(2010)
12. Huang, G., Pan, W., Hu, C., Pan, W., Dai, W.: Energy utilization efficiency of China
considering carbon emissions—based on provincial panel data. Sustainability. 13(2), 877
(2021)
13. Shan, Y., et al.: New provincial CO2 emission inventories in China based on apparent energy
consumption data and updated emission factors. Appl. Energy 184, 742–750 (2016)
The Impact of the Digital Economy on Inclusive
Growth——Empirical Study Based
on the Spatial Econometric Model
Abstract. This paper drew on the “China Digital Economy Index” released by
Caixin Digital Alliance Research Centre, and measured the inclusive growth index
of 30 provinces (cities and districts) in China based on the entropy weighting
method, then used Stata 16.0 to establish the spatial econometric model to explore
the impact of the digital economy on China’s inclusive growth. The results showed
that the digital economy has a positive contribution to inclusive growth in China.
There is a significant positive spatial spillover effect, indicating that the develop-
ment of the digital economy contributes to improving the level of inclusive growth
between regions in China. Therefore, in-depth exploration of the impact of the
digital economy on inclusive growth has important theoretical value and practical
significance for consolidating China’s achievements in poverty eradication and
realizing inclusive growth.
1 Introduction
In December 2019, the Central Economic Work Conference proposed that “Efforts
should be made to promote high-quality development and vigorously develop the dig-
ital economy”, the important assertion clearly defined the basic direction of the trans-
formation of the economic development model. At present, China’s economy is from
high-speed growth to inclusive growth, which is an inevitable choice for economic devel-
opment in the new era. The digital economy can improve economic efficiency by inno-
vating the economic development model and weakening the differences of urban-rural
binary structure, thus improving economic efficiency, which is conducive to maintain-
ing equitable access to the fruits of development for the economy and society, and thus
contributing to sustainable economic development, which is also a basic requirement for
inclusive growth.
2 Literature Review
The《G20 Digital Economy Development and Cooperation Agreement》was signed at
the G20 Hangzhou Summit held in September 2016, which comprehensively elaborated
the concept and significance of the inclusive growth, i.e. the digital economy refers to
a series of economic activities that use digital knowledge and information as factors
of production, modern information networks as microcarriers, and the effective use of
Information and Communication Technology (ICT) as an important driving force for
efficiency improvement and economic structure optimization. Research on the digital
economy has also been discussed and explored conceptually and theoretically by scholars
at home and abroad. Negroponte (1996) proposed that the digital economy is an eco-
nomic system that uses ICT extensively to influence people’s lives through digitization,
informatization, and networking [1]. Wang et al. (2019) argued that the digital econ-
omy is a new economic form based on information and knowledge with new-generation
information technology as the means, information industry as the mainstay, and informa-
tion products and services as the main content [2]. Zhang and Liu (2019) summarised
the characteristics of the digital economy, namely that the production factor is data,
its development model is industrial integration, development is characterized by green
sustainability [3].
The ADB first introduced the concept of inclusive growth in its report “Shared
Growth for Social Harmony” in 2007, which argued that “inclusive growth” was growth
with equal opportunities, emphasized that economic growth should not only focus on
aggregate growth but also on reducing regional economic development gaps and enhanc-
ing equality of opportunity in the process of economic development. Although the aca-
demic community has not yet formed a unified concept of inclusive growth, a review of
the existing literature has shown that the meaning of inclusive growth mainly includes
three aspects: economic growth, equal opportunities, and income distribution. Rauni-
yar and Kanbur (2010) considered inclusive growth as growth with reduced inequal-
ity and emphasized income distribution’s relative equity [4]. In measuring inclusive
growth, Zhang and Wan et al.(2019) included income growth and income distribution
together in constructing the inclusive growth system [5]. Wu and Zhou (2018) mea-
sured China’s inclusive growth index from the perspective of green development, which
mainly involved economic growth, income distribution, fairness in opportunity, and envi-
ronmental optimization [6]. Yu and Wang (2012) measured China’s inclusive growth in
four aspects: sustainability of economic growth, fairness of development opportunities,
income disparity, and basic social security, respectively [7].
In terms of the impact of the digital economy on economic growth, scholars in China
have reached a more consistent conclusion that the digital economy has a positive impact
on economic growth. Li and Yang (2021) proposed that the development of the digital
economy effectively promoted the growth of total factor productivity (TFP), and the
effect of the digital economy has a positive spatial spillover effect, contributing to raising
the TFP level of neighboring cities [8]. Xu (2004) constructed a dynamic production
function to explore the impact of the digital economy on economic growth, and the study
showed that the digital economy was positively correlated with the GDP growth index
[9]. Han (2018) established a dynamic panel model, concluded that the software and
information technology service industry can promote China’s economic growth [10]. Xia
et al. (2018) concluded that the digital economy has a significant contribution to China’s
economic growth by constructing a non-competitive employment input-occupied output
model [11]. Zhang and Shi (2019) argued that the impact of the digital economy on the
The Impact of the Digital Economy on Inclusive Growth 37
national economy is mainly through the substitution effect, penetration effect, innovation
effect, and industrial association effect [12]. Tian and Chen (2021) suggested that the
digital economy can effectively break the constraints on economic development caused
by the limited supply of production factors, improve TFP, and play a potential and
momentum in promoting industrial transformation and optimizing economic structure
[13].
Throughout the existing literature, scholars at home and abroad have explored the
digital economy, mainly focusing on the digital economy’s impact on economic growth.
What is the relationship between the digital economy and inclusive growth? Few scholars
have conducted empirical analyses on this issue. In this context, this paper explored the
impact of the digital economy on China’s inclusive growth based on a spatial econometric
model, which will improve China’s digital economy-related policies and further improve
China’s inclusive growth level.
Table 1. (continued)
Table 1. (continued)
Xitj − min(Xitj )
Yitj = (1)
max(Xitj ) − min(Xitj )
If Xitj is a negative indicator:
max(Xitj ) − Xitj
Yitj = (2)
max(Xitj ) − min(Xitj )
In Eqs. (1) and (2), i represents the province (cities and districts), t represents the
year, j represents the indicator, Xitj represents the initial indicator value, Yitj represents
the indicator value after dimensionless processing.
40 F. Yao and X. Wang
In the third step, the coefficient of variation Gi for the indicator is calculated:
Gi = 1 − Ej (5)
Table 2. (continued)
Explained Variable. Inclusive growth (ig), this paper selected three primary indicators,
11 secondary indicators and 18 tertiary indicators to build China’s inclusive growth
system, then used the entropy weighting method to calculate the inclusive growth index
as the explained variable.
logarithm of them) and the degree of openness to the outside world (open) and price
level (pl, and take the logarithm of them).
A description of the relevant variables selected for this paper is shown in Table 3:
In this paper, the panel data of 30 provinces (cities and districts) in China from 2012 to
2018 were selected as the research samples. Among them, the digital economy (dige) was
obtained from the “China Digital Economy Index” released by Caixin Digital Alliance
Research Centre, while the original data of other variables were obtained from the China
Statistical Yearbook, the statistical yearbooks of various provinces (cities and districts)
and the websites of the National Bureau of Statistics, etc. Since the data of Tibet, Hong
Kong, Macao, and Taiwan are discontinuous and seriously missing, they are not included
in the scope of the study.
The Impact of the Digital Economy on Inclusive Growth 43
To ensure that the panel data of each variable in the paper is steady and to avoid
pseudo-regression, robustness tests were conducted before the regression analysis, and
the LLC test and Fisher-PP test were used in this paper.
As can be seen from Table 5, all variables were robust and therefore brought into the
model for regression analysis.
n
Igit = β0 + λ wij Igjt + β1 Digeit + β2 Controlit
j=i
n
n (8)
+η wij Digejt + θ wij Controljt + μi + φi + εit
j=i j=i
44 F. Yao and X. Wang
In model (7) and model (8), I g represents inclusive growth, Dige represents digital
economy, Control represents control variables, β0 represents constant terms, ρ and λ
represent spatial regression coefficients of SAR and SDM models respectively, Wij are
spatial weight matrices, ε is the random error term, μ is the regional fixed effect, ϕ is
the time fixed effect, i and j represent provinces, t represents years.
In terms of spatial weight matrix settings, this paper chose the economic weight
matrix, where economic distance was measured using the difference between the GDP
of each province in 2020.
Drawing on Cui and Wang et al. (2018), this paper chose the M oran’s I index test
to analyze the global spatial autocorrelation of inclusive growth in China [15]. The
calculation method is as follows:
n n
i=1 j = 1 wij (xi − x̄) xj − x̄
M oran’s I = (9)
S 2 ni= 1 nj= 1 wij
SAR SEM
Test-Statistic p-value Test-Statistic p-value
66.05 0.0000 158.93 0.0000
The regression results are shown in Table 8. In terms of spatial reflection coefficient,
the coefficients of SAR and SDM models are positive, and both are significant at the 1%
level. Among them, the spatial reflection coefficient of the spatial autoregressive model
46 F. Yao and X. Wang
ρ is 0.696; the spatial reflection coefficient of the spatial Dobbin model λ is 0.348,
indicating that there is a significant positive spatial dependence on inclusive growth in
China. In terms of the core explanatory variables, the coefficients of the digital economy
for the two models are 0.159 and 0.0948 respectively, which are significantly positive at
the 1% level, indicating that the digital economy still has a positive effect on inclusive
growth after spatial spillover effects are taken into account. The coefficient of the spatial
lagged term of the digital economy (W *dige) is 0.217, which is significantly positive
at the 1% level, i.e. there is a significant spatial spillover effect of the digital economy
on inclusive growth in other provinces, indicating that an increase in the regional digital
economy can improve the level of inclusive growth in the region and also generate pos-
itive externalities to neighboring regions. In terms of control variables, the findings of
the above two models are basically the same, with the coefficients of government expen-
diture level and industrial optimization both being positive and passing the significance
test, indicating that both still have a positive effect on the level of inclusive growth after
the spatial spillover effect is taken into account. The three indicators of trade openness,
foreign direct investment, and price level failed the significance test, indicating that the
spillover effects of these three variables on inclusive growth are not significant.
To further analyze the direct effects and spatial spillover effects of each explanatory
variable, the spatial effects were decomposed based on SAR and SDM models, and the
results of the spatial effects decomposition are shown in Table 9:
As shown in Table 9, the direct effect, spillover effect, and total effect of the SAR
model and SDM model have passed the significance test, indicating that the digital econ-
omy not only contributes to the level of inclusive growth in the local region but also has a
spatial spillover effect that positively affects the inclusive growth of surrounding regions.
In addition, the direct effect coefficient is smaller than the spillover effect coefficient,
indicating that the positive impact of the digital economy development on the local
region’s inclusive growth is smaller than the positive impact on the surrounding region’s
inclusive growth. As for the other control variables, the gov passed the significance test,
indicating that the gov has both a direct effect on inclusive growth in the local region and
a spatial spillover effect on inclusive growth in the surrounding regions; while it only
has a direct effect on inclusive growth in the local region and no spatial spillover effect
on inclusive growth in the surrounding regions, fdi, open and pl failed the significance
test.
48 F. Yao and X. Wang
Table 10 reported the regression results after changing the spatial weight matrix.
Under the geographical distance matrix, the coefficients of the SAR and SDM models
were both positive and passed the significance test, and their values were similar to those
shown in Table 8. In addition, the coefficient values of the other control variables were
also generally consistent with the results shown in Table 8, again verifying the existence
of significant positive spatial dependence of inclusive growth in China, indicating that
the main findings of this paper are robust and consistent with the reality.
(cities and districts) in China based on the entropy weighting method. Secondly, used
Stata 16.0 establishes the spatial econometric model to further analyze the impact of the
digital economy on inclusive growth. The results showed that the digital economy has a
positive contribution to inclusive growth in China. There is a significant positive spatial
spillover effect, indicating that the development of the digital economy contributes to
improving the level of inclusive growth between regions in China. Based on the research
conclusion, the following recommendations are put forward:
6.2 Recommendations
Improve Laws and Regulations, Perfect the Digital Security Supervision System.
China’s digital economy has developed relatively late, and the corresponding laws, reg-
ulations, and regulatory systems are not yet perfect. Therefore, improving laws and
regulations and perfecting the digital security regulatory system is the primary task to
achieve inclusive economic growth in China. Firstly, regulators should develop a reason-
able and effective digital security system and improve the security system of the digital
economy market in order to ensure the steady and orderly operation of the digital econ-
omy market. Secondly, regulators should clarify the scope and regulatory requirements
for the digital economy and establish an information disclosure system related to the
digital economy.
Improve Infrastructure Construction, Strengthen the Support of the Talent Team.
The development of the digital economy and the realization of inclusive economic growth
are inseparable from the construction of the digital infrastructure and the construction of
talent teams in the digital field. Firstly, strengthen the construction of digital infrastruc-
ture in regions with backward economic development, and expand the scope of services
provided by the digital economy in various industries. Secondly, implement the policy
of introducing professionals. Governments at all levels take the initiative to absorb and
introduce professionals in the field of the digital economy and increase subsidies for
landing talents.
Innovate Digital Technology, Accelerate the Digitization of Industry Development.
Technological innovation is the key to developing the digital economy. Firstly, China
should increase investment in the research and development of talents in the fields of
big data, cloud computing, the Internet, and artificial intelligence. Secondly, large enter-
prises should be encouraged to establish their own digital technology research center so
that digital technology can be deeply integrated with various industries and accelerate
the digital transformation and upgrading of industrial development. Finally, establish a
long-term mechanism for universities, scientific research institutes, and industrial enter-
prises to jointly innovate the development of digital technology, promote the integrated
development of industry, university, and research, timely transform research results, and
further promote China’s inclusive economic growth.
References
1. Negroponte, N.: Being digital. Yonsei Bus. Rev. 33(1) (1996)
50 F. Yao and X. Wang
2. Wang, W., Wang, J.: Research on the development trend and promotion policies of China’s
digital economy. Econ. Aspect (01), 69–75 (2019)
3. Zhang, H., Liu, Z., Wang, S.: Analysis of China’s high-quality economic development path
under the background of digital economy. Bus. Econ. Res. (23), 183–186 (2019)
4. Rauniyar, G., Kanbur, R.: Inclusive growth and inclusive development: a review and synthesis
of Asian development bank literature. J. Asia Pac. Econ. 15(4), 455–469 (2010)
5. Zhang, X., Wan, G., Zhang, J., He, Z.: Digital economy, inclusive finance, and inclusive
growth. Econ. Res. 54(08), 71–86 (2019)
6. Wu, W., Zhou, X.: Measurement, evaluation, and influencing factors of China’s inclusive
green growth. Soc. Sci. Res. 01, 27–37 (2018)
7. Yu, M., Wang, X.: Inclusive growth of china’s economy: measurement and evaluation. Econ.
Rev. 03, 30–38 (2012)
8. Li, Z., Yang, Q.: How does the digital economy affect china’s high-quality economic
development? Disc. Mod. Econ. (07), 10–19 (2021).
9. Xu, S., Mao, X.: Analysis of the contribution of the information industry to economic growth.
Manag. World (08), 75–80 (2004)
10. Han, B., Li, S.: Software and Information technology service industry and china’s economic
growth. Res. Quant. Econ. Technol. Econ. 35(11), 128–141 (2018)
11. Xia, Y., Wang, H., Zhang, F., Guo, J.: Research on the impact of the digital economy on
China’s economic growth and non-agricultural employment —based on input occupancy
output model. J. Chin. Acad. Sci. 33(07), 707–716 (2018)
12. Zhang, H., Shi, L.: Digital economy: new power in the new era. J. Beijing Jiaotong Univ.
(Soc. Sci. Ed.) 18(02), 10–22 (2019)
13. Tian, G., Chen, F.: Xi Jinping’s historical logic, core meaning, and its value in the development
of the digital economy. Theory Guide (01), 4–9 (2021)
14. Tang, Y., long, Y., Zheng, Z.: Research on inclusive growth effect of digital inclusive finance
— an empirical analysis based on 12 provinces in Western China. Southwest Finan. (09),
60–73 (2020)
15. Cui, Q., Wang, W., Zhang, S.: Financial deepening, industrial structure upgrading and tech-
nological innovation—an empirical analysis based on spatial dobbin model. Ind. Technol.
Econ. 37(02), 42–50 (2018)
16. Paul Elhorst, J.: Matlab software for spatial panels. Int. Reg. Sci. Rev. 37(3), 389–405 (2014)
Evolutionary Game Analysis of Digital
Innovation Ecosystem Governance
Abstract. Under the background of the rapid development of the digital econ-
omy, managing the digital innovation ecosystem is an important topic. Based on
the evolutionary game method, this paper studies the evolutionary game strategy
selection of the participants in the governance of the digital innovation ecosys-
tem. It is found that: ➀ The participants in the governance of the digital inno-
vation ecosystem include three aspects: the government, the platform, and other
participants. The roles in the process of the evolutionary game are different. The
government plays the role of macro policy guidance, the platform strengthening its
governance is the key, and other participants play an auxiliary role. By construct-
ing the “government-platform-participant” tripartite evolutionary game model,
this paper simulates and analyzes the evolutionary strategy of the three players
in the governance to achieve the balance and stability of the governance effect
and puts forward some suggestions to promote the governance of the digital inno-
vation ecosystem. ➁ Governance principle of the digital innovation ecosystem:
establish an effective incentive and punishment mechanism, follow the principle
of “promoting innovation, subject fairness and welfare maximization,” and coor-
dinate the relationship between relevant stakeholders and their benefit distribution
through legislation and other policy levels, to maximize the welfare of the whole
digital economy.
1 Introduction
In recent years, the digital economy has become an important part of the national econ-
omy of all countries. It is facing new challenges in the process of digital innovation. The
governance problem has become a research hotspot in economics, management, soci-
ology, and so on. Through the research on the evolutionary game of the participants in
the governance of the digital innovation ecosystem, this paper discusses the goal of pro-
moting the healthy development of the digital innovation ecosystem with a sustainable
competitive advantage and realizing the balance of interests of all participants. Digital
innovation refers to the combination of information, computing, communication, and
connectivity technologies used in the innovation process, including product innovation,
production process improvement, organizational model change, and business model
innovation [1–3]. The innovation ecosystem is a complex system composed of platform
enterprises, consumers, markets and their natural, social and economic environment.
Core enterprises in Apple, IBM, Wal Mart, and other innovation ecosystems enable
other organizational members to create value by establishing service, tool, technology,
and value network platforms. At the same time, it strengthens the overall innovation and
productivity improvement of the ecosystem [4]. With the symbiotic evolution of mem-
bers, the participants of the innovation ecosystem follow the unified platform standards
and rules and create value that a single enterprise cannot create through open innova-
tion [5]. The digital innovation ecosystem involves many participants. Relying solely
on ecosystem platforms or government governance is not enough to regulate transaction
behavior. It is necessary to design a governance mechanism with the participation of
all stakeholders [6]. The existing research results of innovation ecosystem governance
are rich and diverse. Some scholars study ecosystem governance from the perspective
of innovation process governance, analyzing innovation elements, network relations,
Embeddedness, and collaborative innovation. Some analyze the internal relationship
of the network and the governance of value activities from the perspective of network
nodes [7, 8]. Some analyze the impact of knowledge and resource transfer governance
on innovation performance from the perspective of the output of the innovation ecosys-
tem [9, 10]. But there are three research gaps: lack of attention to the game between
governance subjects, less discussion on government supervision and incentive role, and
lack of analysis of other relevant participants.
This article aims to analyze the strategy choice of the evolutionary game between the
governance subjects of the digital innovation ecosystem, explore the optimal strategy
and effective governance mechanism, and the role of government, platform, and other
participants in governance.
process. Mainly system supply and administrative services; Second, Internet platform
enterprises; Third, platform service providers; Fourth, consumers; Fifth, social institu-
tions, including universities, third-party institutions, and other non-governmental orga-
nizations, can not be ignored in promoting governance. They usually play a bridge role
in legal services, news disclosure, educational assistance, and responsibility supervision
and have positive functions in government failure and market failure. For the convenience
of analysis, the last three parts are collectively referred to as the participants. This paper
divides the participants of digital innovation ecosystem governance into three parts,
including government regulators, platforms, and participants, as shown in Fig. 1.
Participants
Government Platform (Service providers, social
institutions, consumers, etc)
In the governance of the digital innovation ecosystem, the participants have information
asymmetry and irrational decision-makers. The governance effect is closely related to
the strategic choice of the government, platform, and participants.
The Needs of Participants Are Different. The government is the maker of policies
and regulations and has the responsibility of supervision. Governance is the ability of the
government to formulate and implement rules and provide services [15]. At the same
time, governance is an ability and a specific and complex interactive network across
different actors [16].
As the resource integration and manager of the innovation ecosystem, the platform
has the responsibility to manage the operation order of the platform. However, the profit-
seeking motivation of the platform supervises the participants to a certain extent. Plat-
form governance can control the opportunistic behavior of members of the innovation
ecosystem and ensure the effective operation of the innovation ecosystem, such as the
reputation mechanism of member star rating and the seller rating of Taobao. At the same
time, the platform may also use information asymmetry to take opportunistic behav-
ior to misappropriate the interests of its members. For example, in Apple’s innovation
ecosystem, technological innovation platforms such as iPhone and iPad dominate. There
are opportunistic behaviors that infringe on the interests of supporting enterprises such
as application software developers. Participants participate in governance for their own
needs.
54 L. Kewen and L. Junji
There Is a Conflict of Interest Among the Three Participants. Most platforms are
for profit. If the platform has violations of laws and regulations, including monopoly
and other situations that disrupt the market order and cause losses to participants, the
platform will face legal sanctions; Relevant government departments are responsible for
formulating laws and regulations to protect the interests of all participants; Participants
cannot distinguish information due to information asymmetry. They must obtain public
information such as platform transaction specifications.
The Governance Effect Depends on the Balance of the Three Parties’ Strategy
Choice. The government is responsible for formulating regulations to regulate all mem-
bers and coordinating the relationship between all participants. The platform needs to
use appropriate technical means to strengthen standardization and ensure compliance
operation; Participants need to improve self-discipline.
The platform owner is responsible for maintaining the long-term prosperity and
development of the entire ecosystem [17]. It can be seen that platform leaders have great
power, and there are many sources and manifestations of power. But the other side of
power is a responsibility, which comes with power and responsibility.
Simultaneous Eqs. (13), (14) and (15) obtain eight equilibrium points at which all
three parties of the game reach equilibrium, namely (0, 0, 0), (1, 0, 0), (0, 1, 0), (0, 0,
1), (1, 1, 0), (1, 0, 1), (0, 1, 1) and (1, 1, 1). In the digital innovation ecosystem, the
government, the platform, and the participants play the game and learn by imitation.
Over time, the three parties optimize their strategies and finally achieve the equilibrium
of the game. According to hypothesis (3), it is meaningful for any initial point and its
evolved point to meet v = {(x, y, z) | 0 ≤ x ≤ 1, 0 ≤ y ≤ 1, 0 ≤ Z ≤ 1}. Therefore, the
eight equilibrium points are meaningful.
(1) If the platform implements the governance strategy, the initial value of revenue is
set to P1 = 100, 110, and 120, respectively, as shown in Fig. 2. With the increase
of the initial income of the platform, the three curves converge quickly, and all
three parties choose governance as the final strategy. With the increasing revenue
of the platform and the strengthening of platform governance, the government’s
governance strategy is stable in a small range, and the change of participants is small.
Figure 2 shows that driven by interests, and the platform tends to choose the strategy
Evolutionary Game Analysis of Digital Innovation Ecosystem Governance 57
Fig. 2. Dynamic evolution diagram of platform revenue adopting governance strategy under
different initial conditions
(2) If the government implements the governance strategy, the initial value of income
is set to Pg1 = 40, 50, and 60, respectively, as shown in Fig. 3. It can be found
in Fig. 3 that when the initial income of the government is low, it is easier to
promote the platform to adopt governance strategies, and it is more suitable to
take the middle state of the three curves. Therefore, to increase the overall benefits
of the digital innovation ecosystem, the government needs to formulate policies
and regulations to guide the platform to participate in governance actively. At the
same time, the platform attracts participants to participate in platform governance
Fig. 3. Dynamic evolution diagram of government revenue adopting governance strategy under
different initial conditions
58 L. Kewen and L. Junji
through operation to achieve the balance of the tripartite game. In addition, the
government’s increased punishment on the platform can not promote governance,
as shown in Fig. 4.
Fig. 4. Dynamic evolution diagram under different initial conditions of government punishment
on the platform
(3) Participants choose to participate in the governance strategy, and the initial income
value is set to Pu1 = 30, 45, and 60, respectively, as shown in Fig. 5. It can be
found in Fig. 5 that the lower-income of participants is conducive to promoting
the implementation of governance of the platform, reflecting that the platform and
participants are a mutually beneficial and symbiotic community of interests, which
is the essential feature of the digital innovation ecosystem.
References
1. Yoo, Y., Henfridsson, O., Lyytinen, K.: Research commentary–the new organizing logic of
digital innovation: an agenda for information systems research. Inf. Syst. Res. 21(4), 724–735
(2010)
2. Yoo, Y., Boland, R.J., Jr., Lyytinen, K., Majchrzak, A.: Organizing for Innovation in the
Digitized World. Organ. Sci. 23(5), 1398–1408 (2012)
3. Nambisan, S., Lyytinen, K., Majchrzak, A., Song, M.: Digital innovation management:
reinventing innovation management research in a digital world. MIS Q. 41(1), 223–238 (2017)
4. Moore, J.F.: Predators and prey: a new ecology of competition. Harv. Bus. Rev. 71(3), 75–86
(1993)
5. Iansiti, M., Richards, G.L.: Information technology ecosystem: structure, health and perfor-
mance. Antitrust Bull. 51(1), 77–110 (2006)
60 L. Kewen and L. Junji
6. Schmeiss, J., Hoelzle, K., Tech, R.P.: Designing governance mechanisms in platform ecosys-
tems: addressing the paradox of openness through blockchain technology. California Manag.
Rev. 62(1), 121–143 (2019)
7. Rong, K., Lin, Y., Shi, Y., Yu, J.: Linking business ecosystem lifecycle with platform strategy: a
triple view of technology, application and organization. Int. J. Technol. Manage. 62(1), 75–94
(2013)
8. Chen, S.H., Lin, W.T.: The dynamic role of universities in developing an emerging sector: a
case study of the biotechnology sector. Technol. Forecast. Soc. Chang. 123, 283–297 (2016)
9. Borgh, M.V.D., Cloodt, M., Romme, A.G.L.: Value creation by knowledge-based ecosystems:
evidence from a field study. R&D Manag. 42(2), 150–169 (2012)
10. Amitrano, C.C., Coppola, M., Tregua, M.: Knowledge sharing in innovation ecosystems: a
focus on functional food industry. Int. J. Innov. Technol. Manag. 14(5), 1–18 (2017)
11. Adner, R.: Ecosystem as structure: an actionable construct for strategy. J. Manag. 43(1), 39–58
(2017)
12. Sawhney, M., Verona, G., Prandelli, E.: Collaborating to create: the internet as a platform for
customer engagement in product innovation. J. Interact. Mark. 19(4), 4–17 (2005)
13. Ansari, S., Garud, R., Kumaraswamy, A.: The disruptor’s dilemma: TiVo and the US television
ecosystem. Strateg. Manag. J. 37(9), 1829–1853 (2016)
14. Hanelt, A., Bohnsack, R., Marz, D., et al.: A systematic review of the literature on digital
transformation: insights and implications for strategy and organizational change. J. Manag.
Stud. (2020). https://doi.org/10.1111/joms.12639(2020)
15. Fukuyama, F.: What is governance? Governance 26(3), 347–368 (2013)
16. Gorwa, R.: What is platform governance? Inf. Commun. Soc. (2019). https://doi.org/10.1080/
1369118X.2019.1573914
17. Boudreau, K.J., Hagiu, A.: Platform Rules: Multi-Sided Platforms as Regulators. Working
Paper. Harvard University (2008)
Measurement and Comparison of Digital
Economy Development of Cities in China
Abstract. As a new growth pole, the digital economy has become an important
force for competition among cities. Based on the existing research results, this
paper defines the connotation of the digital economy. Following the principles of
scientific, normative, effective, and conciseness, the index system of the digital
economy is constructed from four aspects: digital infrastructure, digital industri-
alization, industrial digitization, and the development environment of the digital
economy. Then measure and compare the digital economic development level of
15 sub-provincial cities from 2012–2019. The results show that the overall digi-
tal economy shows a steady upward trend, but there are great differences among
the 15 sub-provincial cities. The level of industrial digitization and development
environment of the digital economy is high, the level of digital infrastructure and
digital industrialization needs to be improved. According to the above conclusions,
suggestions are put forward to improve the development of the digital economy,
and these suggestions guide policy formulation in different cities.
1 Introduction
In 2018, China’s digital economy reached 31.3 trillion yuan, accounting for 34.8% of
GDP. The digital economy plays a core role in China’s economic growth. With the deep
integration of the new generation of information technology and urban construction,
smart and digital cities have become an important trend in future development, which
provides new opportunities for urban transformation. Digital development reflects in
industry, social governance, and so on. In 2018, the urban digital economy accounted
for more than 90% of the total digital economy. Shenzhen promotes the development
of 5G and industrial Internet based on the advantages of the electronic information
industry and software industry. Hangzhou leads the transformation of the service industry
by relying on e-commerce and other industrial digitization advantages and leading the
forefront of digital industrialization such as cloud computing. Each city closely combines
its industrial foundation, resource endowment with digital economic development and
builds a digital development road with its characteristics. sub-provincial cities are the
core cities in the region. Their level of digital economy development is between ordinary
cities and cities such as Beijing and Shanghai. These cities are likely to become a solid
force in promoting regional digital economy development in the future. However, there
are few studies on the digital economy development of sub-provincial cities. Therefore,
it is necessary to accurately grasp the level of digital economy development of sub-
provincial cities and understand their difference, advantages, and disadvantages.
2 Literature Review
At present, the digital economy is a hot topic in academic circles. There are many research
results, which can be summarized into three categories. The first category is the impact of
the digital economy on other research objects, such as the impact of the digital economy
on the competitiveness of export products [1]. The second category is the research on the
digital economy itself, including connotation definition [2], statistical classification [3],
development difficulties [4], and other aspects. The third category is the research on new
formats, such as platform economy. There are few types of research on constructing index
systems of the digital economy and measuring the level of digital economy development
in different regions. At present, the construction of a digital economy index system has
two dimensions.
Constructing digital economy index system based on digital influence path. BEA
constructs an index system from four dimensions: the degree of digitization in the eco-
nomic field, the impact of digitization in economic activities, the new digital field, and
economic indicators [5]. OECD measures the level of digital economy development
from four parts: infrastructure, social use, related innovation, and the impact on the
economy and employment [6]. IDI constructs an index system from three aspects: ICT
access, ICT use, and ICT skills. CAICT constructs an index system from three aspects:
basic conditions of digital economy development, digital industrialization, and industrial
digitization.
Constructing a digital economy index system based on the external driving force of
the digital economy and its influence. WEF believes that the environment, readiness,
and application are driving forces for digital economy development, and they affect
the digital development of the economy and society. Therefore, WEF constructs an
index system from four aspects: environment, readiness, application, and influence.
Chuanhui Liu and Zhipeng Yang(2021) [7] constructs an index system from five aspects:
digital infrastructure, digital industrialization, scientific education, human resources, and
economic development.
The research on the construction and measurement of digital economy index systems
has progressed through literature review. Most of them take countries and provinces as
the research objects for relevant measurement, and few take cities as the research objects.
From the perspective of connotation, this paper defines the connotation of the digital
economy as follows: digital infrastructure is the premise of development, a digital indus-
try that provides technical support, is the leading industry, the traditional industry realizes
digital development and will deepen the depth and breadth in a good development envi-
ronment. The contributions of this paper are as follows: First, based on the connotation
of definition, the index system is constructed from four dimensions: digital infrastruc-
ture, digital industrialization, industrial digitization, and development environment. The
available index data are used for relevant measurement. Second, the development level
of the digital economy is analyzed from the perspective of sub-provincial cities.
Measurement and Comparison of Digital Economy Development 63
Guided by the scientific, normative, effective, and concise principles, taking into account
data availability. According to the connotation of the digital economy, the index system
is constructed from four dimensions: digital infrastructure, industrial digitization, digital
industrialization, and development environment. As shown in Table 1.
In this paper, research object is 15 sub-provincial cities. They are Shenyang, Dalian,
Ningbo, Xiamen, Changchun, Harbin, Jinan, Qingdao, Nanjing, Hangzhou, Wuhan,
Chengdu, Xi’an, Guangzhou, Shenzhen.
Table 1. (continued)
In this paper, the entropy weight method is used to weight index. This method can
ensure the full use of original data information, make the evaluation results more objec-
tive. Based on the basic data of 15 sub-provincial cities from 2012–2019, we can obtain
the index weight in Table 1 and the digital economy development index in Table 2 and
Table 3.
Standardization
Due to the obvious dimension difference between the different indicators, in order to
ensure accuracy, this paper uses the extreme value method to standardize the original
data. The formula is as follows.
Positive indexes:
xij −min{xj }
xij = max{xj }−min{xj }
(1)
Negative indexes:
max{xj }−xij
xij = max{xj }−min{xj }
(2)
xij is the standardized data of the j-th index of the i-th research object, i = 1, 2, …,
m; j = 1, 2, …, n.
m
ej = −k ∗ i=1 pij lnpij (4)
Measurement and Comparison of Digital Economy Development 65
1
k=
lnm
gj = 1 − ej (5)
g
wj = n j (6)
j = 1 gj
n
Si = j = 1 wj∗ xij (7)
Based on the relevant data of 15 sub-provincial cities from 2012–2019, this paper uses
MATLAB software to realize the entropy weight method. The comprehensive develop-
ment index for 15 sub-provincial cities is shown in Table 2. The analysis found that the
overall development level of the digital economy in each city shows a steady growth
trend from 2012–2019. This shows that the digital economy has gradually penetrated
all aspects of social life and is supported by the outside world. The development trend
is good. Guangzhou has been ranked first in the quality of digital economy develop-
ment from 2012 to 2019. The comprehensive development index of the digital economy
increased from 0.427 in 2012 to 0.6738 in 2019. These fully illustrate the strong digital
economy competitiveness of Guangzhou. From the average ranking, the average value
of the comprehensive development index of the digital economy is 0.4849, 0.4773, and
0.4718 in Shenzhen, Chengdu, and Wuhan from 2012 to 2019. Shenzhen, Chengdu,
and Wuhan rank second, third and fourth, respectively. These three cities have obvious
advantages over other cities. Shenzhen and Chengdu develop rapidly. Their compre-
hensive development index of the digital economy increased from 0.2723 and 0.2868
in 2012 to 0.6738 and 0.6456 in 2019, respectively. Wuhan has a good foundation for
digital economy development. Its comprehensive development index of the digital econ-
omy was 0.3773 in 2012, ranking second among 15 sub-provincial cities. With steady
growth, its comprehensive development index is 0.5899 in 2019.
66 P. Han and J. Li
City 2012 2013 2014 2015 2016 2017 2018 2019 Mean Rank
Shenyang 0.1944 0.2009 0.2306 0.2774 0.2306 0.3048 0.3448 0.3614 0.2681 12
Dalian 0.1944 0.1991 0.1959 0.2334 0.2624 0.2559 0.3312 0.3217 0.2486 14
Changchun 0.1487 0.1728 0.2076 0.2404 0.2745 0.2775 0.321 0.3207 0.2465 15
Harbin 0.1936 0.2414 0.2644 0.3167 0.3321 0.2678 0.2975 0.3641 0.2847 11
Nanjing 0.271 0.3176 0.355 0.4019 0.4588 0.4825 0.5121 0.5683 0.4209 5
Hangzhou 0.2864 0.3102 0.3473 0.382 0.4546 0.491 0.518 0.569 0.4198 6
Ningbo 0.1804 0.21 0.2489 0.2756 0.3291 0.3583 0.381 0.423 0.3009 10
Xiamen 0.1456 0.1641 0.203 0.2296 0.2775 0.3056 0.3414 0.3743 0.2551 13
Jinan 0.2225 0.27 0.2988 0.3308 0.3682 0.3612 0.3722 0.4652 0.3361 8
Qingdao 0.189 0.2205 0.3202 0.2826 0.3244 0.3602 0.3884 0.3998 0.3106 9
Wuhan 0.3773 0.3744 0.4114 0.4436 0.5048 0.5312 0.542 0.5899 0.4718 4
Guangzhou 0.427 0.4956 0.5103 0.5195 0.5658 0.6332 0.6891 0.7388 0.5724 1
Shenzhen 0.2723 0.3271 0.3673 0.4422 0.5675 0.5791 0.6084 0.6456 0.4773 2
Chengdu 0.2868 0.3214 0.3673 0.4422 0.5675 0.5791 0.6084 0.6456 0.4773 3
Xi’an 0.2689 0.3278 0.3428 0.3799 0.4242 0.4635 0.4554 0.5111 0.3967 7
the construction of “new infrastructure” project, and establish an efficient sharing mech-
anism of digital infrastructure in “Measures to Accelerate the Creation of a Leading City
for Digital Economy Innovation”. However, Harbin and Xiamen lack clear goals for the
construction of digital infrastructure. Although Changchun and Shenyang put forward
to promote digital infrastructure construction, they lack funds and technical investment.
Therefore, the ranking of the digital infrastructure index is relatively backward.
From digital industrialization, 15 cities are in a fluctuating upward trend from 2012
to 2019. The digital industrialization index of Chengdu, Shenzhen, Guangzhou, and
Xiamen increased rapidly, from 0.005, 0.17, 0.128, and 0.001 in 2012 to 0.032, 0.042,
0.033, and 0.017 in 2019. The average ranking of Shenzhen, Chengdu, and Guangzhou
from 2012 to 2019 is in the top 3. Guangzhou is located in China’s first array of sub-
division fields of the digital industry, such as the new generation communication and
satellite navigation industry, artificial intelligence industry, software, information tech-
nology service industry, big data industry, etc. Shenzhen gathers a group of high-end
software vendors, such as Jindie Tianyan, Tongxin software. Chengdu builds an inno-
vative development pilot zone to promote industrial Internet, software, big data, and
other industries. The average ranking of Harbin, Shenyang, Changchun, Qingdao, and
Ningbo from 2012 to 2019 is in the bottom 5. Their index grows slowly, from 0.003,
0.003, 0.002, 0.001 and 0.002 in 2012 to 0.006, 0.003, 0.003, 0.003 and 0.003 in 2019.
From industrial digitization, 15 cities are in a fluctuating upward trend from 2012 to
2019. The gap between cities is not very prominent. This shows that the role of the digital
economy is becoming more and more obvious in cultivating new momentum, improving
product and service quality. Cities are laying out to grab the commanding heights of a
new round of competition. The average ranking of Chengdu, Wuhan, Shenzhen, Nanjing,
and Hangzhou from 2012 to 2019 is the top 5, respectively 0.101, 0.095, 0.093, 0.09,
and 0.087. The average ranking of Harbin, Changchun, Xiamen, Dalian and Shenyang
from 2012 to 2019 is the bottom 5, respectively 0.081, 0.08, 0.078, 0.078, and 0.073.
From the development environment perspective, 15 cities are in a steady upward trend
from 2012–2019. This shows that the development environment stimulates the potential
of digital economy and injects vitality into the development of digital economy. The
average ranking of Guangzhou, Wuhan, Nanjing, Hangzhou and Shenzhen from 2012 to
2019 is the top 5. Their index grows from 0.186, 0.171, 0.138, 0.132 and 0.079 in 2012 to
0.451, 0.4, 0.369, 0.349 and 0.402 in 2019. There is a significant gap between these cities
and those with backward rankings. These cities attach great importance to personnel
training and technological innovation. Hangzhou, Shenzhen, Wuhan, Guangzhou and
Nanjing are all among the top 10 in science and technology innovation rankings in
“Science and Technology Innovation Competitiveness Report of Chinese Cities in 2020”.
These cities are strong in technological innovation and have little difference. The average
ranking of Shenyang, Ningbo, Xiamen, Dalian and Changchun from 2012 to 2019 is
the bottom 5. Their index grows from 0.068, 0.058, 0.063, 0.061 and 0.046 in 2012 to
0.231, 0.257, 0.246, 0.215 and 0.218 in 2019.
68 P. Han and J. Li
Table 3. Development index and mean value of four dimensions of 2012–2019 digital economy.
Based on the above research conclusions, the following implications can be obtained.
The digital economy, as a new economy form, is deeply integrated with the real economy.
It provides new opportunities for urban development, such as industrial digitization, dig-
ital governance, application, etc. However, there is still room for further improvement
in the development of the urban digital economy. The development differences between
cities can develop a digital economy according to factor endowments and economic
conditions. Specifically, the government of the Northeast region should give relevant
policies to promote infrastructure construction, including 5G base stations, cloud com-
puting centers, and industrial Internet. In addition, the government should also build
digital infrastructure to improve the digital transformation of traditional industries and
stimulate more potential users. More importantly, the government should also pay atten-
tion to scientific and technological innovation, because science and technology is the
core driving force for the development of the digital economy. The digital industry foun-
dation of eastern coastal cities is good, and these cities should focus on key technological
breakthroughs. In addition, cities in the Northeast need to strengthen cooperation with
developed cities to drive the innovative development of the digital economy.
References
1. Wang, Y.: Research on The impact of digital economy on the competitiveness of export prod-
ucts—taking the Yangtze river delta city group as an example. Price Theory Pract. 15183(2),
149–153 (2021)
2. Li, H.J., Zhang, J.L.: Some understandings on the definition of digital economy. Enterpr.
Econ. 20279(7), 13–22 (2021)
3. Liu, W., Xu, X.C., Xiong, Z.Q.: the international progress of digital economy classification
and China’s exploration. Finan. Trade Econ. 8299(7), 1–17 (2021)
4. Yu, Y.W., Chen, G.X.: Related issues and policy recommendations for the development of
China’s digital economy. Southwest Finan. 12472(7), 39–49 (2021)
5. Barefoot, K., Curtis, D., Jolliff, W., Nicholson, J.R., Omohundro, R.: Defining and Mea-
suring the Digital Economy. US Department of Commerce Bureau of Economic Analysis.
Washington, DC, 15 (2018)
6. Colecchia, A., et al.: Measuring the digital economy: a new perspective. In: OECD (2014)
7. Liu, C.H., Yang, Z.P.: Analysis of the characteristics of the temporal and spatial differences in
the measurement of urban agglomeration digital economy: taking six urban agglomerations
as examples. Mod. Manag. Sci. 9924(4), 92–111 (2021)
8. Wang, J., Zhu, J., Luo, Q.: China’s digital economy development level and evolution
measurement. Quant. Econ. Tech. Econ. Res. 7626(7), 26–42 (2021)
9. Liu, J., Yang, Y.J., Zhang, S.F.: Research on the measurement and driving factors of China’s
digital economy. Shanghai Econ. Res. 6648(6), 81–96 (2020)
10. Zhong, W., Zheng, M.G.: The influence and mechanism of digital economy on regional
coordinated development. J. Shenzhen Univ. (Human. Social Sci. Ed.) 4664(4), 79–87 (2021)
11. Zhao, T., Zhang, Z., Liang, S.K.: Digital economy, entrepreneurial activity and high-quality
development: empirical evidence from Chinese cities. Manag. World 8482(10), 65–76 (2020)
12. Guo, F., Wang, J.Y., Wang, F., Kong, T., Zhang, X., Cheng, Z.Y.: Measuring the development
of China’s digital financial inclusion: index compilation and spatial characteristics. Econ. (Q.)
1458(4), 1401–1418 (2020)
Occupational Risks and Guarantee Path
of the Delivery Man in Digital Economy
Abstract. With the digital economy thriving, there are gaps and problems in
protecting the occupational rights and interests of the delivery man that should be
solved urgently. Under Institutional Analysis and Development (IAD) framework,
some deeper reasons lead to risks about safety, health, stability, and insurance,
including digitized involution of platform companies, evaluation replacement in
social networks, the dilemma of livelihood, and problems of social security cov-
erage. In order to remove occupational risks and protect the occupational rights
and interests of the delivery man effectively, the government and forms should
upgrade working conditions and workflow, improve wages and provide training,
design evaluation mechanism, and industry guidelines, strengthen the basic social
insurance, and occupational injury insurance exploration.
1 Introduction
Digital economy research is the current hot topic, and the research of digital economy
and people’s welfare is the hot topic of China in the future [1]. Under the digital econ-
omy, platform and sharing are closely related to employment and social security, among
which the practitioners of new employment forms have new characteristics and needs.
The COVID-19 caused the development of online consumption and the digital economy
at a faster speed. Therefore the number of new employment forms of employees soared,
such as food delivery and online ride-hailing. And the protection of their rights and inter-
ests has got more and more attention. In recent years, scholars from many countries have
paid attention to the guarantee of practitioners of new employment forms. Kannan K.P.
and Papola T.S. (2007) proposed to improve the quality of employment in the informal
sector by improving income, working conditions, employment, and social protection [2].
But Rosenblat A. and Stark L. (2016), who researched Uber, denied classifying flexible
ride-hailing drivers as employees for social security, believing that if platforms classified
the drivers as employees, they would suffer great pressure. Their business model would
also be greatly impacted [3]. Harris S.D. and Krueger A.B. (2015) also studied and
argued that emerging working relationships in the Internet economy prevented indepen-
dent workers from obtaining the rights of traditional employees [4]. Chang K. (2019)
pointed out that the government should confirm the basic framework of the employment
relationship, strengthen and improve the rules of labor laws in order to realize the internal
unity of workers’ basic rights and interests and sustainable development of the Internet
economy [5]. In conclusion, scholars from all countries have studied the security of the
practitioners of new employment forms from the perspectives of social security and labor
relations. However, the delivery Man, which is the subdivision of these practitioners,
is still under discussion, and there is also a lack of specific research on using policy
tools. As one of the main new employment forms at present, food delivery workers face
occupational risks and difficulties under an immature industry system, and the protec-
tion of rights and interests is insufficient. From the digital economy and industry policy
perspective, the institutional situation, occupational risks of takeout workers, the path
of rights and interests protection can be analyzed using policy tools.
Ostrom E., the 2009 Nobel Prize winner in Economics, proposed Institutional Analy-
sis and Development framework (IAD framework), which reversed the view that the
institutional arrangements should be set by single part between government and market,
proved that common-pool pond resources can be governed better through reasonable
institutional arrangements by two parts of government and market, and this framework
has great value in economic governance and public resource governance [6]. As shown
in Fig. 1, the catering service in the digital economy as a new form is still in the explo-
ration of governance policy. The IAD framework can be used to study the independent
governance situation of catering industry market share resources in the digital economy,
analyze the system environment of takeout workers, and find the access point to analyze
takeaway workers’ occupational risk guarantee path.
Finally, external factors, action stage, interaction mode, evaluation criteria and results
influence each other, thus forming the catering industry system Institutional Analysis and
Development framework where food delivery workers are located. In this framework,
the market profitability of catering service public resources is the material condition, the
user consumer groups and the common market space of the catering service industry is
the community attribute, and the application principles are the confine, identity, choice,
information, aggregation, pay and range, which are put forward by Ostrom E.. In the
action stage, catering service enterprises, delivery workers, merchants, customers, the
government and other actors continue to interact [7]. In the face of relatively stable
market returns, catering service enterprises that have main control power among actors,
such as Meituan and Ele.me, must inevitably use digital platforms and algorithms to
compete for user groups and market space, and enterprises and takeout workers continue
to combat each other. As this emerging industry is in a stage of rapid development and the
standards in the industry have not been fully established, timely government intervention
and governance that can be integrated with market is particularly important.
72 J. Daizhi and W. Han
Fig. 1. The catering industry system Institutional Analysis and Development framework where
food delivery workers are located
Under the action scenario and interaction, delivery workers in the digital economy face
various risks, including safety, health, stability, insurance, etc.
Firstly, due to the low safety factor of electric bicycles and motorcycles, delivery rid-
ers often face high risks in the delivery process. Secondly, under the promotion and
requirement of platform enterprises, these workers must complete the food delivery
tasks intensively as soon as possible, which makes them have relatively great cycling
safety risks. In lunch and dinner time in big cities, dangerous scenes such as rushing
for red lights, speeding, and sudden turning without warning can be seen everywhere.
Therefore, the possibility of traffic accidents of the delivery man is high [8]. Thirdly,
the professional nature of food delivery workers has the characteristics of long working
time and large working intensity, which makes it difficult to ensure travel safety with full
energy, resulting in huge safety risks. Finally, owing to the urgent work, it is possible to
conflict with customers, security guards and other groups, and the delivery workers also
face the risk of malignant injury.
Over labor phenomenon is increasingly prominent. The delivery often work late at night
or have no rest off, and at the same time, their three meals a day can not be guaranteed.
Irregular work conditions for a long time may cause sub-health problems such as sleep
deprivation and heavy physical burden. Long-term excessive labor will not only lead to
physical fatigue but also may even cause karoshi. In addition, frequent abuse and bad
comments may affect the mental health of food delivery workers.
Occupational Risks and Guarantee Path of the Delivery Man 73
Takeaway workers have flexible work, unstable income, and unstable labor relations.
In the delivery work dominated by the algorithm, takeaway workers often face passive
work conditions. Their lives and workplaces often change, and their work income also
changes greatly, making them unstable and making delivery workers’ family care to face
great uncertainty.
At present, the vast majority of delivery workers are not included in industrial injury
insurance. Some large-scale platform enterprises, such as Meituan, provide accident
insurance for takeout riders. Still, this guarantee has the characteristics of large arbitrari-
ness, small guarantee scope, and low guarantee level. Due to the differences in benefits
and systems, large platform enterprises can provide commercial insurance for food deliv-
ery workers. In contrast, small enterprises do not pay attention to the occupational risks
of food delivery workers and take some measures [9]. These phenomena reflect that the
new forms of business practitioners, such as delivery workers have not been covered in
the current guarantee of basic social insurance, and their worries have not been relieved.
In the behavioral situation, Meituan, Ele. me and other food delivery platforms hold
asymmetric control power and information, and obtain the high proportion of revenue.
In order to obtain market share and gain greater benefit, the market would compress the
delivery time, capture the best route, and reduce the delivery cost of consumers. With
extreme competition, the system setting to promote the optimal allocation of market
resources causes the involution, then the risk falls on the single merchant and delivery
man. The delivery situation after struggles of speed and time has gradually become a
general requirement. And the short rest after finding a new route has been automatically
compressed by the optimization of the platform route, so the income of each order
has been decreasing in recent years. Under such digital innovation, delivery workers
are involved in the platform competition and can only face the occupational risks with
digital requirements.
74 J. Daizhi and W. Han
4.2 The Replacement of the Evaluation Right in the Social Networks Leads
to the Increased Uncertainty of Occupational Income Risk
In the evaluation system of the work of delivery workers, there are four parts, includ-
ing platforms, merchants, delivery workers, and customers. Still, the evaluation right is
assigned to platforms and customers, which constitute an interactive platform pattern
that differs from the single evaluation of workers by enterprises in the traditional labor
relationship. Under the platform’s operation, the integration of the expectations of cus-
tomers and the platform’s requirements has elevated time requirements, and merchants
and delivery workers accept the time restrictions together. In the end, because delivery
workers have contact with customers, the overtime evaluation risk may be more calcu-
lated on the delivery workers. And uncertainties such as customer mood and possible
misunderstanding reflected in the platform also affect the work evaluation and income of
the delivery man directly. As customer feedback will directly affect the income and per-
formance, the excessive evaluation rights increase the occupational risk of the delivery
workers.
4.3 The Dilemma Under the Living Income and Expenditure Leading
to the Situation that the Delivery Man Can Not Avoid Risks by Themselves
As one of the actors of the catering industry system Institutional Analysis and Devel-
opment framework, delivery workers can not stabilize their income and expenditure
and cannot resist professional risks. According to Amartya Sen’s discussion on the lack
of capability, delivery workers have capability problems such as impaired economic
conditions and obstruction of interest expression channels [10]. As an individual under
the market economy, delivery workers face the balance of income and expenditure,
family life maintenance, and savings. At the same time, they also face the continuous
improvement of platform enterprises and strict punishment mechanisms. Under the high
pressure at both ends, the delivery man can only speed up and increase the quantity in
this dilemma, otherwise, they would seek another job. Living income and expenditure
require delivery workers to work faster and more. Even under the obstacles of complex
transportation environment, urban housing construction layout, and hot or cold weather,
delivery workers have no more choices. In short, risks attached to the profession can not
be avoided or selected.
4.4 Problems of Social Security Coverage Leads to the Situation that Worries
of Occupational Risks Can Not Be Relieved
Practitioners of new business is not covered by social security network built by govern-
ment which is one of the actors. With the guidance of CPCCC and actively exploration of
provinces, the security of these workers is developing, but there are still some coverage
problems because of labor relations restrictions, household registration restrictions, the
threshold of payment rate, immature occupational injury insurance pilots exploration.
Basic social insurance such as endowment insurance, medical insurance and industrial
injury insurance are mostly bound to labor relations, and it is difficult to define the labor
Occupational Risks and Guarantee Path of the Delivery Man 75
relationship between practitioners (like delivery workers) and platforms under new busi-
ness forms, so that finding the fit point under the institutional threshold is hard; payment
rates are too high for low-income delivery workers, which makes the willingness to
participate in the insurance reduced; remote settlement, transferring and other handling
problems exist, in the meantime, the workplace of the delivery man is changing all the
time, and the household registration may be inconsistent with the workplace, so the
contradiction between these two conditions hinders the implementation of the social
security rights of some delivery workers who have entered the social insurance network;
although the occupational injury insurance pilots has achieved some results, the number
of insured people is small, and this insurance is linked with social insurance, so further
exploration is still needed.
5.2 Improve the Welfare of Delivery Workers and Start Employment Training
to Empower ‘actors’
The special minimum wage guarantee system of takeout workers should be formulated in
combination with features of employment and income to prevent takeout workers from
being trapped under the algorithm of the platform economy. On the premise of ensuring
the minimum wage, the firms should also ensure the normal diets of three meals a day,
adopt shift and zoning, provide subsidies of hot and cold days, and ensure the normal
holidays. In addition, the work training of the whole delivery process and psychological
counseling should be carried out to reduce conflicts and difficulties in work, improve
work satisfaction, maintain mental health, and resist the health risks efficiently.
76 J. Daizhi and W. Han
5.3 From the Perspective of ‘control Power’ Balance, the Evaluation Mechanism
and Industry Guidelines Should be Formulated
To solve the problems of the stability risks, the government should intervene in the
catering industry to reduce corporate control power in action scenarios. In the current
evaluation way, data statistics under platforms and customers make work performance
anamorphic. The unreasonable evaluation mechanism caused by inconsistent overtime
standards, unreasonable time compression, and subjective arbitrariness of customers
affects work effect and damages delivery man’s professional rights and interests. The
government should cooperate with the top enterprises of the platform to jointly formulate
a sustainable work evaluation mechanism that is in line with the rights of the delivery
man and the business model of the platform. Removing force majeure from overtime
reasons, and ensuring no direct income deduction from negative comments should be
considered. After establishing government and firms, the reasonable evaluation mecha-
nism for delivery workers could be adjusted according to the applicable situation, and
smaller platform enterprises can also adopt it. In addition, based on in-depth research, the
government can make ranking lists of catering service firms termly by work satisfaction
of delivery man and customer satisfaction to prevent monopoly and vicious competi-
tion of platform head enterprises. And these enterprises should set distribution schemes
according to the normal work intensity and time instead of squeezing distribution time
and route of the delivery man and challenging their body limit [11].
5.4 With the Scope of Public Policy of the Government Which Is One of ‘actors’,
Strengthen the Expansion of Basic Social Insurance and the Exploration
of Occupational Injury Insurance
To strengthen the coverage of basic social insurance and to explore occupational injury
insurance systems constantly can resist insurance risk of delivery workers. Social secu-
rity is the basic institutional arrangement of the state in the field of risk management,
aiming to ensure the basic survival, basic development and rights and interests of all
members of society [12]. At present, China has built the world’s largest social security
system. Social security will move from ‘wide cover’ to ‘full cover’ in the 14th Five-Year
Plan period, [13] ‘Improve the universal social security system, promote the development
of sustainable development’ is the overall goal of China’s 14th Five-Year period. Pro-
moting the delivery man to be with social insurance is the inevitable requirement of ‘full
cover’ [14]. Now, the reverse choice problem is prominent in new forms employment
groups such as employment group did not pay basic endowment insurance and medical
insurance. The government should focus on the nature of these groups, actual income
level to determine the reasonable social security rate and provide a convenient payment
method, which can make most workers be included in the scope of the social security
system. Furthermore, in terms of occupational injury insurance currently being tested,
under the principle of being pilot first before promotion, all provinces should strengthen
the top-level design, increase the pilot cities constantly, and explore the experience.
Occupational Risks and Guarantee Path of the Delivery Man 77
6 Conclusion
The delivery man is an important medium for catering enterprises and consumers. Reduc-
ing their professional risks and improving the protection of their rights and interests are
the key points to making the catering service industry vibrant and improving consumers’
satisfaction. Continuing to implement the guarantee mechanism and gradually reducing
the occupational risks of delivery workers can effectively improve the work satisfaction,
bring more benefits to catering enterprises, and let the majority of consumers enjoy con-
venient and fast delivery services cheerfully to improve the well-being of people and
promote the sustainable and healthy development of the digital economy.
References
1. Pu, D.X., Huo, H.F.: Hot spots, trends and prospects of digital economy research. Stat. Decis.-
Mak. 15, 9–13 (2021)
2. Kannan, K.P., Papola, T.S.: Workers in the informal sector: initiatives by India’s national
commission for enterprises in the unorganized sector (NCEUS). Int. Labour Rev. 146(3–4),
321–329 (2007)
3. Rosenblat, A., Stark, L.: algorithmic labor: a case study of uber’s drivers. Social Sci. Electron.
Publ. 10(27), 3758–3784 (2016)
4. Harris, S.D., Krueger, A.B.: A proposal for modernizing labor laws for 21th century work:
the independent worker. Discuss. Paper (12), 10 (2015)
5. Chang, K., Zheng, X.J.: An employment relationship or a partnership? an analysis on the
nature of labor relations in internet economy. J. Renmin Univ. China 33(02), 78–88 (2019)
6. Wang, Q.: Review of the ostrom institutional analysis and development framework. Econ.
Dyn. 04, 137–142 (2010)
7. Ostrom, E., Gardner, R., Walker, G.: Rules, Games, and Common Pool Resources. University
of Michigan Press, Ann Arbor (1994)
8. Hu, F.Z., Bai, Z.Y., Yao, Y.R.: The dilemma and countermeasures of excessive labor and
occupational injury guarantee: sell riders for example. China Econ. Trade J. (China) 02,
142–144 (2021)
9. Zhou, Y.B.: Thoughts on the occupational injury protection problem of new business form
practitioners. China Human Res. Social Secur. 10, 9–11 (2020)
10. Deng, D.S., Wu, F., Xiong, Y.: Study on the occupational injury guarantee of the new
generation of migrant workers in China. Jiangxi Social Sci. 34(11), 193–197 (2014)
11. Guan, B., Wang, Z.: Protection of The rights and interests of new employment youth: dilemma,
adjustment and problem. Chinese Youth Stud. 04, 22–28 (2021)
12. He, W.J.: Promote a fairer, more sustainable, and more efficient social security. People’s
Forum 31, 37–39 (2019)
13. He, W.J.: From ‘wide cover’ to ‘full cover.’ China Social Secur. 05, 42–43 (2013)
14. Han, Y.: Adhere to the attribute of social insurance, and expand the pilot project of occupational
injury security. Human Res. Social Secur. China (04), 58 (2021)
Research on E-commerce Recommended
Algorithm Based on Knowledge Graph
Yu Zhang1,3,4(B) , Jingming Ye1 , Xin Yue2 , Sifan Wei1 , Yu Wang1 , and Wanli Ren1
1 Harbin University of Commerce, Harbin 150028, China
2 Heilongjiang University of Science and Technology, Harbin 150022, China
3 Heilongjiang Provincial Key Laboratory of Electronic Commerce and Information Processing,
Harbin 150028, China
4 Heilongjiang Cultural Big Data Theory Application Research Center, Harbin 150028, China
1 Introduction
E-commerce [1] is a model to complete transactions through the web platform. The
amount of commodity information increases with its wide application, which leads to
information overload. Commodity recommendation technology, meanwhile, was born
and is developing. It aims to extract and analyze the relevant information of goods
searched and purchased by users in the past, then recommend goods of interest back
to them through algorithms. Due to the traditional one’s is difficult to meet the needs
of users. This paper proposes an improved recommendation algorithm [2] based on
knowledge graph [3], improving the algorithm’s accuracy and enhancing user experience.
algorithms was on the rise from 2016 to 2018 and peaked in 2017. However, during the
outbreak of COVID-19, namely, from the end of 2019 to the end of 2020, the number of
published papers decreased significantly, indicating that scholars in this field paid less
attention due to the epidemic. Since 2021, the number of published papers has been on
the rise again, indicating that scholars have re-engaged in the research in this field again.
Therefore, it can be preliminarily estimated that the epidemic has a negative impact
on the research of e-commerce recommendation algorithm. The number of published
papers on “recommendation algorithm” in the e-commerce field is 254, accounting for
19.39% of the total published papers, indicating a hot research topic in recent years.
Through the network knowledge graph analysis of keywords of electronic com-
merce recommendation algorithm, the number of occurrences of keywords such as “rec-
ommended system”, “collaborative filtering”, “personalized recommendation”, “recom-
mendation algorithm”, “electronic commerce”, “matrix decomposition”, “similarity” is
at the forefront, which is the hot research topic on electronic commerce recommenda-
tion algorithm. The keywords with high centrality [4] are “recommendation algorithm”,
“recommendation system”, “personalized recommendation” and “collaborative filter-
ing", in centrality of 0.45, 0.34, 0.32 and 0.27, respectively. In general, the keywords
with a centrality greater than 0.2 indicate that the keywords play the role of research
supporting point in the whole research field and are the key research object.
Based on the keyword clustering analysis of the research of e-commerce data, the
keywords are clustered into 9 categories, namely, personalized recommendation, recom-
mendation algorithm, matrix decomposition, collaborative filtering, Spark, deep learn-
ing, trust, clustering, and similarity. The Q value of the cluster module is 0.3746, and the
average contour S value of the cluster is 0.6835. It is generally believed that the cluster
structure is significant when the Q value is greater than 0.3, and that the clustering results
are convincing when the S value is greater than 0.6 [4].
According to the analysis of 1036 selected literature, it is found that after the outbreak
of the epidemic, Chinese scholars’ enthusiasm for the study of e-commerce recommen-
dation algorithm has decreased significantly, and the number of published papers has
continued to decline and even reached the lowest point in recent years. However, rec-
ommendation algorithm is still a hot research topic in the e-commerce field. In order to
mitigate the impact of the epidemic on the e-commerce field, the author improved the
existing recommendation algorithm by combining the knowledge graph to help solve
the problem of information overload and make the e-commerce field develop better.
The traditional recommendation algorithms mainly include collaborative filtering
algorithm [5], content-based recommendation algorithm [6], and hybrid recommenda-
tion algorithm [7]. With the development of the “Internet” era and the constant increase
of data and information day by day, the demands of users are constantly changing, so it
becomes more difficult to find the required information in the massive data. Therefore,
a perfect and efficient recommendation system can play a great role. However, there
still exists many problems of recommendation system, such as content-based recom-
mendation algorithm that has the problems of feature extraction difficulty, unable to
recommend new items and cold start problem, the collaborative filtering algorithm that
has the data sparseness problem and cold start problem, and the hybrid recommendation
algorithm that has the problem of increased complexity and increased operating cost.
80 Y. Zhang et al.
3.1 Integrate Commodity Entities and User Comments in the Knowledge Graph
The feature words of the commodity include the performance, cost performance, appear-
ance, user experience, and other features directly related to the commodity and describe
the attributes of the commodity. The sentiment words include the polarity and strength
judgment of the user emotion on the commodity. By extracting a large number of feature
words and sentiment words, the general opinion of the users on the commodity can be
obtained.
Extract Commodity Features and Sentiment Words. The traditional semantic rela-
tion LDA [8] has disadvantages as it is difficult to extract commodity feature words
with low word frequency and meet the allocation requirements of fine-grained words.
Therefore, the SRC-LDA [9] model is used to extract the feature words and sentiment
words.
Firstly, the author fused the user comments with the commodity feature set and
sentiment word set in the knowledge graph and pre-processed the data. The commodity
entity is W = {w1 ,w2 ,…,wn }, the total feature set is F = {f 1 ,f 2 ,…,f n }, and the total
sentiment word set is L = {l1 ,l 2 ,…,l n }, where wi is the ith commodity entity, fi is
the feature word set of the ith commodity, and li is the sentiment word set of the ith
commodity.
Building Semantic Relationships. In this paper, we constrained the feature words and
sentiment words of commodities to extract more accurate feature words and sentiment
words that meet the demands of users, introduced the semantic constraint of must-link
[10] so that the words that meet the must-link relation are allocated to the same set, to
find more fine-grained feature words and emotion words by exploring the correlation
between different words through the semantic relation constraint.
Obtain the Must-Link Between Feature Words. There are words with the same semantic
meanings in users’ comments, such as “cheap”, “cost-effective”, and others. These words
have a strong must-link relationship and can replace each other, that is, this type of words
should be allocated to the same set as far as possible. The semantic relationship between
Research on E-commerce Recommended Algorithm Based on Knowledge Graph 81
feature words can be obtained by formula (1). Where S(f i , f j ) is equal to 1, representing
that there is a must-link relationship between commodity entity wi and wj , namely, high
synonymy. The same relationship between the feature words can be obtained by referring
to Corpus Annotation Expanded Edition.
1, if fi and fj have must − link
S(fi , fj ) = (1)
0 else
Must-link between Feature Words and Emotion Words. The judgment rules are as
follows: if a single sentence satisfies the basic subject-verb relationship, it can be roughly
concluded that the sentence’s subject is the feature word, and the corresponding adjective
is the emotion word. Based on the semantic relationship strength formula OES-PMI [9],
the semantic relationship between feature words and sentiment words can be obtained by
formula (2). Where, ζ 1 is the frequency threshold of feature words, ζ 2 is the frequency
threshold of emotion words, q(f i ) is the word frequency of candidate feature words,
q(l i ) is the word frequency of candidate emotion words, qc (f i , l i ) is the co-occurrence
frequency of candidate feature words and emotion words.
lg qc (fI , li )
OES − PMI (fI , li ) = , q(f ) < ξ1 , q(l ) < ξ2 (2)
lg q(fi ) lg(q(li ) − qc (fI , li ) i i
Selecting an appropriate threshold number, and taking the relational words of the
candidate feature words and sentiment words with the relational value is greater than a
certain threshold value, to form the feature word-sentiment word set, denoted as q = {q1 ,
q2 , …}, qi indicates the stronger semantic relationship between sentiment words and
feature words. To sum up, a feature word set based on sentiment words can be obtained,
denoted as F’ = {(f 1 , q1 ), (f 2 , q2 ),…}.
Pearson Coefficient. Let r(u, i) and r(v, i) be the evaluation values of the feature value
i of commodities u and v respectively, r(u) and r(v) are the evaluation values of the
average feature value of commodities u and v respectively, then P(u, v) is the similarity
of commodities u and v, calculated according to the Pearson correlation coefficient,
seeing formula (3). According to formula (3), the larger P value (u, v) is, the greater the
similarity between commodities u and v.
n
i=1 (r(u,i)−r(u))(r(v,i)−r(v))
P(u, v) = √n √
n (3)
i=1 (r(u,i)−r(u)) i=1 (r(v,i)−r(v))
2 2
82 Y. Zhang et al.
Sij
tij = m (4)
k=1 Sik
The similarity of the goods is taken as the edge weight of the random walk model
[13]. Assume r is the column vector of the goods, where each element represents the
probability of the goods being accessed. Formula (5) is the corresponding formula of
the random walk strategy.
r n = c × T × r n−1 + (1 − c) × r 0 (5)
Where, c is the probability of the random walk model moving to a certain commodity
and its nearest commodity, 1-c is the probability of moving to the next step and returning
to the initial node, r n is the probability of reaching other commodities at the n th step,
and r 0 is the initial probability distribution.
The essence of commodity random walk probability is commodity weight which
directly relates to feature weight, and commodity random walk probability can represent
the weight of feature weight. The vector of commodity i is Qi = {I 1 ,I 2 ,…I i ,…}T , if there
is feature i in f i , then I i = 1, otherwise I i = 0. Assume there are n pieces of goods, the
total goods matrix is I = {Q1 ,Q2 ,…,Qn }T , that is, the feature weight can be expressed
as B = (bi )n = (I)−1 × r n . Then, the commodity feature set with the commodity random
walk weight added is F W = {(f 1 , q1 , b1 ), (f 2 , q2 , b2 ), …}, where bi is the feature weight
of commodity i.
For commodity m, the recommendation value of the commodity to relevant users can
be calculated after the sentiment word-feature word set and the random walk weight set
of the commodity are obtained. The recommendation value of the commodity can be
obtained by formula (6).
V = q×b (6)
(f ,q,b)∈Fw
Research on E-commerce Recommended Algorithm Based on Knowledge Graph 83
After the recommendation value of each commodity to user is obtained, the recom-
mended commodity value is sorted from large to small, and the top N commodities in
terms of value are provided to the user as required.
The new recommendation algorithm proposed in this paper can be described as:
Input: Commodity entities correspond to user comments and knowledge graph.
The experimental data set was collected from the user data of Taobao mall (www.taobao.
com). By searching for the keyword “shoes”, the author collected 10,000 commodities
and 100,000 data comments. The author conducted knowledge extraction of Internet
data under the big data environment in the knowledge graph, and then carried out data
pre-processing.
Evaluation Index. This paper uses the indexes of precision P, recall rate R and F value
[14] to compare the recommendation algorithms. The indexes are calculated by formula
(7). Where, R(u) represents N goods recommended to user u, T (u) is the commodity
set that users are interested in the test set, and F value is the index to be considered
comprehensively when there is a contradiction between P and R indexes. The P, R and F
value are important indexes for evaluating recommendation systems. In this experiment,
the P refers to the ratio of recommended goods to the total number of goods. The R
refers to the ratio of extracted related goods to the total number of goods. F value is the
comprehensive average index of these two indexes. The value of P and R is between 0
and 1, and the closer the value is to 1, the higher the accuracy is.
Explore the Impact of the Number of Comments on the Algorithm Performance (Set
the Number of Commodities to 5000). As can be seen from Fig. 4, Fig. 5 and Fig. 6,
with the increase of the number of comments, the P, R and F value of the algorithm
proposed are constantly rising. The accuracy is obviously better than that of the other
three algorithms when the number of comments is greater than 40000, and the R and F
values are both higher than the other three algorithms.
5 Conclusion
A recommendation algorithm integrating knowledge graphs and user comments is pro-
posed in this paper. The similarity matrix is built by combining knowledge graph and
random walk model, and the commodity features are taken as the research object to make
the recommendation more accurate. The experimental results show that compared with
the traditional recommendation algorithm, the commodity recommendation algorithm
proposed in this paper achieves a better recommendation effect and solves data scarcity
and poor recommendation diversity to a certain extent.
References
1. Hu, Y.H., et al.: Recognizing the same commodity entities in big data. J. Comput. Res. Dev.
52(08), 1794–1805 (2015)
2. Gu, Q.Y., Wu, B., Hu, Q.Q., Sun, Z.Y.: Social network user interest points recommenda-
tion algorithm based on multidimensional feature fusion. J. Commun. Netw. Initial Address
(2021). http://kns.cnki.net/kcms/detail/11.2102.tn.20210401.0909.002.html
3. Wang, J.F., Wen, X.L., Yang, X., Zhang, Q.L.: A bias based graph attention neural network
recommender algorithm. control and decision. Netw. Initial Address (2021). https://doi.org/
10.13195/j.kzyjc.2020.1626, Accessed 31 July 2021
4. Niu, J.F., Li, C.L., Chen, Z.G.: Visualization analysis of knowledge mapping in research on
product supply chain. Supply Chain Manag. 2(07), 54–61 (2021)
5. Wang, H.M., Nie, G.H.: Research on collaborative filtering algorithm based on fusing user
and item’s correlative information. J. Wuhan Univ. Technol. (07), 160–163 (2007)
6. Huang, Z.H., Zhang, J.W., Zhang, B., Yu, J., Xiang, Y., Huang, D.S.: Survey of semantics-
based recommendation algorithms. Acta Electron. Sin. 44(09), 2262–2275 (2016)
7. Zhang, Y.H., Zhu, X.F., Xu, C.Y., Dong, S.D.: Hybrid recommendation approach based on
deep sentiment analysis of user reviews and multi-view collaborative fusion. Chin. J. Comput.
42(06), 1316–1333 (2019)
8. Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent dirichlet allocation. J. Mach. Learn. Res. 3(3),
993–1022 (2003)
9. Peng, Y., Wan, C.X., Jiang, T.J., Liu, D.X., Liu, X.P., Liao, G.Q.: Extracting product aspects
and user opinions based on semantic constrained LDA model. J. Softw. 28(03), 676–693
(2017)
Research on E-commerce Recommended Algorithm Based on Knowledge Graph 87
10. Wagstaff, K., Cardie, C., Rogers, S., Schroedl, S.: Constrained K-means clustering with
background knowledge. In: 18th International Conference on Machine Learning, pp. 577–584
(2001)
11. Bagci, H., Karagoz, P.: Context-aware location recommendation by using a random walk-
based approach. Knowl. Inf. Syst. 47(2), 241–260 (2016)
12. Liu, W.Y.: User network service recommendation method and experimental results analysis
based on PRWDR algorithm. Tech. Autom. Appl. 39(09), 48–51 (2020)
13. Yu, C.H., Liu, X.J., Li, B.: Mobile service recommendation based on context similarity and
social network. Acta Electron. Sin. 45(6), 1530–1536 (2017)
14. Zhu, Y.X., Lv, L.Y.: Evaluation metrics for recommender systems. J. Univ. Electron. Sci.
Technol. China. 41(2), 163–175 (2012)
15. Jiang, M.Y., Zhang, Z.F., Jiang, J.G.: A collaborative filtering recommendation algorithm
based on information theory and bi-clustering. Neural Comput. Appl. 31(12), 8279–8287
(2019)
16. Li, H., Cai, F., Liao, Z.F.: Content-based filtering recommendation algorithm using HMM. In:
Proceedings of 2012 International Conference on Computational and Information Sciences,
pp. 256–263. IEEE Press, Washington D.C. (2012)
17. Zhao, J.K., Liu, Z., Chen, H.M.: Hybrid recommendation algorithms based on convmf deep
learning model. In: Proceedings of 2019 International Conference on Wireless Communi-
cation, Network and Multimedia Engineering, pp. 15–26. IEEE Press, Washington D. C.
(2019)
Research on Regional Heterogeneity
in the Impact of Digital Inclusive Finance
on the Diversification of Household Financial
Asset Allocation
1 Introduction
Inclusive finance is a concept officially introduced by the United Nations in 2005. From
the coverage perspective, it mainly focuses on small micro-enterprises, farmers, and
disadvantaged groups with low income. From the perspective of service characteris-
tics, the cost of inclusive digital finance is lower than that of traditional finance. It
uses artificial intelligence and other technologies to control the risk of data, improving
the efficiency of capital utilization through “accurate portraits”. It enhances financial
risk prevention, thereby reducing transaction costs and providing financial services to
more people. Through digital means, digital inclusive finance benefits people in back-
ward areas, rural areas, and other areas which can not get traditional financial services.
Internet finance enriches investment channels, increases investment diversification, and
brings higher and more stable returns. Family asset investments are no longer limited
to traditional bank deposits but also have various options such as securities, funds, and
financial products. Due to different levels of economic development and financial market
innovation, residents in different regions are exposed to different types of financial prod-
ucts and businesses. Financial exclusion prevents disadvantaged residents from financial
services, which seriously affects the rationalization of household asset allocation.
With the implementation of the rural revitalization strategy and the achievement of
poverty eradication by 2020, the wealth accumulation of residents has been increasing.
To achieve asset preservation and appreciation for families with “surplus wealth,” it is
necessary to make reasonable investments and financial management. This paper first
investigates the impact of digital financial inclusion development on household finan-
cial asset allocation diversification through empirical tests. In addition, this paper will
investigate the heterogeneity of this impact across regions. It illustrates the importance
of promoting inclusive digital finance on the financial investment environment at the
macro level. At the micro-level, it reveals that improving residents’ household financial
literacy can help optimize household financial asset allocation decisions.
2 Literature Review
The concept of digital financial inclusion was first introduced in China in 2016. Relevant
studies mainly focus on the impact of Internet finance on rural financial exclusion. As
the research progresses, some scholars turn to the impact of inclusive digital finance
on economic development and residents’ asset allocation. Many scholars believe that
the development and popularization of inclusive digital finance positively affects the
economy, employment, and social welfare. Tang, Yu et al. (2020) found that the broad
coverage level of inclusive digital finance could enhance investment opportunities and
strengthen financial services’ equity, thus realizing the economic source of long-lasting
growth [1]. Wang, Yannan et al. (2020) found that inclusive digital finance could enhance
social security by affecting income and employment. On this basis, some scholars study
the impact of inclusive digital finance on different industries and industrial structures [2].
Ding Rijia et al. (2020) found that inclusive digital finance has a significant contribution
to the development of the service industry, and the impact is manifested in different
degrees among different regions [3]. Tian Juanjuan and Ma Xiaolin (2020) found that
inclusive digital finance could lead to a more rational and efficient agricultural, industrial
structure and enhance the economic benefits of agriculture [4]. Regarding the impact
of inclusive digital finance on residents’ asset allocation, scholars have found that the
development of digital finance could not only inject vitality into China’s financial marke
[5, 6] but also effectively promote the optimization of household asset allocation (Deng
Shanshan, 2019) [7].
Through a brief literature review, we found that after the concept of inclusive digital
finance was introduced in China, most studies focused on the impact on economic
development, income growth, and residents’ consumption [8]. Most of them have used
provincial data to conduct empirical tests. However, few articles studied the impact of
inclusive digital finance on household financial asset allocation at the household level.
90 S. Jiang and X.-H. Zhao
This paper will use data from the China Household Tracking Survey (CHFS) of South-
west University of Finance and Economics and the Digital Inclusive Finance Index
compiled by the Digital Finance Research Center of Peking University. The China
Household Finance Survey Center (CHFS) has conducted the China Household Finance
Survey every two years since 2009, and its data are nationally representative. The Digital
Inclusive Finance Research Center of Peking University, using microdata on inclusive
digital finance from Ant Financial, compiled the Digital Inclusive Finance Index from
2011–2018, covering three levels: provincial, city, and county. In this paper, households
participating in the survey in 2011, 2013, 2015, and 2017 are selected as the sam-
ple [9]. According to the National Development and Reform Commission, the policy
on dividing the east-central and western regions. The eastern provinces include Bei-
jing, Tianjin, Hebei, Liaoning, Shanghai, Jiangsu, Zhejiang, Fujian, Shandong, Guang-
dong and Hainan. The western region includes Sichuan, Guizhou, Yunnan, Shaanxi,
Gansu, Ningxia, and Qinghai. The central region includes Shanxi, Inner Mongolia, Jilin,
Heilongjiang, Anhui, Jiangxi, Henan, Hubei, and Hunan [10].
Many factors are affecting the allocation of financial assets of residential households. We
need to add factors with a high degree of influence as control variables in the model to
ensure the model’s accuracy. The panel data model can better respond to the heterogeneity
in time and space. In order to reduce the impact of data heteroscedasticity on empirical
test results, we took the logarithm of all variables. The final model is built as follows
Among them, i and t represents the province i and year t, respectively. εi,t is the
random disturbance term. αi is the individual fixed effect, and αt is the year fixed
effect. Investdiv is the explanatory variable in the model, which denotes the number
of household financial asset allocations. It includes demand deposits, time deposits,
stocks, bonds, funds, financial derivatives, bank financial products, other financial prod-
ucts, gold, foreign investment products, and cash. The core explanatory variables are the
total digital financial inclusion index (Digital), depth of coverage (Depth), breadth of
coverage (Cover), payment index (Payment), credit index (Credit), and insurance index
(Insur). Control variables include the age of household head (Age), an education level
(Edu), marital status (Marry: Married 1, Unmarried 0), gender (Gender: Male 0, Female
1), number of household members (Num), and presence of housing assets (House: Yes
1, No 0).
Research on Regional Heterogeneity in the Impact of Digital Inclusive Finance 91
4 Empirical Tests
4.1 The Impact of Total Digital Inclusion Indicators on Household Financial
Asset Allocation Diversification
As can be seen from Table 1. The digital inclusive finance index and each control vari-
able have significant effects on household financial asset allocation. (1) The popularity
of inclusive digital finance expands investment channels for residents, promoting their
understanding of financial information, and increasing household financial asset allo-
cation diversification. (2) The education level of household heads positively affects the
diversification of household financial asset allocation. The higher the education level
of the household head, the more financial information he gets and the more rational he
considers the risks of financial products. He can use digital inclusive finance channels
more rationally to purchase investment products and allocate household assets when
allocating assets. (3) The age of the household head has a weak effect on the household
financial asset allocation. This may be the result of offsetting. On the one hand, as the
household head gets older, he acquires more financial knowledge to make more sensible
decisions on household investments and have a more rational asset allocation. On the
other hand, some older householders do not keep up with the times due to their low
education level, conservative thinking, and lacking knowledge about Internet finance.
He will prefer the single asset allocation method of traditional finance. (4) The increase
in household size has a negative effect. This may be due to the increase in the number
of teenagers and older adults in the household. Most of them are risk-averse and more
conservative in financial investment. It eventually makes the financial asset allocation
tend to be homogeneous. (5) Owning home equity plays a positive role. It may be that
households with housing assets do not need to set aside funds for house purchase or
rent. When basic living conditions are guaranteed, they are willing to put aside more
funds to make financial investments to enrich their household financial asset allocation.
Because of the uncertainty of the future, the homeless people need more precautionary
savings funds. Therefore, they have less funds for financial investment. (6) The gender of
the household head has a positive effect. Compared with male household heads, female
household heads tend to acquire multiple investment products when making financial
asset decisions. (7) Married heads tend to have more diversified financial assets than
unmarried heads.
Table 1. Test results of the impact of total digital financial inclusion indicators on the diversifi-
cation of household financial asset allocation
The test results are shown in Table 2. Different dimensions of digital financial inclusion
development (breadth of coverage and depth of use) and different sub-indicators of busi-
ness types (payment, credit, and insurance) significantly influence household financial
asset allocation. The breadth of coverage refers to the number of users who apply to reg-
ister their investment accounts and tie up their physical bank cards through websites. The
greater the breadth of coverage, the more users are exposed to inclusive digital finance,
and the residents have more opportunities to try different kinds of investment products.
It will lead to a more rational allocation of residents’ financial assets and more diver-
sified investments. Depth of use has a greater impact on the number of financial assets
allocated than the breadth of coverage because the depth of use reflects actual usage
and transactions better. For residents, it is more effective to contact Internet financial
services than to accept financial publicity and promotion. Payments and credit indices
have a greater impact on asset allocation diversity than insurance. This may be because
payment and credit services are basic businesses among traditional financial institutions
with lower thresholds and higher coverage. As a more advanced and complex financial
service, insurance has higher business thresholds. These businesses allow residents to
have more liquid assets and increase the way money flows. They reduce the income
pressure caused by the consumption so that residents have more funds available for
household financial asset allocation.
Table 2. Test results on the impact of digital inclusive finance sub-indicators on the diversification
of household financial asset allocation
Table 3. The results of a test of regional heterogeneity in the impact of digital financial inclusion
on the diversification of household financial asset allocation
Table 4. The differential impact of digital inclusive finance on the diversity of financial
asset allocation of urban and rural households
is the proportion of each investment in cash, demand deposits, time deposits, bonds,
stocks, gold, funds, derivatives, foreign currency assets, bank financial products, and
other financial products to the total investment in household financial assets. For this
part of the robustness test, this paper replaces the indicator measuring household financial
asset allocation dispersion. The model setup, control variables, and regression methods
are identical to those in Sect. 3.2. The formula for calculating the indicator is as follows:
n
Investshare = 1 − s2 , if n > 0 (2)
1
Under the premise of the same asset types, if the difference between the proportions
of assets is greater, it shows that households hold a high proportion of certain financial
assets. On the contrary, if the difference between the proportion of each asset is smaller,
it indicates that households are diversified into multiple assets when allocating financial
assets. The results of the test shown in Table 5 and Table 6 are more favorable. Invest share
will decrease as the diversification of household financial asset allocation increases. Thus
the negative effect coefficient indicates that the increase of digital inclusion makes Invest
share decrease. The total digital inclusion index has a positive effect on the diversification
of household financial asset allocation.
96 S. Jiang and X.-H. Zhao
5 Conclusion
This paper uses a dataset formed by data from the China Household Finance Survey
(CHFS) and the Digital Finance Research Center of Peking University to build a fixed-
effects model. This paper studies the impact of digital inclusion on regional differences
in household financial asset allocation diversity through empirical testing. The study
concludes
The main reason for this result is the different levels of regional economic devel-
opment and financial market sophistication. Based on the above research results, this
paper proposes the following suggestions: First, relying on the current financial institu-
tions that already exist, we should accelerate the introduction of inclusive digital finance
and upgrade various dimensions of inclusive digital finance. Second, the government
and companies should step up their efforts to promote digital financial inclusion and
enhance residents’ financial literacy and Internet thinking. Third, it is necessary to adapt
to local conditions when developing inclusive digital finance. For the western region,
the development of inclusive digital finance should focus on matching infrastructure
construction and the symmetry of financial products and services. For the eastern region
with a higher degree of financial development, the regulator should strengthen financial
supervision and create a fair financing environment.
References
1. Tang Y., Long Y., Zheng Z.: Study on the inclusive economic growth effect of digital inclusive
metals. An empirical analysis based on 12 provinces in Western China. Southwest Finan.
41(09), 60–73 (2020)
2. Wang, Y., Tan, Z., Zheng, L.: Study on the impact of digital inclu-sive metals on social
security. J. Quant. Tech. Econ. 37(07), 92–112 (2020)
3. Ding, R., Liu, R., Zhang, Q.: Research on the impact and mechanism of digital inclusive
finance on service industry development an empirical analysis based on inter-provincial panel
data. J. Finan. Econ. 40(7), 4–10 (2019)
4. Tian, J., Ma, X.: Analysis of the effect of digital inclusive finance to promote agricultural
transformation and upgrading–empirical evidence based on inter-provincial panel data. Credit
Refer. 38(07), 87–92 (2020)
5. Wang, S., Yang, B.: The marriage of rural finance and internet finance: impact, innovation,
challenges and trends. Rural Finan. Res. 38(08), 19–24 (2017)
6. Hang, Y., Huang, Z.: Digital financial development in China: present and future. China
Econom. Q. 17(04), 1489–1502 (2018)
Research on Regional Heterogeneity in the Impact of Digital Inclusive Finance 97
7. Deng, S.: A Study on the impact of digital inclusive finance on SME financing. Think Tank
Era. 3(37), 62+82 (2019)
8. Corrado, G., Corrado, L.: Inclusive finance for inclusive growth and development. Current
Opinion Environ. Sustain. 24,19–23 (2017)
9. Zhou, Y., He, G.: The impact of digital inclusive finance development on financial
asset allocation of farm households. Mod. Econom. Sci. 42(03), 92–105 (2020)
10. Jiang, S., Zhao, X.: The influence of shadow banking on China’s credit transmission
mechanism. Bus. Econom. 40(08), 166–169 (2021)
Research on the Countermeasures
for the Development of Agricultural and Rural
Digital Economy
Abstract. The digital economy has become an important means to promote the
high-quality development of agriculture and realize the modernization of agricul-
ture. It has a significant role in digital technology empowerment, HP sharing role,
and innovation promotion role. Based on the research on the development status of
the agricultural and rural digital economy, this paper found that in the development
of agricultural and rural digital economy in my country, there are problems such
as imperfect communication infrastructure construction, the low utilization rate of
farmers’ information, and inadequate and unbalanced development of digital agri-
cultural economy, which seriously restrict agriculture sustainable development of
the rural digital economy. Therefore, it is proposed to strengthen the construc-
tion of digital infrastructure and the development and promotion of agricultural
information platforms, and accelerate the development of digitalization of the agri-
cultural industry, so that the new round of scientific and technological revolution
can better benefit agriculture and rural farmers, and realize the modernization of
agriculture and rural development.
1 Introduction
Since the beginning of the new century, the Party’s No. 1 Document has focused on the
issues of agriculture, rural areas and farmers for 18 consecutive years. The moderniza-
tion of agriculture can effectively help the country to modernize. The themes of agricul-
tural and rural development in different periods differ, from the initial improvement of
comprehensive production capacity and the construction of agricultural infrastructure.
Overall planning of urban and rural development, to the construction of modern agricul-
ture that incorporates more technical elements, to start the comprehensive construction
of a well-off society and the development of agricultural and rural areas is listed as a
priority for development. A variety of factors have been incorporated in the new era to
accelerate the modernization of agriculture and rural areas.
According to the report of the “White Paper on China’s Digital Economy Devel-
opment (2020)” released by the China Academy of Information and Communications
Technology, the value-added of the digital economy has reached 35.8 trillion by the
end of 2019, and the proportion of the digital economy in GDP has risen to 36.2%.
The development of the digital economy has become an important driving force for the
development of various fields, the integration of rural economy and modern informa-
tion technology, using the advantages of knowledge and information to drive agriculture
to produce significant spillover effects, playing the role of Hewlett-Packard and ener-
gizing the domestic agricultural industry [1] effectively stimulating economic growth
and realizing industrial transformation and upgrading. The digital economy achieves
the high-quality development of the rural economy by innovating the agricultural devel-
opment model, breaking the dual barriers between urban and rural areas, maintaining
the stability of agricultural production, and promoting the sustainable development of
the agricultural economy [2]. In 2019, China issued the “Digital Village Development
Strategy Outline” to coordinate the development of digital villages, build smart cities,
and promote the construction of digital villages. In the No. 1 document of the Central
Committee of 2021, it is pointed out that at the beginning of China’s “14th Five-Year
Plan”, the construction of a socialist modern country is to be realized, and the construc-
tion of digital villages is an important breakthrough to break the revitalization of the
countryside.
Agricultural and rural digital economy refers to advanced digital information tech-
nology to promote agricultural economic development in rural areas based on upgrading
digital infrastructure [3]. Regarding the research of agricultural and rural digital econ-
omy, the existing literature mainly discusses the integration mechanism of the digital
economy and rural industry [4] farmers’ participation in the agricultural and rural digital
economy [5], the future development trend of the agricultural and rural digital economy,
etc. [6]. However, few scholars have analyzed the current situation and existing prob-
lems of developing my country’s agricultural and rural digital economy. After China
entered the new era, it strengthened the construction of information infrastructure to
realize networked operation and production. It now has the conditions to implement the
construction of digital villages [7]. However, there are still a series of problems that need
to be solved urgently to develop China’s agricultural and rural digital economy. There-
fore, this article combs the current status and problems of agricultural and rural digital
economy development and puts forward reasonable suggestions to provide a reference
for further research on the development of the agricultural and rural digital economy.
universal service pilot tasks will be deployed in poor villages, exceeding the “Thirteenth
Five-Year Plan” target, and achieving broadband network coverage of more than 90%
of poor villages.
60.00%
50.00%
40.00%
30.00%
20.00%
10.00%
0.00%
2013.12 2014.12 2015.12 2016.12 2017.12 2018.12 2019.6 2020.12
Fig. 1. Internet penetration rate in rural areas (Data source: China Internet Network Information
Center)
2
1.8
1.6
1.4
Unit:Trillion
1.2
1
0.8
0.6
0.4
0.2
0
2015 2016 2017 2018 2019 2020
education ratio of 89.32%. The low level of education hinders farmers’ from learning and
using information. The second is that the information cannot be transmitted due to the
reliance on traditional production and sales experience, which leads to information block-
ing and unable to be transmitted in time. Farmers have a short time to contact and learn
to use the Internet, and they have little knowledge of agricultural information websites.
The channels for obtaining information on agricultural production are mainly through
traditional channels, such as television media, acquaintances, and nearby businesses.
The transmission and acquisition of information depend on interpersonal communica-
tion and exchange transactions—lack of awareness of using agricultural information
websites and lack of professional talents’ guidance. The third is the low coverage of
intelligent communication equipment due to population size, population distribution,
and economic development level restrictions.
2000
1800
1600
1400
1200
1000
800
600
400
200
0
zhejiang
hebei
sanxi
shandong
fujian
henan
hubei
anhui
shanghai
shanxi
guangxi
hunan
jiangxi
guangdong
jiangsu
beijing
gansu
tianjin
sichuan
chongqing
liaoning
yunnan
jilin
guizhou
xinjiang
heilongjiang
hainan
ningxia
Fig. 3. Number of Taobao villages in each province (Data source: Ali Research Institute)
Digital talent training is not perfect. The use of information technology poses a certain
threshold and difficulty for farmers, and faces the serious problem of rural aging, which
deepens the difficulty of training. The training process usually takes the form of large-
class lectures, and it lacks pertinence for farmers of different education levels, different
Research on the Countermeasures for the Development of Agricultural 103
age groups, different production capabilities, and scales, and the actual application effect
is poor.
The construction of rural logistics needs to be further strengthened. Due to geo-
graphical conditions, the logistics infrastructure in some areas is very scarce, and the
high logistics costs have seriously hindered the development of rural e-commerce. With
the rapid development of e-commerce, third-party logistics has also developed. How-
ever, in some remote areas, especially the western region, the low return on investment in
logistics construction hinders the extension of third-party logistics transportation routes
to rural areas due to the vastness and sparse population.
To build the development of agricultural and rural digital economy, it is necessary to lay
a solid foundation for digital infrastructure construction. The government should play
a leading role and encourage operators to carry out digital infrastructure construction
through cooperation and subsidies to make up for the shortcomings of digital infrastruc-
ture construction. Prioritize the development of basic communication network construc-
tion in remote and backward areas to ensure that farmers access basic digital facilities
to meet their daily needs. Then, according to the development potential and positioning
of different regions, implement the policy of first piloting and then promoting. In-depth
study of the application of big data, cloud computing, and the Internet of Things in
agricultural products’ production and sales process to ensure efficient use of resources
and realize smart agriculture. According to the basic conditions of the residents, such
as the number of residents, age, and labor capacity, guide farmers to change and merge
villages, increase population concentration, and avoid new “digital divides.” Maximize
the sharing range of digital infrastructure, ensure the timeliness and economy of digi-
tal infrastructure, avoid waste of resources and funds caused by repeated construction,
and establish and improve construction standards to avoid easy damage due to quality
reasons.
Improve farmers’ mastery of the use of digital infrastructure, increase farmers’ aware-
ness of the digital economy, increase farmers’ acceptance and discrimination of digital
products, and promote better use of agricultural information platforms. In addition,
the government should actively build a community service platform for farmers, speed
up the exchange of information between farmers, break the dependence of farmers on
acquaintances, can effectively speed up the exchange and dissemination of agricultural
information between farmers, exert spillover effects, and effectively promote farmers’
acquisition and utilization of agricultural knowledge and agricultural technology.
Accelerate the training of digital talents, build a complete and targeted digital talent
training system, and provide agricultural digital production and sales services. Carry
out classified and batch training for different age groups and different levels of edu-
cation, provide professional guidance on Internet operation technology and operation
technology, realize the application of what has been learned, and effectively drive other
farmers to break traditional sales methods and participate in the development of new
agricultural business models middle. Aiming at areas where the development of new
agricultural businesses is slow, speed up the introduction of digital talents to fill the gap
and accelerate the process of realizing the development model of new rural businesses.
Accelerate the digital construction of rural logistics. Use big data and other tech-
nologies to scientifically analyze the geographic location of rural residents, establish
the optimal distribution network, monitor the entire logistics process, and improve the
efficiency of the circulation. Agricultural products have the characteristics of being eas-
ily lost and perishable during transportation. Therefore, the construction of rural cold
chain logistics should be strengthened, and the value of agricultural products should
be preserved through cold chain transportation. Strengthen cooperation with third-party
logistics, and reach a long-term contract with third-party logistics in order farming to
reduce transportation costs and search costs.
Develop a multi-format agricultural and rural digital economy. The government pro-
vides key support for areas with abundant agricultural products resources and lagging
rural e-commerce development, adopting a “scaled” policy guidance method, coordi-
nating financial, fiscal, and tax incentives, and lowering the entry threshold of the rural
e-commerce industry. Accelerate the construction of the agricultural industry chain,
accelerate the development of the processing chain, service chain, and functional chain
in the agricultural industry chain, cross-industry integration and innovation of the agri-
cultural industry chain integration model, and extend the industry chain to the front and
back ends. For example, the front-end variety cultivation, the back-end use short video,
live broadcast, and other forms for promotion and sales. Build agricultural digital indus-
trial parks, improve agricultural production, picking, sorting, and packaging standards,
give play to the agglomeration effect of digital agricultural development, and realize the
spillover of knowledge and technology.
Research on the Countermeasures for the Development of Agricultural 105
Acknowledgment. This work was supported by the grants of the Heilongjiang Province Phi-
losophy and Social Science Research Planning Project “Study on Symbiosis of Entrepreneur-
ship Ecosystem” (17GLE300) and Harbin University of Commerce “Young Innovative Talents”
Support Program Project (2020CX18).
References
1. Cai, Z.Z., Su, X.D.: The opportunities, obstacles and strategies of rural economy in the field
of digital economy. Agric. Econ. 07, 35–37 (2021)
2. Qi, W.H., Zhang, Y.J.: Promote the high-quality development of rural economy with digital
economy. Theoret. Explor. 03, 93–99 (2021)
3. Mu, J., Ma, L.P.: Measurement and regional differences of China’s agricultural and rural digital
economy development index. J. South China Agric. Univ. (Soc. Sci. Edn). 20(04), 90–98 (2021)
4. Kou, S.: Study on the effective integration of my country’s rural digital economy and
agricultural economy under the guidance of technology. Agric. Econ. 06, 12–14 (2021)
5. Su, L.L., Peng, Y.L.: A study on the evaluation of farmers’ practice participation and driving
factors from the perspective of digital village construction. J. Huazhong Agric. Univ. (Soc. Sci.
Edn.). 5, 1–12 (2021)
6. Liang, B., et al.: Discussion on the development trend of large-scale digital agriculture in rural
areas, Taking Xinjiang production and construction corps as an example. Agric. Econ. 12,
36–38 (2020)
7. Yang, B.Z., Wu, M.Q.: Advantages, Difficulties and ways of rural economic development under
the background of digital rural strategy. Agric. Econ. 2021(07), 38–40 (2021)
8. Zou, H.: The dilemma of the development of rural digital economy and its solution. Agric.
Econ. 02, 46–47 (2021)
Research on the Countermeasures of Digital
Intelligence Transformation and Upgrading
of Innovative Manufacturing Enterprises
Abstract. In China, the rapid development of the digital economy has brought
transformation opportunities to innovative manufacturing enterprises. New-
generation information technologies such as big data, 5G, and artificial intelligence
provide powerful tools for transforming digital intelligence and upgrading inno-
vative manufacturing enterprises. In this paper, the concept and characteristics of
innovative manufacturing enterprises are introduced, and the economic transfor-
mation process of innovative manufacturing enterprises is divided into five main
stages. Secondly, from the data-driven perspective, starting from the development
needs of innovative manufacturing enterprises, a digital intelligence platform and
data mining framework are built for them. Finally, the transformation counter-
measures of innovative manufacturing enterprises are given. This paper helps the
innovative manufacturing enterprises scientifically complete the transformation
and upgrading from digitization to digital intelligence. It provides a new perspec-
tive and ideas for the deep integration of the digital economy and real economies’
deep integration.
1 Introduction
New-generation information technologies such as big data, 5G, and artificial intelligence
are opening a new era of the digital economy. Building a digital China is a new national
informatization development strategy in the new era. The rapid development of the digital
economy injects new power into China’s economy, brings revolutionary influence to
the innovative manufacturing industry, and provides unprecedented opportunities for its
development and upgrading from digitization to digital intelligence. In the 14th five year
plan for national economic and social development of the people’s Republic of China and
the outline of long-term objectives for 2035 [1], it is pointed out that the optimization
and upgrading of the manufacturing industry should be promoted, the high-end and
intelligent manufacturing industry should be promoted, and intelligent manufacturing
should be deeply implemented.
Data is the fifth largest factor of production after land, labor, capital, and technology.
According to IDC (2018), more than 70% of China’s top 1000 enterprises take digital
transformation as the company’s strategic core. It expenditure that 65% of China’s GDP
will be related to data by 2022 [2]. The digital transformation and upgrading of innovative
manufacturing enterprises in China have achieved phased results. However, there are
still many problems in high-quality economic development, such as insufficient digital
service capacity of industrial supporting facilities, lack of data integration and sharing
platform, lack of high value-added data mining solutions, etc. Making full use of big
data technology and data mining methods to build a value creation system with data as
the core driving factor and realize the leap from digitization to digital intelligence is the
inevitable trend of developing innovative manufacturing enterprises in today’s digital
economy.
At present, the research on digital intelligence transformation abroad has been rel-
atively mature. The research content covers all aspects of intelligent manufacturing,
showing a multi-perspective, dynamic trend, and interdisciplinary integration. In terms
of research methods have gradually shifted from qualitative research methods such as
early concept elaboration and theoretical discussion to experimental methods and quan-
titative analysis methods such as computational simulation, data investigation, and case
study. Foreign research hotspots on digital intelligence transformation mainly focus on
intelligent design [3], intelligent production [4], intelligent manufacturing service [5],
and intelligent management [6].
Compared with foreign countries, the research on digital intelligence transformation
in China started later. Domestic-related research is mainly reflected in the following
three aspects.
First, research on economic digitization and digital intelligence transformation.
Wang W. [7] proposed that enterprise digital intelligence is based on connection and
data, and intelligence is the final value of digital intelligence. Wang J. [8] shared the
understanding of digital transformation from four dimensions: main motivation, basic
characteristics, path analysis, and intelligent technology. Chen J. [9] pointed out that
China must rely on big data to realize innovation-driven, structural adjustment and
industrial integration in order to promote economic transformation and upgrading under
the new normal.
Second, the economic transformation and upgrading of manufacturing enterprises.
Zhang L. [10] pointed out that data resources have become a key factor in the man-
ufacturing industry’s digital transformation. Jiang J. [11] thought that manufacturing
enterprises should track the latest science and technology trends and use big data to
realize intelligent decision-making. Li L. et al. [12] thought that intelligent manufactur-
ing, as the main direction of manufacturing in China from large to strong, is a strategic
choice for China to comply with a new round of scientific and technological revolution
and industrial reform, reshape new advantages in manufacturing development and build
an innovative country.
Third, the application of big data technology and data mining methods in the manu-
facturing industry. Pan et al. [123] established an analysis model of influencing factors
of shipbuilding cost to achieve the purpose of cost control. Wei et al. [14] proposed a
108 Y. Li et al.
product process adaptive design method to solve insufficient data utilization in manufac-
turing enterprises. Guo et al. [15] constructed a sustainable quality control mechanism
for the heavy vehicle production process using the improved turtle chart and VDA
evaluation model. Zhou et al. [16] proposed a method based on incomplete orthogo-
nal specified element method and SVM information to diagnose concurrent faults of
refrigeration equipment. Ma et al. [17] analyzed the research methods of multi-criteria
decision model and classification model in supplier portrait.
As mentioned above, rich research achievements are made by academia in economic
digitization and digital intelligence transformation, manufacturing transformation and
upgrading, and the application of big data technology and data mining methods in man-
ufacturing. However, most of the data-driven digital intelligence transformation appli-
cations are limited to the macro level, and there are few reports on the application of
big data technology and data mining-specific models and methods to digital intelligence
transformation and upgrading the innovative manufacturing industry. Therefore, from
the perspective of data-driven, starting with the development characteristics and needs
of innovative manufacturing industry, this paper builds a digital intelligence platform
based on big data processing technology and data mining methods, analyzes its digital
intelligence upgrading countermeasures, and helps the innovative manufacturing indus-
try realize the upgrading from digitization to digital intelligence. This has theoretical and
practical significance for exploring digital intelligence, Upgrading Countermeasures of
data-driven innovative manufacturing enterprises, and promoting the deep integration
between the digital and real economies.
(1) Automation stage (started around 1990): the idea of informatization is put forward,
which is mainly characterized by office automation and the installation of various
management software;
(2) Online stage (started around 2005): Internet technology has been developed, mainly
characterized by the development of e-commerce, social networks, and online
payment;
(3) Cloud stage (started around 2015): cloud computing technology tends to mature,
mainly characterized by the cloud of infrastructure and software facilities. At this
stage, the concepts of data middle station and AIOT middle station are put forward,
starting the transformation from business data to data business, and mobile payment
has developed rapidly;
(4) Two-wheel drive stage (started around 2019): the development stage of the digital
economy, which is mainly characterized by the digital pull of the consumer end,
the Digital Collaborative linkage of the supply end, and the two-wheel-drive of
business and technology;
110 Y. Li et al.
(5) Digital intelligence stage (started around 2020): it is characterized by taking con-
sumers as the core, giving full play to big data technology and artificial intelligence
technology, and making the whole link and life cycle intelligent from supply end
to consumption end.
Continuous innovation is the power source driving the development of enterprises, and
the primary demand for the development of innovative manufacturing enterprises is
continuous innovation.
Taking Midea Group, a well-known domestic enterprise, as an example, Midea Group
was established in 1968. After 52 years of development, it has become a mature global
technology group with digitization and intelligence.
In terms of marketing, pressing goods often occur in the household appliance indus-
try, resulting in a large backlog of inventory in the channel, especially in the seasonal
household appliance category of air conditioning. Midea Group has changed the tradi-
tional layered distribution mode through digital technology and innovated the “T + 3”
mode, starting from the order through three cycles of material procurement, manufac-
turing, sales, and delivery. The whole process determines production by sales. Bringing
product planning, marketing management, procurement preparation, manufacturing, and
logistics into the mainline of unified operation promotes the coordinated operation of
the internal value chain, reduces inventory, shortens order delivery cycle, and quickly
responds to market changes and differences. After the “T + 3” conversion, Midea can
organize production before receiving the order, thus eliminating the inventory backlog
of the channel.
In terms of R & D, in order to improve the user experience, Midea has developed
a high-speed distribution network technology, which simplifies the original seven-step
connection to three steps, and takes the lead in reducing the time from equipment distribu-
tion to successful connection to less than 5 s. At the same time, Midea has greatly reduced
the threshold for users to use smart appliances through the “smart touch” intelligent
function and can realize automatic network distribution and one key control.
In terms of management, Midea’s daily production plan involves nearly 2 million
kinds of materials. The amount of each material and who produces it must be arranged
in advance. These arrangements used to take at least 10 h. In 2020, Midea optimized the
material planning algorithm, and the material arrangement took only one hour, saving
90% of the time.
It can be seen that for innovative manufacturing enterprises, digital intelligence is
the only way of innovation. The basis of digital intelligence is data, and the way is data
processing technology. The goal is to improve enterprise efficiency, find out enterprise
profit points and enhance its industry competitiveness.
Research on the Countermeasures of Digital Intelligence Transformation 111
Based on the data-driven concept and big data technology, this paper has built a digital
intelligence platform for innovative manufacturing enterprises. The general architecture
is shown in Fig. 2. The whole platform architecture consists of physical resource layer,
data source layer, big data processing layer, task management layer, data analysis layer
and decision analysis layer.
(1) Physical resource layer mainly includes the underlying physical devices, such as
some distributed computer clusters.
(2) Data source layer: innovative manufacturing enterprises have accumulated a large
amount of data in production, sales, and management. These data may be stored in
traditional databases, files, warehouses, or some distributed file systems HDFS.
112 Y. Li et al.
(3) Big data processing layer: There are two major problems to be solved for massive
data. Big data storage can be completed through HDFS, HBase, cloud database,
NoSQL database. Second, the distributed processing of big data can be completed
through MapReduce, spark, etc.
(4) Task management layer: the task management layer is the core of the platform.
It can connect the analysis function with the background cluster to ensure that
the platform has easy algorithm scaling, supporting task flow scheduling, rational
computing, and storage resource allocation.
(5) Data analysis layer: provides specific data analysis and processing tasks, including
exploratory data analysis, real-time data query, and data mining tasks.
(6) Decision analysis layer: interpret, evaluate and make decisions according to the
results of the data analysis layer.
This hierarchical architecture fully considers the distributed storage and processing
of massive data, the integration of different data mining algorithms, the configuration
and scheduling of multiple data processing tasks, and the support of data analysis results
for intelligent decision-making.
In the platform shown in Fig. 2, data mining is the core and key of digital intelligence
of innovative manufacturing industry, its application framework is shown in Fig. 3.
(1) Decision trees, neural networks, and support vector machines are used for cost
analysis and management.
(2) Multiple regression and model trees are used for product design.
(3) The methods of correlation analysis, neural network, and random forest are used
for production process management.
(4) Frequent pattern mining and multi-criteria decision trees are used for logistics and
supply chain management.
(5) Rough set, fuzzy logic, and other methods are used for intelligent decision support
management.
(6) Using cluster analysis, correlation analysis, and other methods to carry out customer
relationship management and emotional analysis.
(1) Strengthen the training of data analysis talents. Enterprises can join hands with
colleges and universities to cultivate high-quality data analysis talents through col-
leges and universities, which can be transmitted to all departments and links of the
enterprise.
(2) To increase R & D investment, we can combine independent R & D with a foreign
introduction to create core products and leading brands and occupy the market.
(3) Build a technological innovation system combining industry, University, and
research, so that the enterprise has growth and sustainable development.
(4) With the help of a new round of high-tech wave, combined with the big data plat-
form and data mining framework built in this paper, accelerate and expand the
114 Y. Li et al.
application of big data technology and data mining methods, apply them in all links
of production, sales and supply, improve the production efficiency of enterprises
and enhance their industry competitiveness.
(5) Strengthen the protection of enterprise intellectual property rights.
5 Conclusions
This paper analyzes the digital intelligence transformation needs of innovative manufac-
turing enterprises in today’s economic environment and studies its transformation coun-
termeasures. Starting from the characteristics of the innovative manufacturing industry
and its digital intelligence needs, build a data-driven digital intelligence platform for
it, provide a data mining framework, explore the digital intelligence Implementation
Countermeasures of innovative manufacturing enterprises from the perspective of the
specific implementation, open up the data management mechanism through multiple co-
governance, and provide effective solutions for the successful scientific transformation
of innovative manufacturing enterprises.
There is a long way to go in upgrading digital intelligence. Innovative manufacturing
enterprises should give full play to the leading role of transformation, take the lead in
realizing digital intelligence, and lead other manufacturing enterprises to promote the
development of data-driven digital economy jointly, turn intelligence into energy, and
build China into an intelligent manufacturing power as soon as possible.
Acknowledgment. This work was supported by the grants of the Heilongjiang Social Science
(Evolution path and development countermeasures of big data industry ecosystem in Heilongjiang
Province: 20JYB037).
References
1. Chinese government network: The 14th five year plan for national economic and social devel-
opment of the People’s Republic of China and the outline of long-term objectives for 2035
(2021). http://www.gov.cn/xinwen/2021-03/13/content_5592681.htm
2. CCID think tank. Research on digital transformation of traditional enterprises in the devel-
opment of new business formats and models of digital economy. China Information Weekly
(2020)
3. Cao, Y., Jiang, H.: Comparative analysis of China’s equipment manufacturing enterprises
and world-class enterprises based on case study. In: 2021 2nd International Conference on
E-Commerce and Internet Technology (ECIT) (03), pp. 79–82 (2021)
4. Dumitrache, I., Caramihai, S., Stanescu, A.: From mass production to intelligent cyber-
enterprise. In: 19th International Conference on Control Systems and Computer Science
(CSCS) (05), pp. 399–404 (2013)
5. Mejía, G., Lefebvre, D.: Robust scheduling of flexible manufacturing systems with unreliable
operations and resources. Int. J. Prod. Res. 58(21), 6474–6492 (2020)
6. Serrasqueiro, M., Martins, R., Ferreira, H.: Modeling system based on machine learning
approaches for predictive maintenance applications. KnE Eng. 06, 857–871 (2020)
7. Wang, W.: Digital intelligence drives business innovation. Manager 5(322), 10–12 (2021)
Research on the Countermeasures of Digital Intelligence Transformation 115
8. Wang, J.: Analysis on the regular characteristics and development path of digital transforma-
tion. In: Huawei Ecological Conference, Shenzhen, 17–18 May 2021
9. Chen, J.: Big data and China’s economic transformation and upgrading. Econ. Forum 3(584),
26–32 (2019)
10. Zhang, L.: Seizing the opportunity of industrial technological change and accelerating the
digital transformation of manufacturing industry. China electronic news, 18 May 2021
11. Jiang, J.: Intelligent development mode and countermeasures of intelligent manufacturing
industry. Light Ind. Sci. Technol. 35(9), 149–150+152 (2019)
12. Li, L., Shi, A., Liu, J.: 40 years of China’s manufacturing industry: intelligent process and
prospect. China Soft Sci. (01), 1–9+30 (2019)
13. Pan, Y., Li, M., Li, X.: Analysis on association rules of influencing factors of shipbuilding
cost. J. Wuhan Univ. Technol. (Inf. Manage. Eng. Edn.) 39(03), 353–358 (2017)
14. Wei, W., Chen, Z., Yuan, J.: A product process adaptive design method based on manufacturing
big data. China Eng. Sci. 22(04), 42–49 (2020)
15. Guo, H., Zhang, R., Zhu, Y., et al.: Sustainable quality control mechanism of heavy truck
production process for Plant-wide production process. Int. J. Prod. Res. 58(24), 7548–7564
(2020)
16. Zhou, Y., Wu, K., Sun, Y., et al.: Fault diagnosis of refrigeration equipment based on data
mining and information fusion. Vib. Test Diagn. 41(02), 392–398+418 (2021)
17. Ma, M., Pan, S.: Overview of supplier portrait research based on data mining. Wirel. Commun.
Technol. 29(04), 55–60 (2020)
18. Xi, J.: Encyclopedia Dictionary of Scientific Outlook on Development. Shanghai Dictionary
Publishing House, Shanghai (2007)
Data Mining and Association Rules
A Robust Matting Method Combined
with Sparse-Coded Model
Abstract. Digital image matting is widely used in virtual reality, augmented real-
ity, and film and television production. It is mainly used to solve the problem of
separating foreground information from background information. Specifically, it
is to solve the value of the unknown pixels between foreground and background
information. Digital image matting methods mainly include sampling method and
affine method, this paper proposes a sampling method that combines Sparse-Coded
and Robust matting algorithm, named SRMatting. The robust algorithm uses a
local sparse single-point sampling method at the edge. Sparse-Coded sampling
has a lot less color redundancy than dense sampling. Still, local edge sampling
has the problem of insufficient depth and breadth color information collection.
Therefore, this article combines the Sparse-Coded algorithm (global depth sparse
method) with the sampling part of the Robust algorithm to enrich the diversity
of samples and adds the preprocessing steps to the new algorithm SRMatting to
improve the accuracy of the solution and reduce the amount of calculation.
1 Introduction
Digital image matting is different from image segmentation, and it is a classic problem
in a type of image editing and image processing [1]. Image matting adopts the concept
of α in the matting process. The following formula usually solves the unknown pixels:
Ii = αi Fi + (1 − αi )Bi , (1)
where α is the opacity of the pixel, and its value range is [0, 1]. If α = 1 we call it the
absolute foreground, if α = 0, we call it the absolute background, and the other points
are called mixed points. To use this formula to solve α, seven unknowns need to be
estimated from three equations. The problem is ill-posed. Therefore, many algorithms
now solve this problem using user annotations and propose the Trimap template, which is
divided into three areas: absolute foreground F , absolute background B , and unknown
area U , so the image matting problem has gradually evolved into solving the Trimap
problem.
Common matting methods are mainly divided into three categories: 1. Sampling-
based methods, such as Bayesian [2], Shared algorithm [3], etc.; 2. Affine-based methods,
such as Closed-Form [4], KNN [5] algorithm, etc.; 3. The method of combining sampling
and affine, such as the Robust [6, 7] algorithm used in this article. The Sampling-based
method, as the name implies, {Fi , Bi , αi } is calculated by point-by-point calculation. In
this method, the correlation between points is ignored, and each point exists in isolation
so that users can customize more optimized sampling according to their needs. In this
way, some poor effects can be ignored, which is conducive to future improvements. The
Affine-based method uses the correlation between points to recursively find the α value
in the unknown area. It can obtain a closed solution by modeling a mature quadratic
function, and it can avoid the problem of alpha discontinuity caused by the sampling
method. Still, the treatment effect is not good for some discontinuous objects (with
holes). It is divided into two stages, local and non-local. Propagation matting algorithms
based on local criteria such as Closed-Form, which propagate related constraints to the
entire image by minimizing the quadratic cost function, and non-local criteria-based
propagation matting algorithms such as KNN, which solve for the nearest K points of
pixels are used to estimate the pixel value of the position point.
Robust matting considers the proximity of samples to collect samples from the
known nearby foreground and background pixels and selects three sample pairs with
the highest confidence among them to estimate unknown sample pairs. Compared with
pure sampling methods, it combines Closed-Form. The idea of the matting algorithm
is to solve the linear system through the Laplace matrix and conjugate gradient method
to achieve the purpose of α optimization. Since the selection is made by considering
the color distortion, it provides more robust results than the rejection system. However,
since sampling only considers the spatial proximity to unknown pixels, the candidate
set may lack real samples, which reduces the quality of the matting.
The Sparse-Coded algorithm [8] is a matting algorithm based on global sampling.
It is different from the sampling method in the Robust algorithm. It builds dictionary
atoms by collecting the average color of superpixels at the boundary of the three-part
image and directly calculates α. The Sparse-Coded matting algorithm uses probability
segmentation as a clue to mark unknown pixels as high certainty and low certainty.
Samples that can be well separated from the foreground and background information
are marked as high certainty. Otherwise, they are marked as low certainty. The higher the
certainty, the smaller the dictionary. The sparse coding estimation at the pixel is combined
with its feature space neighbors and spatial neighbors in a weighted graph model and
solved in a closed-form to obtain the final. This modification improves performance and
greatly outperforms previous work on the benchmark data set.
2 Previous Work
2.1 Classification of Sampling Methods
Sampling methods can be divided into four categories: 1. Local and global sampling
methods: local sampling is only allowed to sample near unknown points, such as
Bayesian, BP, Robust algorithms, etc., global sampling methods can be performed in
places far away from unknown points sampling, such as Global, SRLO, WCT algorithm,
A Robust Matting Method Combined with Sparse-Coded Model 121
etc. The local type is better than the global type for processing areas with more color
overlap, but it is not good for some images with little change. 2. Sparse and dense sam-
pling methods: sparse sampling includes Robust, Sparse-Coded, and SVR algorithms,
and dense sampling includes BP, Shared, and Global algorithms. Color redundancy is
reduced compared with the sparse and dense types, but some key samples may be missed
if the collected samples are too scattered. 3. Edge and depth type sampling methods: edge
type has Robust, Shared, ISC algorithm, etc., depth type has BP, Sparse-Coded algo-
rithm, etc. The edge type can only sample at the edge of the unknown area, while the
depth type can sample inside the known area. The depth type can use the vertical “depth”
information of the edge of the known area and has more information than the edge type.
4. Parameters and single-point sampling methods: parameter methods include Bayesian,
Comprehensive, and CWCT algorithms and single-point methods include SVR, Robust,
and Sparse-Coded algorithms. The single-point type is better than the parametric type
to handle areas where the local color texture changes greatly, so the parameterization
method is not common.
Considering comprehensively, this paper adopts the sparse single-point sampling
method of local and local mixing. Although Robust is an edge type and Sparse-Coded
is a depth type, I think the two are not conflicting, but α can be solved better.
3 Algorithm Implementation
3.1 Preprocessing
Preprocessing is an important step before calculation. Its purpose is to divide some
points in the unknown area into the absolute foreground and background in advance
122 G. Yao et al.
by analyzing the unknown area in Trimap, setting some spatial thresholds and color
thresholds. This can narrow the range of the unknown area, and it can make the accuracy
of α higher.
The preprocessing step of this algorithm uses the static threshold method. It deter-
mines whether the unknown point in the foreground or the background by the following
formula [8]:
where D(i, j) is the Euclidean distance between the pixel i and j, I i is the color at the
pixel i, and the thresholds of Ethr and Cthr are set to 15 and 4 in this algorithm, respec-
tively. If the pixel i can satisfy the formula, it can be judged as an absolute foreground,
which greatly improves the solution’s efficiency and reduces the workload.
3.2 α Solving
Robust Algorithm Sampling. Robust algorithm [6] noticed that dense sampling is
likely to miss more information on the edge width, and at the same time it is easy
to bring color redundancy in depth, so it places local sampling the foreground edge F
and background edge B , and uses sparse sampling. It finds the sum of the points Fz
and Bz closest to the currently unknown pixel z in the Euclidean distance from the two
areas, and then sums along the two sides F and B from these two points, collecting
a sample point every several distances, in this way, a total of 20 front scenic spots and
20 background points are collected to form 400 sample pairs.
After sampling, high-quality sample points need to be selected. A good sample
should be a linear combination of foreground and background pixels. Specifically, for
unknown foreground and background, the estimated αr should be:
C − Bj F i − Bj
αr = (3)
F i − Bj 2
where C represents the pixel color of the unknown point, and F i , Bj is the
foreground and background sample pair of the unknown pixel.
Then evaluate the sample pair by defining a distance ratio:
C − α F i + 1 − α Bj
Rd F , B =
i j (4)
F i − Bj
where the denominator is the color Euclidean distance between the foreground and
background samples, and the numerator is the error between the color of the unknown
point C and the estimated color of the sample.
In general images, the points belonging to the absolute foreground and the absolute
background occupy the majority, so the following two weights are defined:
2
ω Bj = exp −Bi − C /DB2 (5)
A Robust Matting Method Combined with Sparse-Coded Model 123
2
ω F j = exp −F i − C /DF2 (6)
Where DF and DB are the maximum color distance between it and the foreground
and background samples:
DF = maxi F i − C (7)
DB = maxj Bi − C (8)
Taking these factors into account, the confidence of a pair of samples should be
expressed as:
Rd F i , Bj · ω(F i ) · ω(Bj )
f F , B = exp −
i j
(9)
σ2
Fig. 1. 1-a is the original image, and 1-b shows the super pixels we have marked. It can be seen
that the pixels in each block of super pixels are very similar.
Here C(·) is the RGB color value, which m is the number of samples that are spatially
close to the foreground, m and δ are two constants, usually set to m = 10, δ = 1. pb is
also calculated like this. The sample here is the average vector of the nearest superpixel
in space.
In order to determine the position of each sum atom F and B during sparse coding,
the F and B samples are horizontally spliced into [F1 , F2 , . . . Fn , B1 , B2 , . . . Bn ], and
given a dictionary D of unknown pixels, its sparse code is determined as
where vi is the feature vector at the pixel i, and the sparse code β i is generated
by the Lasso algorithm. The tangent point between the contour line and the constraint
domain is the optimal solution of the objective function [12]. For the Lasso method, the
constraint domain is a square, and there will be a tangent point with the coordinate axis,
which makes the weight of some dimensional features 0, so it is easy to produce sparse
results [13]. Therefore, the Lasso method can achieve the effect of variable selection,
compressing insignificant variable coefficients to 0.
3.3 Post-processing
Using the affine method to smooth the results is a method that many algorithms will use.
The common ones are the Matting Laplacian class and the Nonlocal class. The former
builds a matrix according to the space, the latter builds a matrix according to the color,
and this article uses the former to smooth the result.
Achieving post-processing is to achieve the goal of more absolute pixels than mixed
pixels, so α should be locally smoothed and respect the letters selected for each individual
pixel, that is, satisfy the data and neighborhood constraints.
In a graphical model [7], the initial α and a smooth term are combined as data items,
and then a closed form of coefficient linear equations is used to solve. For example,
the graph model shown in the figure below, virtual nodes F and B represent pure
foreground and pure background. The white nodes represent unknown pixels on the
image grid, and the red and blue nodes are known foreground and background pixels
marked by the user (Fig. 2).
The data weight is the probability that the pixel is the foreground and the background,
which are defined as:
W (i, F) = γ · [fˆi α̂i + 1 − fˆi δ α̂i > 0.5 (14)
W (i, B) = γ · [fˆi 1 − α̂i + 1 − fˆi δ α̂i < 0.5 (15)
where α and f are estimated and confidence values, δ is Boolean functions that return
0 or 1. γ is the parameter that balances the weight and weight of the data, that is, the
confidence value of the pixel. It can be defined as:
where γsprec (i) = e−vi −vi measure the credibility of reconstructing the input
2
2
(i) = e−Ii −[Fi Bi ]α i
rgb rgb
feature vector based on the sparse coefficient (v = Dβ ), γ
i i colrec
measures color distortion.
126 G. Yao et al.
In order to make the result smoother, we need to define the weights between nodes.
The Closed-Form algorithm solves α by establishing a color linear model and combining
with the matting formula. In this algorithm, we also apply this idea to obtain a more
optimized solution [14]. In order to strengthen the smoothing constraint of the image,
the neighborhood term W ij is defined by the sum of all 3 × 3 windows containing pixels
i and j:
(i,j)∈wk 1 ε −1
W (i, j) = 1 + (Ci − μk ) k + I Cj − μk , (17)
k 9 9
where λ is fixed to 100, N is the total number of pixels in the graph model node, which
can ensure that the final α can be consistent with the constraints specified by the user. v is
a set of well-determined foreground and background pixels, it is to ensure that adjacent
pixels share similar values. hi is a user-defined constraint, 0 is a clear background, and 1
is a clear foreground. N i represents the spatial sum and feature neighbors of pixels and
two virtual nodes.
The energy function is written in matrix form:
where H is a N × 1 vector with user constraint values, and the Laplacian matrix L
is defined as [15]:
⎧
⎨ Wii : if i = j
Lij = −Wij : if i and j are neighbors , (20)
⎩
0 : otherwise
where Wii = j Wij , L is a N × N sparse symmetric positive definite matrix, N is
the number of all nodes in the graph, including all pixels in the image plus two virtual
nodes B and F . Finally, the conjugate gradient method is used to solve the linear
system.
programming of VS2019 and Matlab2020a. Regarding the test set, 27 training images
and 8 private images in the online evaluation [16] system are used. Each training image
provides three Trimap templates: Huge\Large\Small. The size of the template is distin-
guished according to the size of the unknown area. It also gives the true α value of each
image, which is convenient for users to evaluate the effect of the algorithm. Since there
are very few test images with true values, it is impossible to compare their parts quanti-
tatively. Therefore, we did not introduce more test images, but these 35 images contain
most of the features contained in the images and cover various pixel transparency from
opaque to completely transparent. We believe that the test result of the new algorithm
on this data set is very convincing.
For some simple images, the effect of each algorithm is not much different, so we
choose a typical part from some complex images for final comparison. This does not
mean that our algorithm has no universal applicability for image matting. Some typical
parts selected here can better illustrate that our algorithm can better achieve separating
the background and the background.
The following is the comparison of Robust, Sparse-Coded, Closed-Form and SRMat-
ting algorithms for α solving. For the convenience of comparison, we have selected four
typical parts from the data set to illustrate the advantages of the new generation algorithm
(Fig. 3).
First of all, for the image GT05 (the first one), the first three algorithms deal with
this part rather fuzzy and do not generate a clear boundary, while the newly generated
algorithm has a better processing effect. Secondly, for the picture DT02 (second picture),
select its hair area. The processing results of the first three algorithms have certain defects,
such as blurring. The processing effect of the new algorithm is closer to the real value.
Then for the picture GT18 (third picture), its special feature is that its foreground is very
close to the background pixels, which easily makes some matting algorithms invalid.
Still, our new algorithm SRMatting can effectively identify the foreground information.
Finally, the image GT10 (fourth) was selected. It has part of the background embedded
in the foreground information and is far from the main background. It can be seen that
the new algorithm can better deal with the hole area than the first three algorithms.
The above is a visual comparison of the processing results of different algorithms. In order
to evaluate the SRMatting algorithm more objectively, the following four algorithms have
passed the two evaluation criteria of SAD and MSE on three different types of Trimap
in 35 images. As shown in the following Tables 1 and 2:
It can be seen that the algorithm we proposed has excellent results in SAD and MSE.
Among the four algorithms, the average value of both MSE and SAD is the smallest,
which shows that the result of our algorithm is closer to the true value. Closed-Form,
A Robust Matting Method Combined with Sparse-Coded Model 129
as a pure affine method, is prone to failure for complex areas. Although Robust uses
Closed-Form to optimize α, it is easy to lose some pixel information for local sampling,
resulting in insufficient accuracy. Sparse-Coded performs global sampling, but it is easy
to fail when the foreground and background overlap is high. At the same time, SRMatting
relies on the innovation of combining Sparse-Coded and Robust algorithms on sampling.
This combination can make the obtained sample set more comprehensive, so the result
can be more accurate.
5 Conclusion
Acknowledgment. This work is supported by the Youth Innovation Talent Support Program of
Harbin University of Commerce (No. 2020CX39).
References
1. Yao, G., Zhao, Z., Liu, H.: A comprehensive survey on sampling-based image matting.
Comput. Graphics Forum 36(8), 1–17 (2017)
2. Chuang, Y.-Y., Curless, B., Szeliski, R.: A bayesian approach to digital matting. In:
Proceedings of IEEECVPR, pp. 264–271 (2001)
3. Gastal, E.S., Oliveira, M.: Shared sampling for real-time alpha matting. Comput. Graphics
Forum 29(2), 575–584 (2010)
4. Levin, A., Lischinski, D., Weiss, Y.: A Closed-form solution to natural image matting. IEEE
Trans. Pattern Anal. Mach. Intell. 30(2), 228–242 (2008)
5. Chen, Q., Li, D., Tang, C.-K.: KNN matting. In: Proceeding of IEEE CVPR, pp. 869–876
(2012)
6. Wang, J., Cohen, M.: Optimized color sampling for robust matting. In: Proceedings of IEEE
Conference on Computer Vision and Pattern Recognition, Minneapolis, USA, pp. 1–8 (2007)
7. Comaniciu, D., Meer, P.: A robust approach toward feature space analysis. IEEE Trans. Pattern
Anal. Mach. Intell. 24(5), 603–619 (2002)
130 G. Yao et al.
8. Johnson, J., et al.: Sparse coding for alpha matting. IEEE Trans. Image Process. 25(7), 3032–
3043 (2016)
9. Fulkerson, B., Vedaldi, A., Soatto, S.: Class segmentation and object localization with
superpixel neighborhoods. In: International Conference on Computer Vision (ICCV) (2009)
10. Achanta, R., et al.: Slic superpixels compared to state-of-the-art superpixel methods. IEEE
Trans. Pattern Anal. Mach. Intell. 34(11), 2274–2282 (2012)
11. Yang, J., Wright, J., Huang, T.S., Ma, Y.: Image super-resolution via sparse representation.
IEEE Trans. Image Process. 19(11), 2861–2873 (2010)
12. Johnson, J., Rajan, D., Cholakkal, H.: Sparse codes as alpha matte. In: Proceeding of BMVC
(2014)
13. Gould, S., Rodgers, J., Cohen, D., Koller, D.: Multi-class segmentation with relative location
prior. Int. J. Comput. Vision (IJCV) 80(3), 300–316 (2008)
14. Levin, A., Lischinski, D., Weiss, Y.: Colorization using optimization. ACM Transactions on
Graphics (2004)
15. Grady, L.: Random walks for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell.
28(11), 1768–1783 (2006)
16. Rhemann, C., et al.: Alpha Matting Evaluation Website (2009). http://www.alphamatting.
com. Accessed Jun 2016
Segmentation of Cervical Cell Cluster
by Multiscale Graph Cut Algorithm
Tao Wang(B)
1 Introduction
Cervical cancer is one of the most common malignant tumors of the female reproductive
system, and it has a huge impact on women’s physical health [1]. In the world, new
cases of cervical cancer reach about 500,000 every year. The high incidence of cervical
cancer is mostly concentrated in the 40–60 age groups. However, in recent years, the
incidence of patients has been getting younger, and the average age has a downward
trend [2, 3]. Pap smear (PAPSMEAR) [4] is one of the earliest and most important
methods of cervical cancer screening to reduce the mortality of women with cervical
cancer. The H&E staining method provides a more stable and accurate image of the
cervical sample for the Pap smear. Currently, cervical screening relies too much on the
clinical experience of the pathologist and requires a lot of time and energy from the
pathologist. Even pathologists with rich clinical experience are prone to misdiagnosis
under the influence of energy and subjective judgment [5].
In order to solve this problem, low-cost, high-efficiency computer-aided diagnosis
technology has begun to develop. Computer-aided diagnosis technology can reduce the
fatigue of manual inspection and reduce the rate of misdiagnosis and missed diagnosis.
Cervical cell detection is mainly divided into cervical sample cytoplasm detection and
nucleus detection. Existing methods have made some progress in the field of cervical
cell detection [9–14]. Algorithms that use cytoplasm as the segmentation target mainly
include morphological methods, k-mean [11, 15, 16], edge detection [14], threshold
method [18], active contour [16, 20], adaptive watershed, which are usually applied to
single-cell segmentation. For segmentation images containing overlapping cells, meth-
ods such as threshold [12, 17–19] and level set [20–22] are usually used. The level set
method obtains a sub-optimal result, likely to converge to a local minimum and obtain a
locally optimal solution. This optimal local solution may be far from the optimal global
solution. Zhang et al. [13] used the multi-threshold Otsu algorithm to segment the cyto-
plasm. However, it could not obtain better results due to the limitations of objective
conditions, such as uneven staining, light differences, and cell overlap.
The active contour model [16, 23], a morphological watershed method [12, 25,
26], a morphological corrosion method can segment the cell nucleus. Zhang et al.
proposed an expectation-maximization algorithm combined with distance transforma-
tion, an unsupervised algorithm that uses ellipse fitting technology for cell nucleus
segmentation.
The graph cut method has attracted the interest of many experts and scholars in cell
segmentation. In binary cell nucleus image segmentation, the Graph cut method obtains
more accurate results than the global threshold. However, the application of binary
images seems to be very limited. Therefore, the proposed algorithm has not attracted
any attention during the ten years. In the late 1990s, new computer vision methods were
proposed to describe applying Min-cut/Max-flow algorithms to solve non-binary image
problems. Applying prior knowledge of cell nucleus shape, manual annotation (seed
node), and local image features on the graph segmentation framework can get more robust
segmentation results. Boykov et al. discussed the theoretical nature of graph construction
in computer vision and gave the conditions that the energy function minimized by the
graph cut algorithm should meet. Song et al. combined a multiscale convolutional neural
network with a graph cut algorithm and obtained relatively ideal results in overlapping
cell nucleus segmentation. At the same time, the concept of superpixel was introduced,
which greatly reduced the computational complexity.
In most cases, since most abnormal cells are in clusters, pathologists need to pay
special attention to clumps. The cell clumps have the characteristics of complex composi-
tion, diverse shapes, and irregular boundaries. The characteristics bring great challenges
to computer-aided diagnosis methods. In the clinical samples, the complicated arrange-
ment of white blood cells, dust, impurities, uneven light, uneven staining make it more
difficult to segment the cell clusters. In the automatic segmentation and identification
of cervical cancer cells, the segmentation algorithm for cervical cell clusters is not yet
mature. Therefore, it is necessary to study the automatic and accurate segmentation
algorithm to segment cervical cell clusters.
Multiscale technology can express and process images at different scales. The reason is
that the characteristics that are not easy to see in one scale or the acquired characteristics
are easy to find or extract in other scales. The pyramid structure is a form of multiscale
Segmentation of Cervical Cell Cluster 133
1 − (x−x0 )2 +2(y−y0 )2
G(x, y) = e 2σ (1)
2π σ 2
Gaussian Pyramid adds Gaussian filtering on the down-sampling of multi-resolution
pyramids. Gaussian blurring is performed with different parameters σ for each layer of
the pyramid so that each layer of the pyramid has multiple Gaussian blurred images, as
shown in Fig. 1.
Ip represents the pixel value of the p pixel in the image P; the k value represents
the weight coefficient of the data item and the interaction item; Lp represents the label
134 T. Wang
assigned by the p pixel; K(p,q) represents the cost of discontinuity; T (Ip = Iq ) is the
indicator function. The specific forms of Lp , K(p,q) , and T (Ip = Iq ) are shown in Formulas
(4), (5), and (6).
255 foreground
Lp = (4)
0 background
= e−(Ip −Iq )
2
K (p,q) (5)
1 Ip = Iq
T = (6)
0 Ip = Iq
For Lp , its value depends on the label (foreground or background) assigned to the
p pixel. It can be seen that the data item of the energy function represents the cost of
assigning the corresponding label to the pixel. The interaction term K(p,q) · T of the
energy model represents the cost of assigning different labels to p pixels and q pixels.
The graph cut algorithm can be applied to solve the global minimum of the energy
function. First, define a directed weight graph G = ν, ε. The directed graph G is com-
posed of a set of vertices ν and a set of directed boundaries ε, and the set of vertices ν
is connected by the set of directed boundaries ε. Normally, each vertex corresponds to a
pixel in the image. The directed graph contains several special vertices called terminal
vertices, and the terminal vertices correspond to the labels assigned by the image pix-
els. In general, only consider the existence of Source(S) terminal vertices and Sink(T )
terminal vertices. In other words, there are only two labels (foreground or background)
for each pixel.
There are two types of terminal boundaries in G = ν, ε: n − link boundary and
t − link boundary. n − link connects the pixels (non-terminal pixels) in the neighborhood
system in the graph, so the weight of the n − link boundary represents the cost of
discontinuity between pixels. t − link connects the pixel point and the terminal vertex,
so the weight of the boundary of t−link represents the cost of assigning the corresponding
label to the pixel point. It can be seen that the weight of the n−link boundary corresponds
to the data item of the energy function, and the t−link weight of the boundary corresponds
to the interaction term of the energy function.
From the directed graph constructed above and the boundary weights, it can be seen
that the problem of solving the energy function can be transformed into the boundary
cutting problem of the graph.
3.1 Pre-processing
Process the given cervical sample image in the RGB color space, and select the channel
according to the strength of the color space contrast. The image of each channel of the
original image is shown in Fig. 2, and the sequence of the image is R channel, G channel,
and B channel. According to the contrast in Fig. 2, this paper selects the G channel image
for processing. Because the cervical sample image is affected by light, staining, and other
noise, the original image is pre-processed using histogram equalization to enhance the
contrast.
The global cell cluster segmentation energy function is shown in Formula (7):
E(I ) = Rp (Ap ) + λ B{p,q} · δ(Ap , Aq ) (7)
p∈P {p,q}∈N
Rp (Ap ) is the data item of the energy function; B{p,q} is the interaction item of the
energy function; λ represents the weight coefficient of the data item and the interaction
item; δ(Ap , Aq ) is the indicator function; Ap is the label of p pixel. On the basis of the seed
node image obtained above, the pixel label distribution probability value is converted
into the data item Rp (·) in the energy function as shown in Formulas (8) and (9).
The interaction cost term is composed of discontinuous cost B{p,q} and indicator
function δ(Ap , Aq ), as shown in Formulas (10) and (11).
(Ip −Iq )2
(− ) 1
B{p,q} = e 2σ 2 · (10)
dist(p, q)
1 Ap = Aq
δ(Ap , Aq ) = (11)
0 otherwise
For the discontinuity cost B{p,q} , the case where adjacent pixels with similar intensity
values (|Ap − Aq | < σ ) are assigned different labels will have a larger discontinuity
cost. On the other hand, if the intensity difference between two adjacent pixels is large
(|Ap − Aq | > σ ), the discontinuity cost of assigning them to different labels is small. In
this paper, the boundary weights of the directed boundary (p, q) and (q, p) are the same.
This paper uses the α − expansion algorithm to complete the energy function solution.
The α − expansion algorithm obtains a global approximate optimal solution through
multiple iterations of the Min-cut/Max-flow algorithm.
This paper collects 20 cervical sample images of 2048 * 2048 size and divides each
image into 4 images of 1024 * 1024 size, a total of 80 images of 1024 * 1024 size.
The cervical specimen image is used in the experiment of this article. All slides were
prepared by manual liquid-based cytology and stained with the H&E method. Among
the 80 sample pictures, 50 images are used for algorithm parameter adjustment, and the
remaining 30 are used as the data set.
Segmentation of Cervical Cell Cluster 137
The algorithm in this paper is higher than the algorithm in literature [36] on the three
performance indicators of DSC, prec and rec. The following Fig. is a picture in the test
set. The cluster cell segmentation results obtained by the method of this paper and the
method of literature [36] are shown in Fig. 4.
5 Conclusions
In this paper, the full-size cervical sample image is transformed into images of different
scales through multiscale transformation. After coarse segmentation, the confidence
region generates a seed node and pixel value probability distribution map and then applies
it to the global cell cluster segmentation graph cut algorithm. The data experiments
of clinical TCT cervical sample images show that the multiscale graph cut algorithm
proposed in this paper can segment cervical cell clusters more accurately. The three
evaluation indicators of DSC accuracy rate and recall rate are better than the threshold
watershed algorithm of literature [36].
Acknowledgment. This work was supported by the project of talented youth reserves funded by
the Harbin University of Commerce.
Segmentation of Cervical Cell Cluster 139
References
1. Song, J., Xiao, L., Lian, Z.: Contour-seed pairs learning-based framework for simultaneously
detecting and segmenting various overlapping cells/nuclei in microscopy images. IEEE Trans.
Image Process. 27(12), 5759–5774 (2018)
2. Arya, M., Mittal, N., Singh, G.: Texture-based feature extraction of smear images for the
detection of cervical cancer. IET Comput. Vision 12(8), 1049–1059 (2018)
3. Pradhan, D.: Clinical significance of atypical glandular cells in Pap tests: an analysis of more
than 3000 cases at a large academic women’s center. Cancer Cytopathol. 124(8), 589–595
(2016)
4. Yu, J., Tan, J., Wang, Y.: Ultrasound speckle reduction by a SUSAN-controlled anisotropic
diffusion method. Patt. Recogn. 43(9), 3083–3092 (2010)
5. Huang, J.J., Wang, T., Zheng, D.Q., He, Y.J.: Nucleus segmentation of cervical cytology
images based on multi-scale fuzzy clustering algorithm. Bioengineered 11(1), 484–501 (2020)
6. Ds, G., Jm, R.: Quantitative immunocytochemistry of hypothalamic and pituitary hormones:
validation of an automated, computerized image analysis system. J. Histochem. Cytochem.
Off. J. Histochem. Soc. 33(1), 11–20 (1985)
7. Bengtsson, E.: Recognizing signs of malignancy - the quest for computer assisted cancer
screening and diagnosis systems. In: Coimbatore: 2010 IEEE International Conference on
Computational Intelligence and Computing Research, pp. 1–6 (2010)
8. Zhang, L., Kong, H.: Segmentation of cytoplasm and nuclei of abnormal cells in cervical
cytology using global and local graph cuts. Comput. Med. Imaging Graph. 38(5), 369–380
(2014)
9. Marinakis, Y.: Pap smear diagnosis using a hybrid intelligent scheme focusing on genetic
algorithm based feature selection and nearest neighbor classification. Comput. Biol. Med.
39(1), 69–78 (2009)
10. Lezoray, O., Cardot, H.: Cooperation of color pixel classification schemes and color
watershed: a study for microscopic images. IEEE Trans. Image Process. 11(7), 783–789
(2002)
11. Meng, H.: Nucleus and cytoplast contour detector of cervical smear image. Patt. Recogn.
Lett. 29(9), 1441–1453 (2008)
12. Genc, M.: Unsupervised segmentation and classification of cervical cell images. Patt. Recogn.
45(12), 4151–4168 (2012)
13. Zhang, L.: Automation-assisted cervical cancer screening in manual liquid-based cytology
with hematoxylin and eosin staining. Cytometry Part A. 85(3), 214–230 (2014)
14. Mao, Y.: Edge enhancement nucleus and cytoplast contour detector of cervical smear images.
IEEE Trans. Syst. Man Cybern. 38(2), 253–366 (2008)
15. Huang, Q.K.: A method based on watershed algorithm for core particles image segmen-
tation. In: 2010 3rd IEEE International Conference on Computer Science and Information
Technology (ICCSIT), pp. 408–410 (2010)
16. Li, K., Lu, Z., Liu, W., Yin, J.: Cytoplasm and nucleus segmentation in cervical smear images
using radiating GVF snake. Patt. Recogn. 45(4), 1255–1264 (2012)
17. Mei, L.: Adaptive image segmentation algorithm under the constraint of edge posterior
probability. IET Comput. Vis. 11(8), 702–709 (2017)
18. Bergmeir, C.: Segmentation of cervical cell nuclei in high-resolution microscopic images:
a new algorithm and a web-based software framework. Comput. Meth. Programs Biomed.
107(3), 497–512 (2012)
19. Aniel, G.B.: Automated marker identification using the radon transform for watershed
segmentation. IET Image Proc. 11(3), 183–189 (2017)
140 T. Wang
20. Harandi, N.M.: An automated method for segmentation of epithelial cervical cells in images
of ThinPrep. J. Med. Syst. 34(6), 1043–1058 (2010)
21. Wang, T., Huang, J.J., Zheng, D.Q., He, Y.J.: Nucleus segmentation of cervical cytology
images based on depth information. IEEE Access. 8, 75846–75859 (2020)
22. Bamford, P., Lovell, B.: Unsupervised cell nucleus segmentation with active contours. Signal
Process. 71(2), 203–213 (1998)
23. Al-Kofahi, Y.: Improved automatic detection and segmentation of cell nuclei in histopathology
images. IEEE Trans Bio-Med Eng. 57, 841–852 (2010)
24. Plissiti, M.E., Nikou, C., Charchanti, A.: Automated detection of cell nuclei in pap smear
images using morphological reconstruction and clustering. IEEE Trans. Inf Technol. Biomed.
15(2), 233–241 (2011)
25. Plissiti, M.E., Nikou, C., Charchanti, A.: Combining shape, texture and intensity features for
cell nuclei extraction in pap smear images. Pattern Recogn. Lett. 32(6), 838–853 (2011)
26. Jung, C.: Unsupervised segmentation of overlapped nuclei using bayesian classification. IEEE
Trans. Biomed. Eng. 57(12), 2825–2832 (2010)
Source Code Author Identification Method
Combining Semantics and Statistical Features
Xu Sun1 , Yutong Sun1 , Leilei Kong2(B) , Yong Han2 , and Hui Ning3
1 Heilongjiang Institute of Technology, Jixi, Heilongjiang, China
2 Foshan University, Foshan, Guangdong, China
3 Harbin Engineering University, Harbin, Heilongjiang, China
Abstract. The globalization of information sharing has made copying easier and
easier. The endless duplication of plagiarism has aroused wide attention in aca-
demic circles, and the related research in plagiarism detection has become a hot
topic in recent years. Taking deep learning-based plagiarism detection modeling
as the research object and improving the performance of the plagiarism detection
system as the research objective, this paper conducts an in-depth study on the task
of source code author identification in internal plagiarism detection. In the task
of source code author identification, text features based on pre-trained language
model can be used to model the semantic information of code fragments. How-
ever, it still lacks in the representation of complex full-text statistical features at
the level of text granularity. Therefore, this paper proposes a source code author
identification method that integrates semantic and full-text statistical features and
combines statistical and semantic features to build a source code writing style
model to realize accurate identification of author identity. Experimental results
on AI-SOCO datasets show that the proposed modeling method is superior to the
statistical and single semantic feature models.
1 Introduce
This paper focuses on source code author identification in internal plagiarism detection.
The focus of its task is to identify and test the authorship of source code text by construct-
ing an author writing style model given a predefined set of source code text and its author
[1]. This paper focuses on the positive role of semantic features in the construction of
document expression and realizes the semantic representation of source code. However,
because of the particularity of the source code programming language, only extracting
semantic features is not enough to model the author’s writing style. Therefore, this paper
proposes a source code author identification method that combines semantic features and
complex full-text statistical features to accurately predict the identity of the source code
author in a deep learning algorithm. Deep learning aims to automatically analyze and
extract the feature representation of deep data to be studied. Based on the working prin-
ciple of deep learning, the researchers proposed the pre-training language model. The
emergence of the pre-trained language model has brought a milestone change to the NLP
research field, making the development of natural language processing technology into
a new era. The BERT pre-training language model [2] released by Google at the end
of 2018 is the most noteworthy, showing amazing performance in 11 KINDS of NLP
tasks. BERT pre-training language model completes the training process in advance on
the massive unsupervised text corpus. The model can capture the semantic, syntactic,
positional structure and other information representation in the natural language text
data and then make subtle adjustments to the model according to fixed research tasks.
This article will introduce the training language model in-depth study, further promote
the plagiarism detection research-related tasks, more rationalization of plagiarism detec-
tion task model, makes the research achievements of this paper can promote plagiarism
detection research and development in the field of natural language processing. It has
certain practical significance.
2 Related Work
Given a predefined set of Source Code texts and their authors, the focus of Source Code
recognition task is to identify the authorship of the written test set text by building an
author style model. In experiments, the identification of source code authors is usually
formalized into a potential plagiarism detection problem, and plagiarism is identified by
extracting text features and analyzing the author’s writing style and personal language
characteristics.
The current work uses different techniques to accurately identify authors by improv-
ing the quality of features captured from source code text. In terms of representation,
most methods still use the AST algorithm, N-gram word frequency statistics, and other
models to transform text to text feature vector and realize the author’s writing style in
a classification model or neural network. For example, Yang et al. used Word N-gram,
Character N-gram (N = 2–5), and Word + Character combination as features and sent
them into logistic regression classifier for modeling.
In addition, some researchers propose that feature extraction methods based on source
code text, coding style, and abstract syntax tree can also represent the author’s writing
characteristics. For example, Pellin et al. [3] used the model to train the abstract syntax
tree (AST) representation extracted from the code. They applied the support vector
machine classifier to determine the source code author. Hugo A et al. [4] chose to
extract the coding features in the source code to find information about the text author to
predict the author’s real identity. Rehana et al. extracted the features based on style and
category from the source code and used a support vector machine, Gaussian function,
M5 algorithm for prediction.
However, the feature representation method based on traditional word frequency statis-
tics and text coding structure is weak in extracting source code text semantic features.
Source Code Author Identification Method 143
It cannot capture the writing characteristics of source code text. With the deepening
of research, researchers gradually introduce deep learning into the source code author
recognition task and use the deep learning model to capture effective features based on
source code text independently. For example, Alsulami et al. [5] used LSTM and BiL-
STM deep learning models to learn and capture the grammatical features contained in
texts and information representation at different levels of granularity. Abuhamad et al.
[6] applied the word frequency-inverse document frequency (TF-IDF) provided by con-
volutional neural networks (CNN) and word embedding representation to identify the
authorial identity of source code text. However, from the actual effect, the prediction
accuracy of the deep learning model is far lower than the expected value, indicating that
the simple deep learning representation model can not solve the problem of low author
recognition performance based on source code text.
Based on the disadvantages of the above methods, this paper will start from the
difference and particularity of source code text compared with natural language text,
and re-model the source code author recognition task, which has a breakthrough to
improve the accuracy of author identification.
3 Method
The focus of source code author identification is given a set of predefined source code text
and its author, and the author identity of the text is tested by building an author writing
style model [1]. At present, most of the mainstream methods use traditional statistical
models to complete the transformation from text to text feature representation vector,
but due to the particularity of programming language compared with natural language,
the feature representation information captured by the model is far from enough for the
author’s unique writing style, resulting in the experimental results far below the expected
value. In view of this, this paper proposes a source code author identification method
that integrates semantic and full-text statistical features to increase the ability to extract
complex full-text statistical features at the text granularity level and accurately predict
the author identity of source code text.
Source code author identification task refers to that given a set of source code documents
and their authors, a model belonging to their own unique writing style is constructed
for each author. In the test data set, the author identity of source code text is identified
according to the matching degree of writing style. Based on the research method of
internal plagiarism detection, this paper formally defines the identification of source code
authors as the problem of determining potential plagiarism. Plagiarism is identified by
analyzing the similarities and differences of source code text writing styles and individual
language characteristics.
This paper attempts to construct a writing style model using full-text semantic fea-
tures, but the experimental results are far lower than expected. The reason is that the
programming language used in source code text has certain particularity compared with
natural language.
144 X. Sun et al.
First of all, there are few keywords in the programming language, and the vocabulary
dictionary formed by it is small in scale. However, the pre-training language model
usually adopts a massive unsupervised training corpus to conduct training operations on
the model. Its corpus covers almost all the linguistic information in the English corpus.
Because the vocabulary set of programming language is far smaller than the number of
English vocabulary learned by the pre-trained language model, the captured semantic
features of the text in the source code feature representation information learned by the
BERT model cannot fully express the author’s writing characteristics.
Second, there are a large number of fixed character sets and expression rules in
programming languages, such as the use of the While loop, which is followed by a
Boolean expression While. = 100), the Mask language model (Mask LM) in the BERT
pre-training language model will randomly Mask 15% of the words in the input corpus
and predict the original masked words depending on the learned context word meaning
and the context information on the left and right sides. But in fact, Boolean expressions
have some randomness, and there is no semantic relationship between Boolean expres-
sions, and While statements, so the model cannot recognize the full masked expression.
Compared with natural language, the semantic relationship between words in the source
text is relatively weak.
Through the above analysis, it is found that although the single semantic feature
model can better model the semantics of code fragments, it ignores the importance of
the statistical features in source code text to establish the author’s unique writing style
model and cannot represent the complex full-text statistical features in the granularity of
text in the aspect of feature extraction. This paper also needs to look for the characteristic
information that can accurately identify the writing style of source code text, solve the
problem of representation selection of source code text, and describe the author’s writing
style and personal language characteristics in a more detailed way.
3.2 Source Code Author Identification Model Integrating Semantic and Full-Text
Statistical Features
Feature Representation Method Integrating Semantics and Full-text Statistics. This paper
proposes that the semantic features, representation information, and full-text statistical
features of source code text should be fully considered in feature construction. The
methods of corresponding feature representation in the source code text will be described
below.
There are some semantic connections between words in the source code. The trans-
former mechanism in the Bert pre-training language model is used to learn and generate
embedded information based on words, words, and sentences in the source code, use its
powerful feature extraction ability to capture the semantic features of the source code,
perform SoftMax operation on the output vector, calculate the probability distribution,
output the prediction probability of the source code author, and extract the maximum
probability value by using the argmax function, Get the predicted author tag.
In this paper, we should integrate various features, including semantic information,
to build the author’s writing style model. Therefore, the author prediction probability
corresponding to each source code text output by the model is expressed as, and N
represents the order of the given author in the source code data set. For all source code
Source Code Author Identification Method 145
texts, a prediction probability matrix is formed. This probability matrix represents the
information as the semantic feature of the source code text.
Full-Text Statistical Features Based on Source Code Structure. The author will
automatically bring in his personal language characteristics and show a certain law in
the use of various sentences and symbols. For example, when writing loop statements,
the author only uses the for loop, and comments (time, name) will be added when writing
code. This paper will conduct quantitative analysis on the source code text through the
method based on mathematical statistics to identify various quantitative laws based on
structure and style in the document [7]. As shown in Table 1.
of the authorship tag and source code text. Set T i as the training text, Ai as the author
identity tag, Y i as the individual writing style model for each author. Then the framework
of source code recognition algorithm can be shown in Fig. 1.
When given a new set of source text, the author identity of source text is predicted
by learning model M by extracting the writing style embodied in the text. The modeling
process for the source code author identification task is described below.
In this paper, full-text statistical features based on vocabulary, full-text statistical fea-
tures based on code structure, and deep semantic features based on source code text are
captured as feature representation information to comprehensively describe the author’s
writing style and personal language characteristics. MLP can overcome the weakness
that single-layer perceptrons cannot identify linear infraction data and has strong gener-
alization ability in terms of prediction. Therefore, this paper uses MLP neural network
to construct the source code author identification model that integrates semantic features
and full-text statistical features.
In order to facilitate the mapping transformation of interpretive text features in neural
networks, this paper simply combines semantic features and full-text statistical features
to make them into a fixed dimension vector, and uses vector to express interpretive text.
In the experiment, the vector representation of the interpretation text was input into the
deep neural network built by MLP to construct the source code author identification
model combining semantic and full-text statistical features.
In the MLP neural network used in this paper, the input layer is used to receive
the input feature representation vector, and the hidden layer is fully connected with the
input layer. By learning the input feature vector, the output feature representation vector
is saved, and then the label probability distribution is calculated through the Softmax
regression function of the output layer to obtain the final matching result. Each layer
adopts different activation functions. The hidden layer is represented by vector x, then
Source Code Author Identification Method 147
its output is y(w1 x + b1 ), w1is the connection coefficient, b1 is the bias, and Sigmoid
function is used as the activation function. The formula is as follows:
The output of the output layer is f (w2 x 1 + b2 ) the output that x 1 represents the hidden
layer Sigmoid(w1 x + b1 ). Function F can use the Softmax function, as shown in Formula
2, where x i represents the output of the hidden layer, and n represents the number of
neurons in the hidden layer.
i n x n
Softmax(x) = ex / e (2)
1
Based on the above analysis, as shown in Fig. 2, a task model of source code author iden-
tification that incorporates semantic and full-text complex statistical features is estab-
lished. When given a new set of source code texts, the author identity of source code
texts can be predicted by judging the matching degree between the modeling author’s
writing style and the source code text’s writing style through MLP neural network.
Predict source code authorship. After feature extraction, the test data set is fed into
MLP neural network, and the author writing style obtained from the test source code
text is matched with the established author style model. The author identity sequence of
the source code is predicted.
The AI-SOCO data set provided by FIRE 2020 evaluation conference was used in the
source code identification experiment. In the data set, 1000 users in the Codeforces
online evaluation system were selected, and 100 compilable source codes submitted by
each user were collected. The source codes were written in C++ language. For each user
(UID), all the collected source code comes from a different problem. The total number
of source documents is 100,000, which are divided into Train (training data set), Dev
(validation data set: for tuning model parameters), and Test (Test data set).
The source code identification model will be evaluated and ranked according to the
Accuracy index, which is defined as follows:
This section reports the experimental results of source code author identification and
analyzes the results in detail. Accuracy was used as the evaluation index to evaluate
the experimental performance. Table 2 shows the prediction results of the text using
only the lexical information based on the N-Gram statistical model as the feature repre-
sentation vector. Table 3 describes the experimental results using different text feature
combinations.
Source Code Author Identification Method 149
Considering that the source code text of programming language cannot model the
author’s writing style model due to the small vocabulary and the particularity of fixed
writing format, this paper hopes that the full-text statistical features based on the text
granularity level can capture the information based on the author’s writing characteristics
besides the deep semantic features.
Therefore, this paper uses AI-MLP3 method to combine semantic features of text with
n-gram features, and investigate the effect of integrating semantic and lexical statistical
features of source code. The experimental results show that the method is improved
compared with MLP1 method. AI-MLP2 combines semantic features with statistical
features of code structure to verify whether statistical features of code structure have
positive effects on experimental performance. When writing code, the author will reflect
certain personal language characteristics. Based on this discovery, this paper extracts the
structural statistical features in the source code and normalizes the indexes. According
to the experimental results, it is about 1.5% higher than the experimental method AI-
MLP1 , which proves the validity of the statistical features of code structure. However,
the effect is weak, and the in-depth quantitative analysis of the text is still needed. The
experimental result of AI-MLP4 method reached 0.9720, which was the highest among
the 4 groups of experiments. The combination of semantic feature, lexical statistic feature
and code structure statistic feature can more fully capture the author’s personal writing
style information and accurately identify the author identification.
150 X. Sun et al.
Finally, Table 4 shows the comparison between the experimental results of the
optimal performance method proposed in this paper and the four baseline methods.
Table 4. Comparison of experimental results between AI-MLP4 method and baseline method
Table 4 compares experimental results between the AI-MLP4 method and four base-
line methods of RoBERTa Tiny, RoBERTa Tiny-96, and UoB. RoBERTa Tiny and
RoBERTa Tiny-96 methods are the baseline methods given by the FIRE evaluators
in 2020. The first two methods adopt BERT single semantic model, and only adjust the
model parameters to distinguish and obtain the final prediction results. The lexical and
structure-based statistical features of the source text are ignored.
The experimental results in Table 4 also verify the view that “the deep semantic
features and full-text statistical features in the source code text play a common role in
constructing the author’s writing style model and author identification”.
5 Conclusion
The semantic text features based on a pre-trained language model can better model
the semantics of code fragments. Still, it cannot represent complex full-text statistical
features at the text granularity.
In this paper, we propose a source code author recognition method that integrates
semantic and statistical features. We pay attention to the positive role of semantic features
in constructing source code representation and consider the importance of statistical
features in fixed character sets of source code for source code recognition according to the
particularity of source code compared with natural language. The author’s writing style
model is constructed by combining statistical information with semantic information, and
the author’s identity of source code can be accurately identified in MLP neural network.
The experimental results show that the accuracy of the AI-SOCO test data set provided
by FIRE 2020 reaches 0.9720, about 2.1% higher than the strong baseline method, which
verifies the effectiveness. The proposed feature fusion model is statistically superior to
several advanced source code author recognition models, including traditional statistical
models and single semantic feature models based on deep learning.
The integration of complex full-text statistical features provides a new idea that
semantic features of modeling cannot fully extract text representation information and
advances the research on the basic issues of internal plagiarism detection.
References
1. Fadel, A., Musleh, H., Tuffaha, I.: Overview of the PAN@FIRE 2020 task on the author-
ship identification of SOurce COde. In: FIRE 2020–12th Forum for Information Retrieval
Evaluation, pp. 649–676 (2020)
2. Devlin, J., Chang, M.W., Lee, K.: BERT: pre-training of deep bidirectional transformers for
language understanding, pp. 4171–4186 (2018)
3. Pellin, B.: Using classification techniques to determine source code authorship. In: Department
of Computer Science (2000)
4. Hugo, A., Bogotá, C.: Personality recognition applying machine learning techniques on source
code metrics. In: FIRE Forum for Information Retrieval Evaluation, pp. 25–29, India (2016)
5. Alsulami, B., Dauber, E., Harang, R.: Source code authorship attribution using long short-
term memory based networks. In: European Symposium on Research in Computer Security,
pp. 65–82 (2017)
6. Abuhamad, M., Rhim, I., AbuHmed, T.: Source code authorship identification using convolu-
tional neural networks. Future Gener. Comput. Syst. 95, 104–115 (2019)
7. Alvi, F., Stevenson, M., Clough, P.D.: Hashing and merging heuristics for text reuse detection.
In: 2014 Cross Language Evaluation Forum Conference, pp. 939–946 (2014)
8. Crosby, A., Tayyar, H., Tayyar, M.H.: UoB at AI-SOCO 2020: approaches to source code
classification and the surprising power of N-grams. In: The 12th meeting of the Forum for
Information Retrieval Evaluation, pp. 677–693 (2020)
The Improvement of Attribute Reduction
Algorithm Based on Information Gain Ratio
in Rough Set Theory
Abstract. Due to the various data sets and the cumbersome and diverse data
types, there must be many redundant attributes in them, which greatly increases
the classification time in the background of rough set theory. In this paper, we
improve the attribute reduction algorithm by information gain ratio. The data sets
obtained after the attribute reduction of this method are used for classification, and
the data sets are directly used for classification and comparison with other common
classification methods. Experimental data verify that the improved algorithm in
this article is effective. It can improve the classification speed greatly and shorten
the time spent.
1 Introduction
2 Basic Theory
In the field of rough sets, knowledge is represented by an information system. Knowledge
is the sum of the knowledge and experience acquired by human beings in the practice of
transforming the objective world [4]. There are mainly two types of knowledge repre-
sentation systems: one is the information table, which is the knowledge representation
system without decision attributes, and the other is the decision table, which is the knowl-
edge representation system with decision attributes. Generally, the columns in the table
mark different attributes, the rows in the table mark the objects of the universe. Decision
table reduction is also called relative reduction of knowledge. Now we will introduce
some important related concepts in some rough sets.
Definition 1 (Sub-cluster). Given a set X , its power set is 2X , then it is called a set
composed of any elements in the power set, that is, each subset C of the power set is
a subset of X . In rough set theory, we are only interested in those sub-clusters that can
form a division or cover in the universe U .
Definition 4 (Relation matrix). Supposed the X is a non-empty finite set and the R is a
equivalence relation. We define a relation matrix M (R):
⎛ ⎞
r11 r 12 · · · r1n
⎜ r21 r22 · · · r2n ⎟
⎜ ⎟
M (R) = ⎜
⎜ .. .. . . ..
⎟
⎟
⎝ . . . . ⎠
rn1 rn2 · · · rnn
3 Improved Algorithm
3.1 Algorithm Principle and Description
In this part, we will give a data case in the algorithm in Table 1. c1 stands for integer,
c2 stands for Boolean, c3 stands for character and the c4 stands for float. They are all
conditional attributes. At last, the D = {d} is the decision attribute.
c1 c2 c3 c4 d
x1 18 False A 7.6 0
x2 13 True B 7.6 1
x3 22 True C 1.0 1
x4 13 True D 6.9 1
x5 16 False E 14.2 1
First, we are supposed to calculate the relation matrix, which including each attribute.
The results are as following:
⎡ ⎤
1 0 0 0 0.1111
⎢ 0 101 0 ⎥
⎢ ⎥
⎢ ⎥
R1 = ⎢ 0 0 1 0 0 ⎥
⎢ ⎥
⎣ 0 101 0 ⎦
0.1111 0 0 0 1
⎡ ⎤
1 0 0 0 1
⎢ ⎥
⎢ 0 1 1 1 0⎥
⎢ ⎥
R2 = ⎢ ⎢ 0 1 1 1 0⎥
⎥
⎢ ⎥
⎣ 0 1 1 1 0⎦
1 0 0 0 1
156 W. Wang et al.
⎡ ⎤
1 0 0 0 0
⎢ ⎥
⎢ 0 1 0 0 0⎥
⎢ ⎥
R3 = ⎢⎢ 0 0 1 0 0⎥
⎥
⎢ ⎥
⎣ 0 0 0 1 0⎦
0 0 0 0 1
⎡ ⎤
1 1 0 0.7878 0
⎢ 1 1 0 0.7878 0 ⎥
⎢ ⎥
⎢ ⎥
R4 = ⎢ 0 0 1 0 0⎥
⎢ ⎥
⎣ 0.7878 0.7878 0 1 0 ⎦
0 0 0 0 1
⎡ ⎤
1 0 0 0 0
⎢ ⎥
⎢ 0 1 1 1 1⎥
⎢ ⎥
d =⎢ ⎢ 0 1 1 1 1⎥
⎥
⎢ ⎥
⎣ 0 1 1 1 1⎦
0 1 1 1 1
Then, we will select the first attribute. We need to calculate the Gain(a, B, D) of
each attribute. Note that B = Ø in this step.
Then, we need to calculate the Gain_Ratio(a, B, D) of each attribute. The gain ratio
of each attribute is:
Gain(c1 , Ø, D) 0.6611
Gain_Ratio(c1 , Ø, D) = = = 0.3552
H ({c1 }) 1.8611
Gain(c2 , Ø, D) 0.3219
Gain_Ratio(c2 , Ø, D) = = = 0.3315
H ({c2 }) 0.9709
Gain(c3 , Ø, D) 0.7219
Gain_Ratio(c3 , Ø, D) = = = 0.3109
H ({c3 }) 2.3219
Gain(c4 , Ø, D) 0.1926
Gain_Ratio(c4 , Ø, D) = = = 0.1322
H ({c4 }) 1.4573
The Improvement of Attribute Reduction Algorithm 157
We need to select the maximum value of the information gain ratio from it, and select
the corresponding attribute. In this example, we should choose c1 and get B = {c1 }. We
continue to choose another attribute.
Gain(c2 , {c1 }, D) 0.0
Gain_Ratio(c2 , {c1 }, D) = = = 0.0
H ({c2 }) 0.9709
Gain(c3 , {c1 }, D) 0.0608
Gain_Ratio(c3 , {c1 }, D) = = = 0.0262
H ({c3 }) 2.3219
Gain(c4 , {c1 }, D) 0.0608
Gain_Ratio(c4 , {c1 }, D) = = = 0.0417
H ({c4 }) 1.4573
We need to select the maximum value of the information gain ratio from it, and select
the corresponding attribute. In this example, we should choose c4 and get B = {c1 , c4 }.
We continue to choose another attribute.
Gain(c2 , {c1 , c4 }, D) 0.0
Gain_Ratio(c2 , {c1 , c4 }, D) = = = 0.0
H ({c2 }) 0.9709
Gain(c3 , {c1 , c4 }, D) 0.0
Gain_Ratio(c3 , {c1 , c4 }, D) = = = 0.0
H ({c3 }) 2.3219
Because all the remaining attribute information gain ratio is 0, we directly output
the set B, that is, get the final result is {c1 , c4 }.
4 Experimental Analysis
This experiment uses four data sets in the UCI machine learning database, and uses
the above attribute reduction algorithm to process separately, compare their reduction
results and the classification accuracy in two case [11].
The Table 2 records the size of each data set, the amounts of attributes, and the
amounts of decision category.
The experimental results are shown in Table 3. The table records amount of after
attribute reduction of each data set [12]. The second and third columns in the table
respectively list the SVM classification accuracy rate before attribute reduction and the
SVM classification accuracy rate after attribute reduction.
158 W. Wang et al.
5 Conclusion
Generally speaking, the attributes in the knowledge base are not equally important, and
even some of the knowledge is unnecessary or redundant [13]. In this article, through
the improvement of the algorithm, it can deal with a variety of data types, which greatly
improves the running efficiency of the algorithm.
Attribute reduction is a hot issue worthy of joint research and efforts of scholars at
home and abroad [14]. We believe that a more beautiful algorithm can be proposed to
guide practical application. In the next step, we continue to apply this method to data
sets with missing values and ther to improve the algorithm’s efficiency in the distributed
environment.
Acknowledgment. The author sincerely thanks the members of the research team for their joint
efforts and the editors and reviewers for their valuable comments.
References
1. Yang, X., Zhang, Q., Sheng, G.: Research on Classification of LBS Service Facilities Based
on Rough Sets Neural Network. Proceedings of the 30th China Conference on Control and
decision-making 30(3), 76–81 (2018).
2. Xiao, Z., Chen, L., Zhong, B.: A model based on rough set theory combined with algebraic
structure and its application: bridges maintenance management evaluation. Expert Syst. Appl.
37(7), 5295–5299 (2010)
3. Wang, M., Deng, S.: Research to e-commerce customers losing predict based on rough set.
Appl. Mech. Mater. 1287(5), 164–170 (2011)
4. Afsoon, M., Vahid, S.: A fuzzy-rough approach for finding various minimal data reductions
using ant colony optimization. J. Intell. Fuzzy Syst. 26(5), 2505–2513 (2014)
5. Li, Z., Liu, Y., Li, Q., Qin, B.: Relationships between knowledge bases and related results.
Knowl. Inf. Syst. 49(1), 171–195 (2016)
6. Liang, F., et al.: Recognition algorithm based on improved FCM and rough sets for meibomian
gland morphology. Appl. Sci. 7(2), 192–208 (2017)
7. Dai, J., Xu, Q.: Attribute selection based on information gain ratio in fuzzy rough set theory
with application to tumor classification. Appl. Soft Comput. J. 13(1), 211–221 (2013)
8. Li, Y., Deng, Q.: Research of rough set theory on forecast of power generation for grid-
connected photovoltaic system. Appl. Mech. Mater. 3147(3147), 885–889 (2014)
9. Liu, J., Hu, Q., Yu, D.: A weighted rough set based method developed for class imbalance
learning. Inf. Sci. 178(4), 1235–1256 (2007)
The Improvement of Attribute Reduction Algorithm 159
10. Zhang, Y., Ding, S., Xu, X., Zhao, H., Xing, W.: An algorithm research for prediction of
extreme learning machines based on rough sets. J. Comput. 8(5), 1335–1342 (2013)
11. Shu, W., Qian, W., Tang, Z.: An efficient uncertainty measure-based attribute reduction app-
roach for interval-valued data with missing values. Int. J. Uncertain. Fuzziness Knowl.-Based
Syst. 27(6), 931–947 (2019)
12. Li, X., Lu, A.: A rough set theory based approach for massive data mining. Int. J. Signal
Process. Image Process. Pattern Recogn. 11(1), 23–26 (2018)
13. Yan, C., Zhang, H., Nan, T.: Attribute reduction based on generalized orthogonal fuzzy rough
sets. Fuzzy Syst. Math. 34(6), 130–139 (2020)
14. Sheng, K., Wang, W., Dong, H., Ma, J.: Incremental attribute reduction algorithm based on
neighborhood discrimination of mixed data. Electron. J. 48(4), 682–696 (2020)
Accurate Teaching Data Analysis of “Special
Effects Production for Digital Films and TV
Programmes” Based on Big Data
Fan Jing(B)
total of 884 documents have been obtained. Specifically, these documents can be divided
into four categories in terms of research directions. The first category of documents
analyze the basic theory and model design of precision teaching from the perspective
of informatization teaching reform combined with smart learning and propose a method
for precise target determination based on recursive thinking [7]; the second category of
documents introduce the basic concepts, operating procedures and application value of
precision teaching from the perspective of theoretical introduction [8]; the third category
of documents propose precise teaching model and its practical strategies under the flipped
classroom [9], and the fourth category of documents use digital media technology to
accurately focus on students by extracting facial expressions and posture features. This
is a technical solution derived from precision teaching [10]. Through research, it is found
that the current research on precision teaching has the following two problems:
(1) There are more theoretical analyses on precision teaching and less precision
teaching analysis through big data;
(2) There are many methods for teaching reconstruction measures, but there are few
analyses on whether there are links between various teaching designs.
3 Learning Analysis
This course is designed for sophomore students majoring in digital media technology.
On the whole, these students have a solid foundation of basic professional knowledge,
strong professional learning motivation, and independent learning ability. They urgently
need comprehensive practical knowledge and skills for future employment and industrial
development. From the individual level, these students have varying levels of curricu-
lum knowledge: some students have relatively high levels of knowledge, while most
students have almost no operational foundation. From the perspective of the online
teaching environment, it is found through pre-class surveys that students are equipped
with basic online teaching conditions such as laptops, mobile phones, software, and
wireless networks, so mixed online and offline teaching can be successfully carried out.
4 Teaching Contents
According to the set teaching goals, the author reconstructs the teaching knowledge
system based on years of teaching experience and the content of other colleges and
universities. Based on the idea of taking students as the center and cultivating learning
ability as the main goal, the teaching content is divided into the following structures
(Fig. 2):
Fig. 2. Teaching contents of the “Special effects production for digital films and tv programmes”
course
6 Teaching Methods
Task-driven teaching: By letting students watch the effect video, the task completion
effect can be reflected intuitively; through the teacher’s question, students can take the
initiative to construct a learning system of inquiry, practice, thinking, application, and
problem-solving.
Problem-oriented inquiry and discussion: students can continuously explore and
innovate in solving problems proposed by both teachers and students.
Group teaching: through group tasks, students’ teamwork ability, language ability
and innovative thinking ability can be improved.
8 Teaching Design
The real subject of “student-centered” teaching process management is students. Stu-
dent behavior occurs not only in the classroom but also throughout the entire teaching
process before, during, and after class. Students should have more time to complete
independent learning and knowledge system construction through practice before and
after class and cultivate their creative thinking, teamwork, and problem-solving ability
in extracurricular practice. As the supervisor of student behavior, teachers should for-
mulate a complete implementation strategy and monitoring system to escort students in
all aspects to complete the teaching tasks.
164 F. Jing
9 Effect Analysis
The teaching evaluation of this course mainly includes the usual grades (50%) and
end-term performance (50%), among which the usual grades include attendance grades
(5%), independent study grades (15%), homework grades (30%); end -term performance
includes group work (40%) and personality test (60%). Normal performance reflects the
result of process evaluation, and end-term performance reflects the result of personal
ability and teamwork ability. The data for this analysis comes from the learning situation
and offline performance of 43 students majoring in digital media technology in the class
of 2019 in the wisdom tree teaching platform.
In the statistical analysis of the data, the P-P diagram can be used to check whether the
data is normally distributed visually. The usual grades, end-term performance, and total
grades of this course satisfy the normal distribution, and the effect is shown in Fig. 3,
Fig. 4, and Fig. 5.
On the basis of normal analysis, this paper analyzes the correlation between nor-
mal performance and end-term performance data to prove the importance of process
evaluation (Table 1 and 2).
Table 1. The pearson correlation between normal performance and end-term performance
As can be seen from the above table, this article uses correlation analysis to study
the correlation between usual performance and end-term performance. It uses Pearson
correlation coefficient to indicate the strength of the correlation. The specific analysis
shows that the correlation coefficient between the usual performance and end-term per-
formance is 0.339. It shows a significant level of 0.05, which shows a significant positive
correlation between usual and end-term performance.
(2) Correlation between group work and individual test in end-term performance
166 F. Jing
Table 2. Correlation between group work and individual test in end-term performance
End-term performance
Group task 0.263
Individual test 1 0.720**
Individual test 2 0.870**
* p < 0.05 ** p < 0.01
As can be seen from the above table, this article uses correlation analysis to study the
correlation between end-term performance and group work, personality test 1, person-
ality test 2 and a total of three items. Pearson correlation coefficient is used to indicate
the strength of the correlation. Specific analysis shows that the correlation coefficient
between end-term performance and group work is 0.263, close to 0. The p value is
0.080 > 0.05, which shows no correlation between end-term performance and group
work. The correlation coefficient between end-term performance and personality test 1
is 0.720. It shows a significance level of 0.01, which shows a significant positive corre-
lation between end-term performance and personality test 1. The correlation coefficient
between end-term performance and personality test 2 is 0.870. It shows a significant level
of 0.01, which shows a significant positive correlation between end-term performance
and personality test 2.
It can be seen from the above table that the attendance score, homework score, and
interactive score are used as independent variables, and end-term performance is used
as the dependent variable for linear regression analysis. As can be seen from the above
Accurate Teaching Data Analysis 167
10 Conclusion
This paper uses normal distribution, Pearson correlation analysis, and linear regression
analysis to analyze the usual grades and end-term performance. It can be seen that the
results with personalized guidance in the teaching process have a significant impact on
the teaching effect, such as the usual homework results and the final personality test.
From the self-learning process data, it can be seen that students can complete the self-
learning content, but the lack of assessment in the teaching management process makes
it impossible to determine the effect of students’ self-learning. Therefore, the diversified
assessment of independent learning content should be added in the teaching process to
further improve the data and complete the precise teaching design.
References
1. Griffin, C.P., Murtagh, L.: Increasing the sight vocabulary and reading fluency of children
requiring reading support: the use of a precision teaching approach. Educ. Psychol. Pract. 2,
186–209 (2015)
2. Binder, C., Watkins, C.L.: Precision teaching and direct instruction: measurably superior
instructional technology in schools. Perform. Improv. Q. 3, 74–96 (1990)
3. Athabasca University. Precision teaching: Concept definition and guiding principles. <https://
psych.athabascau.ca/open/lindsley/concept.php>
4. Downer, A.C.: The national literacy strategy sight recognition programme implemented by
teaching assistants: a precision teaching approach. Educ. Psychol. Pract. 2, 129–143 (2007)
168 F. Jing
5. Gallagher, E.: Improving a mathematical key skill using precision teaching. Irish Educ. Stud.
3, 303–319 (2006)
6. Strømgren, B., Berg-Mortensen, C., Tangen, L.: The use of precision teaching to teach basic
math facts. Eur. J. Behav. Anal. 2, 225–240 (2014)
7. Zhu, Z.T., Peng, H.C.: Efficient knowledge teaching supported by information technology:
energizing precision teaching. China’s Audio-Vis. Educ. 01, 18–25 (2016)
8. Liang, M.F.: Probing into Precision Teaching. Basic Educ. 06, 4–7 (2016)
9. Zhang, L.Z.: A study on the accurate teaching model in the flipped classroom. J. Wuhan
Metall. Manage. Cadre Coll. 2, 50–52 (2016)
10. Zheng, Y.W., Chen, H.X., Bai, Y.H.: An experimental study on the accuracy of students
’attention in classroom teaching based on big data. Mod. Educ. Sci. 2, 54–57 (2016)
An Improved Clustering Routing Algorithm
Based on Leach
1 Introduction
In an era of information and intelligence, the application value and potential value of
Wireless Sensor Network (WSN) have successfully attracted the attention of academia
and industry. Routing algorithm has become an important research content of Wireless
Sensor Network. Normally, the wireless sensor network nodes are randomly distributed
in harsh environments that are even difficult to enter. The exhausted node energy can-
not be replenished, and the failure of some nodes will cause the network to collapse.
So energy is an important factor affecting the further development of wireless sensor
networks. Among the several modules of the network, energy consumption is mainly
concentrated on the communication module. Therefore, it is important to study the
efficient, energy-saving, and extended network life cycle routing algorithm [1].
According to the different network topologies, routing algorithms are mainly divided
into the plane and layered types. The hierarchical routing algorithm is also called the
cluster routing algorithm. The entire network is divided into multiple clusters, and each
cluster has one and only one backbone node, and the rest of the points are ordinary nodes.
Compared with the planar routing algorithm, the network lifetime of the clustering rout-
ing algorithm is longer, and it has become the most widely used routing algorithm [2].
LEACH algorithm [3] is the earliest cluster routing algorithm, which has very important
significance in the routing algorithm of hierarchical structure. LEACH algorithm adopts
the periodic rotation replacement of cluster heads to avoid the problem that excessive
energy loss of fixed cluster heads can lead to premature failure. Once a node with low
remaining energy or a node far from the base station is used as a cluster head, which
will be accelerated death rate of the node. The algorithm proposed in the literature [4]
improves the LEACH threshold calculation formula and selects cluster heads by com-
prehensively considering node energy and surrounding information. But the problem of
uneven clustering may occur. In the literature [5], the cost function is calculated accord-
ing to the node information to determine the cluster head scientifically. In the literature
[6], the particle swarm algorithm divides the network area, clusters first, and selects
the cluster head. The literature [7] proposes that KUCR uses a clustering algorithm to
evenly cluster, prolonging the network life cycle, but not considering later communi-
cation consumption. The literature [8] introduced multi-hop communication based on
non-uniform clustering, effectively balancing network energy consumption. The algo-
rithm in literature [9] proposed a communication method that combines single-hop and
multi-hop during data transmission, which reduces energy loss. So far, most of the opti-
mized routing algorithms proposed by many researchers are based on the basic idea of
the LEACH algorithm. The ultimate common goal is to enhance the energy utilization
rate of nodes and extend the life cycle of the network.
2 System Model
2.1 Network Model
The network model of this article [10] has the following characteristics:
Where, Eelec is the energy consumed per bit of data received or sent by the node,
εfs and εmp are the power amplification energy loss coefficients in the free space model
and the multipath attenuation model respectively, and d0 is the threshold of transmission
distance.
The energy required by the receiving end to receive k-bit data is:
It is very significant to design a reasonable routing algorithm to balance the network load
in WSN. The improved algorithm clusters the network and then selects the cluster head,
and finally performs data communication. The paper will design an energy-saving clus-
tering algorithm from the following three aspects: uniform network clustering through
clustering, selection of cluster heads, and data communication. The research framework
of this article is shown in Fig. 1.
Network Evenly Clustering. There are still many problems to be solved in wireless
sensor networks, among which prolonging network life is an important one that can be
achieved by reducing energy consumption. In recent years, researchers have conducted
experimental studies to prove that clustering algorithms can improve wireless sensor
network survival time. Clustering is the process of classifying objects according to
172 J. Yu and Y. Bai
certain criteria [12]. As a typical algorithm of clustering and clustering, K-Means has
the characteristics of uniform clustering. Due to the random clustering of the traditional
LEACH algorithm, it may cause very large clusters and very small clusters. The density
of cluster heads will also be affected. Therefore, the clustering algorithm is introduced
in the clustering stage of the LEACH algorithm.
Since random nodes are deployed in wireless sensor networks in real life, there
will inevitably be some deviating nodes. Such nodes will affect the clustering results
of the K-means algorithm. In order to avoid these isolated nodes from affecting the
clustering effect of the K-Means algorithm, the concept of truncated average is introduced
here. Firstly, the clustering center is set within a certain distance range. If the node is
not within this range, the node will be regarded as an isolated node. To avoid it from
affecting the judgment of the entire cluster center, the algorithm temporarily removes
it when calculating the clustering center. In the LEACH algorithm, the overall energy
consumption of the network is not only proportional to the square sum of the distance
from the cluster head to the node in the cluster but also proportional to the fourth
power of the distance from the cluster head to the base station [13]. Correspondingly,
the standard measure function is set here as the sum of the square of the Euclidean
distance between the node in the cluster and the cluster center and the sum of the fourth
power of the Euclidean distance between the cluster center and the base station to obtain
the minimum network energy consumption. The improved standard measure function
formula is shown as follows:
k
E= ( εfs ∗ Di2 (p, μi ) + εmp ∗ Di4 (BS, μi )) (5)
i=1 p∈Ci
Where, BS is the base station, μi is the cluster center point of cluster i, p is the
member node of the cluster, Di2 (p, μi ) is the square of the distance between node p and
cluster center point μi , Di4 (BS, μi ) is the fourth power of the distance from the cluster
center μi to the base station.
N sensor nodes are distributed in the area of M × M. Assuming that the network
√ the area of each cluster is M /K,
is uniformly clustered and is an ideal circle, 2 and the
radius of the circle is R, then R = M / π K, the node density ρ is K/M 2 .Then the
following formula can be obtained:
√ √
2π M / πK K 2π M / πK M2
E 2
dtoCH = r ρ(r, θ )rdrd θ = 2
3
r 3 drd θ = (8)
0 0 M 0 0 2π K
The energy consumption of the whole network is as follows:
M2
Enet =KECH + (N − K)Enon−CH = L 2NEelec + NEDA + εmp dtoBS
4
+ N εfs
2π K
(9)
If the derivative of the above equation is zero, the optimal number of cluster heads
K can be obtained, as shown in the following formula:
N εfs M
K= (10)
2π εmp dtoBS2
The choice of cluster head node is the core part of the clustering routing algorithm. The
quality of the cluster head directly determines the prospect of the entire network. The
LEACH algorithm will randomly replace the cluster heads if the energy of the selected
node is not enough for it to complete heavy tasks. The process will speed up the death
rate of the selected node, which is not conducive to the operation of the whole network
and even causes the collapse of the network. When running for cluster heads in a cluster,
multiple factors should be considered comprehensively to ensure that the selected cluster
heads can meet the normal operation of the network. The cluster head that collects and
merges the data has a relatively large load energy consumption. On the contrary, the load
of the nodes in the cluster is not that large. The choice of the cluster head can use the
Hamming network algorithm in the neural network to balance the load of the network.
In the process of node campaign, the nodes take their relative remaining energy, the
distance from the base station, the number of cluster heads, and the number of adjacent
nodes as the input of the competing network. After the continuous learning of the neural
network, the node with more residual energy near the base station and more neighbor
nodes finally become the cluster head. The calculation formula for the relative remaining
energy of sensor node is as follows:
Ec
E= (11)
Em
174 J. Yu and Y. Bai
Where Ec represents the current energy of the node, Em represents the initial energy
of the node.
The specific process of network learning is as follows:
(1) Set network parameters and variables
Take X (n) = [x1 (n), x2 (n), . . . , xN (n)]T as the input vector of the network, where
X (n) is the sensor node participating in the election in the network, the elements in
Xi (n) respectively represent the node’s distance to the base station, the relative remaining
energy of the node, the number of neighboring nodes of the node, and the number of
times that the node is used for the cluster head.
Wi1 (n) = [wi11 (n), w 1 (n), . . . , w 1 (n)]T as the weight vector of the forward subnet-
i2 iN
2
work, and wkl as the weight vector of the competing subnetwork.
The update formula of the weight vector is as follows:
Where η is the adaptive learning efficiency, 0 < η < 1, and n1 is the number of
iterations of the competitive subnetwork.
Y (n) = [y1 (n), y2 (n), . . . , yN (n)]T as the output vector.
(2) Initialization
Initializing the forward subnetwork weight is actually a random assignment to it,
and satisfies the following formula:
N
wij1 = 1 (13)
j=1
The weights of the initialized competing subnetwork are shown in the following
formula. k and l represent the k th and the l th sensor nodes in the network.
1, k = l
wkl =
2
(14)
−δ, k = l
The output function of the forward subnetwork neuron is a linear function, and the
formula is as follows:
⎧
⎨ 0, x < ζ
y = x − ζ, ζ < x < μc (15)
⎩
C, x > μc
(3) Select training samples
After the sensor node distribution is completed, the node needs to send relevant
parameter information to the base station, these information as input vector of training
sample.
(4) Output results of the forward subnetwork are as follows:
N
yk (0) = Vi1 = f1 ( wij1 xj ), i = 1, 2, . . . M (16)
j=1
consumption of the whole communication network. At the same time, the transmission
rate of single-hop communication will also decrease, and the network’s lifetime will
inevitably be shortened. All nodes of the M-Leach protocol [16] network use multi-hop
communication to transmit data. Although the energy consumption of distant nodes is
improved, the power consumption of nodes close to the base station will increase under
this communication mode.
Therefore, the improved algorithm in this paper uses single-hop and multi-hop com-
munication methods to send data to the base station. Firstly, the paper analyzes transmit
data required the energy between the single-hop and multi-hop communication modes.
Assuming that the node’s transmission energy consumption model is a free space chan-
nel model, a cluster head node can reach a base station with a fixed distance of r through
n-1 hops [17]. The energy consumption of single-hop communication for k -bit data
transmission by the cluster head is:
⎧
⎨ ETX (k, n × r)
Edir = k × Eelec + k × εfs × (nr)2 (18)
⎩
k × Eelec + εfs × n × r
2 2
Since Eelec = 50 nJ/bit and εfs = 10 pJ/ (bit.m2 ) are known. Therefore, when the
minimum value of n is 2, that is, r > 70 m, multi-hop transmission consumes less
energy.
On the basis of distance from the cluster head to base station, the data transmission
mode is selected by setting a threshold value. If the distance is greater than 70 m, the
cluster head applies to multi-hop communication method; otherwise, the cluster head
directly applies to single-hop communication method. This can save the energy cost in
the process of network communication.
Parameter Value
Sensor field 100 m × 100 m
Number of nodes 200
Packet size 500 bytes
Initial energy of nodes 0.5 J
Eelec 50 nJ/bit
εfs 10 pJ/(bit·m2 )
Emp 0.0013 pJ/(bit·m4 )
As the number of simulation rounds increases, the change of the remaining energy of
the whole network is shown in Fig. 3. Under the premise of consuming the same energy,
the algorithm in this paper will run more rounds than the LEACH algorithm and KUCR
algorithm. This is because the data transmission stages consume a lot of energy, and the
algorithm in this paper improves the data communication method and saves energy. In
the clustering stage of the network, the clustering algorithm is used to avoid extremely
maximal and minimal clusters. Therefore, the energy consumption of the nodes in the
network is effectively balanced, and the energy utilization rate is improved.
5 Conclusion
This paper proposes an improved algorithm to solve the problems existing in LEACH.
The network clustering uses an improved algorithm based on the K-means algorithm
to make the network clustering uniform. In the cluster head selection process, various
factors are considered comprehensively, which is more conducive to balancing network
energy consumption. In data communication, the data transmission mode is changed,
and the energy utilization rate is improved. The simulation results show that, compared
with the other two algorithms, the proposed algorithm can solve premature node death,
effectively balance the network energy consumption, improve network throughput and
prolong the network life cycle.
The algorithm mentioned in this article was tested on a simulation platform. Due
to the limited experimental conditions, the nodes could not be deployed in the actual
environment, and the impact of unexpected factors was ignored in the actual working
environment. In the future, the algorithm design will focus on research work in this area
and consider more practical factors to make the designed algorithm more perfect.
Acknowledgment. This work was supported by the Plateau Discipline Innovation Team Project
of the Harbin University of Commerce.
References
1. Wang, T.S., et al.: Genetic algorithm for energy-efficient clustering and routing in wireless
sensor networks. Syst. Softw. 146, 196–214 (2018)
180 J. Yu and Y. Bai
2. Pan, J.Q., Feng, Y.Z.: Improved LEACH sensor network cluster routing algorithm. J. Jilin
Univ. (Sci. Edition) 56(6), 1476–1482 (2018)
3. Heinzelman, W.R., Chandrakasan, A., Balakrishnan, H.: Energy-efficient communication
protocol for wireless microsensor networks. In: Proceedings of the 33rd Annual Hawaii
International Conference on System Sciences, Hawaii, pp. 3005–3014. IEEE (2000)
4. Huang, L.X., et al.: Improved LEACH protocol algorithm based on energy balanced and
efficient WSN. J. Commun. 38(S2), 164–169 (2017)
5. Zhu, S.X., Ma, H.F., Sun, G.L.: An energy-efficient wireless sensor network improved LEACH
protocol. J. Harbin Univ. Sci. Technol. 26(3), 91–98 (2021)
6. Singh, S.P., Sharma, S.C.: An improved cluster-based routing algorithm for energy opti-
misation in wireless sensor networks. Int. J. Wireless Mobile Comput. 14(1), 82–89
(2018)
7. Zhang, Y.Q.: Research on uniform clustering routing algorithm for wireless sensor networks
based on K-means. Control Eng. 22(6), 1181–1185 (2015)
8. Senthilkumar, C., Manickam, J.P.: A path-aware clustering mechanism for energy-efficient
routing protocol in wireless sensor networks. J. Comput. Theor. Nanosci. 14(11), 5478–5483
(2017)
9. Li, Y.N., Xu, F.T., Chen, J.X.: Clustering optimization strategy for WSNs Based on LEACH.
Chinese J. Sens. Actuators 27(5), 670–674 (2014)
10. Wu, X.N., et al.: Clustering Routing protocol based on improved particle swarm optimization
algorithm in WSN. J. Commun. 40(12), 114–123 (2019)
11. Huang, W., Ling, Y., Zhou, W.: An improved LEACH routing algorithm for wireless sensor
network. Int. J. Wireless Inf. Netw. 25(3), 323–331 (2018). https://doi.org/10.1007/s10776-
018-0405-4
12. Bano, S., Khan, M.: A survey of data clustering methods. Int. J. Adv. Sci. Technol. 113,
133–142 (2018)
13. Zhang, X.L.: Improved IoT Energy Consumption Balance Routing Algorithm Based on
LEACH Protocol. Jilin University, Jilin (2016)
14. Al-Baz, A., El-Sayed, A.: A new algorithm for cluster head selection in LEACH protocol for
wireless sensor networks. Int. J. Commun. Syst. 31(1), 1–13 (2018)
15. Liu, L.J., et al.: Research on FPGA WSN polling access control protocol. J. Commun. 37(10),
181–187 (2016)
16. Sony, C.T., Sangeetha, C.P., Suriyake, C.D.: Multi-hop LEACH protocol with modified clus-
ter head selection and TDMA schedule for wireless sensor networks. In: Communication
Technologies (GCCT), pp. 539–543. IEEE (2015)
17. Wang, B., Fu, D.S.: Improvement of LEACH routing protocol for wireless sensor networks.
Instrum. Technique Sens. 8, 71–74 (2016)
Association Rules Mining Algorithm Based
on Information Gain Ratio Attribute Reduction
Abstract. In actual association rule mining, data sets collected from enterprises
or real life often have some problems, such as a large amount of data missing
or data redundancy, which greatly increases the spatial complexity of mining
association rules and makes mining efficiency inefficient. Not only that, some
actual data set contain hundreds or even more attributes. Not only does it take too
long to mine association rules, but there are too many association rules obtained,
making it difficult for users to distinguish which is more valuable information
in practical applications. It is difficult to apply these data to actual enterprises
to get greater benefits. In response to these problems, this paper proposes an
association rule algorithm based on the FP-Growth association rule algorithm of
information gain ratio attribute reduction to extract more valuable information and
improve the efficiency of association rule mining. Finally, through experiments and
comparisons, it is verified that the algorithm proposed in this paper can effectively
mine the association rule information of multi-attribute data sets.
1 Introduction
Association rules are a rule-based machine learning method used to discover valuable
internal connections between transaction items in the data set. In 1993, Agrawal proposed
the Apriori algorithm [1]. As an association rule algorithm, the Apriori algorithm aims
to discover interesting connections between products in the customer’s shopping basket,
creating a way to find association rules, and thus the association rules. It has received
extensive attention, and many scholars have studied it. During the research process,
scholars discovered that the Apriori algorithm needs to scan the database multiple times
before generating a complete set of frequent patterns. Many candidate sets are generated
in this process, which makes the mining of association rules inefficient, and it is difficult
to use when the amount of data is large. In 2000, J. Han [2] proposed the FP-Growth
algorithm. The mining efficiency of the algorithm was very high, and it solved some
problems in the Apriori algorithm. Scholars continue to study and put forward many
improved methods, but when these algorithms have more attributes in the data set, the
efficiency of the algorithm’s execution becomes very low.
The appearance of rough set theory provides favorable conditions for mining asso-
ciation rules. Rough set [3–6] is a set of theories proposed by Professor Z. Pawlak of the
Polish University of Technology. Attribute reduction [7, 8] is one of the core contents of
the rough set. Its role is to eliminate redundant attributes under the premise that the nature
of the original decision table does not change to obtain a reduced decision table. As a
typical attribute reduction algorithm, information gain attribute reduction can reduce the
data set’s attributes very well. Still, when the data set is too large, the attribute reduction
algorithm is not ideal. In 2012, Jianhua Dai et al. proposed an attribute appointment
algorithm based on information gain ratio [9], Reducing many attributes when the data
set is large. The reduction efficiency is greatly improved.
In the development of association rules research, no scholar has combined attribute
reduction with the FP-Growth algorithm. Therefore, to deal with the problems of too
many attributes of association rule data sets, and low algorithm execution efficiency,
this paper uses the method of information gain than attribute reduction on the data set
to preprocess the data to clean out the attributes that are irrelevant or unimportant to
the research content, then combined with the FP-Growth algorithm to realize efficient
and accurate mining of association rules. This is a new research direction and provides
research ideas.
Definition 6: [9, 12, 13] Let the information system IS = {U, R, V, f}, P ⊆ R, U/P =
{X 1 , X 2 , …, X n }. The Shannon’s entropy H(P):
n
n
|Xi | |Xi |
H (P) = − p(xi ) log(xi ) = − log (4)
|U | U
i=1 i=1
Definition 7: [9, 12, 13] Taking B in the condition attribute as the condition, the
conditional entropy of decision attribute D:
m Xi ∩ Yj
n
Xi ∩ Yj
H (D|B) = − log (4)
j=1 |U | |Xi |
i=1
Information gain indicates the degree to which this condition reduces information
entropy. See Definition 8 for its calculation formula.
Definition 8: [9, 14] In the decision-making system DS, B = ∅, the gain of attribute a:
Gain(a, B, D) I ({a}; D)
GainRatio = = (7)
H ({a}) H ({a})
Therefore, the algorithm steps of information gain ratio attribute reduction [9] are
as follows:
(1) Calculate the relationship matrix List1 of all attributes in the data set, Calculate the
incidence matrix D of decision attributes;
(2) B = ∅;
(3) Calculate information gain;
(4) ∀a ∈ C − B, calculate the conditional attribute of a and GainRatio ;
(5) Choose the attribute k that maximizes the mutual gain ratio of a. And add the
selected attribute k to B;
(6) if GainRatio >0, then B ← B ∪ {a}, Go to step 4, otherwise go to step 7;
(7) Get the reduced attribute B.
184 T. Han et al.
3.2 Confidence
For a given rule X ⇒ Y , the higher the confidence level, the greater the probability that
Y will be included in X’s transaction. Confidence degree [18] indicates the strength of
the association, or the reliability of the rule:
P(XY ) support(X ∪ Y ) count(X ∪ Y )
Confidence(X ⇒ Y ) = P(Y |X ) = = = (9)
P(X ) support(X ) count(X )
Association Rules Mining Algorithm 185
The minimum confidence [18], denoted as min _conf, represents the minimum
confidence of the association rule.
In 2000, Han Jiawei and others proposed the association rule algorithm FP-Growth algo-
rithm [19]. The algorithm first needs to generate FP-tree for the data set, and secondly,
it continuously mines frequent itemsets from the generated FP-tree [20].
FP-Tree [20] is a special prefix tree structure; each itemset and its frequency are
stored as a string in the prefix tree. Constructing an FP-tree requires establishing a
project header table and two traversals of the data set. After constructing the FP-tree,
you can start mining frequent itemsets from the previously processed data set. First,
obtain the conditional pattern base from the constructed FP-tree, build the conditional
FP-tree, and finally, repeat the first two steps until the tree contains an element item.
Then generate candidate association rules from the obtained frequent itemsets, set the
minimum confidence, and further filter the candidate association rules to finally get the
association rules.
The FP-Growth [21] mining algorithm based on information gain ratio attribute reduction
is mainly divided into two steps:
Step 1: Use information gain ratio attribute reduction algorithm to clean data to
remove redundant or unimportant data.
186 T. Han et al.
Step 2: Use the FP-Growth algorithm to extract valuable association rule information
from the reduced data set, which improves the accuracy of mining efficiency and helps
users better.
Algorithm 1 Information gain ratio attribute reduction algorithm:
The Catalog Cross-Sell data set comes from a multi-department catalog company.
It is a random customer sample with 4998 rows of data and 10 attributes. The attributes
are customer number, clothing department, household goods department, health product
department, automobile department, and personal electronics. Product department, com-
puter department, gardening department, strange gift department, jewelry department.
In the data set, “1” means that 6 products were purchased from a certain catalog in the
department, and “0” means that no purchases were made.
Use the algorithm proposed in this paper and the traditional FP-Growth algorithm to
mine association rules on the Catalog Cross-Sell data set, and compare the two mining
algorithms’ establishment time and mining time. The results are shown in Fig. 2 and
Fig. 3.
It can be seen from the results in Fig. 2 that the traditional FP-Growth algorithm
takes longer to build than this algorithm, and as the data set becomes larger and larger,
the proposed algorithm and the traditional FP-Growth algorithm build time. The gap is
getting bigger and bigger. From the results in Fig. 3, the mining time of the algorithm
proposed in this paper is shorter. The greater the number of data sets, the more obvious
the superiority of the algorithm proposed in this paper.
50
40
30
20
10
0
1000 2000 3000 4000 5000
ACTUAL NUMBER OF EXAMPLES
Fig. 2. Comparison of the establishment time of catalog cross sell data set
188 T. Han et al.
50
MINING TIME/S
40
30
20
10
0
1000 2000 3000 4000 5000
ACTUAL NUMBER OF EXAMPLES
6 Conclusion
This article provides a good research idea. Faced with the problems of too many attributes
in the data set, resulting in too much space complexity and too long time for mining
association rules, this paper proposes an association rule algorithm based on information
gain ratio attribute selection to improve the accuracy of mining association rules and
reduce the space complexity makes the obtained association rules more pertinent, easy
for users to understand and beneficial for merchants to make better use, and the mining
efficiency is greatly improved. And the effectiveness of the algorithm in this paper is
proved through experiments. However, because the data set is static data, the association
rule changes of the increased data set cannot be mined in real-time. In this respect, further
improvement is needed. The next step will improve the algorithm to solve the problem
of dynamic data association rule mining.
References
1. Rakesh, A., Tomasz, I., Arun, S.: Mining association rules between sets of items in large
databases. Manage. Data 22(2), 207–216 (1993)
2. Han, J., Pei, J., Yin, Y.: Mining frequent patterns without candidate generation. In: Proceedings
of the 2000 ACM SIGMOD International Conference on Management of Data, 16–18 May
(2000)
3. Pawlak, Z.: Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers,
Dordrecht (1991)
4. Pawlak, Z.: A rough set perspective. Comput. Intell. 11, 227–232 (1995)
5. Pawlak, Z.: Rough set approach to knowledge-based decision support. Eur. J. Oper. Res. 99,
48–57 (1997)
6. Pawlak, Z.: Rough set theory and its applications to data analysis. Cybernet. Syst. Int. J. 29,
661–688 (1998)
7. Xun, J., Xu, L., Qi, L.: Association rules mining algorithm based on rough set. In: Proceedings
2012 IEEE International Symposium on Information Technology in Medicine and Education,
pp. 361−364 (2012)
Association Rules Mining Algorithm 189
8. Ma, J., Ge, Y., Pu, H.: Survey of attribute reduction methods. Data Anal. Knowl. Discovery
4(1), 40–50 (2020)
9. Dai, J., Xu, Q.: Attribute selection based on information gain ratio in fuzzy rough set theory
with application to tumor classification. Appl. Soft Comput. J. 13(1), 211–221 (2013)
10. Pawlak, Z.: Rough sets. Int. J. Comput. Inf. Sci. 11(5), 341–356 (1982)
11. Liu, D., Li, T., Miao, D.: Three-Way Decision and Granular Computing. Science Press, Beijing
(2013)
12. Hu, Q., Yu, D., Xie, Z., Liu, J.: Fuzzy probabilistic approximation spaces and their information
measures. IEEE Trans. Fuzzy Syst. 14, 191–201 (2006)
13. Lee, T.: An information-theoretic analysis of relational databases-part I: data dependencies
and information metric. IEEE Trans. Softw. Eng. 13, 1049–1061 (1987)
14. Miao, D., Hu, G.: A heuristic algorithm for knowledge reduction. Comput. Res. Dev. 36,
681–684 (1999)
15. Jia, P., Dai, J., Pan, Y., Zhu, M.: Novel algorithm for attribute reduction based on mutual-
information gain ratio. J. Zhejiang Univ. (Engineering Edition) 40, 1041–1044 (2005)
16. Zhang, Q., Yang, J., Yao, L.: Attribute reduction based on rough approximation set in algebra
and information views. IEEE Access 4, 5399–5407 (2016)
17. Wang, Q., Xie, F., Zhao, M.: Analysis of cross-selling model of digital tv package based on
data mining. Radio Telev. Inf. 12, 57–59 (2016)
18. Yue, S.: Research on Association Rule Mining Algorithm Based on Frequent Pattern Tree.
Tianjin Polytechnic University, Tianjin (2019)
19. Ranjith, K., Yang, Z., Caytiles, R., Iyengar, N.: Comparative analysis of association rule
mining algorithms for the distributed data. Int. J. Adv. Sci. Technol. 102, 49–60 (2017)
20. Ma, R., Wu, H.: Research and application of association rules mining based on FP_growth
algorithm. Jo. Taiyuan Normal Univ. (Natural Science Edition) 20(01), 19–22 (2021)
21. Yang, Z., Geng, X.: Research on mining association rules based on multi-granularity attribute
reduction. Comput. Eng. Appl. 55(6), 133–139 (2019)
22. Nick, S., Wolberg, W., Mangasarian, O.: Nuclear feature extraction for breast tumor diagnosis.
In: Biomedical Image Processing and Biomedical Visualization, International Society for
Optics and Photonics, vol. 1905 (1993)
23. Moro, S., Cortez, P., Rita, P.: A data-driven approach to predict the success of bank
telemarketing. Decis. Support Syst. Elsevier 62, 22–31 (2014)
24. Lin, T.Y., Yin, P.: Heuristically fast finding of the shortest reducts. In: Tsumoto, S., Słowiński,
R., Komorowski, J., Grzymała-Busse, J.W. (eds.) RSCTC 2004. LNCS (LNAI), vol. 3066,
pp. 465–470. Springer, Heidelberg (2004). https://doi.org/10.1007/978-3-540-25929-9_55
Based on the Inception and the ResNet Module
Improving VGG16 to Classify Commodity
Images
Abstract. With the advent of the new retail era, intelligent identification of goods
in the shelf image has become an important technology for managing unmanned
supermarkets. In recent years, classification methods based on convolutional neu-
ral networks have been widely used in image classification. In this paper, a deep
neural network based on VGG16-IR is designed to improve the classification
accuracy of low-resolution commodity images. First, Inception and residual ideas
are integrated, and four modules are designed, namely Inception ResNet module
1, Inception ResNet module 2, ResNet module 3, and ResNet module 4. These
modules replace the convolution phase of the original VGG16, and VGG16-IR
replaces the new network. Secondly, a small batch gradient descent algorithm with
momentum is used. Finally, the activation function of ELU is used to avoid the
phenomenon of partial neuron necrosis. The experiment is conducted in Microsoft
PI100 product image training and testing. The experimental results show that the
model’s accuracy in the training set reaches 99.69%, and the accuracy of the test
set reaches 90.63%. Compared with 89.10% of VGG16, the model in this paper
improves 1.5% and increases the generalization ability to a certain extent.
1 Introduction
At present, the unmanned supermarket has become a trend as a new economy. To realize
the function of unmanned supermarkets to query and update commodity information. It
is necessary to analyze commodity information combined with image processing tech-
nology and identify commodity types in commodity images. So automatic acquisition of
commodity categories has become a hot research direction combining image processing
technology and retail industry. The traditional image classification algorithm is mainly
divided into two stages: the first stage adopts the traditional SIFT feature operator [1–3]
and HOG feature operator [1–3] and other artificially designed feature algorithms to
extract the image features; In the second stage, support vector machine [4] (SVM) and
other classifiers are used for classification. Ke Xiao et al. [5] realized the classifica-
tion and retrieval of flower images by using the image multi-feature and support vector
machine method. Liu Chenxi et al. [6] made use of the multi-feature fusion technology
to judge the image. Still, these shallow learning models could not represent complex
functions and could not effectively describe the data itself.
In view of the problems existing in traditional image classification, deep convolu-
tional neural network gradually occupies a dominant position in the image classification
field. In 2014, Visual Geometry Group of The University of Oxford proposed VGGNet
[7], which replaces the 7 × 7 convolution kernel in the original convolution layer with
3 × 3 convolution kernel based on AlexNet [8]. In 2014, GoogLeNet proposed Incep-
tionV1 [9], and then proposed several versions of Inception (V2~V4). After Inception,
the width and depth of the entire network structure were increased, and the max-pooling
layer was used to replace the full connection layer in the last layer, reducing the number
of parameters. In 2014, He Kaiming’s team published ResNet [10], which transmits the
feature map of the upper layer to the next layer through the residual network, which can
make the network layer very deep and get a good classification effect at the same time.
In order to further improve the classification function of VGG16 and solve the prob-
lem that the parameters of VGG16 network layer are not updated. This paper is designed
and improved based on VGG16. In terms of model improvement, the advantages of
Inception and residual ideas are integrated, and four new modules are designed, namely
Inception ResNet module 1, Inception ResNet module 2, ResNet module 3, and ResNet
module 4. Compared with the original module, these new modules have many convo-
lution kernels with different sensing field sizes and channel merging layer and residual
connection. In the forward propagation, the acquisition, fusion and transmission ability
of image features are stronger. Then, these modules are used to replace the convolu-
tion stage of the original VGG16, and the new network designed is named VGG16-IR.
The problem that the parameters of VGG16 network layer are not updated is solved.
The model’s generalization ability and classification accuracy are judged by comparing
the test accuracy between VGG16 and VGG16-IR. In terms of algorithm selection, the
small-batch gradient descent algorithm [11] is adopted to solve the problem of travers-
ing all training data in the stochastic gradient descent algorithm, and the training time
is reduced, and the generalization ability of the model is improved through the batch
method. At the same time, the momentum method is used to prevent the network from
falling into the local minimum when the learning rate is low. The activation function
chooses ELU to solve the problem that the parameters of ReLU activation function can-
not be updated in the negative half interval. In terms of experiments, all experiments
in this paper are carried out under the Microsoft PI100 commodity image. VGG16 and
VGG16-IR are trained and tested 3 times each, and the proportion of training set and test
set is 8:2. The training set and test set of each experiment are selected again. The loss
value is reduced by 0.1093, and the test accuracy is improved by 1.50%, which shows
that VGG16-IR is better than VGG16 in terms of model accuracy and loss.
192 Y. Zhang and Y. Shen
Sl+1 is the number of nodes in layer l + 1. f (•) represents the activation function,
(l+1)
and Wkj represents the weight parameter between the k th node in layer l + 1 and the
j th node in layer l .
Formula (2) is obtained by expanding formula (1) down one level.
sl+1 l+2
(l) (l+1) s (l+2) (l+2) (l+1) (l)
σj = Wkj Wik σi f zk f zj (2)
k=1 i=1
network parameters, the network width and depth are increased, and the generalization
ability of the model is improved. Thirdly, a convolution layer Conv1 of 1 × 1 is added
before the parallel structure, and the convolution nuclear power of 1 × 1 can realize
cross-channel information interaction and information integration, and automatically
realize the dimensionalities’ increase or decrease of the number of feature channels,
and maintain the original input features to a large extent. Fourthly, the actual number of
convolution layer channels in Conv1 is twice that of input channels. The purpose is to
expand the number of input feature maps and better extract image features. The channel
merge layer strengthens the connection between low-level features and high-level fea-
tures and improves model accuracy. Fifth, module 1 continues to use the characteristics
of small convolution kernel, so that the network will go through multiple nonlinear acti-
vation functions, which can make the nonlinear capability of the network stronger and
improve the training accuracy. Sixth, the ResNet module can be used to propagate back
faster and more accurately [14] and update the weight parameters, so that the network
depth can be deepened without affecting the accuracy.
Figure 2 shows Inception 2 ResNet module. The module is the improved version of
module 1 and 2 for VGG16 third stage of convolution, module 2 is to the advantage of
the module 1 is 3 × 3 convolution kernels into two layer 1 × 3 and 3 × 1 series, reduced
the number of arguments increases the nonlinear network, at the same time to obtain
the characteristics of horizontal and vertical, rich network of feature extracting, further
improve the accuracy of the model.
194 Y. Zhang and Y. Shen
Figure 3 shows ResNet module 3, which is designed to act on the fourth convolution
stage of VGG16 by using the idea of ResNet and the convolution characteristic of 1 ×
1. Module 3 adopts the principle that three 3 × 3 convolution kernels and one 7 × 7
convolution kernels have the same size receptive fields. In the case of extracting higher-
level semantic features, image categories can be better differentiated, so that the network
can classify images more effectively.
Fig. 5. VGG16 - IR
Ensemble learning can also reduce the risk of overfitting a single model. However, since
this paper is to study the influence of the new model on the experimental results, in order
to reduce other influencing factors, these methods to reduce overfitting are not used in
this paper.
3 Optimization of Network
3.1 Selection of Optimization Algorithm
The essence of neural network training is optimization, so that the training parameters
can fit the image as much as possible. The most widely used method in neural network
is stochastic gradient descent [16]. The principle is that along the negative gradient
direction, the function drops the fastest. In this paper. The cross-entropy loss function is
calculated with softmax results and the small-batch gradient descent algorithm isused.
Assuming that the batch is a set of m training data x(1) , y(1) , · · · , x(m) , y(m) . The
loss function is shown in formula (3).
m
n
(i) (i)
J (W , b) = − m1 yk ln ok (3)
i=1 k=1
(i)
ok represents the probability that the prediction of sample i belongs to category k,
(i)
and softmax function is used. ok is shown in formula (4).
ef (W ,b,x )
(i)
(i)
ok = m f (W ,b,x(i) ) (4)
i=1 e
(i) (i)
yk is the actual probability (if the true category of the i TH sample is k, then yk = 1,
(i)
otherwise yk = 0). x(i) is the network input value. f W , b, x(i) is the output value of
the last layer of the full connection layer. W is the weight parameter, and b is the bias
parameter.
The optimal parameter (W , b) solved in this paper, according to the principle of
gradient descent method, uses J (W , b) to obtain ∂J ∂W (W ,b)
and ∂J (W
∂b
,b)
by taking partial
derivatives of W and b respectively.
Then, parameters Wt and bt are updated according to their negative gradient direction
to obtain parameters Wt+1 and bt+1 after training. a is the learning rate, as shown in
formula (5) and (6).
Wt+1 = Wt − a ∂J (W
∂Wt
t ,bt )
(5)
198 Y. Zhang and Y. Shen
bt+1 = bt − a ∂J (W t ,bt )
∂bt (6)
Although the gradient descent method is also called the fastest descent method. The
direction of each iteration always advances towards the local minimum value, and it is
easy to cause the local optimal solution. In order to solve this problem, a momentum
[17] term is introduced to accelerate the convergence speed and the convergence curve
is more stable. The weight parameter Wt+1 with momentum is updated, seeing formula
(7) and (8).
Vt = μVt−1 + a ∂J (W
Wt
t ,bt )
(7)
Wt+1 = Wt − Vt (8)
For the update of parameter bt+1 , see formula (9) and (10).
bt+1 = bt − Ut (10)
Compared with the ReLU function, the gradient of the negative half interval of the
ELU activation function is aex , and the whole function does not contain the dead zone.
Based on the Inception 199
When the input is greater than or equal to 0, the gradient is consistent with the ReLU
function. Parameter a is set to 1 to ensure that the gradient at x = 0 is 1. In the case
of x < 0, the gradient in the negative half region is 0~1, and the parameter can also be
updated, which solves the problem that the output result of ReLU function does not have
negative value and the neuron is necrosis.
Commodity
Sample date set
categories
Cup
Electric drill
Eardrop
Guitar
Data preprocessing in this paper mainly includes: manual labelling of data sets; The
resolution of the modified image is 224 × 224. Table 2 shows part of the modified data
set.
200 Y. Zhang and Y. Shen
ntrain−correct
Train-accuracy = 25 , ntrain−correct
is the total number of batch training sam-
1 (i)
m n
(i)
ples that are correctly classified, and train-loss = − 25 yk ln ok . Test-accuracy
i=1 k=1
ntest−correct
= 1000 , ntest−correct
is the total number of Test set samples that have been correctly
1 (i)
m n
(i)
classified, test-loss = − 1000 yk ln ok . The average training accuracy of VGG16
i=1 k=1
is 94.67%. The average test accuracy is 89.10%. The average training loss is 0.0989, and
the average test loss is 0.5243. The average training accuracy of VGG16-IR is 99.69%.
The average test accuracy is 90.63%. The average training loss is 0.0039, and the average
test loss is 0.4150.
It can be seen from Table 3 that the VGG16-IR network used in this paper is better
than VGG16 in terms of accuracy and loss index, indicating that the model and algorithm
adopted in this paper are improved compared with VGG16. This is mainly due to the
Based on the Inception 201
feature extraction capability of Inception residual module, which increases the network’s
ability of feature selection and utilization, and then increases the ability of commodity
image recognition. In addition, through SGDM optimizer and ELU activation function,
the model accuracy is further improved, and the test accuracy is increased by 1.50%,
indicating that the generalization ability of the method in this paper is increased, and the
model performance is improved. The training accuracy increased by 5.02%.
Figure 6 shows the training progress chart of the first 1000 iterations of VGG16 for
three times.
60%
40%
20%
0%
1 70 140 210 280 350 420 490 560 630 700 770 840 910 980
Iteraon
VGG16 experiment 1 VGG16 experiment 2 VGG16 experiment 3
Fig. 6. VGG16 training progress chart for the first 1000 iterations
Figure 7 shows the testing progress chart of the first 1000 iterations of VGG16 for
three times.
202 Y. Zhang and Y. Shen
50.00%
40.00%
30.00%
20.00%
10.00%
0.00%
1 80 160 240 320 400 480 560 640 720 800 880 960
Iteraon
VGG16 experiment 1 VGG16 experiment 2 VGG16 experiment 3
Fig. 7. VGG16 testing progress chart for the first 1000 iterations
Figure 8 shows the training progress chart of the first 1000 iterations of VGG16-IR
for three times.
60%
40%
20%
0%
1 70 140 210 280 350 420 490 560 630 700 770 840 910 980
Iteraon
VGG16-IR experiment 1 VGG16-IR experiment 2
VGG16-IR experiment 3
Fig. 8. VGG16-IR training progress chart for the first 1000 iterations
Figure 9 shows the testing progress chart of the first 1000 iterations of VGG16-IR
for three times.
Based on the Inception 203
60.00%
40.00%
20.00%
0.00%
1 80 160 240 320 400 480 560 640 720 800 880 960
Iteraon
Fig. 9. VGG16-IR testing progress chart for the first 1000 iterations
The x-coordinate represents the number of iterations, and the y-coordinate represents
accuracy. By comparing the training progress chart and test progress chart of VGG16 with
those of VGG16-IR, it can be seen that the training curve of VGG16-IR is steeper, and the
number of iterations needed to reach the stable state is less, mainly because the VGG16-
IR model uses wider and deeper network architecture and residual connection in terms
of model complexity. Residual structure and momentum gradient descent algorithm
accelerate the iteration speed of parameters, resulting in steep training curve. In the same
number of iterations, the test accuracy is higher than that of VGG16. By comparing the
training progress chart and test progress chart of VGG16-IR, it can be seen intuitively
that the generalization performance of VGG16-IR model is improved. Table 4 reflects
the test accuracy of different iterations.
Vgg16-IR model predicts the effect of PI100 products, as shown in Fig. 10. The pre-
diction title consists of prediction label and prediction accuracy. The prediction accuracy
of curtain is 89.74% and that of electric drill is 90.50%. Pendant prediction accuracy is
89.70%. The model predicted commodity categories with an accuracy of about 90%.
204 Y. Zhang and Y. Shen
5 Conclusion
This paper’s main contributions and advantages are as follows: (1) Inception ResNet
module and the ResNet module are adopted in the overall network model, which strength-
ens the network’s ability to capture and utilize image features and improves the ability
to classify commodity images accurately. (2) An ELU activation function is added after
each convolution layer to avoid the phenomenon of gradient disappearance and enable
all neurons to exert effects in the network. (3) In order to avoid the network falling into
the local minimum point, a small batch random gradient descent algorithm with momen-
tum term is used to maintain the true descent direction and accelerate the convergence
speed.
The model used in this paper is tested in the Microsoft PI100 commodity image data
set. The test accuracy is 90.63% in the test set, which significantly improves the clas-
sification ability of commodity image. Compared with the existing general convolution
network model VGG16, the vGG16-IR in this paper has higher classification accuracy.
Because the Inception residual module is proposed in this paper, the training parame-
ters are reduced, the training speed is effectively accelerated, and the model accuracy
is improved. The next research focus is to continue to explore more efficient feature
learning modules.
References
1. Parashivamurthy, R., Naveena, C., Sharath, K.Y.H.: SIFT and HOG features for the retrieval
of ancient Kannada epigraphs. IET Image Process. 14(17), 4657–4662 (2020)
2. Lowe, D.G.: Object recognition from local scale-invariant features. In: Proceedings of the 7th
IEEE International Conference on Computer Vision, vol. 2, pp. 1150–1157(1999)
3. Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: Proceedings
of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition,
vol. 1, no. 1, pp. 886–893(2005)
Based on the Inception 205
4. Akshit, K., Pavan, D., Aarya, V., Manan, S.: A comprehensive comparative study of artificial
neural network and support vector machine on stock forecasting. Ann. Date Sci. 1–26 (2021).
https://doi.org/10.1007/s40745-021-00344-x
5. Ke, X., Chen, X.F., Li, S.Z.: Flower image retrieval based on multi-feature fusion. Comput.
Sci. 37(11), 282–286 (2010)
6. Liu, C.X., Chu, J.H., Lu, W.: Image salient judgment based on multi-feature fusion. Comput.
Eng. Appl. 50(9), 150–154 (2014)
7. Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolution
neural networks. Commun. ACM. 60(6), 84–90 (2017)
8. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image
recognition. Comput. Sci. (2014)
9. Szededy, C., Liu, W., Jia, Y.Q.: Going deeper with convolutions. In: IEEE Conference on
Computer Vision and Pattern Recognition, Boston, pp. 1–9 (2015)
10. He, K.M., Zhang, X.Y., Ren, S.Q., Sun, J.: Spatial pyramid pooling in deep convolutional
network for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 37(9), 1904–1911
(2015)
11. Zhang, H., Fusheng, Y., Sun, J., Shen, X., Li, K.: Deep learning for sea cucumber detection
using stochastic gradient descent algorithm. Eur. J. Remote Sens. 53(1), 53–62 (2020)
12. He, K.M., Zhang, X.Y., Ren, S.Q.: Deep residual learning from image recognition. In:
Conference on Computer Vision and Partner Recognition, pp. 770–778 (2016)
13. Lin, M., Chen, Q., Yan, S.C.: Network in network. In: Proceedings of International Conference
on Learning Representations (2014)
14. Mehmet, A., Özgür, K., Tefaruk, H.: Suspended sediment prediction using two different
feed-forward back-propagation algorithms. Can. J. Civ. Eng. 34(1), 120–125 (2007)
15. Lv, W., Chen, Y.G., Shen, C.: Convergence analysis of backward iterative neural network
algorithm with L2 regularization term. Inf. Technol. Inf. 6, 183–186 (2015)
16. Robbins, H., Monro, S.: A stochastic approximation method. Ann. Math. Stat. 22(3), 400–407
(1951)
17. Qian, N.: On the momentum term in gradient descent learning algorithms. Neural Netw. 12(1),
145–151 (1999)
18. Bai, Y., Jiang, D.M.: Application of improved ELU convolutional Neural network in SAR
image ship detection. Bull. Surveying Mapp. 1, 125–128 (2018)
19. Glomt, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural
networks. J. Mach. Learn. Res. 9, 249–256 (2010)
Improved Algorithm of Multiple Minimum
Support Association Rules Based on Can Tree
Abstract. To improve the mining efficiency of Can Tree in association rules, this
study found that the header table in the traditional Can Tree mining algorithm
repeatedly traverses the path, causing excessive overhead. And the single support
will lead to invalid rules. The improved Can Tree mining algorithm performs
reduction processing and removes the header table, making it more suitable for
the idea of Can tree construction. Traversing the path only once avoids multiple
traversals, and the addition of multiple minimum support does not introduce too
much overhead. Experiments have proved that the improved Can tree mining is
better than the traditional mining algorithm.
1 Introduction
Association rules have been an important field of data mining since they were proposed by
Agrawal [1]. With fixed data sets and support, scholars continue to mine static association
rules. The classical algorithms include Apriori [2], FP-growth [3], Ecalt [4], etc.
With the development of The Times, people are concerned about how to better mine
data and update the original data mining results with the least cost, namely dynamic asso-
ciation rule mining. Dynamic association rules refer to the mining of association rules
when the database is modified or the minimum support changes. Incremental mining has
been a research topic in dynamic mining since Cheung et al. first proposed the incremen-
tal mining idea –FUP [5, 6] algorithm based on the Apriori framework in order to solve
the incremental updating problem. Afterward, under the tree structure framework, Koh
et al. proposed an incremental mining algorithm based on FP tree (AFPIM) [7] algo-
rithm. To achieve the effect of the building once, mining many times, and to overcome
the shortcomings of constantly scanning the database, Cheung et al. improved the tree
structure and proposed the compressed and arranged transaction sequence tree (CATS
tree) [8] and the mining algorithm FELINE through continuous sorting to improve the
compressed storage. Aiming at cats’ shortcomings, Leung et al. Proposed a natural ordi-
nal tree (Can Tree) sorted according to a certain normative order [9]. Compared with
incremental mining, it is found that CATS trees are more suitable for interactive mining
because the insertion of transaction items needs to adjust the position according to the
frequency. Can Tree is especially suitable for incremental mining because of its stability.
In practical application, the single support confidence framework cannot fully express
the differences of each component of the mining object in its characteristics and actual
occurrence frequency, so conventional information or some useless information is often
mined. In order to satisfy the characteristics of each project and fully excavate each
project, researchers propose association rule mining with multi-minimum support [10–
13].
2 Can Tree
2.1 Related Theories of Can Tree Construction
Can Tree (Canonical-order Tree) adopts the prefix tree model, arranges the transaction
items in a certain order, the user sets the order. Generally it can be lexicographical
or alphabetical order and the user’s interest. The sorted transaction items are directly
inserted into the Tree with “null” as the root node. The Can Tree can have multiple
subtrees. Once the order is determined, the transaction tree is unique, and the insertion
of a transaction item does not require a large-scale search to find the insertion position.
Only Fig. 1 shows the establishment process. The minimum support degree does not
participate in the construction of the Can Tree but only participates in the mining process
after the Tree is established. The algorithm can perform frequent pattern mining under
different support degrees without rebuilding the tree structure.
As for tree mining, the proposer of Can Tree did not specifically describe it, but
suggested one-way mining with the idea of divide and rule like FP tree, as shown in
Fig. 2. Because of the unique nature of Can Tree, scholars are attracted to continue
in-depth research. Zou Li-Kun et al. [14] used the structure of Can Tree to make up for
the deficiency of FP tree in 2008 and put forward an algorithm of rapidly generating
a conditional mode tree, which improved the operation efficiency. Chen Gang et al.
proposed a fast construction algorithm based on Can-Tree (FCCAN) in 2014 [15]. In
this algorithm, an auxiliary storage structure based on a hash table is added based on a
rapidly generating conditional schema tree to reduce the search time of items. In 2014,
Hu Junequan proposed a Can [16] Tree based on data quantity sorting, which reduced
the size of Can Tree.
208 X. Zhao and S. Y. Ning
For example, as shown in the figure below, the traversal starts from the root node “Null”,
and the depth traversal path {a, d, b} is pushed onto the stack. At this time, the elements
of the stack are < a, d, b >, b is the leaf node, and b is popped out. Extract the remaining
elements in the stack as {(a, d): 1} as the conditional pattern base of b, backtracking
to d, at this time d has no untraversed child nodes, and d pops the stack to extract the
conditional pattern base {a: 3}, backtracking To a, at this time, a has unprocessed child
nodes and continue to traverse down deeply (Fig. 3).
At present, the algorithms for the minimum support research mainly focus on the tradi-
tional Apriori and FP-Growth algorithms. The Can Tree is originally constructed accord-
ing to user-defined order and contains all data items, which leads to long mining process-
ing time. The additional minimum support will increase the processing time. The Can
Tree needs to process each node on the path when traversing the Tree. The processing
time will not increase when the multi-minimum support is regarded as a single minimum
support of multiple values.
In order to record the minimum support of each item, add a column to store the
minimum support in the transaction item set table. Although the minimum support is
set separately for each project, it still faces the same problem as the single support, that
is, it is unreasonable to set the minimum support too high or too low. In order to solve
the deviation of the minimum support of manual assignment, the minimum support of
the item is determined by the number of times the item appears in the database and the
maximum support set. The formula is as follows.
maxminsupport minsupport
= (1)
firstsupport isupport
Maxminsupport is a manual set of the maximum and minimum support, first support
is the support for the most supported data item, minSupport is the minimum support for
the current data item to be set, and support is the support for the current item.
If we write the left-hand side of the formula as α, the multimin supports calculated by
this formula in descending order must be equal to the multimin supports in descending
210 X. Zhao and S. Y. Ning
order. In this way, in the tree building stage, sorting by support order is sorting by multiple
minimum support order. For a Can tree constructed in this order, assuming a path from
the root node to the leaf node, if the support degree of the leaf node is greater than its own
minimum support degree, then the path must be frequent. Because the minimum support
of leaf nodes is the minimum value of multiple minimum support of all nodes on this
path, that is, minsupport(leaf) = min{min (I1), min (I2),……, min(In)}, I1-In is the node
on this path. Similarly, when extracting the conditional pattern base for a node, it is not
necessary to look at whether the support of the minimum support term in the conditional
pattern base is greater than its own minimum support degree, but only need to look at the
support of this node and this path. In this way, the problem is simplified to the traditional
algorithm mining under multiple minimum support models. Although the maximum
and minimum support are manually set as a whole, the minimum support of other items
is indirectly determined by the maximum and minimum support. To a large extent, it
avoids the non-standard error caused by the manual setting of the minimum support
for each project, and it will be more reasonable to determine the minimum support for
other projects according to the proportion of the maximum project in the database. In
conclusion, the frequent item set algorithm expanded and mined according to formula
(1) must have recursion, and must be sorted downward closed, because all subsets of
frequent candidate item sets in the frequent item set after the recursive production of L2
are frequent (Fig. 4).
4 Algorithm Process
According to the theorem combined with the characteristics of Can Tree, scan the
database twice. During the first scan, all item sets are sorted in order of support from
large to small to make a sequence table of item sets, and the minimum support of each
item is calculated. Scan the database for the second time, sort each transaction according
to the item set order table, and insert the sorted data items directly into the Can Tree
with “null” as the root node to build the Tree. In the process of tree construction, search
for common items layer by layer from the root node. If there are any common items,
add one to the count of support degree of data items under the path. The pseudo code is
as follows.
Improved Algorithm of Multiple Minimum Support Association 211
1. If thisitem.getChildren() ! = null
2. If tmpRoot = subTreeRoot.findChild(record.peek())) ! = null
3. tmpRoot.countIncrement(1);
4. subTreeRoot = tmpRoot;
5. record.poll();
6. Else Node = new Node();
7. Node.count = 1;
8. addNode();
After building the Tree, initialize the stack, traverse the Tree in depth first, and press
the items on the stack at the same time; On the premise of descending order of support
degree and multi-minimum support degree as the measurement standard, Once the item’s
support is less than its minimum support, the path is no longer traversed down the stack,
If there are other path nodes at the top of the stack that haven’t been traversed, traversing
other child paths, Each time out of the stack, the conditional mode base of the stack item
is extracted and saved. The pseudo code is as follows.
1. if(nowNode.getName()! = null)
2. stack.push(nowNode.getName());
3. if (nowNode.getChildren()! = null){
4. for(TreeNodechild:nowNode.getChildren())
5. if(child.getCount > itsmin)
6. CPBtable = traversal(child, stack,singlefequentlist);
7. if(stack.empty() = = false)
8. stack.pop();
5 Experiment
Fig. 5. The results of a simple can tree running Fig. 6. Experimental results of can tree
experiment mining after traversal optimization
mode base for memory storage data items, which occupies a larger memory space.
And it is necessary to judge whether all nodes belong to a frequent 1 item set, which
also increases time consumption. In the incremental aspect, when adding new data,
although the Can Tree does not need to be reconstructed like the FP algorithm, it still
needs to process all nodes when mining. Constantly add quantitative data to simulate
the formation of incremental databases.
6 Conclusion
Based on the selection of Can trees, throwing away the constraints of the initial ideas, and
using a reduced Can tree construction method; Using stack to assist traversal simplifies
the mining process, avoiding the time waste caused by repeated traversal path, greatly
shortens the running time. The multi minimum support is integrated into the Can Tree to
solve the defect of inaccurate rules under the single support confidence framework. At
the same time, the formula is used to avoid the defect of manually setting the minimum
support of each item. The comparison of experimental results proves that the improved
algorithm can greatly improve the mining efficiency and has very good practicability.
References
1. Agrawal, R., Imieliński, T., Swami, A.: Mining association rules between sets of items in
large databases. ACM SIGMOD Rec. 22(2), 207–216 (1993)
2. Agrawal, R., Srikant, R.: Fast algorithms for mining association rules. In: International
Conference on Very Large Data Bases, San Francisco, pp. 487–499 (1994)
3. Han, J., Pei, J.: Mining frequent patterns by pattern-growth: methodology and implications.
ACM SIGKDD Explor. Newsl. 2, 14–20 (2000)
4. Zaki, M., Parthasarathy, S., Ogihara, M., et al.: New algorithms for fast discovery of associa-
tion rules. In: 7th International Workshop Research Issues on Data Engineering, pp. 283–286
(1997)
5. Cheung W.L., Lee S.D., Kao, B.: A general incremental technique for maintaining dis-
covered association rules In: International Conference on Database Systems for Advanced
Applications, Singapore, pp. 185–194 (1997)
Improved Algorithm of Multiple Minimum Support Association 213
6. Han, J., Pei, J., Yin, Y., Mao, R.: Mining frequent patterns without candidate generation: a
frequent-pattern tree approach. Data Min. Knowl. Disc. 8, 533–587 (2004)
7. Koh, J.-L., Shieh, S.-F.: An efficient approach for maintaining association rules based on
adjusting FP-tree structures. In: Lee, Y., Li, J., Whang, K.Y., Lee, D. (eds.) Database Systems
for Advanced Applications, pp. 417–424. Springer, Heidelberg (2004). https://doi.org/10.
1007/978-3-540-24571-1_38
8. Cheung, W., Za¨ıane, O.: Incremental mining of frequent patterns without candidate
generation or support constraint. In: Proceedings of IDEAS, pp. 111–116 (2003)
9. Leung, C. K.-S., Khan, Q.I., Li, Z., Hoque, T.: CanTree: a canonical-order tree for incremental
frequent-pattern mining. Knowl. Inf. Syst. 11(3), 287–311 (2007)
10. Zhao, C.J., et al.: Algorithm for mining association rules with multiple minimum supports
based on FP-Tree. NZ J. Agric. Res. 50(5), 1375–1381 (2010)
11. Alhusaini, N.Q., Hu, M., Hawbani, A.: Improved P-tree Mine Rare Association Rule with
Multiple Minimum Support, Hong Kong (2015)
12. Liu, Y.C., Cheng, C.P., Tseng, V.S.: Discovering relational-based association rules with mul-
tiple minimum supports on microarray datasets. Bioinformatics (Oxford, England). 27(22),
3142–3148 (2011)
13. Huang, T., Cheng, K.: Discovery of fuzzy quantitative sequential patterns with multiple
minimum supports and adjustable membership functions. Inf. Sci. 222, 126–146 (2012)
14. Cheung, D.W., Han, J., NG, V.T., et al.: Maintenance of discovered association rules in large
databases: an incremental updating technique. In: Twelfth International Conference on Data
Engineering, pp. 106–144 (1996)
15. Zou, L.K., Zhang, Q.S.: Efficient incremental association rules mining algorithm based on
CANTree. Comput. Eng. 34(3), 29–31 (2008)
16. Cheng, G., Yan, Y.Z., Liu, B.Q.: A fast construction algorithm based on CAN-Tree.
Microelectron. Comput. 1, 76–82 (2014)
Mining Association Rules of Breast Cancer
Based on Fuzzy Rough Set
1 Overview
According to statistics, the incidence and mortality of breast cancer are among the best
among all kinds of malignant tumors in women, one of the leading killers threatening the
life and health of all human women. Therefore, breast cancer screening is an essential
strategy that enables people to find breast cancer earlier and ensure better treatment with
greater probability. Our goal is that the data collected from routine consultation and
blood analysis can predict breast cancer robustly. So the critical data selection method
is the key to predicting breast cancer.
With the deepening of scientists’ research, more and more data mining algorithms are
available, and it is the main difficulty to find a suitable algorithm for the problems studied.
A.Bourouis and others [1] use an artificial neural network algorithm to analyze the retinal
images collected by microscope films and identify retinopathy. Research shows that BP
neural network has a slow convergence speed and average prediction accuracy. Wang
Cheng [2] optimized the traditional random forest algorithm and proposed an improved
random forest algorithm based on decision tree clustering reduction, which improved
the accuracy and efficiency of the conventional random forest algorithm. Still, it was
difficult for decision trees to find rules based on the combination of multiple variables.
EI-Alfy [3] has studied the parallel attribute reduction method of large-scale data sets
oriented to rough decision-making sets based on a genetic algorithm. Still, it is difficult
for rough sets to deal with continuous attributes directly, so attributes must be discretized
first.
To sum up, through combing the literature, it is found that any single data mining
method has certain limitations, especially in the face of various data. It is often neces-
sary to integrate and apply multiple ways to obtain more practical results [4, 5]. Fuzzy
rough sets are involved. Because breast cancer data are all continuous attribute values,
If the classical coarse mass theory is used directly for data mining, the data needs to be
discretized, which will destroy some relations among attributes and may not get some
valuable ties [6–9]. Therefore, this paper uses fuzzy rough set theory based on the gain
ratio to reduce the data, which can delete some conditional attributes that have nothing
to do with breast cancer. Finally, the Apriori algorithm is used to get the rules that can
predict breast cancer, and the extracted rules are expressed productively so that people
can use them themselves.
Definition 2 [8, 10]. There is a resolution table DS = ‹U, C ∪ D, V›, where C is the
definite set and D is the substantial value, B ⊆ C. This is the definition of information
between b and d.
Definition 4 [11]. ∀a ∈ C – B, and the mutual information gain ratio of detail a is defined
as
Gain(a, B, D) I (B ∪ {a}; D) − I (B; D))
GainRatio(a,B,D) = = (5)
H ({a}) H ({a})
You can calculate the equivalence relation using the following two similarity
functions:
|x −x | |x −x |
1 − 4 ∗ |amaxi −ajmin | , |amaxi −ajmin | ≤ 0.25
rij = (7)
0, otherwise
xj − (xi − σa ) (xi + σa ) − xj
rij = max min , ,0 (8)
xi − (xi − σa ) (xi + σa ) − xi
Definition 6 [12, 13]. U can be regarded as the universe. According to the fuzzy
equivalence relations r, u can be divided into fuzzy partitions, which are defined as
n
U /R̃ = [xi ]R̃ i=1 (9)
Definition 8 [12, 13]. There is a fuzzy information system FIS = <U, A, V, f>, storing
attribute sets. P and Q are two subsets of A, and the joint entropy of P and Q is defined
as
1 n [x ]
i R̃ ∩ [xi Q̃
]
H̃ (PQ) = H Rp RQ = − log (11)
n i=1 n
Mining Association Rules of Breast Cancer 217
Definition 9 [12, 13]. There is a fuzzy decision system FDS = <U, A, V, f>, The
conditional entropy of D conditioned on B is defined as
1 n [xi ]R̃ ∩ [xi ]Q̃
H̃ (D|B) = − log (12)
n i=1 |[xi ]R̃ |
B, D) = Ĩ (B ∪ {a}; D) − Ĩ (B; D)
Gain(a, (14)
Definition 13. In FDS, the expected input ratio of an attribute can be defined as
B, D)
Gain(a,
Gain_Ratio(a, B, D) = (15)
H̃ ({a})
Then the selection of attributes based on the Gain is an algorithm.
GAIN_RATIO_AS_FRS.
Definition 15. The degree of S support in the union rules indicates the X ∪ Y trades
percentage in the trading database D.
Definition 16. Confidence c in connection rules refers to the ratio of the number of
trades containing X ∪ Y to the number containing x in the trading database.
support(X ∪ Y )
confidence(X ⇒ Y) = (16)
supoort(X )
218 M. Guo et al.
Usually, the breast cancer type selection rules are multidimensional association rules.
The antecedents of the rules contain at least two items, that is, frequent itemsets with
more than three items. Find out the itemset Lk (k > 3) containing breast cancer types
from frequent itemsets, take breast cancer as the afterpart of the rule, and other attributes
as the antecedents of the government, and calculate the confidence of each rule Y ⇒Z.
The strict control of breast cancer type selection meets the confidence threshold.
4 Case Analysis
For the convenience of analysis, the breast cancer data in the UCI data set is simplified.
It is assumed that only the following five attributes are included, namely, Age, Glucose,
HOMA, Resistin. The following is an example of some of the original data (Table 1).
We use fuzzy approximate group theory to remove features unrelated to breast cancer
from the dataset. Table 1 can be regarded as an information system. The universe U
includes six objects, namely U = {x1 , x2 , x3 , x4 , x5 , x6 }. The conditional attribute set
includes four attributes: Age, glucose, Homa, and resistance, which are marked as C1 ,
C2 , C3, and C4 respectively, that is, C = {C1 , C2 , C3 , C4 }, and the decision attribute
set D has only one attribute of disease type. First, we use Formula 7 to calculate the
relational matrix for each property.
220 M. Guo et al.
Next, we need to calculate the Gain of each Gain characteristic Gain(a, B, D) and
B, D). And you need to select B = Ø.
the gain ratio Gain(a,
1 , ∅, D) = Ĩ ({C1 }; D) = 0.34507503
Gain(C
Gain(C2 , ∅, D) = Ĩ ({C2 }; D) = 0.5162319
3 , ∅, D) = Ĩ ({C3 }; D) = 0.1187528
Gain(C
4 , ∅, D) = Ĩ ({C4 }; D) = 0.5362271
Gain(C
Next, we need to calculate the gain ratio Gain_Ratio(a, B, D) of each attribute.
Gain(C1 ,∅, D)
Gain_Ratio(C1 , ∅, D) = = 0.19556396
H̃({C1 })
Gain(C2 ,∅, D)
Gain Ratio (C2 , ∅, D) = = 0.31164397
H̃({C2 })
Gain(C3 ,∅, D)
Gain Ratio (C3 , ∅, D) = = 0.16571818
H̃({C3 })
Mining Association Rules of Breast Cancer 221
Gain(C4 ,∅, D)
Gain_Ratio(C4 , ∅, D) = = 0.32315529
H̃({C4 })
Then we choose C4 as the standard of gain ratio. And let B = {C4 }. and then continue
to select other attributes.
Gain(C1 ,{C4 },D)
Gain_Ratio(C1 , {C4 }, D) = = 0.0833209
H̃({C1 })
Gain(C2 ,{C4 },D)
Gain_Ratio(C2 , {C4 }, D) = = 0.0936883
H̃({C2 })
Gain(C3 ,{C4 },D)
Gain_Ratio(C3 , {C4 }, D) = = 0.0
H̃({C3 })
Then we choose C2 as the standard of gain ratio. And let B = {C4 , C2 }. and then
continue to select other attributes.
Gain(C1 ,{C4 ,C2 },D)
Gain_Ratio(C1 , {C4 , C2 }, D) = = 0.1285770
H̃({C1 })
Gain(C3 ,{C4 ,C2 },D)
Gain_Ratio(C3 , {C4 , C2 }, D) = = 0.0
H̃({C3 })
Then we choose C1 as the standard of gain ratio. And let B = {C4 , C2 , C1 } and then
continue to select other attributes.
Gain(C3 ,{C4 ,C2 ,C1 }, D)
Gain Ratio (C3 , {C4 , C2 , C1 }, D) = = 0.0
H̃({C3 })
Therefore, after the attribute reduction of the fuzzy rough set, the attribute that
influences breast cancer is {C4 , C2 , C1 }, that is {Age, Glucose, Resistin}.
In this paper, according to the routine blood information table in physical examination,
referring to related papers [4], and consulting doctors and experts, the critical value of the
range of discrete attributes is determined. Age (A1: 20–40, A2: 41–65, A3: 66), Glucose
(B1: 70, B2: 70–108, B3: 108), Resistin (C1: 18.57, C2: 18.75–28.07, C314-16] (Table
2).
rules from frequent itemsets L3, L3 can extract 9 rules. As shown below, the association
rules with patients as the successor of rules are all strong association rules with 100%
confidence.
According to the above rules, age, the values of MCP-1, glucose, resistin, and
adiponectin in blood all have significant effects on breast cancer. Middle-aged women
Mining Association Rules of Breast Cancer 223
aged 41–65 are more likely to suffer from breast cancer than other ages. When the glu-
cose in the blood is within the normal range and resistin is on the high side, they will
have breast cancer. When adiponectin and glucose are typical for middle-aged women,
but MCP-1 in the blood is high, they will have breast cancer. Therefore, more attention
should be paid to women aged 41–65. When MCP-1 and resistin are on the high side in
physical examination results, doctors should be consulted in time, and early detection
and treatment should be done.
6 Conclusion
In this paper, the attribute reduction of the fuzzy rough set can delete the attributes
that do not influence breast cancer, improve the accuracy of subsequent rules, reduce
the calculation amount of rule extraction, and improve the efficiency of experiments.
Using the Apriori algorithm to extract the association rules between routine blood data
and breast cancer can strengthen prevention according to physical examination results,
which is very practical.
Acknowledgment. The author sincerely thanks the research group members for their efforts on
this paper and thanks to the editors and reviewers for their valuable comments.
References
1. Bourouis, A., Feham, M., Hossain, M.A.: An intelligent mobile based decision support system
for retinal disease diagnosis. Decis. Support Syst. 59, 341–350 (2014)
2. Wang, R., Xuan, Y.: A research on the improvement of dual optimization on BP neural
network. IEEE Computer Society (2017)
3. El-alfy, E.M., Alshammari, M.A.: Towards scalable rough set based attribute subset selection
for intrusion detection using parallel genetic algorithm in MapReduce. Simul. Model. Pract.
Theory 64, 18–29 (2016)
4. Sun, M., Zhao, S., Duan, Y.: GLUT1 participates in breast cancer cells through autophagy
regulation. Naunyn Schmiedebergs Arch. Pharmacol. 394, 205–216 (2021)
5. Hu, Q., Yu, D., Xie, Z.: Information-preserving hybrid data reduction based on fuzzy-rough
techniques. Pattern Recogn. Lett. 27, 414–423 (2006)
6. Dai, J., Xu, Q.: Attribute selection based on information gain ratio in fuzzy rough set theory
with application to tumor classification. Appl. Soft Comput. 13, 211–221 (2013)
7. Wang, W., Xu, L., Shen, C.Y.: Effects of traditional Chinese medicine in treatment of breast
cancer patients after mastectomy: a meta-analysis. Cell Biochem. Biophys. 71, 1299–1306
(2015)
8. Song, J., Lyu, Y., Wang, M.: Treatment of human urinary kallidinogenase combined with
maixuekang capsule promotes good functional outcome in ischemic stroke. Front. Physiol.
9, 84 (2018)
9. Lee, T.T.: An information-theoretic analysis of relational databases-part I: data dependencies
and information metric. IEEE Trans. Softw. Eng. 13, 1049–1061 (1987)
10. Miao, D., Hu, G.: A heuristic algorithm for knowledge reduction. Comput. Res. Dev. 36,
681–684 (1999)
224 M. Guo et al.
11. Jia, P., Dai, J., Pan, Y., Zhu, M.: Novel algorithm for attribute reduction based on mutual-
information gain ratio. J. Zhejiang Univ. 40, 1041–1044 (2005)
12. Hu, Q., Yu, D., Xie, Z.: Information-preserving hybrid data reduction based on fuzzy-rough
techniques. Pattern Recogn. Lett. 27, 414–423 (2006)
13. Hu, Q., Yu, D., Xie, Z.: Fuzzy probabilistic approximation spaces and their information
measures. IEEE Trans. Fuzzy Syst. 14, 191–201 (2006)
Multi-task Model for Named Entity Recognition
dqzheng@hrbcu.edu.cn
Abstract. Although the model based on deep learning significantly improves the
Natural Language Processing (NLP) model, compared with the real conversation
scene, Natural Language Understanding (NLU) task is still limited by training
data and training tasks. In this paper, we propose a Bi-directional Long Short-
Term Memory network model with multi-task learning (Multi-Bi-LSTM + CRF)
of Classification and named entity recognition, which combines multiple task
training to NLU. Multi-Bi-LSTM + CRF uses a large number of cross task data
and benefits from the regularization effect so that that word vectors can be more
commonly expressed and NLU tasks can be understood more deeply. Based on
Bidirectional LSTM + CRF models (Bi-LSTM + CRF), we propose a new neu-
ral network model, which can simultaneously carry out multi-task learning of
intention recognition and named entity recognition, focusing on named named
entity recognition training. In addition, we use different optimizers to control and
update task parameters to adapt to the loss. Compared with the baseline model,
our proposed model achieves better results.
1 Introduction
If computers want to communicate with humans without obstacles, they need to under-
stand the structural characteristics of human language. Natural language understanding
(NLU) aims to analyze and process human language, parse a sentence or paragraph with
certain rules, the foundation for other tasks, such as relationship extraction, machine
translation, question answering system, etc. Named entity recognition (NER) is a basic
task in natural language processing (NLP) [2]. The basic task of NER is to intercept
one or more short sub-segments in a given information segment (audio or text), and the
intersection between sub-segments is empty. Suppose we think that text is the carrier of
information and can represent the speaker’s intention and attitude. Therefore, the types
of entities vary according to the speaker and intention. For example, the speaker’s entity
usually changes with the location, time and emotion, etc. Human beings can naturally
judge the entity or nonentity in the sentence and know the reason. Still, it is difficult
for computers to use the speaker’s information other than information fragments. For
example, in the text “Wang Wu goes to see the sunrise at five o’clock today” and “Wang
Wu says the sunrise will be at five o’clock today”, if other processing is not done to the
text, the two paragraphs of text may be cut out to obtain the same entity, which are [0,
‘Wang Wu’, 1, ‘sunrise], similarly, In the “welcome to Nanjing Yangtze River Bridge”,
it is impossible to identify whether it is the “Yangtze River Bridge” of the “mayor of
Nanjing” or the “Nanjing Yangtze River Bridge”. If you do not understand the speaker’s
intention or background, it is difficult to correct judgment. In order to solve these real
problems, we next introduce a sequence-based neural network model for named entity
recognition.
This paper proposes a joint multi-tasking training learning direction to construct
a named entity recognition system considering the speaker’s intention. The proposed
multi-task learning model does not aim to improve the accuracy of intention recognition.
Different from the work of Liu et al. [3], on this basis, we determine the main task and
auxiliary task. We expect to learn to share complementary information between similar
tasks and serve the main task. Considering that the proportion of intention recognition
in multi-task learning is too large, compared with Tatsuya et al. [4] controlling the
proportion of tasks with weights, we have explored the improvement of proportion
balance by updating the parameters of the task model with different Optimizers and
learning rates. In conclusion, the main contribution of this paper is to show how our neural
network model deals with the situation of multiple factors affecting entity recognition.
2 Related Work
In this part, in order to review and summarize more comprehensively, we introduce the
work in the natural language processing model of intention recognition, named entity
recognition, and multi-task learning.
Named Entity Recognition: Named entity recognition aims to identify the subject with
specific meaning for an input text, such as person name, place name, organization name,
etc. entity recognition is divided into character level, phrase level, and text level, depend-
ing on the task requirements. Natural language processing is usually used as the first step
of natural language understanding for many subsequent NLP tasks, such as entity linking,
relationship extraction, question answering system, etc. Its performance directly affects
the subsequent NLP tasks. Huang et al. [1] proposed using bidirectional long-term and
short-term memory networks to solve mission entity recognition in 2015. Conditional
random fields are added to solve the label dependency. Subsequently, Chiu et al. [15]
added CNN to learn the character-level features of words, significantly improving the
effectiveness of person names and place names. However, it has no significant effect on
organization names or nested entities with equal lengths of multiple words. At the same
time, Yin et al. [16] used word mixing (char) to learn more semantic features but ignored
global semantic information.
Multi-task Learning: For a given text or sentence, learning its vector space representa-
tion, the two most effective methods are multi-task learning and the language pre-training
model. Some of the knowledge we learned from the auxiliary task helps the main task.
For example, a person who studies English can learn German at the same time. In this
way, some of the knowledge in German can be applied to English learning and help him
master English. Similarly, joint learning of multiple tasks is also useful. For example,
Liu et al. [3] used the multi-task deep neural network (MT-DNN) model to learn multi-
ple tasks, which determined excellent results on 8 of the 10 NLG tasks, and multi-task
learning can eliminate the over-fitting effect caused by over adaptation to specific tasks,
which is also a way of regularization. Thus, the Multi-task learned is common among
tasks.
3 Model
3.1 Overview
Our model takes named entity recognition and intention classification as classification
tasks implemented by two classifiers, respectively. Through multi-task learning, named
entity recognition and intention classification are learned at the same time. Multi-task
learning usually involves several similar tasks because they can share information to
improve each task’s performance. However, our multi-task method aims to improve
named entity recognition, and the performance of intention classification is not affected.
This is different from general multi-tasking learning.
Our model is based on Bi-LSTM + CRF model [1]. It includes Bi-LSTM, CRF,
and a linear layer to solve entity naming recognition and intention classification. In our
experiment, the CRF layer is used for named entity recognition and the linear layer
is used for intention classification. Its architecture is shown in Fig. 1. Its architecture
is shown in the figure. The lower layer is shared by all tasks, while the upper layer
represents the output of a specific task. For the input X = {t0 , t1 ,…,tn-1 , tn } is a sequence
of tokens of length n. Firstly, the word embedding vector is obtained by word embedding
228 D. Zheng et al.
matrix, the output is expressed as Xemb = {temb 0 , temb 1 ,…, temb n-1 , temb n }, Then enter the
Bi-LSTM encoder to capture the context representation information of each word and
obtain a sequence context word embedding Xlstm = {tlstm 0 , tlstm 1 ,.., tlstm n-1 , tlstm n }. This
is represented by the shared parameters of multi-task target training. In the following
content, we will elaborate the model.
Fig. 1. The Bi-directional Long Short-Term Memory with multi-task learning model
Bi-LSTM Encoder: We use a bidirectional LSTM neural network structure to map the
input representation vector into a string of context embedding vectors Xemb ∈ Rn × m .
This part of the vector is shared across different tasks. The bidirectional LSTM uses
both the previous hidden state and the latter hidden state in calculating the hidden layer
and backpropagation:
−
→ ← −
ht = ht : ht + LSTM xt , ht−1 ; LSTM xt , ht+1 ] (1)
−
→ ← −
where ht , ht is the forward, back propagation hidden layer parameter. xt is the
vector of a unit.
Intention Classification Output: We take the last vector tlstm n of Xlstm for intention
classification, the probability that tlstm
n is marked as class C after passing through a linear
layer:
P(c|tlstm
n ) = softmax W T
·
intent nt lstm
(2)
where WTintent is a specific classifier parameter matrix, and its dimension is WTintent ∈
R2024× 13
Named Entity Recognition Output: After obtaining the final vector representation
Xlstm of Bi-LSTM output vector, enter the CRF layer. The last layer is the linear-chain
Multi-task Model for Named Entity Recognition 229
conditional random field (CRF), which is parameterized as shown in formulas: the CRF
adopts a linear structure(linear chain conditional random field), and uses the label infor-
mation before and after to predict the current label. Finally, the vector representation of
CRF is obtained.
P(y|x) = z(x)
1
exp I,k λk tk yi−1 , yI , x, i + I,l μl sl yI , x, i (3)
Z(x) = y exp I,k λk tk yi−1 , yI , x, i + I,l μl sl yI , x, I (4)
where sl is the feature function and is only related to the current node, and the features
of the current node are extracted. tk is a context dependent feature function to extract
the information of nearby nodes. λk and μl are their weight coefficients. The CRF layer
output this is the sequence with the greatest probability.
3.2 Training
Intention Classification: For the classification task, the cross entropy loss is used as
the target. If the correct label is y, linear layer classification output of the model is ŷ.
Entity Recognition: In CRF training, the CRF loss function only needs two scores: the
score of the real path and the total score of all possible paths. Among the scores of all
possible paths, the proportion of real path scores will gradually increase. Our goal is to
maximize the proportion of real paths and define the loss function:
LossFunction = RealPath
P1+P2+...+PN (6)
where the total path score remains unchanged, we finally optimize the real path, and
the final loss function:
4 Experiments
4.1 Datasets
This section briefly describes the data sets used in this experiment. The data set is con-
structed by Harbin University of technology and marked with data for Winter Olympics.
There are 19891 named entity identification data sets and 17010 intention identification
data sets. We selected 17010 intersection parts to form a new data set. Each data consists
of three parts, namely input statement, entity and intention. The newly established data
set is divided into 11010 training sets, 3000 development sets and 3000 test sets.
There are 13 kinds of intentions, namely “ask the mascot”, “ask the participating
countries” and “ask the competition venues of the Olympic Games”, “Ask about the
events of the Olympic Games”, “ask about the sub events of a certain category”, “ask
about the first place of a sub event”, “ask about the second place of a sub event”, “ask
about the third place of a sub event”, “ask about the achievements of an athlete”, “ask
what competitions an athlete has participated in”, “ask about which Winter Olympics
an athlete has participated in”, “Ask about the competitions held in the venues”, “ask
about the number of Winter Olympic Games held in the venues”.
Our model code is implemented based on pytorch1.5. Adam and SGD are optimizers- 1
and optimizer-2, respectively. The model learning rates are lr1 = 0.0001 and lr2 = 0.015
respectively. The dimension of LSTM neuron implicit state is dimension_size = 1024
and the dimension of size word vector is embedded_size = 300, reserved parameter
scale in dropout layer d = 0.5, and appropriate learning rates are adopted for different
optimizers, the batch size is batch_ size = 8, and the maximum number of model training
epoch is 10. In order to prevent gradient explosion, the gradient is limited to 5.
Multi-task Model for Named Entity Recognition 231
4.3 Results
We use different systems to evaluate the quality of named entity recognition, and use
recall rate: the proportion of correctly labeled Winter Olympic entities in labeled entities
in this experiment, accuracy rate: the proportion of correctly labeled Winter Olympic
entities in predicted entities in this experiment, and the average harmonic value of the
two to test the performance of the system. The experimental results are shown in Table 1.
Where p is the accuracy rate, which indicates how many positive samples are predicted
to be positive, r is the recall rate, which indicates how many positive examples in the
sample are correct, and F1 is the harmonic average of them.
Comparative experiments analyze the experimental results of the named entity recog-
nition model based on Bi-LSTM and Bi-LSTM + CRF. It can be found that the effect
of named entity recognition is greatly improved after adding the CRF layer, indicating
that adding CRF can indeed learn the dependency between tags from the training data,
To improve the accuracy of label prediction.
Through the naming of named experimental entity results based on Bi-LSTM + CRF
and with the word mixing analysis, it can be found that the effect of the model is better
after incorporating the word information obviously improved, indicating that integrating
word vector information into word vector can effectively alleviate the problem of data
sparsity in training data,enriches the meaning of word vector, to improve the effect of
named entity recognition.
Through the named entity recognition model of word mixing and the named entity
recognition model integrating CNN with word information. The experimental results
show that adding Convolutional Neural Network can be a slight improvement in the effect
of the model. It shows that the context feature of window size extracted by Convolutional
Neural Network is useful for named entity recognition model beneficial.
By analyzing the experimental results of the multi-task learning named entity recog-
nition model and the other four existing named entity recognition models, it can be
found that after adding multi-task hybrid training, the named entity recognition effect is
effectively improved, And get the best score on such indicators as Precision it is close
to the best score in the index of Recall and F1, which shows that multi-task learning can
indeed learn complementary information from data and improve the accuracy of label
prediction.
Model p r F1
Bi-LSTM 80.25 81.06 80.65
Bi-LSTM+CRF 93.99 91.78 92.87
Bi-LSTM+CRF+char 95.47 95.65 95.56
Bi-LSTM+CRF+char+cnn 95.98 95.39 95.68
Multi-Bi-LSTM+CRF 96.45 94.57 95.50
232 D. Zheng et al.
5 Conclusion
This paper proposes a Multi-Bi-LSTM + CRF model, which integrates multi-task learn-
ing and named entity recognition to better understand language features. Multi-Bi-LSTM
+ CRF has achieved excellent results and demonstrated strong generalization ability
through four baseline criteria: Bi-LSTM, Bi-LSTM + CRF, Bi-LSTM + CRF + char,
Bi-LSTM + CRF + char + CNN. There are still many areas to be explored in the future
to improve the multi-Bi-LSTM + CRF model, including a deeper understanding of the
model sharing structure and the loss weight relationship between tasks, More training
correlations between tasks, or use fine-tuning and pre-training models. Finally, we also
want to verify whether the Multi-Bi-LSTM + CRF model has resilience against attacks.
References
1. Huang, Z., Xu, W., Yu, K.: Bidirectional LSTM-CRF models for sequence tagging. arxiv:
1508.01991 (2015)
2. Wen, Y., Fan, C., Chen, G., Chen, X., Chen, M.: A survey on named entity recognition. In:
Liang, Q., Wang, W., Liu, X., Na, Z., Jia, M., Zhang, B. (eds.) CSPS 2019. LNEE, vol. 571,
pp. 1803–1810. Springer, Singapore (2020). https://doi.org/10.1007/978-981-13-9409-6_218
3. Liu, X., He, P., Chen, W.: Multi-task deep neural networks for natural language understanding.
In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics
(2019)
4. Ide, T., Kawahara, D.: Multi-task learning of generation and classification for emotion-aware
dialogue response generation. In: Proceedings of the 2021 Conference of the North American
Chapter of the Association for Computational Linguistics: Student Research Workshop (2021)
5. Ramanand, J., Bhavsar, K., Pedanekar, N.: Finding suggestions and ‘buy’ wishes from product
reviews. In: Proceedings of the NAACL HLT 2010 Workshop on Computational Approaches
to Analysis and Generation of Emotion in Text, pp. 54–61. Association for Computational
Linguistics, Stroudsburg, PA (2010)
6. Li, X., Dan, R.: Learning question classifiers: the role of semantic information. Nat. Lang.
Eng. 12(3), 229–249 (2015)
7. Bhatta, J., Shrestha, D., Nepal, S.: Efficient estimation of Nepali word representations in
vector space. J. Innov. Eng. Educ. 3, 71–77 (2020)
8. Kim, Y.: Convolutional neural networks for sentence classification. arXiv https://arxiv.org/
abs/1408.5882 (2014)
9. Technicolor, T., Related, S.: ImageNet classification with deep convolutional neural networks.
Commun. ACM 60(6), 84–90 (2017). https://doi.org/10.1145/3065386
10. Wang, R., Li, Z., Cao, J.: Convolutional recurrent neural networks for text classification. In:
International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, July 14–19
2019. https://doi.org/10.1109/IJCNN.2019.8852406
11. Zhou, C., Sun, C., Liu, Z.: A C-LSTM neural network for text classification. Comput. Sci.
1(4), 39–44 (2015)
12. Chung, J., Gulcehre, C., Chok, H.: Empirical evaluation of gated recurrent neural networks
on sequence modeling. https://arxiv.org/abs/1412.3555 (2014)
13. Vaswani, A., Shazeer, N., Parmar, N., et al.: Attention is all you need. In: 31st Conference on
Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA (2017)
14. Devlin, J., Chang, M., Lee, K.: BERT: pre-training of deep bidirectional transformers for
language understanding https://arxiv.org/abs/1810.04805 (2018)
Multi-task Model for Named Entity Recognition 233
15. Chiu, J., Nichols, E.: Sequential Labeling with Bidirectional LSTM-CNNs, pp. 937–940.
https://anlp.jp/proceedings/annual_meeting/2016/pdf_dir/D6-2.pdf
16. Yin, Z., L., X.Z., Huang, D.: Chinese named entity recognition ensembled with character. J.
Chin. Inf. Process. (2019)
Research on Non-pooling Convolutional Text
Classification Technology Combined
with Attention Mechanism
1 Introduction
As an effective means of information management and application, text classification
is mainly classified into different categories according to the text content and topics. In
recent years, the realization of text classification tasks through deep learning technology
has become the mainstream solution for researching this application. Deep learning is
based on low-level features combined to form abstract high-level features for text classifi-
cation tasks [1]. Text classification technology is used on many occasions, automatically
classifying work orders, providing personalized news [2], and analyzing customers’
intentions and moods [3].
With scholars’ continuous in-depth research on text classification technology at home
and abroad, many classification algorithms have emerged since artificial intelligence
has become a hot topic. The current more popular text classification methods can be
roughly divided into traditional machine learning-based text classification methods such
as Naive Bayes [4], random forest [5] model, support vector machine (SVM) [6], and
so on. The other is a text classification method based on deep learning. Most of the deep
learning models commonly used in text classification tasks are based on neural networks,
used text representation method converts the text into one-hot encoding and uses TF-
IDF and its extension method for feature extraction. The above method is suitable for
datasets containing only a small amount of data because under normal circumstances, the
data volume of the corpus is over one million, and using one-hot encoding method will
cause the vector-matrix dimensionality and sparsity to be too high, leading to the curse
of dimensionality. Meanwhile, the relationship between contexts is ignored because the
one-hot encoding has only two values of 0 and 1. Therefore, in recent years, word vector
technology has gradually replaced the traditional text representation method. In 2013,
Mikolov et al. proposed word2vec [12], based on the prediction method, whose core idea
is to use the context of the current word to predict the target word, while the Skip-gram
method is the opposite of the CBOW method, that is, the word that occurs is used to
predict the word in its context.
According to the literature [7], a fast text model for Fast text classification [13]
appeared. This model is based on the CBOW model. The biggest feature is that the
model is simple, the training speed is fast, and it can handle tasks with many samples
and multiple category labels.
This article uses Microsoft’s open-source word2vec tool for word vector training.
The Skip-gram method in word2vec is used for word vector pre-training after removing
stop words from the Chinese text dataset by the Jieba package. The Skip-gram model
trains semantic embeddings by predicting target words in context and also captures
semantic relationships between words. Assuming that the length of a sentence S is N,
the entire text can be vectorized into the following formula:
S = x1 , x2 , x3 , · · · · · · , xN (1)
The text classification model uses a non-pooling layer convolutional neural network
combined with the attention mechanism. Zhu previously proposed a convolutional neu-
ral network classification algorithm based on the attention model in his paper [14] and
verified the attention and CNN. However, this paper still uses the traditional convolu-
tional neural network structure. It is believed that the pooling layer can indirectly reduce
the number of neural network parameters in the next layer, increase the calculation speed
and reduce overfitting. Based on the reference [15], this paper discards the pooling layer
on the existing CNN structure, and uses the attention mechanism to replace the pooling
layer to complete the feature extraction operation.
Research on Non-pooling Convolutional Text Classification Technology 237
3 Model Structure
In 2014, Kim proposed the Text CNN model, which was the first application of CNN
in natural language processing. After that, some text classification models with CNN as
the main structure have emerged one after another, but because CNN could not capture
the correlation of location information, CNN has not received much attention in natural
language processing. In order to improve the accuracy and efficiency of Chinese text
classification tasks, this paper proposes a non-pooling layer convolutional neural network
combined with an attention mechanism called PA-CNN (Pure and Attention CNN). In
the attention layer, the important features of each sentence are obtained by extracting the
maximum weight of each dimension feature of the sentence to complete the classification
task. The model in this paper is mainly composed of the input layer, convolution layer,
attention layer, and fully connected layer. The model structure is shown in Fig. 2, and
the various components of the model are introduced below.
semantics used when calculating the output of each item remains unchanged. Therefore,
this traditional coding structure has two disadvantages: first, if there are noise words
in the input sequence that are weak in distinguishing ability or not helpful for text
classification, it will affect the intermediate semantics and then affect the classification
ability. Secondly, because the value of y1 will influence the subsequent output solution,
the order of the word input sequence is very important.
To make up for the shortcomings of the codec framework, Bahadanau et al. [10]
proposed a model based on the attention mechanism. As shown in Fig. 4, unlike RNN,
the output result of the attention mechanism depends on the hidden status of the last
input. Each input will also affect the output with different weights α i,j [18]. The features
that impact the text classification are assigned a larger weight, and the weight is reduced
on the features with less distinguishing significance. The attention mechanism adjusts
the weight by continuously training the input data, and finally achieves the purpose of
improving the classification effect. This operation can further refine the feature words
after the convolutional layer operation to simplify the model parameters.
In this model, the output of the attention layer is transmitted to the full connection layer
as an input. In order to improve the performance of the convolutional neural network,
the full connection layer generally selects the Relu function as the activation function.
However, the softmax function is generally used for multi-classification problems.
The softmax function, also known as the normalized exponential function, promotes the
two-class function sigmoid in multi-classification. Its function is to output the results
of multi-classification in the form of probability values. The biggest feature is that the
prediction probability is non-negative, and the prediction result is equal 1. The calculation
formula is:
exp(Wy .x)
P(y|x) =
C (3)
exp(Wc .x)
c=1
4 Analysis of Results
4.1 Experiment Preparation
This article is generated by Tsinghua University based on historical data of the Sina
News RSS subscription channel from 2005 to 2011. The dataset contains a training set
of 50,000 samples, a validation set of 5,000 samples, a test set of 10,000 samples, and a
vocabulary list of 5000. The text contains ten categories: sports, finance, real estate, home
furnishing, education, technology, fashion, current affairs, games, and entertainment.
The experimental configuration is shown in Table 1.
Research on Non-pooling Convolutional Text Classification Technology 241
(1) Accuracy: The ratio of the correct prediction result to the sum of all prediction
results, which is:
TP + TN
accuracy = (5)
TP + TN + FP + FN
The training time and accuracy of this model are shown in Fig. 5.
(2) Precision: All correctly predicted positive samples are divided by all positive sample
predictions, which is:
TP
precision = (6)
TP + FP
242 H. Li et al.
(3) Recall: Divide the correctly predicted positive samples by all the positive sample
predictions, which is:
TP
recall = (7)
TP + FN
(4) F1 value (F1-score): is the weighted average of precision and recall, which is:
2 × recall × precision
F1 = (8)
recall + precision
1,000 pieces of data of each type in the test sample are intercepted, and there are
a total of 10,000 pieces of data in the test set. The experimental results are shown in
Table 3.
It can be seen from the experimental results that the accuracy of the text classification
task processed by the model in this paper is the highest, up to 94.68%, compared with the
CNN mentioned in the Zhu [13] paper, the accuracy of the attention model has increased
by 2.7%. At the same time, the training time of the latter under the same word vector
Research on Non-pooling Convolutional Text Classification Technology 243
dimension is 1h, and the training time of the model in this paper is 38 min, which has
nearly doubled in speed. Combined with the analysis of the experimental results of the
Tu paper, it can be considered that the pooling layer in the convolutional neural network
structure has a limited effect in processing text classification problems. The training
time of the model with the pooling layer is four times that of the model without the
pooling layer. Furthermore, because pooling is used to reduce dimensionality to extract
features, this feature is similar to the attention layer. The coexistence of both may cause
some relatively important content to be discarded as noise, which affects classification
accuracy. Therefore, it is believed that the convolutional neural network without the
pooling layer can effectively improve the efficiency of model training while ensuring
classification accuracy when dealing with text classification problems.
5 Conclusion
Based on the traditional convolutional neural network model and previous research
results, this paper proposes a reconstruction of the convolutional neural network structure
by deleting the pooling layer and introducing the attention layer after the convolutional
layer. By verifying the Chinese news dataset, the experimental results show that the
model proposed in this paper can improve the accuracy of text classification tasks. It
can be seen that although the convolutional neural network has always been considered
unable to capture the contextual and sequential characteristics of text information, the
attention layer can make up for this shortcoming and achieve similar effects as RNN and
the LSTM. At the same time, it is proved that the pooling layer is unnecessary in the
convolutional neural network model structure. The existence of the pooling layer has
almost no effect on the accuracy of the experimental results and the training time of the
model. The convolutional neural network model without a pooling layer even has some
degree of improvement in efficiency. This proves that the structure of the convolutional
neural network is not fixed. It can be combined with other technologies to change its
internal structure to obtain better performance, providing new directions and ideas for the
further research of neural networks. Through experiments, it can be concluded that the
attention mechanism as a feature extractor can extract features more efficiently than the
pooling operation and can be used in the feature selection stage of the deep learning task.
Therefore, in the following fields of image recognition and natural language processing,
attention mechanisms combined with relevant models can be considered to obtain the
required results more efficiently and accurately.
References
1. Wan, J., Wu, Y.: Review of text classification research based on deep learning. J. TianJin
Univ. Technol. 37(02), 41–47 (2021)
244 H. Li et al.
2. Antonellis, I., Bouras, C., Poulopoulos, V.: Personalized news categorization through scalable
text classification. In: Zhou, X., Li, J., Shen, H.T., Kitsuregawa, M., Zhang, Y. (eds.) APWeb
2006. LNCS, vol. 3841, pp. 391–401. Springer, Heidelberg (2006). https://doi.org/10.1007/
11610113_35
3. Cai, X., Lou, J.: Sentiment analysis of telecom official micro-blog users based on LSTM deep
learning model. Telecommun. Sci. 33(12), 136–141 (2017)
4. Li, D.: The study of Chinese text categorization based on Naive Bayes. Hebei University
(2011)
5. Sun, Y., Li, Y., Bian, Y.: Application of random forest algorithm for book subject classification.
Comput. Technol. Dev. 30(06), 65–70 (2020)
6. Yan, P., Tang, W.: Chinese text clasification based on improved BERT. 33(07), 108–110+112
(2020)
7. Kim, Y.: Convolutional neural networks for sentence classification. In: Proceedings of the
2014 Conference on Empirical Methods in Natural Language Processing, EMNLP, Doha,
Qatar, pp. 1746–1751 (2014)
8. Ruderman, A., Rabinowitz, N.C., Morcos, A.S., et al.: Pooling is neither necessary nor
sufficient for appropriate deformation stability in CNNs. arXiv (2018)
9. Tu, W., Yuan, Z., Yu, K.: Convolutional neural networks without pooling layer for Chinese
word segmentation. Comput. Eng. Appl. 56(02), 120–126 (2020)
10. Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align
and translate. Computer Science (2014)
11. Vaswani, A. Shazeer, N., Parmar, N., et al.: Attention is all your need. In: Advances in Neural
Information Processing Systems, pp. 5998–6008 (2017)
12. Mikolov, T., Corrado, G., Chen, K., et al.: Efficient estimation of word representations in
vector space. In: Proceedings of the International Conference on Learning Representations,
ICLR, Scottsdale, Arizona, USA, pp. 1–12 (2013)
13. Joulin, A., Grave, E., Bojanowski, P., et al.: Bag of tricks for efficient text classification.
In: Proceedings of the 15th Conference of the European Chapter of the Association for
Computational Linguistics: Volume 2, Short Papers (2017)
14. Zhu, M.: Research and implementation of Chinese text classification algorithm based on
machine learning. Beijing University of Posts and Telecommunications (2019)
15. He, K.: Reasearch and application of text classification based on natural language processing.
Nanjing University of Posts and Telecommunications (2020)
16. Wu, M., Wen, L., Sun, M.: Attention Mechanism Image understanding algorithm of ocean
scene. Computer Engineering and Applications, pp. 1–11. http://kns.cnki.net/kcms/detail/11.
2127.TP.20210317.1008.002.html.2021/07/11
17. Zhang, P., Xu, Z., Hu, P.: Segmentation of intracranial hemorrhage fusing dense connection
and attention mechanism. J. Chin. Mini-Micro Comput. Syst. 42(07), 1458–1463 (2021)
18. Liu, T., Zhu, W., Liu, G.: Advances in deep learning based text classification. Electr. Power
ICT 16(03), 1–7 (2018)
Network Security and Blockchain
A New Credit Data Evaluation Scheme Based
on Blockchain and Smart Contract
1 Introduction
The credit data records students’ primary data during college, including course tran-
scripts, exceptional certificates, graduation certificates, rewards, and punishment infor-
mation. The credit data can help students continue their education or seek jobs. However,
in the traditional data storage model, the data is often centrally processed and stored in
the internal platform of educational institutions. The credit data will be lost with the cen-
tral node fails to varying degrees. In addition, attackers can easily steal data by attacking
the central node.
In a worldwide report, falsification of academic data in the job market is rampant,
that millions of forged academic documents and certificates are produced every year
[1]. Then the enterprise suffered many losses with improper employees. The survey by
Career Builder showed that 33% of respondents had forged their certificates [2]. With
the increasing number of advanced talents flooding into society and the limitation of the
verification ability of enterprises or colleges, it will be more difficult for the centralized
management model to deliver accurate credit data. The database is only for internal
staff of educational institutions to view and access, and there is no interoperability for
students [3]. In this paper, combined with these problems, we confirmed that the main
reason is the centralized storage of credit data.
The emerging blockchain technology proposed by Satoshi Nakamoto can effectively
fill this gap [4]. Blockchain technology can provide a trusted platform for exchanging
value with the chained storage structure, the mathematical encryption algorithm, and
the consensus mechanism. The chain structure and encryption algorithms ensure user
privacy and data security, and the consensus mechanism provides a foundation for trust
in a decentralized environment. Blockchain technology has also been applied to varying
degrees in the field of education.
The University of Glasgow has built an intelligent campus system based on Etherum’s
blockchain platform to store students’ academic information such as grades and degrees
[5]. The Maribor University has built EduCTX, higher education credit and score plat-
form based on the open-source Ark blockchain platform, which supports recording
and managing students’ scores [6]. The University of Texas at Dallas used blockchain
technology to manage students’ academic records, including transcripts, digital badges,
certificates, recommendation letters, award records, and other information [7]. The Bit-
Degree platform provided in-service employees with learning opportunities, records
and verifies information such as test scores, homework, and problem-solving skills, and
supports the payment of course fees and application for scholarships with Bitcoin [8].
Despite the above successful cases, the following points are still not considered in the
existing studies: i) the selection of consensus mechanism, and ii) whether the utterly
decentralized model is appropriate. Taking the PoW (Proof-of-Work) consensus mech-
anism adopted by most existing studies as an example, when all the student users on the
chain agree to a branch chain of forged data, it will make the data more accessible to forge
than the traditional model. When 51% of the nodes synchronize the profit-maximizing
account, the entire network synchronously updates the account in PoW. The number of
students is far greater than 51%, and the choice of the branch chain can make their credit
data better under universities cannot repair the blockchain. Moreover, because of the
complexity of the data, PoW requires much more computing power than the traditional
model.
To solve these problems, this paper proposes a credit data storage scheme based
on blockchain, PoA (Proof-of-Authority) consensus mechanism, and smart contract by
combining the centralization and partial decentralization ideas. This program aims to
provide references for the application of blockchain in various sub-fields of education
and promote the actual implementation of the project.
The remaining sections are arranged as follows: Sect. 2 introduces the theoretical
knowledge used in this paper. Section 3 describes the specific process of the whole
scheme. Section 4 is the experimental analysis. Section 5 is the summary of this paper.
A New Credit Data Evaluation Scheme Based on Blockchain 249
2 Technical Background
The blockchain is a new distributed database that integrates multiple technologies, where
the core characteristics are distributed data storage, data tamper-proofing, and data trace-
ability. “Distributed data storage” describes that data is not subject to centralized con-
trol without reliance on third-party intermediaries. “Data tamper-proofing” refers to each
block records a group of tree transaction status information composed of a hash algo-
rithm. The generated hash value is the unique identifier of the block, and the transaction
data of the block can be stored permanently. “Data traceability” means the traceability of
historical block information through the chain storage structure at the current height [9].
2.2 Proof-of-Authority
The consensus mechanism drives peer nodes to maintain the distributed database by
establishing a unified system and rewarding them. The PoA (proof-of-authority) con-
sensus mechanism is a consensual mechanism based on permissions and whose identity
must be publicly verified. The ordinary node is only responsible for initiating trans-
actions but not verifying them. In PoA, nodes can reasonably obtain the bookkeeping
rights instead of relying on computing power to compete for bookkeeping rights. This
can effectively reduce energy consumption and increase the speed of block generation.
The working steps of the PoA consensus mechanism are as follows: i) the founder
of the blockchain sets rules and specifies a set of authorized nodes in the genesis block
according to these rules, ii) when the transaction set reaches a specific number, the set of
authorized nodes digitally signs the transaction set and then broadcasts the transaction
set to the blockchain network. The network randomly marks one of the authorized nodes
in each block as in turn and the rest as no turn. The mining difficulty of the authorized
nodes with the status of “in turn” is lower than that of other nodes. Their packaged
blocks have priority access to the blockchain, and iii) according to the root hash value,
the remaining nodes should calculate whether the address of the authorization node
responsible for the block packing is a member of the authorization set in this round of
mining. If the above steps are correct, then the block is legal, and the full network node
updates the distributed ledger [11].
In addition, the nodes applying to join the authorization set publish broadcast mes-
sages in the blockchain network and are voted by all the authorized nodes. Suppose
250 Y. Wang et al.
the number of authorized nodes that agree to authorize the node exceeds 50%. In that
case, the node will be added to the authorization set, and the authorization set will be
synchronized and updated by the network. The same applies to deleting authorization
nodes.
1) Firstly, identify the initial PoA consensus committee’s member nodes where autho-
rized members are the educational administration node with high-level authority and
the general teacher node. The committee members vote for the student nodes that
meet the preset conditions in the follow-up process.
A New Credit Data Evaluation Scheme Based on Blockchain 251
2) According to the categories of credit data, member nodes with corresponding per-
missions are responsible for the corresponding smart contracts. Such as awards at
the school level can only be published and recorded by the more advanced academic
administration node.
3) With the continuous triggering of the credit event, the committee nodes of PoA
maintain the credit data.
4) The high-level authority node grants identification certificates to students who meet
the graduation conditions.
5) After obtaining the permission node’s permission, the enterprise can access the full
life cycle of students’ credit data in school.
Advanced Permission
Node
Students Node
Students complete
the credit event
Submitting an access
request
Student users
Returns credit data
Enterprises or
universities
Students seek help or access
other students' records
Blockchain network
In the section, we will describe the main smart contract algorithms involved in storing
credit records, certificate management, and external access to certificates.
The address of the advanced permission node is seNode_address, the address list of
the teacher node to be authorized is teaList_address, the function triggered by the event
is authorize_tea, and permission is the permission type of the authorization. Then, as
shown in algorithm 1, we can grant permission to low-privilege nodes through high-
privilege nodes to further ensure the authenticity of the trust data, and any smart contract
issued by unauthorized nodes will be regarded as invalid. If the initiator msg.sender of
the current contract is seNode_address, and all addresses are granted permission, that
is, k == n, event emit(authorize_tea(teaListNode_address)) is triggered.
252 Y. Wang et al.
As shown in Algorithm 2, in the current grade, the credit data entry contract is only
deployed once, where its effect is the duration of the grade. The contract is destroyed
until all credit data recording events of the current grade are triggered and completed.
This avoids the repeated deployment of smart contracts with similar functionality, which
can further reduce the cost of Gas. The address list of the node that has permission is
allRight_address, and allRight_address = teaListNode_address ∪ seNode_address. The
data is defined according to the type of the credit data.
1 Initialization the deployment time of contract y, and the deployment duration k=0.
2 Initialize the total number of credit events v_m that the current node is responsible
for entering, and the number of credit events that have been entered n=0.
3 Initialization the list of the credit data types responsible for the corresponding author-
ized node tealistnode_address.data_name.
4 Establish the mapping rule data_variety(address=>uint) between the permission
node and the credit data categories.
5 Establish the mapping rule data_property(uint=>type(data))between the types of
credit data and their attributes.
6 If (k!=y) Then
7 While (msg.sender==seNode_address) do
8 If (z!=v_m) Then
9 For (i=1 to M)
10 data_variety[seNode_address].data_property[i].data=input_creditData[i];
11 emit(luru(input_creditData[i]));
12 End for
13 End while
14 k=k+1;
15 End if
16 If (k==y) Then
17 selfdestruct(seNode_address) // Destruction of the smart contract
18 End if
Algorithm 2. Credit data entry contract.
A New Credit Data Evaluation Scheme Based on Blockchain 253
As shown in Algorithm 3, if the students meet the preset conditions for each year
during all periods of school, and the students are deemed to meet the graduation require-
ments, then the corresponding high-level authority node will grant certificates and sig-
nature certificates with the unique identifier. If the conditions are met, the smart contract
will be triggered automatically, and the signed certificate will be given. For the infor-
mation with modified records, the advanced permission node will review and decide
whether to sign. The contract is deployed only once for students at the same grade.
1 Initialize the number of qualified student credit data entry contracts n=0.
2 Initialize the modification times m of students’ academic information.
3 Initialize the total number of credit data entry contracts to be completed by students k.
4 Initialize the attributes of graduation conditions stu[address[num]].graduation.
5 Initialize the number of all students in the current grade M.
6 For (i=1 to M)
7 If (k==n&&m==0&&msg.sender==seNode_address) Then
8 stu[address[i]].graduation = true;
9 emit(sign(msg.sender));
10 End if
11 If (k==n&&m!=0&&msg.sender==seNode_address) Then
12 stu[address[i]].graduation = input[i];
13 emit(sign(msg.sender));
14 End if
15 End for
Algorithm 3. Graduation signature contract.
4 Experimental Analysis
Based on the theoretical framework in the above sections, we developed a data man-
agement system and tested its performance to make a comparative analysis between the
proposed scheme and the traditional centralized scheme form the rational analysis. The
experimental environment configuration are Windows 10, 16 GB of memory, and core
I7 CPU. The entire project is accessible at link https://github.com/LitSpirits/credit-data.
In Fig. 2, the process of PoA authorization node combined with the maintenance
of learning information data. When students or teachers apply for login, the event is
triggered and the transaction request is sent to the blockchain network until the data
maintenance is completed and the user enters the system. This means that the user needs
to use their own address registered in the client to perform related operations and trigger
events. In Fig. 2, based on the characteristics of blockchain, untrusted data can be traced
through login records.
As shown in Fig. 3, unauthorized nodes cannot control the credit data in other con-
tracts beyond their authority, and the idea of partial decentralization can ensure the
security of data in a fully decentralized environment. Compared with the traditional
model, which cannot provide effective data proof, the scheme proposed in this paper can
help enterprises to easily obtain students’ accurate and effective credit data and credit
records. Moreover, we show how to use the management of credential data to provide a
reference for other research fields such as certificate anti-counterfeit verification here.
Fig. 3. The permission id of the unauthorized node does not match the stored data in the
blockchain, and the contract time cannot be executed.
In summary, centralized data storage means that the data is changed and read by a
centralized client, and if the hacker attack is not noticed, the changed data state is perma-
nently recorded. With the scheme proposed in this paper, as long as the node obtains the
permission, it can obtain the required data at will and write the data within the permis-
sion. Students only need to obtain permission from others to gain valuable experience. It
should not be ignored that not all students are willing to disclose their proprietary data
thoroughly, and how to provide such a platform is challenging to achieve in a centralized
storage system. Comparison with traditional schemes is shown in Table 1.
We find that consensus mechanisms need to be jointly selected or improved accord-
ing to credibility, transaction throughput, and maintenance cost by selecting different
A New Credit Data Evaluation Scheme Based on Blockchain 255
Table 1. The comparison between the proposed scheme and the centralization scheme.
consensus mechanisms. Credibility is the first criterion for assessing consensus mech-
anisms, but its maintenance costs should also be considered. The data is so complex
that the PoW consensus mechanism will lead to an exponential increase in costs. Nev-
ertheless, it will consume far more resources than traditional schemes. However, in the
PoS (Proof-of-Stack) consensus mechanism, the restriction of maintenance authority
through the amount of currency will lead to difficulty for students to obtain profits. In
the PoA consensus mechanism, the maintenance cost is lower. The teacher node and
student node jointly maintain the network, which can relieve the pressure of teachers
when facing the credit data and motivate students to obtain benefits.
The cost of us scheme is mainly related to the updatable state, the amount of data,
the functions complexity, and the state of the network. The smart contracts inherit from
each other, so the Gas cost is higher than the deployment of a single deployed. However,
this method can control the cost value of Gas below the GasLimit value stipulated by
Ethereum (12 million), which is conducive to the further development of the system in
the later stage. However, the deployment project consumed 7104,672 Gas, and the actual
cost was much lower than the cost of using and maintaining a traditional database. The
cost would be further reduced if educational institutions set their own trading rules. And
we assess trade throughput by making the trade generator randomly generate trades by
changing the frequency. Each time 200–1000 transactions are sent to the network, and
transactions on the blockchain are recorded in the unit. The stable network transaction
throughput is obtained by using the method of averaging, which the range of us program
between 130–190tps. The transaction throughput has yet to be improved for jumbled
credit data.
5 Conclusion
Based on the decentralized blockchain system and part of the decentralized smart contract
permission control mechanism, this paper can ensure the authenticity and traceability
of the credit data and effectively avoid the problems such as data loss in the traditional
credit data storage mode. To promote the practical application of blockchain in education,
we have developed the credit data management system, which can provide a reference
for other sub-fields in education. This can provide a credible academic record for the
256 Y. Wang et al.
business. Therefore, the new data evaluation scheme proposed in this paper has certain
practicability. The performance of the consensus mechanism will determine the cost and
the degree of utilization of decentralization. The formal verification of a smart contract
ensures the anti-attack ability of the contract, which protects the rights and interests of
users. We will study the formal verification of smart contracts and the improvement of
consensus mechanisms in the future.
References
1. Turkanović, M., Hölbl, M., Košič, K., et al.: EduCTX: a blockchain-based higher education
credit platform. IEEE access 6, 5112–5127 (2018)
2. LNCS. https://www.careerbuilder.com/share/aboutus/pressreleasesdetail.aspx?sd=8/7/
2014&id=pr837&ed=12/31/2014. Accessed 15 July 2021
3. Rasool, S., Saleem, A., Iqbal, M., Dagiuklas, T., et al.: Docschain: blockchain-based IoT
solution for verification of degree documents. IEEE Trans. Comput. Soc. Syst. 7(3), 827–837
(2020)
4. Wang, S., Zhang, Y., Zhang, Y.: A blockchain-based framework for data sharing with fine-
grained access control in decentralized storage systems. IEEE Access 6, 38437–38450 (2018)
5. Rooksby, J., Dimitrov, K.: Trustless education? A blockchain system for university grades 1.
Ubiquity J. Pervasive Media 6(1), 83–88 (2019)
6. Xie, J., Tang, H., Huang, T., Yu, F.R., et al.: A survey of blockchain technology applied to
smart cities: research issues and challenges. IEEE Commun. Surv. Tutor. 21(3), 2794–2830
(2019)
7. Kamišalić, A., Turkanović, M., Mrdović, S., Heričko, M.: A preliminary review of blockchain-
based solutions in higher education. In: Uden, L., Liberona, D., Sanchez, G., Rodríguez-
González, S. (eds.) LTEC 2019. CCIS, vol. 1011, pp. 114–124. Springer, Cham (2019).
https://doi.org/10.1007/978-3-030-20798-4_11
8. LNCS. https://ipfs.io/ipfs/QmdEnhQcWHTY4ndotRs1YRpeTdjZGAyisHz8buqfYidkP2.
accessed 15 July 2021
9. Mingxiao, D., Xiaofeng, M., Zhe, Z., et al.: A review on consensus algorithm of blockchain.
In: 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), China,
pp. 2567–2572. IEEE (2018)
10. Leng, K., Bi, Y., Jing, L., et al.: Research on agricultural supply chain system with double
chain architecture based on blockchain technology. Future Gener. Comput. Syst. 86, 641–649
(2018)
11. Salah, K., Rehman, M.H.U., et al.: Blockchain for AI: review and open research challenges.
IEEE Access 7, 10127–10149 (2019)
12. Norta, A.: Creation of smart-contracting collaborations for decentralized autonomous orga-
nizations. In: Matulevičius, R., Dumas, M. (eds.) BIR 2015. LNBIP, vol. 229, pp. 3–17.
Springer, Cham (2015). https://doi.org/10.1007/978-3-319-21915-8_1
An Algorithm of Image Encryption Based
on Bessel Self-feedback Chaotic Neural Network
1 Introduction
Thriving on the increasing rise of big data, networking, and informatization will be
the general trend of the development and progress of human society. The amount of
information and speed of image data transmitted through the network is amazing. To
prevent the image data from being intercepted, stolen, or destroyed by others in the
process of transmission, it is particularly important to propose a secure and efficient
image encryption algorithm.
In the references, Chen-Aihara proved that neural networks with chaotic could
become stable gradually [1] and introduced the chaotic simulated annealing algorithm
by exponentially reducing the connection weight of the self-feedback. It confirmed that
the traversal search of chaotic was related to the self-feedback connection terms. The
self-feedback connection terms with different properties will produce a different chaotic
traversal search. In this paper, the self-feedback connection term of the Chen-Aihara net-
work is improved [2], and the Bessel function is introduced based on linear self-feedback
[3]. Therefore, a chaotic self-feedback neural network based on the Bessel function is
proposed. The inverse bifurcation and Lyapunov exponential evolution diagram of a
chaotic dynamic system are researched, the network model and the corresponding image
encryption algorithm are applied to the image encryption algorithm.
Bessel function chaotic neural network with self-feedback model is described as below:
1
xi (t) = (1)
1 + exp(−yi (t) ε0 )
⎡ ⎤
n
yi (t + 1) = kyi (t) + α ⎣ wij xj (t) + Ii ⎦ − zi (t)g(xi (t) − I0 ) (2)
j=1,j=i
∞
1 cu 2k
J0 (cu) = (−1)k (4)
k!(k + 1) 2
k=0
In the formula, at time t, yi (t) is neuronal internal state and xi (t) is the neuronal
output, k is in the range (0,1) represents the ability of a neuron to retain its internal
state, ε0 is the gradient parameter of the excitation function, g(xi (t) − I0 ) is the network
own self-feedback term, zi (t) is its own feedback connection item of the neuron i, λ
is combination coefficient, c is the expansion coefficient of Bessel function, β is the
annealing parameter.
When g(u) = u, the self-feedback term exhibits a linear state, and above Chaotic neu-
ral network model degenerates into Chen’s model. For this reason, a self-feedback term
based on nonlinear Bessel function is proposed. When g(u) = λJ0 (cu), this nonlinear
term makes the system show a new chaotic dynamic behavior (Fig. 1).
Fig. 1. The curves of Linear self - feedback and Bessel self – feedback.
An Algorithm of Image Encryption Based on Bessel Self-feedback 259
Through the above single-neuron inverted bifurcation diagram and the maximum
Lyapunov exponential evolution diagram, it can be reflected that it’s a chaotic neural
network with transient chaos dynamics behavior, which reflected the chaos searching
ability of the network to some extent. In the meantime, we can change the four parameters
of the dynamic system of a single neuron to display the dynamic behavior of chaotic
systems [4].
When three of the parameters keep a certain value, we can change one parameter to
observe the impact on the SNDS dynamic behavior and draw the corresponding inverted
bifurcation diagram and Lyapunov exponential evolution diagram (Figs. 3, 4, 5 and 6).
260 Y. Xu et al.
Fig. 3. Bifurcation and the Lyapunov exponents of maximal index diagram of single parameter
x versus z for I0 = 0.6, ε0 = 0.004, k = 0.2.
Fig. 4. Bifurcation and the Lyapunov exponents of maximal index diagram of single parameter
x versus k for I0 = 0.6, ε0 = 0.004, z(1) = −0.8.
Fig. 5. Bifurcation and the Lyapunov exponents of maximal index diagram of single parameter
x versus ε0 for I0 = 0.6, k = 0.2, z(1) = −0.8.
Fig. 6. Bifurcation and the Lyapunov exponents of maximal index diagram of single parameter
x versus I0 for k = 0.2, ε0 = 0.004, z(1) = −0.8.
We gain the value of the parameter xi (t) after do an intermediate calculation to the
value of xi (t). In this paper, n is set to a fixed value of 18.
The binary number used for the test is generated by the parameter’s value xi (t) in the
formula (12). For each value of xi (t), the former 10 decimal digits was diacarded and
compare the result with 0.5, the process is shown as Eqs. (14) and (15):
0, 0 ≤ mi ≤ 0.5
ni = (15)
1, 0.5 ≤ mi ≤ 1
In the test, there are 15 subsets should be considered, and each subset will out-
put a p-value. If p-value is better than 0.01, the sequences is considered to be uniformly
distributed. We tested 100 sets of 1,000,000 bits binary sequences, in this process, param-
eters of the initial value are set a fixed value: ε0 = 0.004, I0 = 0.6, z(1) = −0.8, y(1) =
0.8, k = 0.2. Since the p-value is better than 0.01 for each of fifteen subsets, accept the
sequence as random [8] (Table 1).
262 Y. Xu et al.
Step 1: The original grayscale image is read as an M × N matrix A (M, N), where each
element of matrix A corresponds to the pixel value at the image point (M, N).
Step 2: The chaotic sequence obtained by the iteration of the ESNDS is encrypted,
and the four parameters of the Bessel self-feedback chaotic neuron are used as the
security key of the chaotic sequence. We can iterate the chaotic neural network (M × N
+ M + N + H) times, then discard the first H data, and finally, get a length (M × N +
M + N) of the new sequence.
Step 3: Make the former M elements sequence a, the nest N elements are sequence
b, and the rest are sequence c. The following formula (16) is calculated for the two
sequences:
a = floor(a × M ) + 1
(16)
b = floor(b × M ) + 1
Step 4: Take the calculated sequences a and b to perform row and column
permutation for the original matrix A (M, N). Specific permutation methods in [4].
Step 5: Convert the transpose matrix into a one-dimensional matrix P, and sort the
matrix P in the order of matrix c to obtain the matrix P prime.
An Algorithm of Image Encryption Based on Bessel Self-feedback 263
Step 6: By the Eqs. (17) and (18), we can obtain the diffused matrix Q from the
matrix P and the sequence c. Finally, convert Q into an encrypted image of size M × N.
Q = P ⊕ c (18)
The decryption is a reversion process of encryption. Using initial the correct key
reversal encryption process can recover the original image.
Fig. 7. Result of Simulation. (a) (b) the original images; (c) (d) the encrypted images; (e) (f) the
decrypted images.
P(xi ) represents the proportion appearance for grayscale symbol xi , and n represents
the grayscale of the image. In theory, the value of information entropy should be closed
to 8. As Table 2 shows, the encryption can change information from order to disorder.
In the image, gray histogram analysis reflects the occurrence times of each gray value
[10], and the distribution is used to reflect the image pixel values. In the Fig. 8, pixel
value of encryption image distributes uniformity that can use to resist statistical attack.
Correlation analysis is to eliminate the impact of accidental factors through the obser-
vation of amount digital data. Correlation analysis reflects the related degree between
inter-adjacent pixels of an image through the correlation coefficient. The correlation
coefficient is closer to 1, and the correlation will be greater between two adjacent pixels.
Reducing the related degree can reduce the possibility of statistical image attack [11].
Figure 9 shows the correlation of the original and encrypted image in the horizontal,
vertical, and diagonal directions calculated by the below formulas:
1
N
E(x) = (xi ) (20)
N
i=1
An Algorithm of Image Encryption Based on Bessel Self-feedback 265
1
N
D(x) = (xi − E(xi ))2 (21)
N
i=1
1
N
cov(x, y) = (xi − E(xi ))(yi − E(yi )) (22)
N
i=1
cov(x, y)
rxy = √ (23)
D(x) D(y)
Fig. 9. Correlation analysis images in original and encrypted. (a) (b) (c) the original image; (d)
(e) (f) the encrypted image.
The number of bits usually counts encryption keys. The longer the number of bits,
the bigger the keyspace and the more powerful the resistance to exhaustion, clearly
illustrating the security and efficiency of this encryption algorithm (Table 3).
266 Y. Xu et al.
In this paper, an initial value and four parameters are selected as the key in the encryp-
tion algorithm. Other parameters of the Bessel self-feedback chaotic neural network can
also be used to expand the keyspace enough to meet the needs of encryption.
Key sensitivity refers to the key sequence generator or iterative function keys gen-
erated accordingly [12] in the encryption process when the initial key occurred small
changes. It can’t give a correct decryption process after changing the entire encrypted
image dramatically, to ensure the security of the algorithm.
4 Conclusion
Bessel function in this paper as the feedback is introduced into the chaotic neural net-
work, the network model for different parameters reflected different dynamics behavior.
Based on Bessel self-feedback chaotic neural network, an image encryption algorithm
is proposed. Through the experimental analysis from information entropy, histogram,
keyspace, and correlation, the effectiveness and security of the encryption algorithm are
verified. It is worth noting that the single-neuron chaotic dynamical system has a large
development space, and this applicability can make the system participate in more fields.
Acknowledgement. This work was supported by the Nature Science Foundation of Heilongjiang
Province (LH2021F035).
References
1. Chen, L., Aihara, K.: Chaos and asymptotical stability in discrete-time neural networks.
Neural Netw. 13, 731–744 (2002)
2. Xu, Y., He, S.: Chaotic neural network with self-feedback of trigonometric function and its
application. In: Proceedings of the 2009 China Intelligent Automation Conference (2009)
3. Ye, Y., Xu, Y.: A chaotic neural network with bessel function. J. Harbin Univ. Commer. (Nat.
Sci.) 29, 561–565 (2013)
4. Xu, X., Chen, S.: Single neuronal dynamical system in self-feedbacked hopfield networks
and its application in image encryption. Entropy 23, 456 (2021)
5. Abd El-Latif, A.A., Niu, X.: A hybrid chaotic system and cyclic elliptic curve for image
encryption. AEU-Int. J. Electron. 67, 136–143 (2013)
6. Moysis, L., Volos, C.: Modification of the logistic map using fuzzy numbers with application
to pseudorandom number generation and image encryption. Entropy 22, 474 (2020)
7. Meranza-Castillón, M., Murillo-Escobar, M., López-Gutiérrez, R., Cruz-Hernández, C.:
Pseudo random number generator based on enhanced HÉNon map and its implementation.
AEU-Int. J. Electron. Commun. 107, 239–251 (2019)
8. Bassham, L.E., III, Rukhin, A.L., Soto, J., Nechvatal, J.R.: Sp 800-22 Rev. 1a. a Statistical Test
Suite for Random and Pseudorandom Number Generators for Cryptographic Applications,
National Institute of Standards & Technology, Gaithersburg, MD, USA (2010)
9. Wu, Y., Zhou, Y., Saveriades, G.: Local shannon entropy measure with statistical tests for
image randomness. Inf. Sci. 222, 323–342 (2013)
10. Jeffrey, J.R., Christopher, C.Y.: High-resolution histogram modification of color images.
Graph. Models Image Process. 57(5), 432 (1995)
11. Behnia, S., Akhshani, A., Mahmodi, H., Akhavan, A.: A novel algorithm for image encryption
based on mixture of chaotic maps. Chaos Solitons Fractals 35, 408–419 (2018)
12. Zhou, Y., Li, S.J.: BP neural network modeling with sensitivity analysis on monotonicity
based spearman coefficient. Chemom. Intell. Lab. Syst. 200, 103977 (2020)
Evaluation of the Effect of Blockchain
Technology Application in the International
Trade
Abstract. Blockchain technology is widely used in many fields due to its non-
tamper resistance, openness, transparency, and decentralization characteristics.
The application of blockchain technology in international trade mainly includes
five aspects: product traceability, trade payment, customs procedures, trade financ-
ing, and trade supply chain. Evaluation of the effect of blockchain technology
application in the above aspects by a method of analytic hierarchy process (AHP)
shows that the effect on trade payment and trade supply chain is better than others.
At the same time, it has a less obvious effect on product traceability. Currently,
there are still some barriers to applying blockchain technology in the field of
international trade. Blockchain technology still has certain technical risks and
lacks legal supervision. Better application of blockchain technology can be real-
ized by cultivating high-quality blockchain technical talents and improving the
supervision mechanism of blockchain technology.
1 Introduction
Blockchain is a distributed ledger with non-tamper resistance, openness, transparency,
and decentralization characteristics, making it widely used in many fields. Blockchain
technology also provides a new impetus for the development of international trade. It
helps it to solve some long-standing problems, such as information asymmetry, high
trade communication cost, and high transaction cost [1]. Based on the application of
blockchain technology in international trade, this paper uses the analytic hierarchy pro-
cess (AHP) to evaluate the application effects. Accordingly, it puts forward targeted
policy suggestions to promote the high-quality development of international trade.
storage, sales, and transportation. In case of product quality and safety problems, the
source of the product can be tracked and recalled in time to minimize the loss. Prod-
uct traceability can effectively prevent product fraud, guarantee food safety, improve
product competitiveness in the international market, and promote commodity circula-
tion and export trade [2]. With the improvement of living standards, people have higher
requirements for product quality and pay more attention to environmental protection.
Product traceability can make consumers feel safe to buy products. Therefore, product
traceability plays an important role in international trade, especially as the COVID-19
epidemic situation still exists in most countries in the world. Product traceability can
greatly contribute to the prevention and control of the epidemic and effectively prevent
major public health events [3].
Applying blockchain technology to product traceability in international trade can
solve the problem of product fraud in origin. Traditional anti-counterfeiting technology
generally uses TWO-DIMENSIONAL code or bar code to record information, prevents
product fraud by adding coating or printing in the bottle cap. But it cannot solve the
problem of counterfeiting by copying and transferring anti-counterfeiting logos. The
blockchain has the characteristics of decentralization and publicly records all informa-
tion on the account book without management by a third-party organization. All the data
recorded on the blockchain are timestamped and cannot be tampered with. The unique-
ness of a commodity can be ensured after every link from raw material purchase to sales
of a commodity is recorded on the blockchain, and the information of fake commodi-
ties cannot enter the blockchain system [4]. At the same time, blockchain technology
can realize the real-time traceability of commodity information. The information in the
blockchain is formed by packaging the data into blocks, adding timestamps, and finally
forming a chain. Therefore, a commodity’s whole production process can be recorded
in chronological order, and the information on the blockchain will also be arranged in
chronological order. It can be traced in real-time [5]. In product traceability, international
standards such as ISO9000 and ISO14000 can be combined with blockchain technol-
ogy. Blockchain technology records all the information of products on the block. This
information is authentic and cannot be tampered with, promoting export manufacturers
to produce and export products according to international standards and meet the quality
and environmental protection requirements of export products.
Applying blockchain technology to international trade payment can make up for the
shortcomings of traditional international trade payment and promote the high-quality
development of international trade. Firstly, the application of blockchain technology
can reduce the risks of international trade payment. When using blockchain technology
for international trade payment, the ledger is updated through the consensus mechanism,
and the information of each transaction is open and transparent. The transaction data can-
not be tampered with [7]. At the same time, the cryptographic technology of blockchain
can also ensure the security and authenticity of transaction data and reduce the possibility
of illegal operations by both parties. Secondly, the application of blockchain technol-
ogy can improve the efficiency of transactions. Traditional international trade payment
procedures are cumbersome and time-consuming. However, blockchain technology can
realize the information sharing between the buyer and the seller. Their transaction infor-
mation will be published in the payment system, and their information exchange could
be realized on the platform without the participation of a third party [8].
Meanwhile, the trading platform based on blockchain technology can realize the
direct transaction between the two parties by simplifying the transaction process in sec-
onds or hours. Thirdly, the application of blockchain technology can reduce transaction
costs. Trading platforms based on blockchain technology do not rely on traditional inter-
mediaries to provide credit certificates, which can reduce the related commission fees,
be more convenient to communicate, and reduce the cost of communication.
Customs procedure refers to the procedure for imported goods to pass through the cus-
toms, which usually includes four steps: customs clearance, classification, valuation,
and taxation. In dealing with the documents and records generated in international trade
business, the customs need to invest a lot of money, and the work efficiency is not high
enough. At the same time, because these documents come from different participants,
there is information asymmetry, and fraud may occur.
Using blockchain technology can effectively solve the problems of high cost and low
efficiency in current customs procedures. Firstly, blockchain technology packages the
relevant data involved in the customs procedures into blocks and finally into chains. The
timestamp technology can ensure the authenticity and security of the relevant data in the
transmission and will not be tampered with. It can also help the Customs Department
confirm the goods’ destination and source more accurately and effectively track down
and provide evidence for the goods that may have problems [9]. Secondly, the imple-
mentation of smart contracts can realize the digitalization of trade contracts and bills,
effectively solve file damage and loss problems, reduce labor and information transmis-
sion costs, and simplify the work process [10]. Finally, the decentralized data sharing
model can enhance trust between enterprises and the customs supervision department,
help suppliers realize real-time monitoring of goods, and the Customs Department com-
plete all work with high quality and efficiency. Applying smart contract technology to
customs tariff management can avoid tax evasion to a certain extent, reduce the burden
of Customs departments, and improve the work efficiency of Customs departments.
270 J. Zhang and Y. Fan
3.1 Build an Index System for Evaluating the Effect of Blockchain Technology
Application in the International Trade
According to the construction requirements of the AHP index system, this paper sets
up five primary indicators, including product traceability, trade payment, customs pro-
cedures, trade financing, and trade supply chain, to evaluate the effect of blockchain
technology in international trade. Product traceability includes three secondary indexes:
anti-counterfeiting ability, real-time traceability, and consumer trust. Trade payment
includes three secondary indexes: payment security, payment efficiency, and cost. The
customs procedures include three secondary indexes: time to pass the customs, probabil-
ity of fraud and handling fee. Trade financing includes three secondary indexes: length
of audit, human capital costs, financing expenses. The trade supply chain includes three
secondary indexes: information privacy, data transparency and information sharing. The
above index system is shown in Fig. 1.
Fig. 1. The evaluation system of the effect of blockchain technology application in international
trade.
According to the proportion of each aspect, trade payment accounts for 32.25%, trade
supply chain 31.34%, trade financing 19.03%, customs procedures 12.89%, and product
traceability 4.48%. This shows the effect on trade payment and trade supply chain are
better than others, while it has a less obvious effect on product traceability.
Trade payment is an important component of international trade. Based on a char-
acteristic of strong confidentiality, blockchain technology provides a more secure and
efficient environment and makes up for the defects of traditional international trade
payment. Therefore, trade payment is a key field of all applications of blockchain tech-
nology in international trade and should fully combine with blockchain technology to
innovate the mode of payment. There is still more development space for applying
blockchain technology in the trade supply chain, trade financing, and customs proce-
dures. Blockchain technology has the characteristics of openness, transparency, and
decentralization. However, current applications do not take full advantage of these mer-
its yet, and only solve partial problems existing in the trade supply chain, trade financing,
and customs procedures. Therefore, these three aspects should deepen integration with
blockchain technology and make full use of its characteristics. Blockchain technology
has a great space and a certain advantage of application in product traceability. However,
at present, it is not widely used in product traceability. Manufacturers have an insuffi-
cient understanding of the importance of product traceability. They should strengthen
their understanding of product traceability and further promote it.
5 Conclusion
As blockchain technology emerges, many industries and fields set off a technological
revolution and bring about technological changes. Due to the limited application of
blockchain technology in international trade, this paper only researches the application of
blockchain technology in trade payment, customs procedures, product traceability, trade
financing, and trade supply chain. Because blockchain technology is gradually improved,
its application in international trade needs to be further explored. At present, there are
still some problems in applying blockchain technology in the field of international trade,
such as technical risks and legal application of blockchain, which look forward to further
study by scholars.
Subject Source
National Social Science Foundation of China (18BJL094).
References
1. Chen, W.G., Yuan, J.: Integration of blockchain technology into global economic governance:
paradigm innovation and regulatory challenges. Tianjin Soc. Sci. 06, 91–99 (2020)
2. Akash, T., Arun, S., Richa, K., Anand, N., Sudeep, T., Neeraj, K.: Blockchain-based efficient
communication for food supply chain industry: transparency and traceability analysis for
sustainable business. Int. J. Commun. Syst. 34(4), e4696.1–e4696.20 (2021)
3. Sheng, S.Y.: Research on construction of supply Chain information resource sharing model
based on blockchain technology. Inf. Sci. 39(07), 162–168 (2021)
4. Yanling, C., Eleftherios, I., Weidong, S.: Blockchain in global supply chains and cross border
trade: a critical synthesis of the state-of-the-art, challenges and opportunities. Int. J. Prod.
Res. 58(7a8), 2082–2099 (2020)
5. Lin, X., Chang, S.C., Chou, T.H., Chen, S.C., Ruangkanjanases, A.: Consumers’ intention to
adopt blockchain food traceability technology towards organic food products. Int. J. Environ.
Res. Public Health 18(3), 912 (2021)
6. Wang, X.D., Sun, Z.Y.: Discussion on using blockchain technology to carry out international
settlement. Foreign Econ. Trade Pract. 07, 63–65 (2017)
7. Liang, X., Zhang, H.: Analysis on the application mode of blockchain technology in
international payment. Friends Account. 02, 155–160 (2021)
8. Zhao, Z.K.: Blockchain creates a new cross-border payment model for international trade.
Enterp. Econ. 36(09), 163–168 (2017)
9. Hong, T., Cheng, L.: Application of blockchain in global trade and finance. Int. Trade 10,
63–67 (2018)
10. A decentralized and secure blockchain platform for open fair data trading. Concurr. Pract.
Exp. 32(7), e5578.1–e5578.11 (2020)
11. Li, J.J., Wang, Z.W.: Application mode, risks and challenges and policy suggestions of supply
chain finance based on blockchain technology. New Financ. 01, 48–55 (2021)
12. Jia, X.L.: Research on supply chain management under the development of blockchain and
big data. Bus. Econ. Res. 10, 49–51 (2020)
13. Zhang, Z.P., Ma, Y.G.: The development of blockchain + supply chain finance in China:
models, challenges and countermeasures. Res. Financ. Dev. 08, 48–54 (2020)
Fraud Network Identification Model
for Insurance Industry
Jiaqiu Wang1(B) , Yining Jin2 , Zaiyu Jiang1 , Xueyong Hu1 , and Peng Wang1
1 Beijing CHINA-POWER INFORMATION TECHNOLOGY Co., Ltd., Haidian District,
Beijing, China
wangjiaqiu@alu.hit.edu.cn
2 Harbin University of Commerce, Songbei District, Harbin, China
1 Introduction
In recent years, with the rapid development of the economy, diversified development
fields have emerged one after another. The insurance field has been recognized and
received wide popularity to reduce property losses—for example, medical insurance,
vehicle insurance, life insurance, financial insurance, and so on. However, because the
anti-fraud work in China started relatively late, the anti-fraud system in all aspects is not
perfect, insurance fraud occurs from time to time.
To reduce the economic losses caused by insurance fraud, some insurance compa-
nies have developed anti-fraud platforms to detect fraud. For example, China Insurance
Information Technology Management Co., Ltd. developed a vehicle insurance infor-
mation platform to detect insurance fraud. However, the existing insurance anti-fraud
platform cannot actively provide effective information and still rely on manual checking.
This situation leads to the low efficiency of anti-fraud work, and the practical applica-
tion cannot achieve good results. At the same time, with the individual fraud gradually
identified, insurance fraud transforms into gang fraud. This makes it difficult for some
platforms to analyze fraudulent conspiracies that could cause huge losses effectively.
Considering that auto insurance fraud presents the trend of syndication, the detection
and identification of fraud gangs play an important role in the healthy development of
auto insurance.
Therefore, to cope with the gang fraud of vehicle insurance, this paper processes
the data of the vehicle insurance industry, merges them into the attribute graph database
Neo4j, and aggregates two hundred thousand nodes into four fraud network models
through Cypher language to assist insurance personnel in identifying the endless gang
fraud.
This article is organized as follows. The related work is described in Sect. 2. Section 3
introduces the background knowledge of insurance fraud identification. Section 4
describes the proposed model design in detail. Section 5 introduces the case and shows
the corresponding knowledge graph structure. Section 6 summarizes the paper and looks
forward to future work.
2 Related Work
This section introduces the work related to insurance fraud. At the same time, this section
introduces the relevant research of insurance fraud of vehicle insurance.
With the rapid development of the insurance industry, Insurance fraud is a growing
problem in different areas., but the detection technology of insurance fraud is relatively
backward. Therefore, there is an urgent need to identify insurance fraud-related technol-
ogy to improve the environment in the insurance field. For example, Li et al. [1] analyzed
the three new features of insurance fraud in detail from the three components: the fraud-
ulent behavior of insurance fraud, the subjective fault of the perpetrators, and the causal
relationship between the fraudulent conduct and insurance compensation. Xu et al. [2]
summarized the main risk types of large-sum insurance based on large-sum insurance
fraud. At the same time, based on the Apriori algorithm, a big data intelligent anti-fraud
system model for large insurance is proposed, focusing on the analysis of statistical
identification, insurance fraud law mining, insurance fraud behavior identification.
Graph database technology can also be effective in identifying mob-style insurance
fraud related to critical illness insurance. For example, Zhou and Huang et al. [3] use the
graph database method to analyze the compensation information of a large life insurance
company in the past ten years and explore and verify the feasibility of the graph database
technology in the identification of serious disease insurance gang fraud. Wu et al. [4] use
Logistic regression and K-means clustering analysis to identify claims risk from the two
aspects of category characteristics and individual characteristics and find fraud factors.
Nui et al. [5] constructed a medical insurance anti-fraud warning model of medical credit
rating from the perspective of fraud through the health level, income level, and other
specific data of medical reimbursement subjects.
278 J. Wang et al.
In the era of big data, the entities involved in insurance fraud are increasingly com-
plicated. For example, Car insurance fraud from a single personnel’s accident, sudden
change to organized and planning criminal gang. Insurance companies often lack evi-
dence, the exploration of neglect or malicious complaints brought under the industry
regulatory pressure and the risk of litigation, the unreasonable compensation to the
deceiver, this brings the massive economic losses to the insurance company [6]. In this
case, a lot of scholars began to study vehicle insurance fraud. Joint Task Force on vehicle
Anti-fraud [7] analyzes the new trends and characteristics of auto insurance fraud, from
the insurance company management, information sharing, and the external environment,
laws, and regulations, and many other aspects analyze the problem. It also puts forward
targeted suggestions to build a multi-party cooperative prevention and control system
with big data as the core.
In addition, Macedo et al. [8] conducted on-site semi-structured interviews with a
convenience sample of 20 auto repair workshops in Portugal to explore the role of vehicle
repair workshops in auto insurance fraud and how they might contribute to the reduction
or increase of such crimes. The research found that auto repair shop, as one of the main
actors participating in insurance fraud, has the potency to prevent these crimes. Yang [9]
designs and implements a graph-based visual analysis system for auto insurance fraud,
which mainly analyzes fraudulent behavior from three perspectives: element analysis,
suspicious target, and fraud gang, and provides decision support for professionals. Yu
and Feng et al. [10] study the anti-fraud detection method of auto insurance. They apply
micro-gang modeling to Motor vehicle insurance fraud detection and adopt matrix-based
similarity calculation, rank ranking, and transformation algorithm to identify gangs with
minimal probability but highly suspicious. The experimental results show that the method
is more accurate and efficient than the traditional method.
To sum up, no matter in the field of financial insurance, medical insurance, or vehicle
insurance, insurance fraud is affected by a variety of subjects and factors. In order to
better identify insurance fraud, this paper takes vehicle insurance fraud as the research
object and establishes a network identification model based on vehicle insurance fraud.
The model consists of four parts: abnormal human-vehicle relationship group, vehicle
crossover group, driver ring relationship group, and vehicle recurrence group, which
fully analyzes the different relationships between people and vehicles, and analyzes and
identifies vehicle insurance fraud according to different situations.
Insurance fraud seriously infringes on the legitimate rights and interests of insurance
consumers, destroys the normal order of the insurance market, and has a negative impact
on many aspects of society. It is mainly introduced from the following three aspects.
Insurance fraud results in losses to insurance companies. The insured fraud claims would
be resulting in the insurance company spending unnecessary additional funds, which will
bring financial losses to the insurance company. With the increase of the cases of the
insured defrauding the insurance company, insurance fraud cases will inevitably bring
loss of reputation to the insurance company.
Harm spillover poses a threat to the whole society. The increase of insurance fraud will
corrupt social morals and disturb the stability and development of the national economy.
Other honest policyholders suffered losses. Insurance fraud will lead to an increase in
the insurance rate, the degree of protection of honest policy-holders will be weakened,
and the insurance interests will be damaged.
4 Model Designment
This section introduces the proposed fraud network identification model for the insurance
industry and the methods to identify fraud based on the model. This section contains
280 J. Wang et al.
six parts, which are the introduction of knowledge graph, insurance entity relationship
based on knowledge graph, model design, data introduction, data preprocessing and
node relationship import, data modeling and group generation.
Knowledge graph is proposed by Google on May 17, 2012. Its original intention is to
improve the capability of the search engine, improve the user’s search quality and search
experience [13]. The current artificial intelligence technology can be simply divided into
perceptual intelligence (image, video, speech, text recognition, etc.) and cognitive intel-
ligence (knowledge reasoning, causal analysis, etc.). Knowledge mapping technology
is regarded as the core technology in the field of cognitive intelligence and an impor-
tant part of artificial intelligence. Its powerful semantic processing and interconnection
organization ability provide the basis for intelligent information application [14].
As one of the important branches of artificial intelligence technology, knowledge
graph technology can display the relationship between entities and the attribute informa-
tion of entities and associations in a visual way [15]. Moreover, the knowledge graph can
discover new relational information, new unstructured data modes and new knowledge
faster and more accurately, achieving insight customer and reducing business transaction
risks.
Report a
People hurt accident The case crime Informant
Car
own
accident accident owner
crash
policy-
Three car Mark car insure insure
holder
Insurance
policy
insurabl
The
insured
survey e benefit insured
survey repair
repair
drive
The
driver
Survey
member The garage
loss assessment
The person
Verify the
Part fee damage who verified
the damage
Data Introduction. In this model, the insurance policy information, report information,
and repair shop information of a certain insurance company are used as the experimental
data, and different persons are combined with the certificate number. Different cars
are combined with the license plate number. After the data is imported into the graph
database, 29 kinds of nodes (people, cars, cases, etc.) and 21 kinds of relationships
(driving, insurance, reporting, etc.) are generated. Each node and relationship has its
attributes. Because natural persons may have multiple labels, we set up a specific entity
classification, as shown in Fig. 2. To protect the privacy of users, the process of data
modeling in this experiment adopts the method of encryption and desensitization.
282 J. Wang et al.
Data Preprocessing and Node Relationship Import. Since this data is based on the
policy and report as the main key, it is necessary to merge the personnel information
according to the certificate number of the driver (subject, three, insured, insured), internal
personnel (survey, damage, damage) and other data. Finally, import it into the Neo4j
graph database.
Data Modeling and Group Generation. In Neo4j, the four fraud network models in
a stage of model design are constructed by Cypher. The claims adjuster can determine
whether the case is a fraud by using the model and the unfold node data.
The Implementation Process. With Neo4j and Cypher language, a program can be
designed to automatically detect all potentially abnormal vehicle insurance behaviors.
Implementing Apoc plug-in based on Neo4j and developing stored procedures through
Java-Neo4J. (1) All nodes of the fraud model are queried through Cypher, node rela-
tionship IDS in each subgraph are aggregated into a new group node through traversal,
and then are marked with the identification of the model and are stored in neo4j. (2)
The Cypher of this stored procedure can be automatically detected by executing the
scheduled task. (3) Storing the aggregated nodes into the node relation ID of the model,
and setting the model type identifier. You can expand the model through the group node
or summarize the classified groups. Figure 3 shows the stored procedure that queries,
aggregates, and stores the fraud model into Neo4j.
A fraud network model is a designed group according to the abnormal occurrence of
nodes and relationships, which discusses suspected fraud groups. Insurance personnel
is required to further verify group status, reducing most of the screening workload. In
addition, the fraud network model may give false-positive results.
Fraud Network Identification Model for Insurance Industry 283
A fraud network
model that queries a
model using Cypher
Is there a subgraph
Yes
The node ID, relationship ID, and
attribute value of the subgraph are
found
Check whether
Generate a new
the groupld exists on group ID
the node No
Yes
Yes
Combined group
End
Fig. 3. The fraud model queries, aggregates, and stores stored procedures in Neo4J.
284 J. Wang et al.
5 Cases
This section introduces the case of using the fraud network identification model and the
corresponding knowledge graph, which consists of four parts: abnormal human-vehicle
relationship group, vehicle crossover group, driver ring relationship group, and the target
three-vehicle recurrence group.
The driver loop relationship group is an abnormal group of people - people. The driver
(the driver of the target, three car drivers) is out of danger together, connecting into a ring.
For example, Fig. 6 shows possible gang fraud. In the picture, Mr. Tan and Mr. Liu are
doubt as gang frauds. Insurance personnel can expand the nodes of the knowledge graph
or through its associated cases, policies, and related institutions to determine whether to
be cheated for insurance.
The three vehicles in the target appear in repeated groups are the same car and all three
cars in more than two cases, so this is a car-to-car model. Figure 7 shows a sample of
this model. The blue target car and the yellow three cars appear at a time in multiple
cases, and the claims adjuster can further track the two vehicles to protect the rights and
interests of the company.
Aiming at fraud network detection in the insurance industry, this paper uses graph
database tool and knowledge fusion, merging entity methods to study related issues. The
detection processes can identify the group fraud through the relationship between the
entities. The case presented in this paper is the data of the actual insurance company,
including the insurance policy information, report information and repair shop infor-
mation of the insurance company. Because the data of people, cars, repair shops and
insurance policies are sensitive. Therefore, the data shown in this paper are encrypted,
but it does not affect the identification of the actual auto insurance gang fraud.
In future work, in terms of model improvement, identifying models that have identi-
fied fraud aims to reduce the cost and time of anti-fraud. As for the application function
method, based on the Apoc plug-in of Neo4j, the stored procedures and groups searched
by Cypher would be aggregated into a new node, and attributes such as the amount and
associated cases are added to the new group node. Then the group node development
can realize more customized visualization.
Acknowledgment. The authors thank the editor and the anonymous reviewer for their helpful
comments and suggestions, which improved this paper.
References
1. Li, Y.Q., Qiao, S.: New characteristics and risk prevention path of insurance fraud under the
background of big finance. Insur. Res. (04), 121–127 (2021)
2. Xu, Q.M., Zhang, M.R.: Design of intelligent anti-fraud system based on big data of large
insurance. Comput. Age (07), 117–120 (2021)
3. Zhou, X.N., Huang, L., Wang, F.Y., Chu, M., Huang, T.: Research on the application of graph
database in the identification of serious illness insurance gang fraud. Insur. Res. (09), 92–104
(2020)
4. Wu, J.T., Zhang, Y.L.: Research on anti-fraud identification model of insurance. Natl. Circ.
Econ. (26), 152–154 (2020)
5. Nui, X.F.: Research on the design of anti-fraud warning mechanism in medical insurance. J.
Account. (18), 104–108 (2020). (in Chinese)
6. Ji, W., Cheng, G.: Application of traffic accident trace identification in anti-fraud of auto
insurance. Law Soc. (17), 77–78 (2021)
7. Auto Insurance Anti-fraud Joint Research Group: A study on fraud and anti-fraud in auto
insurance. Insur. Res. (06), 3–10 (2021)
8. Macedo, A.M., Viana, C.C., Marques, N.J.S., da Costa Brás da Cunha C.A.: Car insurance
fraud: the role of vehicle repair workshops. Int. J. Law Crime Justice 65, 65–75 (2021)
9. Yang, S.: Design and Implementation of Visual Analysis System of Auto Insurance Fraud
Based on Graph. Beijing University of Posts and Telecommunications (2019)
10. Yu, W., Feng, G.F., Zhang, W.J.: Research on fraud detection system and gang identification
of motor vehicle insurance. Insur. Res. (02), 63–73 (2017)
11. http://xw.sinoins.com/2019-06/25/content_295813.htm
12. Stijn, V., Guido, D.: Insurance fraud: issues and challenges. Geneva Pap. Risk Insur. Issues
Pract. 29(2), 313–333 (2004). https://doi.org/10.1111/j.1468-0440.2004.00290.x
13. Liu, Q., Li, Y., Duan, H., Liu, Y., Qin, Z.G.: A review of knowledge graph construction. J.
Comput. Res. Dev. 53(03), 582–600 (2016)
14. Liu, J., Li, Y., Duan, H., Liu, Y., Qin, Z.G.: Knowledge graph construction: a review. J.
Comput. Res. Dev. 53(03), 582–600 (2016)
15. Du, R.S., Chen, H.Y.: Building knowledge graph to realize knowledge fusion: a review of
knowledge graph. J. Shanxi Univ. Finance Econ. 43(05), 127 (2021)
Global Supply Chain Information
Compensation Model Based on Free Trade Port
Blockchain Information Platform
Abstract. In the research on the construction of the global supply chain of the
Hainan Free Trade Port (FTP), it is found that sharing supply chain information is
an important issue to improve the performance of the global supply chain built on
the FTP. With the help of blockchain technology, solutions to information sharing
problems can be proposed. During the construction of the FTP, it is necessary
to build a Free Trade Port Blockchain Information Platform (FTPBIP), in which
the information compensation service can improve the effectiveness of supply
chain information sharing. By discussing the operation mechanism of information
compensation in the Free Trade Port Supply Chain (FTPSC) based on blockchain,
a set of information compensation models is constructed to provide a theoretical
basis for further improving the construction of the FTPBIP.
1 Introduction
As the process of economic globalization continues to accelerate, more and more com-
panies have begun to rely on FTP to participate in the construction of global supply chain
networks. Many studies have shown that the construction of my country’s FTP has an
important impact on coordinating the global supply chain network. The FTP informa-
tion service plays a particularly prominent role in the coordination of the global supply
chain. However, the information service function of the FTP is still in the exploratory
stage. Therefore, this subject considers the use of blockchain technology to cut in from
the perspective of information sharing, trying to reveal the information compensation
operation mechanism of the FTP for the innovation and coordination of the global supply
chain.
This research will take the free trade zone with Chinese characteristics as the node
of the supply chain network as the research object. Based on the perspective of the inno-
vation of information compensation mechanism in my country’s free trade zone and the
expansion of global supply chain coordination methods under the blockchain technol-
ogy, combined with the investigation and analysis of specific case areas, With the help
of regional economics, management and system dynamics theories and methods, based
The origin of the research on information sharing can be traced back to Srinivasan [1]
in the just-in-time environment, who studied the effect of electronic data interchange
technology on the supplier’s shipment performance. Hau et al. [2] researched the distor-
tion of supply chain information, who proposed the bullwhip effect. Domestic research
on supply chain information sharing began in the early 21st century. Wang et al. [3]
discussed the role of information flow in supply chain management for the first time.
Later, due to the country’s strategic considerations, a series of deepening reforms have
been carried out, gradually extending the domestic supply chain network. In recent
years, many scholars have researched supply chain information from the perspective of
different disciplines.
Since the supply chain network involves issues such as the internal connections and
interactions between stakeholders, information sharing research has become the focus of
attention in supply chain coordination research. Domestic and foreign-related research
mainly focuses on supply chain performance, coordination, and mechanism, empirical
research, etc.
2.2 Blockchain Has Gradually Become a Hot Technology in Global Supply Chain
Management Research
Research. The free trade zone’s role in coordinating global supply chain information
sharing is still in the exploratory stage in related research fields. Still, in recent years,
blockchain technology has been used as a decentralized and intelligent management
method, and scholars have used it to carry out related issues of supply chain coordi-
nation. Given the application of blockchain technology in supply chain information
sharing and coordination, some scholars at home and abroad have begun to study and
try. For example, Mitsuaki et al. [4] proposed a blockchain solution to solve information
asymmetry and double marginalization in supply chain management. Kuntal et al. [5]
discussed the prerequisites for adopting blockchain in the manufacturing supply chain.
Naif et al. [6] used blockchain technology to build a block supply chain to solve the anti-
counterfeiting problem in the supply chain. Vishal et al. [7] used blockchain and Internet
of Things technology to build an information-sharing supply chain management system
to solve information transmission efficiency. Li et al. [8] used the blockchain governance
mechanism system to develop supply chain intelligent governance mechanisms to solve
opportunistic risks and trust issues. Yang et al. [9] established a supply chain information
290 S. Tian et al.
platform through blockchain technology to solve information sharing issues. Zhang et al.
[10] use blockchain technology to build application scenarios to optimize supply chain
management. He et al. [11] believe that the data authenticity of information sharing in
the supply chain is worthy of attention, and blockchain technology can achieve coordi-
nation. These research results can help the free trade zone further explore the theory and
practice of global supply chain information sharing and coordination.
In summary, the current domestic research pays little attention to the role of free
trade zones in coordinating global supply chains. It seldom considers the influence of
free trade zones on the structure of global supply chains, interest relationships, and
information compensation.
to global supply chain information sharing and coordination, the internal mechanism of
the information compensation operation process and its reasonable implementation path,
etc., can promote free trade Expansion and in-depth research on regional information
compensation and global supply chain optimization coordination.
The FTPBIP service is the main body of the FTPGSC information compensation.
The FTPBIP service attracts and incorporates companies into the FTPGSC network
through relevant policies. The network members include manufacturers, manufacturers,
distributors, retailers, and consumers from various countries. These members will use
the platform to carry out related trade activities, generating an FTP blockchain data
warehouse. With the accumulation of data and information and the formation of trust in
the FTP’s blockchain, information services can provide necessary information compen-
sation for members of the FTPGSC network to achieve information sharing. Therefore,
the FTP blockchain information service will play an increasingly important role as the
number of participating companies increases.
The global supply chain enterprises relying on the FTP are the objects of the infor-
mation compensation. The construction of the Hainan FTP is guided by Xi Jinping’s
thoughts on socialism with Chinese characteristics in a new era. China’s further integra-
tion into global economic cooperation is an important practice. Under the background
of the new round of reform and opening up, enterprises from various countries will use
the FTP platform to carry out cross-border trade with China. At the same time, the FTP
can also be used as an important transit point for trade with Asia. At this time, relying
on the FTPGSC companies will obtain the FTP’s blockchain information services and
help them obtain the necessary data and information for information sharing through
information compensation, thereby improving the performance of the FTPGSC.
The construction of FTP is the beginning of China’s further opening up and high-
quality economic development. The FTP will become an important hub connecting
mainland China and overseas. With the support of the FTP policy, enterprises will bring
294 S. Tian et al.
huge profit margins through processing, storage, transportation, and sales of zero-tariff
imported goods. Come to the huge market space in Asia. Therefore, if the company is
strategically positioned in the Asian market, relying on the FTP to build a global supply
chain is of great significance to the company. At this time, companies need to make
reasonable plans and judgments. Faced with mainland China and even the “One Belt
One Road” market, they can consider using FTP as a supply chain node to reconstruct the
global supply chain, increase the level of supply chain information sharing, and improve
supply. Chain performance.
China has entered a new period of historical development. The Communist Party of
China with Chairman Xi Jinping at the core, guided by Xi Jinping’s Thought on Socialism
with Chinese Characteristics for a New Era, has led China out of the path of socialism
with Chinese characteristics for a new era. Under the influence of the new crown virus
in the global economy, China has assumed its due responsibilities to support global
economic development and welcomes all countries to carry out win-win cooperation
with China to tide over the difficulties together. This requires that the enterprises joining
the FTP also have the spirit of responsibility, peace, and sustainable development and
mutually agree with the ideology of socialism with Chinese characteristics in the new
era. Only in this way can the FTP blockchain information service platform provide better
information compensation services.
(3) The consensus of the FTPGSC enterprises on the information on the chain
Since the construction of the FTP is China’s latest open model, companies from all
over the world can participate in it. At this time, due to differences in ideology, ideol-
ogy, politics, religious beliefs, policies and regulations, etc., there will be awareness of
appropriate information on the chain difference. After studying in the FTP, in order to
better play the role of the FTPGSC and achieve corporate strategic goals, the chain enter-
prises will use this as a basis to reach a consensus. The consensus that the appropriate
information in the blockchain is on-chain is the guarantee of the FTP’s blockchain infor-
mation platform for the stable progress of information compensation services between
enterprises in the FTPGSC. If there is no consensus on information on the chain when
joining the FTP, information compensation services will be difficult to develop, which
will affect the role of the FTPGSC.
blockchain information platform will inevitably lead the self-built enterprise platform
from the perspective of perfection, and the services launched will also reflect Chinese
characteristics, thereby ensuring the development of the FTPGSC information compen-
sation service and also improving The brand effect of the FTP blockchain information
platform. Therefore, the completeness of the Hainan FTP’s blockchain information plat-
form will directly affect the information compensation services enjoyed by the FTPGSC
enterprises.
Relying on the FTP to build a global supply chain will form a complex network, and
the network will include all walks of life, and the information flows formed by different
industries will also be intertwined. This requires measuring the degree of demand for
various types of information by supply chain companies. The information of the FTP
blockchain information platform originates from the enterprises that have joined the FTP,
so the degree of contribution of the enterprise to the information will become its bargain-
ing chip for information compensation. Here is a definition of an “FTP data currency”.
The data currency can be used to obtain corresponding information compensation, which
has reached the data exchange of the FTPGSC enterprises. The corresponding data cur-
rency will be given according to the big data 5V model to evaluate the company’s
information supply situation. At the same time, the data currency can also be traded to
give full play to the liquidity of data and currency to give full play to the role of data.
The information in the blockchain information platform of the FTP is composed of global
supply chain companies that rely on the FTP to contribute their data to the appropriate
chain. With the support of blockchain technology, supply chain information sharing
can be better realized. The service of information compensation in the FTPGSC can
help companies optimize upstream supply and downstream distribution and improve the
overall efficiency of the company and the supply chain. Figure 3 shows a schematic
diagram of the information compensation mechanism of the FTPGSC.
Ensure fair transactions between enterprises. The FTPGSC includes suppliers, man-
ufacturers, manufacturers, distributors, and retailers. The information flow formed can
include supply information provided by suppliers, manufacturers, and manufacturers
and demand information provided by distributors and retailers. Because it is an informa-
tion platform built with blockchain technology, the storage and use of this information
are anonymous, encrypted, available and invisible. The uniqueness of the product can
be guaranteed through the information sharing of the data on the chain. When a global
supply chain company built on the FTP joins the blockchain information platform, the
companies on the chain can use the platform certification system to verify the validity
of the information on the supply chain, such as commodities, funds, and logistics.
296 S. Tian et al.
The global supply chain formed between enterprises relying on the FTP can better
share data based on trust and cooperation. To obtain higher benefits for the FTPGSC, the
information compensation service provided by the FTP blockchain information platform
can be used to compensate for the missing information in upstream and downstream of
the FTPGSC. At this time, companies are required to make corresponding contributions
to the platform, that is, to analyze corporate data to obtain “FTP data currency.” After
obtaining a certain amount of data currency, you can enjoy the data compensation service
brought by the platform. These compensation data can be data on the supply chain,
industry data, or the supply chain network. In this way, the data of the entire platform
can flow quickly, forming an FTP data circulation system. This can expand the scale
of platform enterprise users and expand the scale of their information and realize the
common growth of the platform and users in a virtuous circle.
When the blockchain information service platform operates normally, it can be
extended to other related industries based on information sharing in the FTPGSC. For
example, the blockchain information service platform can provide information compen-
sation services for supply chain information sharing and provide financial service insti-
tutions with complementary information compensation services to ensure the credibility
and feasibility of corporate financing. For another example, the blockchain informa-
tion service platform can provide necessary information compensation for safe produc-
tion, allowing insurance institutions to obtain necessary safe production data based on
the information compensation service to determine whether the company’s production
behavior is compliant.
Therefore, when the FTP blockchain information platform gathers a sufficient num-
ber of companies and their related data, it can enter the self-generation mode to form vari-
ous information compensation paths and become a powerful assistant in the construction
and development of the FTP.
Global Supply Chain Information Compensation Model Based on FTPBIP 297
According to the information compensation service content, this paper designs a three-
layer model of information compensation for the blockchain information service plat-
form. The three-layer structure of the model is the infrastructure layer, the core busi-
ness layer, and the entity application layer. Figure 4 shows the three-layer model of
information compensation.
The infrastructure layer of the information compensation model is the data acquisition
and storage of business flow, capital flow, logistics, and information flow of each node in
the FTPGSC. The data source is mainly the data voluntarily shared by the enterprises on
the chain after joining the FTPBIP. The data to be shared is minimized on the chain using
the FTPBIP. The enterprise stores four streams of information as a distributed account
298 S. Tian et al.
to ensure the authenticity of the data. Because the blockchain has the characteristics
of information that cannot be tampered with, it can help consumers and enterprises to
verify products through the FTPBIP. It can enhance the credibility and brand value of
the enterprise.
Through the infrastructure layer, suppliers, manufacturers, manufacturers, distribu-
tors, and retailers can be effectively linked to form an FTPGSC and realize information
sharing.
The core business layer of the information compensation model is the integration of
data processing, data currency exchange, data transaction, and information compensa-
tion. For the FTPSC companies that have joined the FTPBIP, the information they share
mainly includes upstream and downstream logistics information, transaction informa-
tion, production information, credit information, etc. Some of the information is private
to the company. The above information can be effectively shared under the principle that
the data is appropriately placed on the blockchain without affecting the core competi-
tiveness of the enterprise. The FTPBIP will use data processing technology to process
the data as anonymous, immutable, and encrypted and then upload it to the blockchain.
The information provider uses blockchain data to exchange the corresponding data cur-
rency. Enterprises can use data currency for information transactions and information
compensation services. With the continuous accumulation of data in the supply chain
and industry data, the FTPBIP can develop data derivative services. For example, the
development of information compensation services in the financial industry can help
companies measure and judge the credibility of partners more accurately.
The entity application layer of the information compensation model integrates supply
chain enterprises’ applications and related industries’ applications. The application of
supply chain enterprises is mainly to improve the performance of the supply chain
and provide information compensation services to the upstream and downstream of the
supply chain. Corresponding applications include information query, statistical informa-
tion analysis, data currency exchange, data transaction, and information compensation.
FTPSC companies can use corresponding apps to carry out specific businesses. The
applications of related industries are mainly aimed at the financial industry, insurance
industry, security management, human resource management, etc. The feature of the
application is to provide corresponding information compensation based on the needs
of different industries based on ensuring sufficient data and to help it better serve the
FTPSC. With the help of blockchain technology and information compensation services,
information sharing in the FTPSC will likely be resolved. By breaking through all sup-
ply chain links and inter-industry data barriers, FTPBIP can provide FTPSC enterprises
with more valuable services.
Global Supply Chain Information Compensation Model Based on FTPBIP 299
5 Conclusion
This paper proposes a blockchain-based information compensation model for the
FTPGSC under the construction of the FTP. In order to improve the overall perfor-
mance of the supply chain, information compensation can be used to compensate supply
chain companies for solving the problem of supply chain information sharing. Through
research, the factors and influencing factors of information compensation in the FTPGSC
are obtained. On this basis, the mechanism of information compensation is discussed.
Through the construction of a three-tier model of information compensation, it systemat-
ically describes the information compensation system of the FTPGSC. It provides some
theoretical basis for the reconstruction of the FTPGSC.
In the construction of the FTP, the application research of blockchain technology
is still in its infancy, and related applications are still being improved. The informa-
tion compensation model we constructed still has shortcomings, and further research is
needed in the future. For example, the issuance and use of data currency has not yet been
studied, and interested scholars can do in-depth research in this area.
Acknowledgments. This work was supported by the 2021 Philosophy and Social Science Plan-
ning Project of Hainan Province of Chain (HNSK(ZX)21-87, Research on the safe and orderly flow
mechanism of Hainan Free Trade Port data under the blockchain).This work was also supported
by Hainan Provincial Natural Science Foundation of China (720RC567, Research on Information
Compensation Mechanism of Free Trade Zone to Global Supply Chain Innovation Coordination
under BlockChain Technology, and 720RC569, Tourism Value Chain Distribution and Ecological
Optimization Mechanism of Hainan International Tourism Consumption Center Based on System
Dynamics). This work was supported by the Humanities and Social Sciences Research Innovation
Team of Hainan University, Hainan Free Trade Port Cross-border E-commerce Service Innovation
Research Team (HDSKTD202025).
Authors’ Contributions. Shaoqing Tian was responsible for proposing the overall idea and
framework of the manuscript. Fan Jiang was responsible for the writing of the first draft of the
manuscript and translation. Chongli Huang was responsible for the revision of the manuscript and
proofreading. Shaoqing Tian, Fan Jiang, and Chongli Huang contributed equally to this paper and
are regarded as the first authors.
References
1. Srinivasan, K., Kekre, S., Mukhopadhyay, T.: Impact of electronic data interchange technology
on JIT shipments. Manag. Sci. 40, 1291–1304 (1994)
2. Lee, H.L., Padmanabhan, V., Whang, S.: The bullwhip effect in supply chain. Manag. Sci.
38(3 R), 93–102 (1997)
3. Wang, C.G.: Logistics and information flow management in the supply chain. Chin. Manag.
Sci. (04), 17–24 (2000)
4. Nakasumi, M.: Information sharing for supply chain management based on block chain
technology. In: 2017 IEEE 19th Conference on Business Informatics, pp. 140–149 (2017)
5. Bhattacharyya, K., Smith, N.: Antecedents to the success of blockchain technology adoption
in manufacturing supply chains. In: 2018 2nd International Conference on Business and
Information Management, pp. 64–67 (2018)
300 S. Tian et al.
6. Alzahrani, N., Bulusu, N.: Block-supply chain: a new anti-counterfeiting supply chain using
NFC and blockchain. In: CRYBLOCK 2018-Part of MobiSys, pp. 30–35 (2018)
7. Naidu, V., Mudliar, K., Naik, A., Bhavathankar, P.: A fully observable supply chain man-
agement system using blockchain and IOT. In: 2018 3rd International Conference for
Convergence in Technology, pp. 84–89 (2018)
8. Li, X., L, Z.G.: Intelligent governance mechanism of supply chain based on blockchain
technology. China Circ. Econ. 31(11), 34–44 (2017)
9. Yang, H.Q., Sun, L., Z, X.C.: Building mutual trust and win-win supply chain information
platform based on blockchain technology. Sci. Technol. Prog. Policy 35(05), 21–31 (2018)
10. Zhang, X.Q.: Optimization of supply chain management model based on blockchain. China
Circ. Econ. 32(08), 42–50 (2018)
11. He, C., L, Y.F.: Research on a new supply chain model integrating blockchain. Manag.
Modernization 40(01), 84–87 (2020)
12. Hoberg, K., Thonemann, U.W.: Modeling and analyzing information delays in supply chains
using transfer functions. Int. J. Prod. Econ. 156, 132–145 (2014)
13. Nie, J.J.: The impact of retailer information sharing on the recycling mode of closed-loop
supply chain. J. Manag. Sci. 16(05), 69–82 (2013)
14. Shi, C.L., Nie, J.J.: The influence of network externalities on the information sharing of
dual-channel supply chain. Chin. Manag. Sci. 27(08), 142–150 (2019)
15. Zhang, Z.: From “information compensation” to “conforming translation”. Foreign Lang.
Stud. (01), 63–64 (1998)
16. Cao, M.L.: “Translation studies also need translation”——re-discussion on misreading and
mistranslation in the process of introducing western translation theories. Foreign Lang. Stud.
(03), 67–74 (2012)
17. Shang, W.G., Wu, H., Zhao, B.: Mobile user group mining algorithm with information
compensation. Comput. Eng. Appl. 46(11), 226–229 (2010)
18. Jin, Z.Y., Han, L.G., Hu, Y., Ge, Q.X., Zhang, P.: Data-driven Marchenko imaging based on
low-frequency information compensation. Chin. J. Geophys. 60(09), 3601–3615 (2017)
Multi-layer Intrusion Detection Method Using
Graph Neural Network
Ling Ma(B)
Abstract. Aiming at the problem that many unknown intrusion behaviors are dif-
ficult to detect, resulting in the low security of deep belief network, a multi-layer
intrusion detection method based on graph neural network is proposed. Firstly,
the elm classifier is designed to obtain multi-layer intrusion data with inconsistent
range and uneven distribution; secondly, the deep belief network intrusion detec-
tion model is constructed by combining the monitoring, collection, preprocessing,
classification, and response modules; finally, the multi-dimensional state vector is
set, and the model detection results are divided based on the graph neural network
to realize the classification of multi-layer intrusion data and optimize the deep
belief network The result of network multi-layer intrusion detection. Experimen-
tal results show that the proposed method can identify all unknown multi-layer
intrusion behaviors with low false alarm rates and improve network security. The
detection process is more targeted to optimize the false alarm rate and accuracy
of the detection results.
1 Introduction
As a public space, many personal information leakage events have occurred in recent
years. Therefore, deep intrusion detection of the network has become an important
premise for the development of information technology at this stage [1, 2]. Aiming at
the problem of network intrusion, the concept of intrusion detection was first proposed
abroad in 1980. Intrusion refers to all acts that attempt to destroy the privacy of network
resources and network security mechanism [3, 4]. Therefore, Dorothy et al. designed an
intrusion detection expert system ides from 1980 to 1990. Based on machine learning
algorithm and deep learning algorithm, a domestic network intrusion detection method is
established. Since the beginning of the 21st century, the network coverage has gradually
increased, and the network vulnerability data is also increasing year by year.
Literature [5] constructs a rotating forest algorithm based on SPCA, which increases
the integration diversity by introducing rotation, and solves the insufficient strength of the
base classifier. Literature [6, 7] combines the number of layers of neural network hidden
layer and the number of neuron nodes in each layer to improve the weak adaptability
of the detection algorithm. It adopts the strategy of layer by greedy layer training to
establish a feature extraction model to detect network intrusion data. Literature [8]
discusses in detail the main progress and shortcomings of existing network intrusion
detection research from three aspects: undirected unit graph recommendation, undirected
binary graph recommendation, and undirected multivariate graph recommendation, and
defines the interpretability of GNN recommendation and other research directions of
GNN recommendation in the future.
Then the corresponding output matrix is obtained according to the above calculation,
and the formula is:
⎡ ⎤
g(ω1 x1 + u1 ), g(ω2 x1 + u2 ), . . . , g(ωk x1 + uk )
⎢ g(ω1 x2 + u1 ), g(ω2 x2 + u2 ), . . . , g(ωk x2 + uk ) ⎥
A=⎢ ⎣
⎥
⎦ (3)
...
g(ω1 xN + u1 ), g(ω2 xN + u2 ), . . . , g(ωk xN + uk ) N ×K
Multi-layer Intrusion Detection Method Using Graph Neural Network 303
The optimized elm classifier is used to identify the basic patterns of intrusion behavior,
and a deep belief network intrusion detection model is constructed with the goal of
synthesizing a few category samples, as shown in Fig. 1.
Aiming at the problem of uneven distribution of data sets, the classifier elm is used
to synthesize a few category samples, improve the category balance, and complete the
preprocessing and classification of intrusion data. The response module obtains feed-
back according to the data classification results, including active response and passive
response.
1. Assuming that some markers are represented in graph I (P, E), the unlabeled nodes
are predicted according to the labeled nodes. Assuming that the vector of each node
is Hb under the v-dimensional state vector, there is:
Hb = f (Fb , Fa , Hc , Fc ) (5)
2. Let f (Fb , Fa , Hc , Fc ) be the mapping function to map the input node features to
the v-dimensional vector space. According to the above formula, the v-dimensional
state vectors of all nodes are obtained to describe the characteristics of connecting
nodes and the interaction information between other nodes.
3. It is calculated that the output signal of the interaction task Z of a single connection
node is:
⎡ ⎤
Z11 Z12 Z13 · · · Z1k
⎢ Z21 Z22 Z23 · · · Z2k ⎥
⎢ ⎥
⎢ ⎥
Z = Z1 Z2 Z3 · · · Zn = ⎢ Z31 Z32 Z33 · · · Z3k ⎥
Y
(6)
⎢. . . . ⎥
.
⎣. . .. . .
. ⎦
Zn1 Zn2 Zn3 · · · Znk
According to the above calculation, the control output signal is related to the output
signal of the functional section and the synchronization difference between adjacent dif-
ferent functional sections in the interaction process, so that the synchronization difference
between adjacent different functional sections is w.
assumptions:
Wn = {Wn1 , Wn2 , Wn3 , · · · , Wnk } (7)
4. At the same time, since formula (5) obtains the unique solution, iterative formula
(5) obtains.
Then the evaluation results of the accuracy rate s1 , detection rate s2 and false alarm
rate s3 of the detection method,
Two groups of traditional detection methods (literature [3], literature [4] and literature
[5]) were selected as the control object. Two test groups were established with the
proposed detection method. The above calculation formula determined the accuracy,
detection rate and false alarm rate of each group of methods, and the differences between
the two groups of methods were compared.
1. Accuracy rate
According to L =
p
i=1 y log(G(x)). When the value of L is less than 0.4, the
feedback intrusion data is classified as class A; When the value of L exceeds 0.4, the
intrusion data is classified as class B, which corresponds to different detection mecha-
nisms respectively. The ratio of the number of relevant samples in the identification or
detection results of the three methods to the total number of samples in the results is
obtained for comparative test.
2. Classification effect
The classification data obtained from the above experimental test results are derived.
Among them, the L value obtained by the experimental group according to formula (9)
is 0.5745, which is a positive number. Therefore, the first quadrant classification data is
taken as the classification effect data sample, the classification effect comparison test is
carried out, and the comparison results are derived.
Taking the method of literature [3] as group A, the method of literature [4] as group B
and the method of literature [5] as group C, the test results of accuracy rate and iteration
times, classification effect and false positive rate were compared and analyzed.
306 L. Ma
Taking the proposed method as the experimental group, Fig. 2 shows the change of
accuracy rate of each test group under different iteration times.
According to the above test results, the accuracy of the detection method in the
experimental group has stabilized at 91.37% since the 92nd iteration; In control group
A, the accuracy rate was controlled at 91.02% from the 167th time; group B and group
C failed to control the accuracy within a reasonable range within 250 iterative tests, and
their fluctuation was large. In order to obtain accurate experimental data, the number
of iterations was expanded to 500. It was found that the accuracy rate of group B was
stable at 90.7% at 307th time.
It can be seen that although the iteration times of the three groups of methods are
inconsistent, the accuracy is relatively close, and can be used as the same comparison
test group.
According to statistics, it is found that the third round of test results of group A are
empty, which is caused by human operation errors after investigation. Therefore, after
removing the third round of test results of group A, the false alarm rate of the three
groups of methods is lower than that of the experimental group, which verifies that the
method in this paper has high practical applicability. It is shown in Table 1.
Multi-layer Intrusion Detection Method Using Graph Neural Network 307
4 Conclusion
This paper constructs a deep trust network intrusion detection model, divides the detec-
tion results of the model, realizes the classification of multi-layer intrusion data, and
further optimizes the detection results of the model. The ELM classifier is used to ran-
domly generate the input weight and hidden layer neuron bias. The ELM classifier is
used to overcome the problems of easy falling into local optimization and slow training
speed, and improve the overall classification effect of the detection method.
However, when the data volume of deep belief network is further expanded, the accu-
racy decreases. Therefore, in the future, we can strengthen the analysis of the correlation
between nodes and get more accurate intrusion information classification results.
References
1. Tang, L., Li, H.X., Yan, C.Q.: Overview of deep neural network structure search. Chin. J.
Image Graph. 26(2), 245–264 (2021)
2. Liu, F., Yang, A.Z., Wu, Z.W.: Adaptive siamese network based UAV target tracking algorithm.
Acta Aeronautica ET Astronautica Sinica 41(1), 248–260 (2020)
3. Yang, T., Ye, X.: Network intrusion detection algorithm SPCA erof. Comput. Eng. Des. 42(2),
356–362 (2021)
4. Song, Y., Hou, B.N., Cai, Z.P.: Network intrusion detection method based on deep learning
feature extraction. J. Huazhong Univ. Sci. Technol. (Nat. Sci. Ed.) 49(2), 115–120 (2021)
5. Wu, G.D., Cha, Z.K., Tu, L.J.: Research progress of neural network recommendation. J. Intell.
Syst. 15(1), 14–24 (2020)
6. Ren, J.D., Liu, X.Q., Wang, Q.: An multi-level intrusion detection method based on KNN
outlier detection and random forests. J. Comput. Res. Dev. 56(3), 566–575 (2019)
7. Zhong, M., Zhou, Y., Chen, G.: Sequential model based intrusion detection system for IoT
servers using deep learning methods. Sensors 21(4), 1113–1116 (2021)
8. Yan, X.L., Wang, B.B.: Fault diagnosis of analog circuits based on elm optimized by an
adaptive wolf pack algorithm. Comput. Eng. Sci. 41(2), 246–252 (2019)
308 L. Ma
9. Wang, X.L., Xie, H.Y., Wang, J.J.: Prediction of dam deformation based on bootstrap and
ICS-MKELM algorithms. J. Hydroelectric Eng. 39(3), 106–120 (2020)
10. Zhang, X.M., Cao, G.Q., Chen, Z.Q.: Landslide displacement prediction based on AdaBoost-
PSO-ELM algorithm. Appl. Electron. Tech. 56(3), 566–575 (2021)
Research on an Image Hiding Algorithm
with Symmetric Means
1 Introduction
With the rapid development of network technology, multimedia technology, and com-
munication technology in the information age, information hiding technologies repre-
sented by steganography, digital watermarking technology, visual passwords, submarine
channels, and concealment protocols have been widely used in the protection of digital
intellectual property rights, Multimedia information authentication, data integrity verifi-
cation, data non-repudiation confirmation, and covert transmission and other information
security fields. In information hiding technology, images as the carrier for the concealed
transmission of information has become a popular, secure communication technology.
In recent years, people have proposed various image information hiding technologies
and methods, such as: to recover the embedded confidential images completely. Ranzan
Wang et al. proposed an image hiding method through the best LSB replacement and
genetic algorithm in 2001 [1]; YCHu’s image hiding scheme based on vector quan-
tization (VQ) in 2003; however, the secret image embedded in this method cannot be
completely extracted [2]. To improve the quality of the camouflage image, SukLing et al.
[5] proposed a simple LSB replacement method through the optimal pixel adjustment
program. Chengyun Yang et al. [4] proposed a lossless information hiding method based
on wavelet domain spread spectrum in 2004, and this method embeded data into the high-
frequency subband coefficients of the image after integer wavelet transforms according
2 Design of Algorithm
2.1 Overview of Method
The image discussed in this article is an 8-bit standard gray value image. Let C be a
cover-image with a size of Mc × Nc pixel (cover-image), and S indicates a secret image
with a size of Ms × Ns pixel (secret-image), which will cover the image respectively C
and the secret image S are divided into image blocks with I × J pixels, and the carrier
image C and the secret image S can be expressed as:
×Nc ×Ns
C = Ck |0 ≤ k ≤ MIc×J , S = Sk |0 ≤ k ≤ MIs×J (1)
Let Ck be the camouflage image after embedding the secret image. The specific
process is shown in Fig. 1 and Fig. 2:
Research on an Image Hiding Algorithm with Symmetric Means 311
Camouflage image
Image block matching
Mean Compression. Let: Ck = ( Cij)mxn be the cover image; Sk = ( Sij) be the secret
image, (i = 1,2,…,m; j = 1,2,…,n).
Let: mean E( Sk) = N1 Sij,and ij = Sij— E(Sk).
If ij ≥ 0, then Sij = Sij/Max(Sij);
If ij < 0, then Sij = Min(Sij)/Sij.
So the secret image is compressed into a new image: Sk = ( Sij ) mxn (i = 1,2,…,m; j
= 1,2,…,n).
Symmetry.
(1) x-axis symmetry.
The image block Sk = (Sij ) mxn (i = 1,2,…,m; j = 1,2,…,n), expressed in a
rectangular coordinate system as follows:
➀ If m is an even number, let: a = m/2, and subscript i1 belongs to (1., a). The
symmetry point i2 of subscript i1 has a — i1 = i2 — (a + 1), and we get: i2 = 2a—i1
+ 1. (E.g:)
Therefore, the pixels S i1 j1 < —— > S i2 j1 , can be swapped. (i1 = 1, 2, ---, a; i2 =
2a — i1 + 1; j1 = 1, 2, …, n) (Fig. 3).
➁ If m is an odd number, let: a = (m + 1)/2, and the subscript i1 belongs to (1., a).
The symmetry point i2 of subscript i1 has a—i1 = i2—a, and we get: i2 = 2a—i1 + 1.
(E.g:)
Therefore, it is possible to swap pixels S i1 j1 ’ < —— > S i2 j1 ’ (i1 = 1, 2,…, a; i2 =
2a—i1 + 1; j1 = 1, 2,…, n).
➂ The subscripts i2 = m, m − 1,…,a correspond to the subscripts p = 1, 2,…, a
respectively.
Then subscript i1 = a—1, a—2,…, 2, 1 correspond to the subscript p = a + 1, a +
2,…, m respectively.
Then: S p j1 = S i2 j1 ( p = 1, 2,…, a; i2 = n, n − 1,…, a; j1 = 1, 2,…, n);
312 H. Wang et al.
(1) Divide the cover image C and the secret image S into:C 1 , C 2, …, C N ; S1, S2 , …,
SN ;
(2) Calculate the average value of the secret image block E(Sk );
(3) The secret image block Sk is compressed into a new secret image block Sk , namely.
(4) Symmetry: swap the pixels S i1j1 ’ < —— > S i2j1 ’ and S i2j1 ’ < —— > S i2j2 ’. The
new pixel obtained is denoted as SS ij (i = 1, 2,…, m; j = 1, 2,…, n).
Then get the secret image: SS k = (SS ij )mxn.
(5) Order
Ck = α* Ck + β* SSk
Among them α + β = 1 (can choose α = 0.999, β = 0.001).
So C k ’ is a new hidden image with secret image information.
Ck = α ∗ Ck + β ∗ SSk
(1) Solve by
SSk = (Ckα ∗ Ckα)/β.
(2) Solve S k from SS k .
(3) When ij ≥ 0, S ij = S ij ’Max(S ij ).When ij < 0, S ij = Min(S ij )/S ij ’. Then there is
S k = (S ij ).
(4) Restore image S = ∪ S k .
3 Numerical Examples
There are examples to illustrate the process of image hiding and restoration. Firstly 256
× 256 pixels Airplane is the overlay image, and 128 × 128 pixels Lena is the secret
image, dividing them into 8 × 8 pixels blocks. The overlay image will be divided into
256 × 256/8 × 8 = 1024 blocks, and the secret image will be divided into 128 × 128/8
× 8 = 256 blocks.
An 8 × 8 block of the cover image and the secret image is shown, respectively:
⎡ ⎤
127 176 176 176 177 191 177 191
⎢ 127 176 176 176 176 177 176 176 ⎥
⎢ ⎥
⎢ 127 176 177 177 176 176 177 177 ⎥
⎢ ⎥
⎢ ⎥
⎢ 127 159 158 159 159 158 159 159 ⎥
Ck = ⎢ ⎥
⎢ 132 159 159 151 152 152 149 152 ⎥
⎢ ⎥
⎢ 127 176 158 158 159 151 152 151 ⎥
⎢ ⎥
⎣ 121 177 176 159 176 177 159 159 ⎦
118 158 159 177 177 176 176 176
314 H. Wang et al.
⎡ ⎤
133 133 133 132 132 133 130 132
⎢ 133 133 131 133 133 133 133 133 ⎥
⎢ ⎥
⎢ 133 133 133 131 133 133 133 133 ⎥
⎢ ⎥
⎢ ⎥
⎢ 133 133 133 132 133 133 133 133 ⎥
Sk = ⎢ ⎥
⎢ 133 131 133 133 133 133 133 133 ⎥
⎢ ⎥
⎢ 131 133 133 133 133 133 133 133 ⎥
⎢ ⎥
⎣ 133 133 133 133 131 133 133 133 ⎦
133 133 133 133 133 132 133 133
From the significance level α = 0.005, the number of samples n = 64, the look-up
table has tα(n−1) = t 0.005 (64–1) = t 0.005(63) = 2.656. Then the result is shown in
Fig. 4. Because of |t|= 3.568 > 2.656 = t 0.005(63), and the hypothesis test H1: μ μ0
holds.
The mean value of the cover image block Ck,μ 0 = E(Ck) = 162.7656, the variance σ 2
= 300.849. The mean square error σ = 17.345.
The mean value μ = E( Ck ) = 162.3191 of the new hidden image block Ck, the
variance S 2 = 297.953, and the mean square error S = 17.2613. After calculation:
– (1) The difference between the new hidden image block Ck and the covered block
Ck is less than 1 pixel.
– (2) From μ/μ = 0.00275 = 0.275%, it is known that the new hidden image block
Ck contains 99.725% of the information of the cover image block Ck and 0.275%
of the information of the secret image block Sk. Therefore, the newly hidden image
block Ck has a better hiding effect.
Research on an Image Hiding Algorithm with Symmetric Means 317
Fig. 6. Secret image with 128 × 128 pixels Lena Fig. 7. Overlay image after 256 × 256
pixels hidden image
4 Experimental Results
The quality of the camouflage image evaluates the value of PSNR based Eq. (1).
The experimental results are obtained with 256 × 256 pixels Airplane as the overlay
image and 128 × 128 pixels Lena as the secret image (Figs. 5, 6 and 7).
2
MAXI
PSNR = 10 × log10 (1)
MSE
1 m−1 n−1
where MSE = mn i=0 i=0 (I (i, j) − K(i, j)) , MAXI represents the maximum
2
value of image grayscale, I (i, j) represents the original pixel value of the image, K(i,j)
represents the pixel value after embedding information.
318 H. Wang et al.
The result of our calculation is PSNR = 45.01, which is much higher than the
PSNR value in SukLing et al. [5], Zhang et al. [6], and Wang et al. [7], the quality of
the camouflage image is improved. The difference between the cover image and the
camouflage image cannot be seen visually.
5 Conclusions
The algorithm presented in this paper has used the image as the carrier, and a new
image information hiding algorithm (mean symmetric algorithm) is created to realize
information hiding. Less information is embedded in the image, which effectively saves
information space and improves transmission efficiency. The impact on secret images
is reduced as much as possible, and the quality of disguised images is improved. The
difficulty of cracking algorithms is increased, and the security of information hiding is
improved.
Acknowledgments. This work is supported by the Natural Science Foundation of China’s Hei-
longjiang Province (No. YQ2020G002), University Nursing Program for Young Scholars with
Creative Talents in Heilongjiang Province (No. UNPYSCT-2020212).
References
1. Wang, R.Z., Lin, C.F., Lin, J.C.: Image hiding by optimal LSB substitution and genetic
algorithm. Pattern Recogn. 34(3), 671–683 (2001)
2. Hu, Y.C.: Grey-level image hiding scheme based on vector quantization. IEEE Electron. Lett.
39(2), 202–203 (2003)
3. Yang, C.-Y., Guo, E.X., Zheng, Y.: Reversible data hiding based on wavelet spread-spectrum.
Comput. Eng. Appl. 40(14) (2004)
4. Zheng, Y., Yang, C., Xuan, G.: Wavelet-based high-capacity lossless image hiding and its
application. Comput. Eng. 31(1), 56–59 (2005)
5. Li, S.-L., Leung, K.C., Cheng, L.M., Chan, C.-K.: A novel image-hiding scheme based on
block difference. Pattern Recogn. 39(6), 1168–1176 (2006)
6. Zhang, R., Ma, L., Zhang, J.: Images concealment based on interframe matching mean and
variance. J. Image Graph. 15(11), 1578–1582 (2010)
7. Wang, H., Qu, L.: A triangular algorithm of image-hiding. In: The 8th World Congress on
Intelligent Control and Automation, Jinan, Shandong, China, (2010)
8. Wang, H.: A complex number algorithm of image-hiding. In: The 10th World Congress on
Intelligent Control and Automation, Beijing, China (2012)
9. Zhang, X.: Reversible data hiding in encrypted images. IEEE Signal Process. Lett. 18(4),
255–258 (2011)
10. Ali, B.M., Babu, T.R.: A Novel image hiding algorithm using optimal band selection. Int. J.
Innov. Technol. Explor. Eng. 8(5), 1155–1159 (2019)
11. Li, X., Kim, S.-T., Lee, I.-K.: Robustness enhancement for image hiding algorithm in cellular
automata domain. Optics Commun. 356, 186–194 (2015)
12. Pang, H.W.: The technology of digital image hiding based on discrete wavelet transform.
Appl. Mech. Mater. 3822, 718–721 (2015)
13. Yan, Y.: Image information hiding algorithm research of network sensor based on visual
characteristic. Sensors Transd. 160(12), 315–322 (2013)
14. Feng, D., Chen, L.: An improved image hiding algorithm of singular value parity interval
quantization. Sci. Technol. Eng. 19(26), 283–287 (2019)
Image Analysis and Processing
A Detection Network United Local Feature
Points and Components for Fine-Grained Image
Classification
Yong Du1,2 , Bin Yu3 , Peng Hong4 , Wei Pan5 , Yang Wang2 , and Yu Wang6,7(B)
1 Tianjin University, Tianjin 300072, People’s Republic of China
2 Northeast Agricultural University, Harbin 150001, People’s Republic of China
3 Harbin Institute of Technology, Harbin 150001, People’s Republic of China
4 Lanzhou Jiaotong University, Lanzhou 730000, People’s Republic of China
5 Harbin Huade University, Harbin 150025, People’s Republic of China
6 Tianjin University, Tianjin 300072, People’s Republic of China
7 Northeast Agricultural University, Harbin 150001, People’s Republic of China
1 Introduction
Image classification is a fundamental research problem in computer vision, which has
attracted much attention in recent years. With the rapid development of deep learning
technology, this fundamental problem has been solved. Nevertheless, fine-grained image
classification gradually replaces image classification with increasing demand for retrieval
and classification. It has a broad application prospect such as the classification of dish,
clothing, and animal subcategory [1, 2, 3]. The fine-grained image classification is a
subdivision of subcategories based on the classification of essential categories, which is
more complex due to the natural similarity of subcategories. The simplest solution is to
use the general image classification model for training, but the classification performance
is relatively poor that cannot deal with practical situations. The following disadvantages
are as below:
The algorithm based on deep learning technology has gradually shown great poten-
tial in fine-grained classification [4–6]. The fine-grained image classification is very
flexible in manual annotation, component information, component location prediction,
and second-order information. Experts and scholars have proposed various solutions
to this problem. It can be generally divided into four categories: i) integrated network
method, ii) methods based on visual attention mechanism, and iii) feature representation
method based on the high order information.
The synthetic network method uses one or more neural networks to learn different
feature representations of images to achieve better classification performance [7]. The
method based on visual attention mechanism has been widely used in fine-grained image
classification, which does not directly process image information. Still, it completes the
classification by simulating human’s ability to focus on local areas and distinguish [8, 9].
The feature representation method based on high-order information is different from the
low-order description of standard manually constructed features, which construct high-
order details to promote fine-grained classification [10, 11]. In fine-grained classification,
the differences between categories are mainly reflected in minor local differences. There-
fore, the improvement of partial detectors with fine-grained subtask recognition ability
is essential.
From the traditional manual feature extraction method to the deep learning automatic
feature extraction method [12, 13], and then to the integration [14], attention mechanism,
second-order information statistics [15], and other methods [16], the accuracy of fine-
grained image classification is getting higher. However, the localization method of local
fine feature points and local parts is still worth studying and very important. Therefore,
this paper proposes a joint detection network model of local feature points and component
references from this point of view. Experiments on CUB 200-2011 data sets verify the
local feature detection performance of the improved model, which different feature
points and components achieve optimal detection accuracy.
The remaining sections are arranged as follows. Section 2 describes the database
and pre-processing. Section 3 describes the network structure proposed in this paper.
Section 4 is the experiment and analysis. Finally, conclusions and future work are given
in Sect. 5.
is improved and even exceeds the accuracy of human recognition. This section will
introduce a fine-grained image database and the pre-processing of the data.
1 2 3 4 5 6 7 8
Back Beak Belly Breast Crown Forehead Left eye Left leg
9 10 11 12 13 14 15
Left wing Nape Right eye Right leg Right wing Tail Throat
As shown in Table 1, some body parts are close, such as the left and right eye, so
detecting all 15 feature points is unnecessary. At the same time, each feature point cannot
exist in every image due to the different bird poses and the occlusion problems in some
images. If a feature point that does not exist in an image needs to be detected, we will
replace it with an adjacent feature point. In this paper, 15 body parts are divided into
five local feature points according to their relative positions, namely, head, wing, chest,
leg, and tail. The specific corresponding relationship is shown in Table 2.
In this paper, the bird’s body is divided into the head and the torso. For all points
on the head and torso area, take the smallest rectangle that surrounds these points. In
324 Y. Du et al.
this way, each image can be divided into three parts: head, torso, and background. The
corresponding relationship between the two categories of body parts and the body parts
of 15 birds is shown in Table 3.
Table 3. The serial numbers of the bird body parts corresponding to the two types of parts.
Finally, different interpolation methods can be used to map the image input to the
output image. The following equation, k, represents different sampling kernel functions.
H W
Vci = c
Unm k(xis − m; x )k(yis − n; y )∀i ∈ 1...H W ∀c ∈ [1...C] (2)
n m
326 Y. Du et al.
Fig. 3. Training process diagram of component detection network based on STN substructure.
Firstly, the original image is pre-processed to make the size 224 * 224 or 448 * 448,
which is input into the VGG16 network, and the features of the conv5_3 layer, namely
the last layer of the fifth convolution module, are extracted. Then, the features of the
conv5_3 layer were input into the positioning network and grid generator of STN, and
the extracted regional location coordinates (X, Y ) were obtained. After that, the labeled
rectangular box is pre-processed and input into the grid generator to obtain the ground
truth region’s position coordinates (Xg, Yg). Finally, L2 loss was used to regression the
extracted position coordinates and ground truth coordinates, as shown in formula (3).
Some of the annotated information here is only used in the training phase and is not used
in the testing phase.
448 448
loss = 0.5 (xgroundtruth − xi,j )2 + (ygroundtruth − yi,j )2 (3)
i = 1 j=1
The specific structure of the positioning network is shown in Fig. 4. To obtain the six
parameters of the affine transformation, the positioning network uses a fully connected
structural design so that the number of channels is gradually reduced to 6.
The process of using position coordinates and grid generators to generate position
regions is as follows: i) taking the points in the rectangle box’s upper left and lower right
A Detection Network United Local Feature Points 327
corners to get Xmin, Ymin, Xmax, Ymax, ii) scaling these four points to make them match
the pre-processed original image that was input into VGG16, iii) dividing Xmin-Xmax
into w parts and Ymin-Ymax into h parts according to the size of the output image preset
by STN, such as w and h, and iv) mapping the X and Y coordinates as shown in Fig. 4
to obtain a series of positions in regions.
When the training process is completed, the specific application mode of the
component detection branch is shown in Fig. 5.
To observe the effect of using partial annotation and spatial transformation network to
locate the local area, we sampled the bird’s head and torso using the location coordinates
generated by the trained network, as shown in Fig. 3.
The model design of the second option is shown in Fig. 7. Feature points and multiple
components are directly detected in parallel. Therefore, only one detection network
model needs to be trained, an upgraded version of scheme one.
The model design of the third scheme is shown in Fig. 8. Based on program one, the
input of the network is changed from three channels to six channels. The added three
channels are the RGB three channels of the original image extracted through the label
box. The original intention of this design is that the network pays more attention to the
central part of the image so that the multi-task model can predict the feature points and
parts more accurately.
In the next part of the experiment, we will analyze and compare the effect of these
three model designs on the improvement of local feature point detection performance.
To quantitatively compare the effect of joint detection model on feature point detection
with that of single detection, this paper calculates the Euclidean distance between the
predicted value and the actual value for each type of feature point and then calculates
the average of the sum of distances for the same type of feature points. The smaller the
value is, the smaller the distance between the prediction point and the actual value is,
and the more accurate the prediction result is. The experimental calculation results are
shown in Table 4.
A Detection Network United Local Feature Points 329
Table 4. Euclidean distance between the predicted value and true value of different schemes.
It can be seen from the table that the three joint detection models have a better
detection effect on feature points than the local feature point detection branch alone.
Compared with the three joint detection models, the detection effect of feature points
is similar. However, model 2 only needs to train one network model, model 1 needs to
train several network models, and model 3 also uses additional annotations. Therefore,
as shown in Fig. 9, only the feature point detection effect, individual detection effect,
and the truth value of model 2 are visualized on the original image.
Fig. 9. Comparison of prediction results of local feature points between single detection and joint
detection program two.
In Fig. 9, the green dots represent actual values, the red dots represent the results of
individual tests, and the blue dots represent the results of the joint detection model. It
can be seen from the visualization effect that the result of joint detection is closer to the
actual value. The detection effect of eyes and chest is relatively accurate whether it is
detected alone or in combination, but the detection effect of legs, tail, and wings needs
to be improved.
The experimental results show that the component information has a positive effect on
the network structure of detecting local feature points, and the network model designed
in this paper combining local feature points and component detection can obtain more
330 Y. Du et al.
accurate and detailed local location information, which is also suitable for other fine-
grained classification tasks. At the same time, with effective localization of local details,
this model will provide support and help for more extended tasks of fine-grained analysis.
References
1. Liu, C., Liang, Y., Xue, et al.: Food and ingredient joint learning for fine-grained recognition.
IEEE Trans. Circuits Syst. Video Technol. 31(6), 2480–2493 (2021)
2. Seo, Y., Shin, K.S: Image classification of fine-grained fashion image based on style using
pre-trained convolutional neural network. In: 2018 IEEE 3rd International Conference on Big
Data Analysis, pp. 387–390. IEEE, China (2018)
3. Huang, H., Zhang, J., Zhang, J., et al.: Low-rank pairwise alignment bilinear network for
few-shot Fine-grained image classification. IEEE Trans. Multimedia 23, 1666–1680 (2021)
4. Liu, N., Han, J., Yang, M. H.: Picanet: learning pixel-wise contextual attention for saliency
detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 3089–3098. IEEE, USA (2018)
5. Ding, Y., Zhou, Y., Zhu, Y., et al.: Selective sparse sampling for fine-grained image recognition.
In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 6599–
6608. IEEE, USA (2019)
6. Chen, Z., Bei, Y., Rudin, C.: Concept whitening for interpretable image recognition. Nature
Mach. Intell. 2(12), 772–782 (2020)
7. Qian, Q., Jin, R., Zhu, S., et al.: Fine-grained visual categorization via multi-stage metric learn-
ing. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
pp. 3716–3724. IEEE, USA (2015)
8. Jaderberg, M., Simonyan, K., Zisserman, A., et al.: Spatial transformer networks. Adv. Neural.
Inf. Process. Syst. 28, 2017–2025 (2015)
9. Peng, Y., He, X., Zhao, J.: Object-part attention model for fine-grained image classification.
IEEE Trans. Image Process. 27(3), 1487–1500 (2017)
10. Cai, S., Zuo, W., Zhang, L.: Higher-order integration of hierarchical convolutional activations
for fine-grained visual categorization. In: Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, pp. 511–520. IEEE, USA (2017)
A Detection Network United Local Feature Points 331
11. Wang, Q., Li, P., Zhang, L.: G2DeNet: global gaussian distribution embedding network and
its application to visual recognition. In: Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, pp. 2730–2739. IEEE, USA (2017)
12. Girshick, R., Donahue, J., Darrell, T., et al.: Rich feature hierarchies for accurate object
detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, pp. 580–587. IEEE, USA (2014)
13. Zhang, N., Donahue, J., Girshick, R., Darrell, T.: Part-based R-CNNs for fine-grained category
detection. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol.
8689, pp. 834–849. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10590-1_54
14. Gao, Y., Beijbom, O., Zhang, N., et al.: Compact bilinear pooling. In: Proceedings of the
IEEE Conference on Computer Vision and Pattern Recognition, pp. 317–326. IEEE, USA
(2016)
15. Dai, X., Yue-Hei Ng, J., Davis, L.S.: FASON: first and second order information fusion
network for texture recognition. In: Proceedings of the IEEE Conference on Computer Vision
and Pattern Recognition, pp. 7352–7360. IEEE, USA (2017)
16. Zheng, H., Fu, J., Mei, T., et al.: Learning multi-attention convolutional neural network for
fine-grained image recognition. In: Proceedings of the IEEE International Conference on
Computer Vision, pp. 5209–5217. IEEE, USA (2017)
17. Sengupta, A., Ye, Y., Wang, R., et al.: Going deeper in spiking neural networks: VGG and
residual architectures. Front. Neurosci. 13, 95 (2019)
A Random Walks Image Segmentation Method
Combined with KNN Affine Class
Xiaodong Su1,2 , Shizhou Li1,2(B) , Guilin Yao1,2 , Hongyu Liang1,2 , Yurong Zhang1,2 ,
and Shirui Wu1,2
1 Harbin University of Commerce, Harbin 150028, China
suxd@hrbcu.edu.cn
2 Heilongjiang Provincial Key Laboratory of Electronic Commerce and Information Processing,
1 Introduction
In recent years, with the development of artificial intelligence, computer vision has
gradually become the direction of research. As the basis of machine vision, image seg-
mentation is widely used. Image segmentation is the technology and process of dividing
an image into several designated areas with unique properties and extracting objects
of interest. Compared with the current state-of-the-art supervised image segmentation,
unsupervised interactive image segmentation does not require a long-time training pro-
cess and only incorporates the user’s experience and subjective requirements for some
interactive sparse inputs. Typical algorithms include Grab Cut, random walk, k-means
clustering, and other algorithms.
Xia Min [1] uses random walks to segment part of the edge of the image affected by
light changes and then uses an iterative segmentation algorithm to segment the image.
Guo Li [2] and other people use background difference, morphology, and other princi-
ples to obtain the skeleton structure, obtain marker points, and then use random walks
algorithm for image segmentation to solve the problem of slow interaction in the field
of vehicle detection. Mao Zhengchong [3] and other people preliminarily selected the
marker points for random walks through the pattern mining algorithm and then seg-
mented the input image using the random walks algorithm, which can accurately and
automatically segment the image. The random walks algorithm was first proposed by
Grady L and applied to image segmentation [4]. Its advantages are fast calculation speed,
high efficiency, and good noise elimination. The basic idea of the algorithm is to cal-
culate the weighted average of the neighborhood of unknown pixels. The greater the
weight, the greater the correlation between the two. However, when the background
texture is complex, the number of neighborhood sampling points of unknown pixels is
small, leading to inaccurate results and the edge fuzzy. Therefore, how to improve the
accuracy of the foreground probability value of the random walk algorithm is a problem
worthy of study.
The KNN algorithm was jointly proposed by Cover and Hart in 1968 [5]. It is a
very mature algorithm in the field of matting and segmentation. Bai Yang [6] proposed a
matting method combining the KNN algorithm with a general sampling-based method.
Compared with a random walk, when sampling unknown pixels, the KNN algorithm
pays more attention to using points with larger weights (closer colors) to the unknown
pixels to form a gradual relationship between pixel colors, thereby forming an affine
relationship. The calculated foreground probability value will be more accurate.
This paper proposes a random walk image segmentation algorithm combined with
KNN affine class. The neighborhood search space is larger than the original algorithm by
combining the KNN algorithm in the affine class with the random walk algorithm. More
effective sample points closer to the color characteristics of unknown pixels are collected.
The experimental results show that the new algorithm can accurately and meticulously
extract the contours of the target object while integrating the user’s experience knowledge
and subjective requirements and improve the segmentation accuracy. The method in this
paper improves the segmentation accuracy of the KNN algorithm compared with the
traditional method. The random walk algorithm can improve the segmentation effects
and can increase the research values.
the seed pixel is selected as the label of the current pixel to realize the segmentation of
the image. Its basic process can be roughly expressed as: Convert the image into a matrix
composed of individual pixels [7], calculate the weighted average of the unknown pixel
and its neighborhood to obtain the foreground probability value and determine whether
it belongs to the front scenic spot or the background point, weighted average The value
di is expressed as
di = wij (1)
In the formula (1), wij is the weight between the unknown pixel point i and its
neighboring point j. The greater the weight, the closer the relationship between the two.
Random walk algorithms are usually used in binary segmentation, color supplements,
and other fields with low boundary requirements. However, in the field that requires the
accurate calculation of the foreground probability value, there are obvious defects and
cannot reflect the details [8].
αi = ω1 α1 + · · · + ωj αj + · · · + ωk αk i, j = 1, 2, · · · k (2)
The αi of all unknown points i can be obtained by solving the following large sparse
linear equation
(L + θ D)x = θ b (3)
Among them, L is a sparse square matrix whose length and width are both the total
number of image pixels. The coefficient of the K neighborhood corresponding to the row
of the unknown point i is denoted as ωj , j = 1, ···, k, D are diagonal The position of the
diagonal element corresponding to the known point is 1, the unknown point is 0, and θ is
a larger and b is a column vector whose length is equal to the total number of pixels. x is
the foreground probability value of the unknown pixel [10]. Compared with the method
of calculating each point in the random walk algorithm, it emphasizes the connection
of adjacent pixels. Therefore, the result of the final calculated foreground probability
value and the smoothness and the visual experience to the user are obviously better than
random Sampling algorithms such as wandering. This is in sharp contrast with the result
of the foreground probability value obtained by the random walk algorithm [11], which
usually has more noise points and the overall distribution is more messy [12].
A Random Walks Image Segmentation 335
4 Algorithm Combination
4.1 Sparse Matrix of Random Walks
The user first enters the front sights and background points, sets the foreground prob-
ability to 1, and the background probability to 0. Figure 1 shows the front sights and
background points entered by the user. The algorithm judges the unknown points by
solving the unknown pixel points in the sparse matrix. It belongs to the former scenic
spot or the background spot.
In the figure, red and blue are the front scenic spot and background point entered by
the user.
In a sparse matrix, if the number of elements with a value of 0 is far more than the
number of non-zero elements, and the distribution of non-zero elements is irregular, the
matrix is called a sparse matrix. In the random walk algorithm, 0 elements are input by
the user Background points and unknown pixel points, and non-zero elements are the
previous scenic spots entered by the user. Assuming that the image size is m × n, the
sparse matrix Arw of t × t(m × n = t) is constructed. The elements of t/2 rows in the
matrix are the weights between the unknown pixel i and its neighborhood point j, and
the diagonal elements are the sum of the weights between the unknown pixel i and its
four neighborhoods. The foreground seed point and the background seed point are set
corresponding to the rows of the matrix, and the other matrix elements are 0. The weight
formula of the random walks algorithm is
wij = exp −Ii − Ij /σ (4)
In the formula, wij is the weight between the unknown pixel point i and its neighbor
point j, Ii is the pixel value of point i, Ii − Ij is the Euclidean norm between point
i and point j, and σ is the only free parameter in the algorithm. The closer the gray
values of two adjacent pixels, the greater the weight between the two points, and the
greater the connection, when solving Eqs. (4), only the probabilities of foreground points,
background points and points to be solved need to be calculated. The larger the calculated
weight between two adjacent pixels, the closer the connection between the two pixels.
336 X. Su et al.
(1) Select the first K samples with the smallest distance to search for the spatial distance
(K selected 10 in this article).
(2) The random k-d forest algorithm in the FLANN search library is used to select
points with similar features. The key of this algorithm is to find the closest point
in the neighborhood of the unknown pixel. The feature similarity defined by the
Euclidean distance is
d
D(x, y) = Xi2 + Yi2 (5)
i=1
In the formula, Xi represents the feature of the unknown pixel, and Yi represents the
feature of the neighborhood point.
If the obtained D value is smaller, the features between the feature point pairs are
closer, and the degree of similarity is higher. An example of its spatial search is shown
(see Fig. 2).
In the formula, Ri , Gi , Bi are color features, and Xi and Yi are the coordinates of
point i.
A Random Walks Image Segmentation 337
In the formula, xi is the feature of unknown point i; xik is the feature of K neighboring
points; c is a weight coefficient, which is the upper bound of xi − xik .
In the formula, Arw + AKNN is the sparse matrix after the addition of matrices Arw
and AKNN , D is the diagonal matrix, In the diagonal matrix, the diagonal element of the
known seed point (including the foreground and background seed points) is 1, otherwise,
it is 0. b is a column vector, its length is equal to the total number of vectors, 1 at the
foreground seed point, otherwise 0, θ is the constraint parameter, so that the value of x
is between 0 and 1. The x solved by the solution of the linear equation is the foreground
probability value. When x is greater than 0.5, it is judged as the foreground point, and
vice versa, it is judged as the background point, so as to complete the binary segmentation
of the image.
In this formula, m is the width of the image; n is the height of the image; E(i, j) is
the result of segmentation algorithm; H (i, j) is true results for training sets [15].
338 X. Su et al.
This section evaluates the results of image segmentation from a visual perspective. First,
select the full moon image to analyze the segmentation results (see Fig. 3). Ground
Truth is the real foreground image of the full moon, Trimap is the front scenic spot
and background point manually marked (white is the front scenic spot, black is the
background point), the first image in the first line is the image to be segmented; The
three images in the second row are the results of gray images output by KNN, random
walks, and RW_KNN (the method in this paper) algorithm, respectively; The third line is
the binary segmentation image output by three algorithms. The color difference between
the foreground and background of Fig. 3 is large. The hard border of the full moon image
is rounded, and the image segmentation is not difficult. From the output results, it can
be seen that the three algorithms can extract the foreground segmentation of the object.
Still, in the outline for some external boundary details, the performance of the three
algorithms is not the same. KNN algorithm has the worst performance.
The segmented results show over-segmentation and under-segmentation, and the
target image cannot be seen. However, the target contour segmented by the random
walk algorithm is relatively clear, conforming to its fine segmentation effect on the hard
boundary. The RW_KNN algorithm in this paper has the best effect. Compared with the
previous two algorithms, it can accurately segment the edge of the full moon.
Then the bush image with complex background texture was selected for segmentation
analysis (refer to Fig. 4). The feature of this image is that the foreground and background
are both close to the green, and the segmentation difficulty is greater than that of a
full moon image. The trimap in the first row is the three-part image of the original
image, where the foreground and background point probabilities have been marked in
advance, and Ground Truth is the real foreground image of the bush; The three images
in the second row are grayscale images output by the KNN, random walk, and RW-
KNN algorithms; the third row is the binary segmentation image output by the three
algorithms. It can be seen from the figure that the three algorithms can segment the
foreground and background of the object very well. Still, for the segmentation results of
the image edge, the performance of the KNN algorithm and the random walk algorithm is
slightly worse than that of the RW-KNN algorithm. The random walk algorithm has the
worst effect among these three algorithms (the image in the middle of the second row in
Fig. 4). The bottom flowerpot of the bush image divides most of the background into the
foreground. This is also the calculation weight of random walk. The problem was caused
by the limited sampling point range of the unknown point neighborhood. Although the
KNN algorithm can segment the rough outline of this part of the foreground, there are
still many background pixels divided into the foreground, resulting in a decrease in
segmentation accuracy. The algorithm in this paper (RW-KNN) can accurately segment
the background. The algorithm in this paper combines the KNN algorithm in the affine
class when sampling to accurately segment the bushes, which improves the accuracy of
the foreground probability value.
Finally, the person image with the characteristics of soft hair boundary was selected
to analyze the segmentation situation (refer to Fig. 5). The feature of this image is that
there are soft borders of hair in the image, the edge details are more complicated than the
previous two images, and the segmentation is more difficult. The edge details are more
complex than Figs. 3 and 4, and the segmentation is more difficult than the previous
two images. The first line of trimap is a three-parted graph of the original image, and
Ground truth is the person’s true foreground image. The three images in the second
row are the output of the gray image by KNN, random walks and RW_KNN algorithm,
respectively. The third line is the binary segmentation image output by three algorithms.
It can be seen that the performance of the random walk algorithm is the worst, and the
image of the soft boundary causes the image segmented by the random walks algorithm
to be unable to recognize the segmentation target intuitively. The overall contour is
not completely segmented. KNN algorithm performs better and can roughly extract the
contour of the person, but the division error still causes over-segmentation phenomena.
The performance of the proposed algorithm is the best person head contour segmented by
the algorithm is more realistic, which improves the segmentation effect of soft boundary
and can completely divide the foreground region.
340 X. Su et al.
Section 5.1 evaluates the image segmentation results from the visual point of view. This
section used mean MSE and SAD to evaluate the image segmentation results. The MSE
and SAD of the above three images are listed in Table 1 and Table 2. It could be seen
that the mean square error and absolute error of this algorithm were smaller than the
other two algorithms; that is, the results of this algorithm were closer to the real results
than the original algorithm.
In order to evaluate more objectively, this paper calculates the mean MSE and SAD
of the three images and calculates the average value. But to avoid contingency, this
experiment calculates the error of each algorithm for 50 images in the dataset with the
same trimap, and the results are listed in Table 3. The error caused by this algorithm was
less than the other two algorithms.
A Random Walks Image Segmentation 341
6 Conclusions
Based on the random walk algorithm, this paper proposes a random walks image seg-
mentation method combined with KNN affine class. Using the KNN algorithm in the
affine class, you can search for more points in the space that are similar to the color char-
acteristics of the unknown pixel. Combine the random walk algorithm and the sparse
matrix formed by the KNN algorithm with the sparse random walk algorithm. After the
matrices are added, it can make up for the shortcomings of a small search space in the
neighborhood of unknown pixels and less pixel feature information when calculating
the weights of random walks. The experiment in this paper is compared with the KNN
algorithm and the random walk algorithm, which effectively improves the segmentation
accuracy of the random walk algorithm and has a better segmentation effect on the seg-
mented image with soft boundaries such as hair. Compared with the traditional random
walk and KNN algorithm, the segmentation results of the improved random walk algo-
rithm on the Berkeley data set have a greater improvement in vision and accuracy. The
experimental results show that the method in this paper has a good processing effect, has
research value in segmentation accuracy in the image field, and has a certain use-value
in batch processing of image segmentation projects.
References
1. Xia, M.: Iterative segmentation algorithm based on random walk. Digital Technol. Appl.
37(09), 120–122 (2019)
2. Guo, L., Gao, L., Liu, M., et al.: A new random walk multi vehicle detection algorithm.
Chinese J. Image Graph. 3, 392–396 (2010)
3. Mao, Z., Han, Y.: Automatic image segmentation algorithm based on random walk. Sens.
Microsyst. 37(06), 142–145 (2018)
4. Gradyl: Random walks for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell.
28(11), 1768–1783 (2006)
5. Chen, Q., Shen, Q., Zhang, K., Liu, P.: Edge preserving KNN matting. Minicomput. Syst.
36(07), 1591–1596 (2015)
6. Bai, Y., Yao, G.: A robust matting method based on KNN post-processing. Comput. Appl.
Softw. 37(09), 170–175 (2020)
7. Zhou, Y., Chen, J., Kan, Y.: An automatic image segmentation algorithm based on random
walk. Inf. Technol. 2015(12), 75–79 (2015)
A Random Walks Image Segmentation 343
8. Chen, Q., Li, D., Tang, C.: KNN matting. In: Proceedings of the 2012 IEEE Conference on
Computer Vision and Pattern Recognition, pp. 2175–2188. IEEE (2012)
9. Wang, F., Qin, F., Jiang, D., Song, C.: Random walk image segmentation based on visual
attention. Chin. J. Sci. Instr. 38(07), 1772–1781 (2017)
10. Yao, G., Liu, S., Li, M., et al.: Application and analysis of post-processing in digital matting.
Acta Electronica Sinica 45(3), 719–729 (2017)
11. Yao, G., Yao, H.: Overview of image matting algorithm based on affine method. J. Comput.
Aided Design Graph. 28(4), 677–692 (2016)
12. Yu, X., Zhou, N., Zhang, F.: Research on Image automatic classification model based on
KNN. J. Libr. China 01, 74–76 (2007)
13. Nie, D., Wang, L.: Improved KNN matting technology. Minicomputer Syst. 36(06), 1316–
1320 (2015)
14. https://www2eecs.berkeley.edu/Research/Projects/CSision/bsds/.last. Accessed. The Berke-
ley Segmentation Dataset and Benchmark Homepage (2020)
15. Wang, F., Qin, F., Jiang, D., Song, C.: Random walk image segmentation based on visual
attention. J. Instrum. 38(07), 1772–1781 (2017)
Generative Adversarial Network Image
Inpainting Based on Dual Discriminators
1 Introduction
Image inpainting [1–3] refers to the use of known contextual information to repair
missing areas of the picture. It is widely used in restoring images and areas, such as
removing targets and dispersing them. Although many researchers have proposed many
image inpainting methods, both deep and non-depth methods, such as patch-based [4]
Image inpainting methods, context encoder [5] methods, etc., image inpainting is still a
hot task in computer vision and computer graphics today. This is because it is tough to
synthesize textures with more detail and visual clarity and make the computer understand
the semantic information in the image. Based on this phenomenon, we have considered
the skip connection in the network, which allows the contextual information of shallow
images to be fused with deeper semantic information.
Our work is based on the Context Encoder model put forward by Pathak et al. [5]
in 2016, which is the first model that uses generative adversarial network(GAN) [6]
regarding image inpainting. And the method can only process 128*128 images, which
has some limitations.
We use a fully convolutional neural network as the basis of our model and propose
a dual generator that fuses dilated convolution [7] and residual blocks [8] as a way to
repair to obtain higher quality and visually better images. Our model from four networks:
a rough inpainting network, a refinement network, a local discriminator network, and
a global discriminator network. Two generators are used to complete the images, and
both global and local discriminators are auxiliary networks dedicated to training. These
discriminators are used to discriminate whether an image is real or not. The global
discriminator uses the whole image as input to identify consistency across the image,
thus ensuring consistency and continuity in and around the area to be repaired. In contrast,
the local discriminator looks at only a small repair area to discern whether the local details
are perfect.
We evaluate our models against other deep learning-based models in specific street
scenes and faces, and our models both have smaller loss values and higher peak signal-
to-noise ratios.
In summary, in this paper, we put forward that:
• Propose a generative model with skip connection to fuse the contextual information
of the image with higher-level semantic information so that our model can understand
the semantic information better than other models;
• To increase the perceptual field, both generators are using dilated convolution;
• The generator also uses a convolution layer with residual blocks to avoid gradient
disappearance and speed up convergence.
2 Related Work
Currently, image inpainting methods fall into two types: traditional algorithms and deep
learning-based approaches.
Traditional image restoration methods can be mainly classified into diffusion-based
and sample-based methods. Diffusion-based methods mainly include the partial differ-
ential equation-based methods proposed by Bertalmio. With these methods, the user
needs to specify the area to be repaired. The algorithm diffuses the information outside
the contour normal to the boundary of the area to be repaired to the middle pixel to
be repaired along the contour. The algorithm uses the smoothness of the local color to
diffuse along the contour, taking into account anisotropic diffusion to ensure that the
boundary is continuous at the edges, but the method is computationally unstable.
In contrast, sample-based methods assume that known samples can represent miss-
ing regions of an image. These methods mainly include texture synthesis-based image
inpainting algorithms. One way of thinking about texture synthesis-based image inpaint-
ing algorithms is to decompose the image into structural and textured parts. The structural
part is restored using a variational PDE-based method and the textured part filled with
a texture synthesis method. This method can achieve better inpainting results for large
areas of broken images by filling in missing regions of arbitrary size and repairing texture
details in broken parts.
Conventional algorithms tend to have better results when the inpainting area is small
and the texture structure is relatively simple. Once the missing area is relatively (30%
346 J. Sun et al.
or more) large, inpainting results tend to be particularly poor. This is because traditional
image inpainting methods often fail to understand the semantic information in the image
at a deep level. Therefore, with the development of deep learning after 2015, more and
more researchers working in image inpainting are adopting deep learning methods to
obtain a deeper understanding of semantic information and higher quality inpainting
results.
Image inpainting for deep learning is implemented using generative adversarial net-
works. For example, Pathak et al. proposed the Context Encoder (CE) model, which is
similar to the autoencoder network [9] in that it consists of an Encoder-Decoder, with
the generator consisting of the CE network, taking the first five layers of AlexNet [10]
as the encoder. The decoder is implemented by deconvolution to transform from high-
dimensional features to the true size of the image. The discriminator consists of a local
discriminator. Both the generator and discriminator structures of the Context Encoder
model are relatively simple, and the complementary results, although more realistic,
have very unsmooth boundaries and do not satisfy local consistency. Iizuka et al. [11]
improved the Context Encoder model by using double discriminators to overcome this
problem. Traditional image inpainting methods are good at sampling background images,
while CNN models generate new textures. To comprehensively utilize the advantages
of these two methods and make full use of redundant information in pictures, Yu et al.
[12] proposed a contextual attention method for image inpainting, which usually adopts
two steps from coarse to fine. The first step is rough inpainting, and the second step is
to find image blocks similar to those in occluded areas for improvement.
Although the above ways have shown good results in image inpainting tasks, there
are many shortcomings: inaccurate inpainting when the texture information is complex;
blurring at the edges; low semantic consistency.
Therefore, we propose a generative adversarial model incorporating dilated convo-
lution and residual blocks to address the problems of unclear edges, poor texture, and
semantic consistency after inpainting. The model consists of a rough repair model and a
refinement model. The image with the mask is used as the input. After the first repair by
the rough network, the image after the first repair is used as the input of the refinement
network to generate a more detailed and realistic image. The reconstruction loss and the
weighted loss function of the adversarial loss are used to optimize the training results of
the model. The skip connection in the rough network fuses the contextual information of
the image with higher-level semantic information, so the rough and refined networks with
skip connection proposed in this paper can generate images with clear edges, textural
structure, and visually realistic results.
3 Network Model
The model proposed in this paper (shown in Fig. 1) consists of two main components: a
generator and a double discriminator. The generator refers to the rough repair network
with skip connection and dilated convolution (shown in Table 1) and the refined net-
work with dilated convolution and residual blocks (shown in Table 2). In contrast, the
double discriminator refers to the global discriminator (shown in Table 3) and the local
discriminator (shown in Table 4) to ensure global consistency and local consistency. The
Generative Adversarial Network Image Inpainting 347
rough process is as follows: the image with the mask is used as input to generate a coarse
restored image through the first layer (coarse network). Then the coarse restored image
is used as input to generate a finer image through the second layer (refined network). The
global discriminator and local discriminator are then used to determine the integrity of
the whole image and small areas, respectively, to achieve global and local consistency.
Table 1. The rough network architecture, in particular the up-sampling before the thirteenth and
fifteenth convolutional layers, and the convolutional kernel with 128/64 convolutional steps of 3*3,
is similar to the effect of deconvolution, both to increase the size of the image and the following
refinement network has a similar operation.
Table 1. (continued)
This paper uses four dilated convolutions in both generator networks to achieve
different sizes of perceptual fields by setting different void rates (2, 4, 6, 8) to obtain
multi-scale information.
4 Experiment
4.1 Datasets
To evaluate the effectiveness of the inpainting method in this paper, we used images
from two databases to train and evaluate our dual generator model: Paris StreetView and
CelebA. Paris StreetView contains 62,058 high-quality street view images with accurate
GPS coordinates and compass directions for each street view landmark. CelebA contains
202,599 face pictures of 10,177 celebrities. We set the mask to the central region and
the coverage to 25% of the whole image.
In this paper, the effectiveness is evaluated on the Paris StreetView and CelebA datasets. It
is also compared with representative image inpainting methods based on GAN in recent
years (generative adversarial model with context encoder and generative adversarial
model with contextual attention). This paper uses PSNR and SSIM for quantitative
evaluation.
The PSNR is calculated as follows:
MAXX21
PSNR(X1 , X2 ) = 10 lg( ) (4)
MSE
Where X1 and X2 are the two images, MAXX21 is the maximum possible pixel value
of the image. Each pixel is represented by an 8-bit binary, so the maximum pixel is taken
as 256, and MSE represents the mean squared error of X1 and X2 .
The model is trained and evaluated on CelebA and Paris StreetView data sets. The
model is trained by a large amount of data. Finally, the average PSNR and SSIM of the
model on different data sets are tested and calculated and compared with the generative
confrontation model based on context encoder and the generative confrontation model
based on context attention. The average pairs of PSNR and SSIM of the above three
repair methods on data sets are shown in Table 5 and Table 6.
352 J. Sun et al.
From Tables 5 and 6, it can be seen that the mean values of PSNR and SSIM of this
method are higher than those of the other two methods, and the experiments prove that
the restoration results of this algorithm on the CelebA and PSV datasets are better than
those of the other two restoration methods.
The restoration results of the way in this paper on the Paris StreetView and CelebA
datasets are shown in Fig. 3. From Figure., it can be seen that the method based on
Context Encoder can achieve good results on the textures but is too blurred due to the
addition of L2 loss for training. The method based on Contextual Attention with the
addition of an attention mechanism does not work well on the five senses after repair,
and even the glasses are not repaired. However, the image restored by the method in this
paper is the same as the real image in terms of color, structure, and style, and the texture
details of the image are built more accurately and clearly; the boundary between the
missing area and the background area is natural and clear, the visual effect is consistent,
and there is no trace of restoration. This shows that the method in this paper has a good
effect on the image restoration task.
Generative Adversarial Network Image Inpainting 353
5 Conclusion
We propose a GAN with dual discriminators fusing dilated convolution and residual
blocks. The network uses skip connection in the rough repair network after having
two generators and two discriminators, which makes the contextual information fused
with the high-level semantic information to improve the semantic consistency better;
proposes the use of multiple dilated convolutional layers with different dilated rates
in the two generators to obtain various perceptual fields and multi-scale information;
also proposes the use of residual blocks in the fine-tuning network, which To a cer-
tain extent, it avoids the problems of gradient disappearance and too slow convergence
due to the complexity of the model; the loss functions using reconstruction loss and
adversarial loss weighting are trained together with the generative network and the dual
discriminant network to enhance the global and local semantic consistency of the repair
region. Qualitative and quantitative analyses were conducted on the CelebA and Paris
StreetView datasets, respectively. The experimental results show that the model pro-
posed in this paper, after inpainting, results in clear images, well-structured textures,
and high semantic consistency.
Acknowledgments. This work is supported by the grants of the Natural Science Foundation of
Heilongjiang Province of China (No. LH2020F007), Heilongjiang philosophy and Social Sci-
ences project (Grant Nos.18GLB029), and Harbin University of Commerce young creative talents
Support project (Grant Nos. 2019CX02).
References
1. Bertalmio, M., Sapiro, G., Caselles, V.: Image inpainting. In: Proceedings of the 27th Annual
Conference on Computer Graphics and Interactive Techniques, pp. 417–424. New Orleans,
LA, USA (2000)
2. Shen, J.H., Chan, T.F.: Mathematical models for local nontexture inpaintings. SIAM J. Appl.
Math. 62(3), 1019–1043 (2002)
3. Criminisi, A., Pérez, P., Toyama, K.: Region filling and object removal by exemplar-based
image inpainting. IEEE Trans. Image Process. 13(9), 1200–1212 (2004)
4. Huang, J., Kang, S., Narendra, A.: Image completion using planar structure guidance. ACM
Trans. Graph. 33(4), 1–10 (2014)
5. Pathak, D., Krähenbühl, P., Donahue, J.: Context encoders: feature learning by inpainting.
In: Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las
Vegas, NV, USA, pp. 2536–2544 (2016)
6. Goodfellow, I.J., Pouget-Abadie, J., Mirza, M.: Generative adversarial nets. In: Proceedings of
the 27th International Conference on Neural Information Processing Systems, pp. 2672–2680.
Montreal, QC, Canada (2014)
7. Yu, F., Koltun, V.: Multi-scale context aggregation by dilated convolutions. In: ICLR (2016)
8. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016
IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016). https://
doi.org/10.1109/CVPR.2016.90.
9. Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning internal representations by error
propagation. In: Rumenhart, D.E., McCelland, J.L. (eds.) Parallel Distributed Processing:
Explorations in the Microstructure of Cognition, pp. 318–362. MIT Press, Cambridge (1986)
Generative Adversarial Network Image Inpainting 355
10. Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional
neural networks. In: NIPS. Curran Associates Inc. (2012)
11. Iizuka, S., Simo-Serra, E., Ishikawa, H.: Globally and locally consistent image completion.
ACM Trans. Graph. 36(4), 107 (2017)
12. Yu, J.H., Lin, Z., Yang, J.M.: Generative image inpainting with contextual attention. In:
Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition,
Salt Lake City, UT, USA, pp. 5505–5514 (2018)
13. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image
recognition. Computer Science (2014)
14. Kingma,D., Ba, J.: Adam: a method for stochastic optimization. Computer Science (2014)
GrabCut Image Segmentation Based on Local
Sampling
1 Introduction
Image segmentation is an important part of the field of computer vision. It is a process
of dividing several pixels of an image into several specific regions. The purpose of
image segmentation is to classify the pixels of the image. With the development of
artificial intelligence, image segmentation technology based on deep learning has been
widely used in image segmentation. However, traditional image segmentation methods
still occupy a place in image segmentation due to their characteristics. Traditional image
segmentation technology has the following methods: direct segmentation based on image
threshold, segmentation based on the growth of seed pixel area, segmentation based
on edge detection, segmentation based on graph theory, and so on. GrabCut image
segmentation algorithm is an image segmentation technology based on graph theory.
The basic principle of graph theory will be introduced below.
2 Related Work
The image segmentation method based on graph theory [1] maps the image into a
weighted undirected graph. Each pixel in the image is a node of the directed graph,
and the adjacent pixels are the nodes of the undirected graph connected by edges. The
weight of the edge represents the relationship between pixels. The cutting capacity is
represented by the energy function, which is the sum of the weights of all edges in the
cutting. A certain cutting algorithm cuts the graph, and then the graph is divided into a
number of different sets to minimize the energy function. In Graph Cut, two terminal
vertices are additionally added, as shown in Fig. 1.
Fig. 1. A graph-cut has two terminal vertices, a source point and a sink point. Each vertex of the
image has an edge connected to the adjacent vertices and an edge connected to the source point
and the sink point. The blue line is the connection between each vertex and the source point. The
red line is the connection between each vertex and the sink point.
The source point can be expressed as the foreground and background segmentation,
and the sink point can be represented as the background. The weight of the connection
between each vertex and the terminal vertex on the graph can be expressed as the cost of
a vertex belonging to the foreground or the background. The connection between each
vertex reflects the color relationship between pixels. If we can find a cut to break these
edges, the undirected graph will be divided into two disjoint subsets: the foreground and
the background. We need to find a minimum cut set with the smallest energy to minimize
the sum of the weights of the two sets of edges.
In 2001, Boykov [2] proposed the Graph Cut, a segmentation method based on graph
theory. In 2004, Rother [3] proposed the GrabCut algorithm based on Graph Cut. GrabCut
is an interactive image segmentation technology based on Graph Cuts. It transforms the
gray histogram processing into RGB image processing based on graph cut. Firstly, the
GMM model [4] is used to describe the pixel distribution of the image’s global known
foreground and background, and the max flow min cut [5] algorithm is used to cut the
image. Finally, multiple iterations are performed to minimize the energy function to
obtain the final segmentation result. The algorithm can achieve better results when the
foreground and background color distribution is different, but the effect is poor when
the color difference is small. Others improved the algorithm for this problem. Yi Cong-
cong [6] first downsamples the image, calculated GMM parameters with representative
pixels, and then performed watershed segmentation [7] to solve the problem of similar
foreground and background colors. However, morphological operations are still needed.
Ling Bin [8] used the image’s depth information to construct super-pixels [9] with a
linear iterative clustering algorithm to generate a structural graph model. The algorithm
358 G. Yao et al.
improved the segmentation efficiency and result, but it is not suitable for the image with
less depth information.
This paper proposes GrabCut, which collects foreground and background samples
based on local sampling and then conducts local clustering. When collecting samples, we
collect samples according to the distance from unknown pixels to the known foreground
and background pixels so that we can pay more attention to the local information around
unknown pixels and avoid the problem of color information loss to a certain extent when
the foreground and background colors are similar.
When calculating the GMM model of foreground and background, the GrabCut algo-
rithm obtains global foreground and background samples. In the GMM iteration process,
each pixel is assigned to a Gaussian component of the foreground or background GMM.
The weights and parameters of the GMM are updated after the iteration. However, in the
process of clustering and iteration, once some background pixel clusters or foreground
pixel clusters are too close, unknown pixels are easily assigned to the wrong pixel clus-
ters, which leads to inaccurate classification results. In the clustering process, if some
background pixels are not significantly different from other background pixels or are
similar to foreground pixels, the difference information is easily lost in the iterative pro-
cess. Therefore, the Gaussian mixture model will focus on the overall color distribution
and ignore the color information.
In this paper, a sampling method in [10] is proposed. We propose a GrabCut method
based on local sampling. It is considered that when judging whether an unknown pixel
belongs to the foreground or the background, the pixel closer to itself should be found as
the sample. Therefore, the unknown pixel collects surrounding foreground samples or
background samples as a sample set for local clustering instead of constructing a GMM
of global foreground and background pixels. The calculation of the region term remains
unchanged, and the max flow min cut can obtain the image segmentation result without
multiple iterations to minimize the energy.
The search range can be determined by the distance between the sample points and the
unknown points and the number of sample points to be collected when collecting sample
points. Since the color change between adjacent pixels will not be very drastic, adjacent
sample points can be collected at intervals. If the color change is not severe, fewer sample
points represent the local color distribution. The color components in the clustering can
better reflect the details of local color changes in the image. The missing local color
information of the original global GMM model is captured in the local clustering. At the
same time, this also means that foreground and background clustering is established for
each unknown pixel, instead of sharing a foreground GMM and a background GMM for
all unknown pixels like the original GrabCut algorithm. The Gaussian clustering model
based on local sampling does not need to calculate parameters many times, and only one
iteration can obtain good results.
GrabCut Image Segmentation Based on Local Sampling 359
Fig. 2. This is a schematic diagram of global and local sampling, where the red dots represent
the sampling points. The left figure shows the global sampling adopted by the GrabCut algorithm,
which samples all known sample points in the global range. The right figure is the local sampling
adopted by our method, and it samples in the local range around unknown pixels without using
all known sample points.
Local sampling ensures that each color distribution within the sampling range can
be represented in the sample set. We can also weigh the sample points according to the
distance. This is because for unknown pixels close to the boundary, generally, the closer
the boundary, the higher the correlation between pixels in the known area. In addition,
the farther away the known region is from the boundary, the more difficult it is for its
pixels to establish a correlation with any unknown pixels. Therefore, when we want to
collect more comprehensive samples, we can follow the above sampling principles.
First, divide the sample color as the parent node to get its own two child nodes, and
then divide the child nodes. Like its parent node, the child nodes of any node are also
divided into two groups. By generating such a binary tree, the pixels in the sample color
can be represented by the pixels in the leaf nodes. The following describes the specific
color separation method.
Suppose that the pixel group of the initial sample is represented as Cn , xs is the sample
pixel, and by the following equation, Rn is a 3 × 3 matrix, mn is a vector of 3 × 1.
Rn = xs xst (1)
s∈cn
mn = xs (2)
s∈cn
Nn = |Cn | (3)
Since the mean value of the group is the point deviating from the least square, it is
assumed that the quantitative representation of the group is the mean value of the group
qn .
mn
qn = (4)
Nn
Define the covariance of this pixel group as
1
Rn = Rn − mn mtn (5)
Nn
According to the structure of binary tree, the pixel group is divided into two groups,
which are similar to the two children of the parent node. The intuitive idea is to select a
plane that can best divide the color group to divide it. Therefore, to determine the direction
of the largest color change, divide it into two subsets with a plane perpendicular to that
direction and passing through the average value of the group. In order to obtain the
direction vector e of this pixel group, there is the following equation
2
(xs − qn )t e = et Rn e (6)
s∈Cn
Since the direction of the main eigenvector of the matrix is the direction in which
it changes the most. Therefore, the eigenvector solution en of this pixel group Rn is its
principal eigenvector λn .
2
λn = (xs − qn )t en (7)
s∈Cn
When en is solved, Cn is further divided into two subsets C2n and C2n+1 by the
following equation.
C2n = s ∈ Cn : ent xs ≤ ent qn (8)
C2n+1 = s ∈ Cn : ent xs > ent qn (9)
GrabCut Image Segmentation Based on Local Sampling 361
So Cn has its two child nodes C2n and C2n+1 . R2n , m2n and N2n is calculated according
to the Eqs. (1), (2), (3). R2n+1 , m2n+1 and N2n+1 is calculated according to the following
equation.
R2n+1 = Rn − R2n (10)
Fig. 3. After the above steps, the pixel groups with similar colors will be divided into the same
Gaussian cluster. The red points represent sample points, and the blue ellipses represent different
Gaussian clusters.
Finally, all leaf nodes of the binary tree are clusters of sample pixels and the local
Gaussian clustering of sample colors are obtained.
In Eq. (13), U represents the data term, represents the cost of an unknown pixel
belonging to foreground GMM or background GMM. The cost is obtained by the Gaus-
sian mixture model. It can be specifically expressed as the probability that each unknown
pixel is assigned a color component belonging to the foreground GMM or background
GMM.
U(α, k, θ, z) = D αn, kn , θ, zn (14)
n
D(αn , kn , θ, zn ) = −logπ (αn , kn ) + 21 logdet (αn , kn )
(15)
+ 21 [zn − μ(αn , kn )]T (αn , kn )−1 [zn − μ(αn , kn )]
θ = π(α, k), μ(α, k), (α, k), α = 0, 1, k = 1 . . . K (16)
In the above equation, π is the weights,μ is the mean, and the is the covariance
of the Gaussian components.
V in Eq. (13) is the smooth term, which represents the neighborhood relationship
of pixels, and it is obtained from the Euclidean distance between the color spaces of
adjacent pixels.
⎛ ⎞
V(α, z) = γ⎝ [αn = αm ]exp − βzm − zn2 ⎠ (17)
(m,n∈C)
After several iterations, the parameters of Gaussian mixture model are optimized,
and the energy function is minimized by max flow min cut.
At this time, the minimum energy function assigns a label belonging to the foreground
or background to each unknown pixel, so as to obtain the result of image segmentation.
Our Experimental platform is Win 10 Home edition with Microsoft Visual Studio 2019
using OpenCV 4.3.0. CPU is Intel(R) Core(TM) i7-10750H 2.60 GHz. Refer to [13],
our method is implemented by C++ and named GrabCutLocal. In this paper, the exper-
iment is conducted on the Berkeley data set [14]. In order to verify the effect of the
improved algorithm and the original algorithm, an intuitive comparison was made after
the experiment. The experimental results are shown in Fig. 4.
It can be seen that GrabCut has some defects in the segmentation of the corners
in (a) image. GrabCutlocal has a relatively complete split on the edges and corners. In
(b), GrabCut omits part of the tail, and GrabCutLocal splits the entire tail. For (c), the
segmentation effect of GrabCut in helmet and shoes are not very good. In (d), part of the
background is retained in the nose, and part of the brain is regarded as the background
in (E). Therefore, when segmenting these images, GrabCutLocal can clearly distinguish
GrabCut Image Segmentation Based on Local Sampling 363
the foreground from the background. Since the foreground and background colors are
similar in the above images, the GrabCutLocal method shows obvious advantages.
Objectively, the Sum of Absolute Differences (SAD) is used to evaluate the error,
where m and n are the width and height of the image. E(i, j) is the result of the algorithm
segmentation, and H(i, j) is the true value of the segmentation provided by the data set.
m
n
SAD = (E(i, j) − H (i, j)) (19)
i=1 j=1
Fig. 4. Among the five typical images in the experimental results, (a), (b), (c), (d), and (e) are five
sets of images. The first line is the original image in the dataset. The second line is the trimap. The
third line is the segmentation result of the original GrabCut. The fourth line is the segmentation
result of our GrabCutLocal. The red circle is a clear contrast between the two algorithms.
5 Conclusion
Based on the improvement of the GrabCut image segmentation algorithm, this paper
proposes a GrabCut image segmentation algorithm based on local sampling, that is,
foreground and background samples come from local sampling around unknown pixels,
instead of selecting global foreground and background as samples. Since the global
foreground and background are used for modeling, the GrabCut algorithm pays more
attention to the overall color information. It is easy to ignore the local color information.
GrabCut can achieve good results when the foreground and background are significantly
different. Still, it is not ideal when the foreground and background are similar, and there
is much color overlap. This paper proposes a GrabCut image segmentation algorithm
based on local sampling, which uses local samples for clustering to detect more color
details and avoid losing color information. The experimental results show that when
the foreground and background sampling accuracy is higher, and the foreground and
background colors are more similar, the local sampling GrabCut algorithm can achieve
better results than the original algorithm. This method effectively improves the accuracy
of GrabCut image segmentation and has a certain application value.
References
1. Srinivas, B., Manjunathachari, K.: New approach for image segmentation based on graph
cuts. Int. J. Signal Process. Image Process. Pattern Recogn. 10(1), 119–130 (2017)
2. Boykov, Y., Veksler, O., Zabih, R.: Fast approximate energy minimization via graph cuts.
In: Proceedings of the Seventh IEEE International Conference on Computer Vision, vol. 1,
pp. 377–384. IEEE, Kerkyra (1999)
3. Rother, C., Vladimir, K., Blake, A.: “GrabCut”: Interactive foreground extraction using
iterated graph cuts. ACM Trans. Graph. 23(3), 309–314 (2004)
4. Blake, A., Rother, C., Brown, M., Perez, P., Torr, P.: Interactive image segmentation using
an adaptive gmmrf model. In: Pajdla, T., Matas, J. (eds.) ECCV 2004. LNCS, vol. 3021,
pp. 428–441. Springer, Heidelberg (2004). https://doi.org/10.1007/978-3-540-24670-1_33
GrabCut Image Segmentation Based on Local Sampling 365
5. Leighton, T., Rao, S.: Multicommodity max-flow min-cut theorems and their use in designing
approximation algorithms. J. ACM 46(6), 787–832 (1999)
6. Yi, C., Wu, B., Zhang, H.: An improved grabcut image segmentation method. J. Chin. Comput.
Syst. 35(5), 1164–1168 (2014)
7. Xu, Q., Guo, M., Wang, Y.: Fast color image segmentation based on watershed transform and
Graph cuts. Comput. Eng. 35(19), 210–212 (2009)
8. Ling, B., Guo, Y.: GrabCut image segmentation combing color and information. Comput.
Appl. Softw. 37(8), 188–193 (2020)
9. Hu, Z., Guo, M.: Fast segmentation in color image based on SLIC and GrabCut. Comput.
Eng. Appl. 52(2), 186–190 (2016)
10. Yao, G., Zhao, Z., Liu, S.: A comprehensive survey on sampling-based image matting.
Comput. Graph. Forum 36(8), 613–628 (2017)
11. Shahrian, E., Rajan, D., Price, B., Cohen, S.: Improving image matting using comprehensive
sampling sets. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 636–
643. IEEE, Portland (2013)
12. Orchard, M., Bouman, C.: Color quantization of images. IEEE Trans. Signal Process. 39(12),
2677–2690 (1991)
13. Sun, F., Zhang, H.: Studying and implementing GrabCut-digital matting algorithm. J.
TIANJIN Univ. Technol. 24(2), 42–45 (2008)
14. The Berkeley Segmentation Dataset and Benchmark Homepage https://www2.eecs.berkeley.
edu/Research/Projects/CS/vision/bsds/. Accessed 27 Oct 2020
Multi-angle Face Recognition Based on GMRF
Abstract. The general face recognition methods mostly use positive face images
or with a small angle of deflection. Such databases are well established. But in
real life, face recognition problems often do not happen to be positive face state,
especially in monitoring and public security monitoring. The existing effect of
multi-angle face recognition is not ideal, and this paper proposes a face recognition
method based on Gaussian Markov Random Fields (GMRF). GMRF model is a
statistical probability model that can effectively extract image texture information,
which first divides the face image into several sub-blocks in different chunks,
and then, for each sub-block under the block mode, extracts the GMRF feature
after wavelet transformation; finally combines the GMRF features of different
blocking methods, and then classifies the SVM of the Gaussian nuclear function.
Experiments were carried out on the self-built data set, and the proposed method
reached 98.83% for face recognition.
1 Introduction
With the development of computer vision technology and pattern recognition, face recog-
nition has become a research hotspot. Because the face is a relatively insensitive type of
non-contact recognition; however, in the extreme environment of face recognition, such
as monitoring and security and other unconstrained environments, the recognition effect
needs to be improved.
Traditional face recognition algorithms are mostly based on artificially designed
features and machine learning algorithms, such as Local Binary Patterns (LBP) [1],
Histogram of Oriented Gradient (HOG), Principal Component Analysis (PCA) [2], Lin-
ear Discriminant Analysis (LDA), etc. At the same time, the number of face databases
obtained is limited, and the face type is too single. Therefore, the accuracy of early face
recognition algorithms in practical applications is not high. The rapid development of
computer hardware technology and software technology provides a new method to solve
face recognition -- a deep learning method based on a convolutional neural network. A
convolutional neural network (CNN) uses a multi-layer neural network composed of
basic mental elements to learn target features by simulating the learning way of the
human brain [3].
Although deep learning Uncle XI has improved the accuracy of the face recognition
algorithm, there are still many problems in practical application. In the uncontrolled
environment, especially in the surveillance video, most of the face images obtained have
problems such as low resolution, face occlusion, large illumination changes, expression
changes, head posture changes, etc. Facial gestures of the diversity of face recognition
algorithm have great interference effects, under the condition of reality monitoring video
captured most of the face image is the pitch change, spin around or partial first-class
attitude change, even if the depth of using a neural network to extract the face “abnormal”
characteristics, compared with the characteristics of the “standard face” gap is bigger,
will fall into the wrong location in feature space, This, in turn, leads to misjudgment
in identification and verification [4]. These interference factors lead to the decline of
the accuracy of face recognition algorithms, which brings a difficult problem for the
application of face recognition technology. Therefore, the study of multi-angle face
recognition algorithms has a strong practical prospect and research value.
This paper proposes a multi-angle face recognition method based on the combination
of Gaussian Markov Random Field (GMRF). At present, Gaussian Markov random field
model is widely used in the field of image processing. It can express the texture informa-
tion of the image well and has good spatial correlation. This paper adopts the Gaussian
Markov random field feature extraction method combining multi-block features, and
the Gaussian kernel SVM is used for face recognition classification. The recognition
results on self-built data sets show the effectiveness of this method in multi-angle face
recognition.
These databases are shot in the laboratory environment. The postures include −90°,
−60°, −30°, 0°, 30°, 60° and 90°, including different clothes, lowered head, wearing
glasses, hats, and expressions, as shown in Fig. 1.
At the time of the establishment of the self-built database, there were 10 different
faces, each with 30 pictures of different angles and different postures (as shown in Fig. 2)
by a certain proportion of the division. After extracting the characteristics according to
their respective labels, two-thirds of each training session is as a training set and one-third
is a test sample set.
368 S. Huadong et al.
First, according to the nature of Gaussian Markov random field characteristics, the
face image is divided into 9 pieces, 16 pieces, 25 pieces (Fig. 3); then take the R, G, and
B three channels of the image (as shown in Fig. 4).
Multi-angle Face Recognition Based on GMRF 369
3 Feature Extraction
3.1 Gaussian Markov Random Field Model
The Gaussian Markov, Random Field model, is one of the two branches of the Markov
random field. Assuming that the distribution of excitation noise is Gaussian, a differential
equation can be obtained, and the grayscale of the spatial pixel represents the equation.
Therefore, this model becomes a Gauss Markov random field model. In practical appli-
cations, compared to the Markov random field and Gibbs distribution model, the Gauss
Markov random field model has the advantage of a small amount of calculation, so it
has been widely used. Based on the real-time nature of multi-angle face recognition
technology, this paper selects the Gauss Markov random field feature of the extracted
image to classify and recognize the multi-angle face image [5].
The Gauss Markov Random Field Model is a stationary autoregressive process. The
covariance matrix is positive definite, the neighborhood system is symmetric, and the
parameters of the symmetric neighborhood points are equal. When using Gauss Markov
random field features to express the texture characteristics of an image, it can be expressed
in the form of the conditional probability of formula:
Nei Represents asymmetric neighborhood with s as the center and r is the radius (the
symmetric neighborhood does not include the center point s). This formula expresses
that the gray scale y(s) of any pixel s in the image is a function of the grayscale of
neighboring points in all directions of s. The neighborhood relationship of the Gauss
370 S. Huadong et al.
Markov random field model is represented by the structure diagram of the Gauss Markov
random field model, as shown in Fig. 5 [6].
5 4 3 4 5
4 2 1 2 4
6 3 1 S 1 3 6
4 2 1 2 4
5 4 3 4 5
Suppose S is the point set on the M × M network, S = {(i, j), 1 ≤ i, j ≤ M {}}, and
assuming that the given texture y(s), s ∈ S, S = {(i, j), 1 ≤ i, j ≤ M } is a Gaussian
random process with zero means, the GMRF model can be represented by a linear
equation containing multiple unknown parameters.
y(s) = θr max(y(s + r), y(s − r)) + e(s) (2)
r∈NS
Where, NS represents the GMRF neighborhood of the point S, and θr represents the
coefficient, e(s) is a Gaussian noise sequence and the mean value is zero. Write (2) as:
y(s) = θr y1 (s + r) + e(s) (3)
r∈NS
Among them, y1 (s + r) is the point with the larger value in the closed ring area of S.
Applying formula (3) to each point in the area of S, getting M 2 equations about {e(s)}
and {y(s)}:
y(1, 1) = θr y1 ((1, 1) + r) + e(1, 1)
r∈NS
y(1, 2) = θr y1 ((1, 2) + r) + e(1, 2)
r∈NS
y(1, M ) = θr y1 ((1, M ) + r) + e(1, M )
r∈NS
y(M , 1) = θr y1 ((M , 1) + r) + e(M , 1) (4)
r∈NS
y(M , M ) = θr y1 ((M , M ) + r) + e(M , M )
r∈NS
Multi-angle Face Recognition Based on GMRF 371
Use the form of a matrix to represent all the equations formed by y1 (s + r), which
can be written as:
y = QT θ + e (5)
Equation (5) is the linear model of the Gauss Markov random field, QT is the matrix
about all y1 (s + r), and θ is the eigenvector of the model to be estimated. Using the least
square error criterion, the following estimated solution formula can be obtained.
∧ −1
θ = QT Q QT y (6)
The linear autoregressive GMRF model, when the order is relatively low, although it
is convenient to analyze and calculate, has certain limitations when describing complex
image features. As the order increases, the amount of calculation becomes larger, but
it can describe the rich texture information of the image. Choosing the best order can
effectively reflect the texture characteristics of the image. This paper selects the second-
order, fourth-order, and fifth-order models of Gaussian Markov Random Field for multi-
angle face recognition.
good generalization ability and is extremely suitable for small sample classification.
The significance of the method is to find the optimal solution under limited sample
data and convert the task into a binary classification problem to obtain the theoretically
optimal solution. Compared with other traditional algorithms, SVM obtains the optimal
solution from a global perspective, thereby avoiding local extreme values, and it is less
sensitive to the curse of dimensionality. The main explanation behind this attractive
feature is that because of the principle of margin maximization they are based on, it is
not necessary to explicitly estimate the statistical distribution of classes in the super-
dimensional feature space to perform classification tasks. Another important feature is
Multi-angle Face Recognition Based on GMRF 373
their good generalization ability, supported by the r sparse representation of the decision
function. SVM algorithm has a wide range of applications in image processing, computer
vision, speech recognition, etc. [8]. The basic idea of the SVM algorithm: first map the
data to be classified to a high-dimensional space through the constructed kernel function,
and then find a linear function in the high-dimensional space, that is, a hyperplane, so that
it can accurately classify the two types of samples. The point closest to the hyperplane
is the support vector, and the support vector is used to classify the sample.
f (x) = wT x + b = 0 (7)
Where w defines the direction of the hyperplane, which is the normal vector, and b
defines the distance between the origin and the hyperplane, which is the displacement
term. The distance from any point x in the sample space to the hyperplane f (x) is:
T
w x+b
= w
(8)
If the hyperplane f (x) can accurately classify the training samples, for (xi , yi ) ∈ D,
let
wT x + b ≥ +1, yi = +1;
(9)
wT x + b ≤ −1, yi = −1.
When some sample vectors in the d-dimensional sample space satisfy Eq. (9) and
the distance from the hyperplane is the smallest, then these vectors are support vectors,
and all support vectors will form a positive class support surface and a negative class
The supporting surface, the sum of the distances from the two surfaces to the hyperplane
f(x) is:
= 2
w (10)
The algorithm is to find the optimal classification interface, that is, satisfy:
2
max (11)
w,b w
From the above formula, the maximum separation hyperplane is found min 21 w2 .
w,b
In order to solve it, it is transformed into a Lagrange function:
1 m
L(w, b, α) = w2 + αi 1 − yi wT xi + b (12)
2 i=1
374 S. Huadong et al.
Where α is the Lagrange factor, and the partial derivatives of w and b are respectively
calculated:
W = m i=1 αi yi xi (13)
m
0= i=1 αi yi (14)
Incorporating formula (21) into formula (20), the dual problem can be obtained:
m 1 m m
max αi − αi αj yi yj xiT xj (15)
α i=1 2 i=1 j=1
The non-linear problem in the original space cannot be solved in the normal way. In
order to make the original function linearly separable, only the method of mapping is
used to convert the non-linear original function in the low-dimensional space into the
sample function of the high-dimensional space, and then use The kernel function and
the transformed function equation are solved. The commonly used kernel functions are
shown in Table 1:
Let K be the inner product of the mapping function, and meet K (x, z) = (x)·(z),
the functional equation of the regression support vector machine is:
among them, α, α ∗ , γ , γ ∗ ≥ 0, i = 1, 2, . . . , k
Multi-angle Face Recognition Based on GMRF 375
field, which are stitched together into a 480-dimensional vector; Divided into 25 sub-
blocks to extract the features of the fifth-order Gauss Markov random field, the feature
vector is 900 dimensions. The experimental results are as follows:
Experiments have proved that for blocks, the more blocks, the worse the effect, and
the reason is also obvious: the features obtained from more blocks are not obvious, and
the redundant feature extraction will have an impact on the recognition effect. For the
order of Gauss Markov Random Field, order 5 has the highest recognition rate.
Since the previous experiments were performed under the condition of grayscale
images, the color channel information was ignored; therefore, the RGB three-channel
experiment was added to the following experiment; that is, the wavelet transform is per-
formed on the block images of different channels, and after the low-frequency channel,
Gauss Markov random field feature extraction is performed, and then the various fea-
tures are combined. Therefore, the dimensionality of the feature vector extracted under
different block methods and different orders of Gauss Markov random field features is
different. For example, the image is divided into 9 sub-blocks, each of which extracts
the characteristics of the second-order Gauss Markov random field, which are stitched
together into a 324-dimensional vector; the image is divided into 16 sub-blocks, each of
which extracts the characteristics of the fourth -order Gauss Markov random field, which
are stitched together into a 1440-dimensional vector; Divided into 25 sub-blocks and
each of which extracts the characteristics of the fifth-order Gauss Markov random field,
the feature vector obtained is 2700 dimensions. The experimental results are shown in
Table 2:
Multi-angle Face Recognition Based on GMRF 377
The graph shows that in the case of GMRF features of different orders of 2, 4, and
5, the face recognition rate of the algorithm in this paper reaches the highest 98.44%,
which shows the method’s effectiveness.
6 Conclusion
In this paper, the Gauss Markov random field algorithm is used to extract the face image
features. The Gauss Markov random field feature can effectively describe the local texture
information and the spatial position relationship of the image and can effectively describe
the spatial position characteristics of the face. To obtain more detailed information of
the face, this paper also performs block processing on the image, performs wavelet
transformation on the RGB three channels of different blocks, extracts the low-frequency
accessible and extracted GMRF features, and combines them. Finally, the SVM classifier
is used to conduct multi-angle face classification experiments. The results confirm the
method’s validity proposed in this paper.
References
1. Lu, Z., Jiang, X., Kot, A.: A novel LBP-based color descriptor for face recognition. In:
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP),
pp. 1857–1861 (2017)
2. Abbas, E.I., Safi, M. E., Rijab, K.S.: Face recognition rate using different classifier methods
based on PCA. In: 2017 International Conference on Current Research in Computer Science
and Information Technology (ICCIT), pp. 37–40 (2017)
3. Tsai, A.C., Ou, Y.Y., Wang, J.F.: Efficient and effective multi-person and multi-angle face
recognition based on deep CNN architecture. In: 2018 International Conference on Orange
Technologies (ICOT), 4 (2018)
4. Yin, X., Liu, X.: Multi-task convolutional neural network for pose-invariant face recognition.
IEEE Trans. Image Process. 27(2), 964–975 (2017)
5. Coşkun, M., Uçar, A., Yildirim, Ö., Demir, Y.: Face recognition based on convolutional neural
network. In: 2017 International Conference on Modern Electrical and Energy Systems (MEES),
pp. 376–379 (2017)
6. Wang, Y., Wang, H.: Application of rank GMRF in textural description and recognition.
Comput. Eng. Appl. 47(25), 202+204 (2011)
378 S. Huadong et al.
7. Zhang, L., Zhang, L., Zhang, L.: Application research of digital media image technology based
on wavelet transform. J. Image Video Proc. 138 (2018)
8. Yang, L., Chang, H.: Face recognition based on the combination method of multiple classifier.
Int. J. Sign. Process. Image Process. Pattern Recogn. 9(4), 151–164 (2016)
Multi-scale Object Detection Algorithm Based
on Faster R-CNN
Abstract. Most object detection methods have low detection accuracy for small
and dense objects and susceptibility to noise interference. In response to these
problems, this paper designs an improved multi-scale object detection algorithm
based on Faster R-CNN, aiming to improve the detection accuracy of small and
dense objects. This paper introduces depthwise separable Convolution to reduce
the number of network parameters. The amount of calculation uses dilated Con-
volution to improve the feature extraction of small objects and dense objects by
increasing the receptive field, uses Convolution instead of 3 × 3 conventional Con-
volution to solve the problem that the feature extraction of small objects caused
by Convolution is easy to be lost, distorted or the feature extraction redundancy
of large objects is too high. The spatial attention mechanism is introduced to effi-
ciently screen features that are more beneficial to object detection and improve
the model’s ability to deal with large object scale differences in multi-scale object
detection. The network model in this paper was trained and verified on the datasets
VOC2007 and VOC2012, and the mean average precision reached 86.3%.
1 Introduction
Object detection technology is an important research direction in computer vision, widely
used in military, transportation, medical, and other fields. The main task of object detec-
tion is to classify and locate objects. Early object detection technologies mainly include
Haar features [1], Histogram of Gradient [2], Local Binary Pattern [3], and classifiers,
such as Support Vector Machines (SVM) [4] and AdaBoost [5]. The traditional object
detection method uses a sliding window for the region selection, which wastes time and
produces many useless features. Secondly, the traditional object detection method uses
the operator for feature extraction to be manually set and unchanged. Hence, it is impos-
sible to perform effective feature extraction on diversified objects and have occlusion or
complex lighting problems.
In summary, the traditional object detection methods are relatively poor in robustness
and real-time performance [6]. Since the birth of AlexNet [7], the use of Convolution
neural networks for object detection is increasing, and various object detection models
have emerged one after another. The detection performance has also been continuously
improved. However, compared with large-scale objects, small and dense objects have
the characteristics of small size, low resolution, and are easily affected by noise, which
can easily cause missed and incorrect detection. Therefore, improving the detection
capability of multi-scale objects and meeting the requirements for detecting small objects
and dense objects has become the main problem in the object detection field at this stage.
Before classification and positioning, the region proposal is first generated, and then
the generated region proposal is fused with the feature map, and finally the softmax and
regression are used for classification and positioning. The current object detection tech-
nology is mainly divided into two categories, and one is the two-stage object detection
algorithm represented by Faster R-CNN [8]. Therefore, its detection accuracy is high,
but its two-step strategy leads to a substantial increase in the complexity of the model
and reduces the detection speed of the model. The other is the one-stage object detection
algorithm represented by YOLO [9] and SSD [10], which uses a separate CNN model to
implement end-to-end object detection, with great speed advantages. Still, if more than
one object in the same grid is divided, the objects will be missed so that the detection
accuracy will suffer a certain loss.
At present, there are many existing types of research on the use of Faster R-CNN in
multi-scale object detection. However, the current research results show that small and
dense objects’ detection accuracy is still not very high, and the detection speed is slightly
slow. For example, an improved Faster R-CNN algorithm with a split mechanism, which
uses the difference in classification and positioning performance between convolutional
structure and fully connected structure, and the Region Of Interest (ROI) features are
respectively input to the classification network using the fully connected structure and
the bounding box classification network using the convolution structure. By introducing
the convolution structure, the spatial information in the ROI feature is more fully utilized.
ROI Align is used to replace ROI pooling, which cancels the quantization in ROI pooling
and eliminates the bounding box deviation of the Faster R-CNN model itself. Although
this model improves the detection accuracy, the detection speed decreases due to the
increase in the calculation [11]. There is also an improved Faster R-CNN ship surface
multi-scale object detection algorithm. This algorithm adds a multi-scale feature region
proposal generation network based on the original algorithm feature extraction network.
And perform cluster analysis on the ship surface object data set to generate the anchor size
suitable for the object detection in this article, which improves the algorithm’s accuracy
for the detection of ship surface multi-scale objects, especially small objects. Still, the
low detection rate is also a disadvantage of this algorithm. [12] In summary, to improve
the detection ability and detection rate of small and dense objects, this paper uses Faster
R-CNN as the basic model to carry out work. This paper introduces depth-wise separable
Convolution, dilated Convolution, Involution, and spatial attention mechanisms, which
improve the feature extraction ability of the model and increase the model’s detection
rate by reducing the amount of parameters the amount of calculation.
Multi-scale Object Detection Algorithm Based on Faster R-CNN 381
In 2016, Ross Girshick and others proposed Faster R-CNN based on Fast R-CNN, inte-
grating feature extraction, candidate region acquisition, regression, and classification
into a deep network. The speed and accuracy are much higher than Fast R-CNN. How-
ever, Faster R-CNN is an object detection algorithm based on a two-step strategy, it
needs to obtain the region proposal first and then classify each proposal. The amount of
calculation is still relatively large. And due to the limitation of the number of parameters
and calculation, conventional convolution operations generally use a kernel with a size
of 3 × 3 or 5 × 5, which greatly affects the feature extraction of small objects and dense
objects.
The classic Faster R-CNN model comprises the Convolution, Region Proposal Net-
work (RPN), ROI pooling, and classification regression. The convolution layer uses the
VGG16 network model. The main function is to extract the feature maps of the input
image. The convolution layer is composed of convolution, pooling and activation func-
tions. The RPN is the core part of the Faster R-CNN, which replaces using selective
search to obtain region proposals. This method uses the CNN network quickly and more
efficiently when generating a region proposal; anchors are generated. The network uses
a discriminant function to determine whether the anchors are foreground or background
and then uses bounding box regression to make the first anchor adjustment to obtain an
accurate region proposal. Adding the ROI pooling layer mainly solves the problem of
the different sizes of the feature maps that are finally input to the fully connected layer
and obtain a fixed size through upsampling. Classification and regression are used to
determine which class the object belongs to and fine-tune the region proposal’s location
to obtain the final result of object detection.
The Faster R-CNN model can be divided into three steps for object detection. First,
the VGG16 network is used to extract the input image features, and the extracted feature
maps are input to the RPN layer. The RPN layer uses a 3 × 3 sliding window to ergodic
each pixel of the feature map and generates nine anchors of different scales around each
pixel on the feature map. The area of the anchors is set to 1282 , 2562 , and 5122 pixels
of three different sizes, and each anchor with a different area has three ratios of 1:1, 1:2,
and 2:1 [8]. The 256 region proposals are extracted according to the set IoU value. Then
the region proposal is divided into 128 positive samples and 128 negative samples in a
1:1 ratio by the IoU value, and the loss function is optimized through these 256 region
proposals. The 128 positive samples generated by RPN are projected onto the feature
map to obtain the corresponding feature matrix, and then each feature matrix is scaled
into a 7 × 7 feature map through the ROI polling layer, and then the feature map is
flattened, and the prediction result is obtained through a series of fully connected layers.
The model structure of Faster R-CNN [8] is shown in Fig. 1.
382 X. Su et al.
σ = ReLU + BN (1)
C
yreduce = σ W 0 X i,j yreduce ∈ R1×1× /r (2)
Y = sum mul Hi,j , Xunfold Y ∈ R1×1×C (6)
Among them, Xunfold represents the output feature map obtained after unfolding the
feature map. mul represents the product operation of the obtained kernel and the output
feature map. sum represents the sum of each channel of the feature map obtained after
the operation. C represents the number of channels of the input feature map. And G
represents the number of groups into which the input feature map is divided.
On the basis of Feature Pyramid Networks (FPN) [18], depthwise separable convolu-
tion, dilated convolution, spatial attention mechanism (SAM) and Involution operators
are introduced. The resnet50 feature extraction network obtains c2, c3, c4 and c5 feature
maps with different scales from bottom to top, as shown in Fig. 4. After inputting these
feature maps into the FPN network, the DDS network module performs a series of con-
volution and pooling operations on the four feature maps, and performs 2x upsampling
on the obtained feature maps through the nearest neighbor interpolation algorithm. The
obtained feature map is input downward and merged with the feature map processed by
the DDS network in the upper layer, and the fused feature map is then output through a
3 × 3 involution operation. The DSI-FPN network structure is shown in Fig. 5.
Take a feature map c2 output by ResNet as an example to introduce the DDS network
module. Perform two-way depth-wise separable Convolution on the c2 feature map
simultaneously. One way uses the dilated Convolution with kernel_size = 1, dilation
rate = 1 and stride = 1, and the other uses the dilated Convolution with kernel_size
= 3, dilation rate = 3 and stride = 1. And channel fusion is performed on the feature
maps after two convolutions, and then do a 1 × 1 convolution to change the number of
channels to 256. Finally, the two feature maps with 256 channels are subjected to pixel
multiplication and output. The DDS network structure is shown in Fig. 6.
386 X. Su et al.
The dataset used in this article is a fusion of the VOC2007 and VOC2012 officially
released by PASCAL, and data enhancement through methods such as horizontal flip-
ping. The dataset used by object detection has one background category and 20 object
categories. It contains 14,895 training images and 5823 verification images, and both
the training set and the verification set disclose the corresponding label files.
In this experiment, Mean Average Precision (mAP) is used as a measurement index,
and mAP is a commonly used index to measure the performance in object detection.
The calculation method is formula (7) and (8):
1
AP = 11 r∈(0,0.1...1) Pint erp(r) (7)
sum(AP)
mAP = N
(8)
Among them, Pint erp(r) is the interpolated precision, AP is the average precision of
the 11-point interpolated precision of this category, and the AP values of the 20 categories
are summed and averaged to obtain the mAP value of the experiment.
This experiment uses the method of transfer learning to freeze the weights of the
ResNet network, a total of 40 rounds of iterative training, the batch-size is set to 2, the
lr is set to 0.005, the gamma of lr is set to 0.33, the attenuation strategy is set to lr =
lr × gramma, and a new lr is generated every five rounds of attenuation. The model
optimization method is the stochastic gradient descent method.
This experiment was performed on the data sets VOC2007 and VOC2012. The exper-
imental results show that the mAP value reached 86.3%, which is 15.9% higher than
the classic Faster R-CNN algorithm and 8.7% higher than the R-FCN algorithm which
Multi-scale Object Detection Algorithm Based on Faster R-CNN 387
the highest accuracy among the classic algorithms. It can be seen from Table 1 that
the model has the highest detection accuracy on the three categories of bus, person and
car, reaching 93%. It can be seen from the comparison with the AP value of the Faster
R-CNN + FPN model in different categories that the AP values of all categories of the
model are higher than the Faster R-CNN + FPN model. Comparing with YOLO v3, it
can be seen that, except for the two categories of train and horse, the AP values of the
experimental model in the other categories are all higher than YOLO v3 [19]. Compared
with YOLO v4 [20], it can be seen that there are 11 categories of AP values in this
experimental model that are higher than YOLO v4. It can be seen from Table 2 that
because the model in this paper introduces the dilated Convolution, involution operator
and spatial attention mechanism on the basis of Faster R-CNN, it improves the model’s
ability to extract image features, so the mAP value of the model in this experiment is
higher than other models.
Fig. 7. Change of parameters, where (a) represents the change of mAP during the experiment,
(b) represents the change of loss and lr during the experiment
Table 1. (continued)
Model mAP/%
R-CNN 53.3
Fast R-CNN 68.4
Faster R-CNN 70.4
YOLO v2 73.4
YOLO v3 74.6
SSD 74.9
R-FCN [21] 77.6
Faster R-CNN + FPN 81.8
Ours 86.3
5 Conclusion
This experimental model is based on the classic Faster R-CNN model and mainly
improves the FPN network. It combines depth-wise separable and dilated Convolu-
tion to replace the 1 × 1 Convolution in FPN, and uses Involution to replace the 3 × 3
volume in FPN. At the same time, the spatial attention mechanism is introduced, which
reduces the number of parameters and calculations and improves the feature extraction
ability for small and dense objects. In addition, the anchors generated by the RPN model
based on each pixel are expanded from nine to 15, adding two small scales of 322 and
642 , and each scale is still in three ratios of 1:1, 1:2 and 2:1, it is helpful to improve the
390 X. Su et al.
detection ability of small objects and dense objects. The test results of this model on the
dataset VOC2012 show that the detection accuracy of small objects and dense objects
have been significantly improved. Although the model improves the detection speed by
reducing the number of parameters and the amount of calculation, it still cannot real-
ize the function of real-time detection. Therefore, how to further improve the detection
speed of the model is a further research topic.
References
1. Panning, A., Al-Hamadi, A.K., Niese, R., Michaelis, B.: Facial expression recognition based
on Haar-like feature detection. Pattern Recogn. Image Anal. 18(3) (2018)
2. Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: IEEE
Computer Society Conference on Computer Vision & Pattern Recognition. IEEE (2005)
3. Ojala, T., Pietikainen, M., Maenpaa, T.: Multiresolution gray-scale and rotation invariant
texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 24(7),
971–987 (2002)
4. Burges, CJC., Fayyad, U.: A Tutorial on Support Vector Machines for Pattern Recognition.
Data Mining and Knowledge Discovery (1998)
5. Zhu, J., Arbor, A., Hastie, T.: Multi-class AdaBoost. Stat. Interf. 2(3), 349–360 (2006)
6. Ma, Y., Zhao, Z.H., Yin, Z.Y., Fan, C., Chai, A.Y., Li, C.M.: Segmented deconvolution
improves the object detection algorithm of SSD [J/OL]. Minicomputer System, pp. 1–7 (2021)
7. Technicolor, T., Related, S., Technicolor, T., et al.: ImageNet Classification with Deep
Convolutional Neural Networks
8. Ren, S., He, K., Girshick, R., et al.: Faster R-CNN: towards real-time object detection with
region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137–1149 (2017)
9. Redmon, J., Divvala, S., Girshick, R., et al.: You Only Look Once: Unified, Real-Time Object
Detection. IEEE (2016)
10. Liu, W., Anguelov, D., Erhan, D., et al.: SSD: Single Shot MultiBox Detector. Springer, Cham
(2016)
11. Wang, X.B., Zhu, X.Y., Yao, M.H.: Object detection method based on improved Faster RCNN.
High Technol. Lett. 31(05), 489–499 (2021)
12. Fan, J.L., Tian, S.B., Huang, K., Zhu, X.D.: Multi-scale target detection algorithm for air-
craft carrier surface based on Faster R-CNN [J/OL]. System Engineering and Electronic
Technology, pp. 1–10 (2021)
13. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778 (2016)
14. Chollet, F.: Exception: deep learning with depth wise separable convolutions. In: 2017 IEEE
Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2017)
15. Yu, F., Koltun, V.: Multi-scale context aggregation by dilated convolutions. In: ICLR (2016)
16. Zhu, X., Cheng, D, Zhang, Z., et al.: An empirical study of spatial attention mechanisms in
deep networks. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
IEEE (2020)
17. Li, D., Hu, J., Wang, C., et al.: Involution: Inverting the Inherence of Convolution for Visual
Recognition (2021)
Multi-scale Object Detection Algorithm Based on Faster R-CNN 391
18. Lin, T.Y., Dollar, P., Girshick, R., et al.: Feature pyramid networks for object detection.
In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE
Computer Society (2017)
19. Redmon, J., Farhadi, A.: YOLOv3: An Incremental Improvement. arXiv e-prints (2018)
20. Bochkovskiy, A., Wang, C.Y., Liao, H.: YOLOv4: optimal speed and accuracy of object
detection (2020)
21. Dai, J., Li, Y., He, K., et al.: R-FCN: object detection via region-based fully convolutional
networks. In: Advances in Neural Information Processing Systems. Curran Associates Inc.
(2016)
Persistent Homology Apply in Digital Images
Abstract. We use the method of persistent homology and complex filter homol-
ogy group calculation to study digital images. 2D digital images are composed
of pixels, 3D images are composed of voxels. Both of them own a natural cubi-
cal construction. The complex cubical numbers are more suitable than simple,
complex numbers in the field of digital images. At the same time, the way to
compute simplicial complex can also apply to study cubical complex and its per-
sistent homology. We construct a cubical complex of this space under different
parameters, then calculate the homology of the cubical complex to obtain the per-
sistent diagram, and use it to get the corresponding geometric structure message
and the image’s topological characteristics. In the last, we discuss the application
of persistent homology to image classification on the support vector machine and
random forest method.
1 Introduction
Image similarity analysis is the basic research goal of computer vision and a key step
in image processing research. Recently, many scholars have made a lot of progress in
image retrieval, image classification, target recognition, and other research works. At
present, the more traditional image retrieval, classification, and recognition generally
have two research directions: one is the research method based on image features, and
the other is the research method based on gray correlation.
In the past, the research methods of extracting image features based on image con-
tent mainly extracted the color feature, texture feature, shape feature of the image, and
the local feature of the image adopted in recent years. When the above features are
processed, the image has certain limitations if they undergo rotation transformation,
non-linear transformation, and projection transformation. We study the method of per-
sistent homology to extract the topological features of the image and the geometric
structure features of the image shape and combine the extracted features to perform
image classification in this paper.
There are many methods for topological data analysis (TDA) [1]. Among them, we
use persistent homology (PH) [2] to study the qualitative characteristics of data that
persist across multiple scales. Persistent homology is robust to the perturbation of input
data. It offers a concise expression of the qualitative features of the input data. Persistent
homology is very useful for applications. It is based on algebraic topology. It provides a
more comprehensible academic structure to learn the qualitative characteristics of data
with complex structures. It can be calculated by linear algebra. When encountering small
input data, interference is powerful. This concept was first proposed by ROSINI and
LNDI in 1999 when they were studying scale theory [3]. Later, they applied persistent
homology to analyze various data and achieved good results. Now it has been widely
used in various fields. Carlsson G studies the pixel blocks of 3 × 3 images by selecting
the local structure of natural images [4]. Moo and Kim offer a newfangled framework
that uses computational algebraic topology techniques to characterize signals in images
[5]. Paul and his team used the robustness of persistent homology and employed relevant
tools to study 3D images of plant roots [6].
2 Relate Work
2.1 Homology
If the given data is located in the metric space, for example, a subset of the Euclidean
space with inherited distance functions. Under many circumstances, we don’t want to
know the accurate geometry of spaces. Still, we try to figure out the existence of voids,
holes or compositions, and some other basic characteristics. Algebraic topology catches
essential features by computing them or correlating vector spaces or more complex
algebraic structures. We want to understand homology, which correlated one vector
space Hi (X) to a space X for each number i ∈ {0, 1, 2, …}. The dimension of H0 (X)
computes the amount of path components in X, the dimension of H1 (X) is a count of
the number of holes, and the dimension of H2 (X) is a count of the number of voids.
These algebraic structures don’t transform when the underlying spaces are deformed by
stretching, bending. In other words, the spaces are homotropy invariant and robust in
professional terms. Calculating the homology of arbitrary topological spaces may be a
little difficult. Therefore, we apply a combination structure called ‘simplicial complex’
to approach our spaces, and then use related algorithms to calculate its homology.
Consider that the experimental data we give is in the form of a set S. points or vectors are
denoting the measured values in this set. Some distance functions on the points or vector
sets given by correlation or dissimilarity measures. Here, we do not consider whether the
set S belongs to a sample of the underlying topological space. If you want to restore the
attributes of the above-mentioned underlying space, you can use a method that is robust
to little disturbances in the dataset S. If S is a part of Euclidean space, you can consider
“thickening” the trumpet of S according to the combination of a sphere with a certain
fixed radius to give ε around its point, the calculation of Čech is complex. We attempt to
construct a Čech complex [7] for the selected value and compute its simple homology
to compute the qualitative characteristics of the finite metric space S. There is a problem
in this method. The parameter that we choose to construct the Čech complex is not
394 S. Huadong et al.
explicitly chosen a priori. Using the knowledge of persistent homology, we may need to
consider several possible values of the parameter to be able to extract the information
we want from the data. As the value of increases, the simplex is gradually added to the
complex. Use persistent homology to study how the homology of the complex changes
with changes in parameters. Finally, detect which features “continue to exist” even if
the parameter values are changing.
Simplicial complex [8] is a set K of nonvoid subsets of a collection K0 , {v} ∈ K for all
v ∈ K0 , and τ ⊂ σ, σ ∈ K ensures that τ ∈ K. The elements of K are called simplices,
the elements of K0 are called vertices of K. If a simplex has a cardinality of p+1, we call
that it is a p-simplex or it has dimension p.
The set of p-simplices can be represented by Kp . The k-skeleton of K is the union
of the collections Kp for all p ∈ {0, 1,…,k}. We say τ a face of σ, if τ, σ are simplices,
and τ ⊂ σ. In the case that the dimensions of τ and σ differ by k , we call that τ is a face
of σ of codimension k . The dimension of K is the maximum of the dimensions of its
simplices. A map of simplicial complexes, f : K → L, f :K0 → L0 , f (σ ) ∈ L and σ ∈ K.
We now define homology for simplicial complexes. 2 represents a domain with two
elements. Given a simplicial complex K, let Cp (K) denote the F2 -vector space with basis
given by the p-simplices of K. p ∈ {1, 2, …}, we define the following map:
dp : Cp (K)σ → Cp − 1(K)
σ → τ
τ ⊂ σ,τ ∈ Kp −1
Since the borders of the borders are usually empty, the map dp has the following
property: for p ∈ {0, 1, 2, …}, dp ◦ dp+1 = 0.
3.1 Barcodes
In the barcode [9], the starting point of the straight line indicates the time when the
hole turns up, and the endpoint indicates the time when the hole fades away. In each
dimension, the length of straight lines is various. If the length of a straight line is longer,
the topological feature is more stable; the shorter the length, the topological feature has a
shorter duration, and it may also represent random noise. Furthermore, there is a straight
line with arrows in the figure below. We name it “infinite interval”. It means that this
topological feature is generated from a certain point in time and will not die out in the
end (Fig. 1).
Persistent Homology Apply in Digital Images 395
Fig. 2. In the persistence diagram of beer mug, the red part represents 0 dimension, and the blue
part represents 1 dimension.
396 S. Huadong et al.
4 Experiment
4.1 Data
Provided that given an image made up of N voxels or pixels, we treat the digital image
as a C × N dimensional space, and a vector of length c may be stored at each coordinate.
The vector stands for the color of the pixel or the voxel. We define a suitable distance
function in such a space and allow a group of images to be treated as limited metric space.
These images have N pixels or voxels. Therefore, the way to learn a limited metric space
is to use persistent homology. It can also apply in the research about image data sets.
The digital image has a cubical structure. Simply, a cubical complex is a space made
up of corners, edges, squares, cubes, and some other things. For a two-dimensional digital
image, building cubical complexes involves specifying a vertex for each pixel, then using
edges to connect the vertices corresponding to adjacent pixels, filling the resulting square
at the end. We can conduct a similar technique to deal with three-dimensional images.
Label each vertex with an integer relating to the pixel’s gray value, and then use the
maximum value of adjacent vertices to mark the corresponding edge.
For n ∈ Z, [n, n+1] is called non-degenerate interval, [n, n]is called a degenerate
interval. A boundary of a elementary interval is a chain ∂[n, n+1] = [n+1, n+1] − [n,
n] in case of non-degenerated elementary interval and ∂[n, n] = 0 in case of degenerate
elementary interval. An elementary cube C is a product of elementary intervals, C =
I1 × ... × In . Embedding dimension of a cube is n, the number of elementary intervals
(degenerate or not) in the product. A dimension of a cube C = I1 × ...×In is the number
of non degenerate elementary intervals in the product. A boundary of a cube C = I1 ×
...×In is a chain obtained in the following way:
The boundaries of each cube are in the set. If the cube C in the cubic complex number
K is not within the boundary of any other cube in K, then the cube C is the largest. The
support of the cube C is the collection occupied by C in Rn .
In the case where the cube may be equipped with filter values, we filter the cubical
complex. Although the scope of filtering could be a collection of two elements, we still
have to consider that all cubical complexes are filtered cubic complexes (Figs. 4 and 5).
Fig. 4. A cubical complex in a map f : [−2, 2]2 → R. f(x, y) is a distance from (x, y) to a unit
circle x2 + y2 = 1.
Fig. 5. Obtain the filter of the cubical complex through the grayscale image and related barcode.
Since all the new pixel is invariably concatenated to the previous pixels, we know the only one
connected component which emerges at first time and doesn’t disappear till the end.
398 S. Huadong et al.
The MNIST data set is composed of 28 × 28 handwritten digital images. It has a total
of 50,000 images. In this paper, we use topological knowledge to study data set, and
classify the numbers in it.
According to the number of rings in the data set, we can divide the numbers into three
classifications: {1, 2, 3, 5, 7}, {0, 4, 6, 9}, {8}. Use the position and scale information
of topological features that persistent homology can provide to classify the MNIST data
set.
In the following experiments, we use two methods to classify the numbers according
to topological features (Tables 1 and 2).
0 1 2 3 4 5 6 7 8 9
0 0.966 0.795 0.939 0.924 0.937 0.616 0.961 0.855 0.629
1 0.749 0.560 0.580 0.593 0.940 0.525 0.991 0.939
2 0.698 0.677 0.681 0.711 0.726 0.871 0.747
3 0.535 0.537 0.905 0.533 0.970 0.896
4 0.537 0.892 0.548 0.950 0.882
5 0.900 0.562 0.966 0.892
6 0.933 0.847 0.565
7 0.988 0.928
8 0.856
9
0 1 2 3 4 5 6 7 8 9
0 0.975 0.795 0.941 0.922 0.930 0.629 0.968 0.869 0.612
1 0.750 0.584 0.583 0.602 0.950 0.519 0.992 0.950
2 0.711 0.692 0.698 0.691 0.745 0.856 0.750
3 0.522 0.516 0.904 0.563 0.965 0.902
4 0.515 0.887 0.573 0.941 0.891
5 0.899 0.586 0.964 0.894
6 0.939 0.846 0.578
7 0.987 0.945
8 0.872
9
From the above experiments, we can see that the classification effect of (1, 7) is the
worst, the accuracy is 51.9%. The classification effect of (1, 8) is the best, the accuracy
is 99.2% (Fig. 6).
Persistent Homology Apply in Digital Images 399
Fig. 6. Grayscale the numbers 1, 7, 8 in the MNIST data set, and then we get their persistence
diagrams.
From the persistence diagrams, we can see that numbers 1 and 7 have only 0-
dimensional topological features. They have similar dimensions, so this group has bad
classification results. The number 8 has a one-dimensional feature, the group (1, 8) has
the best result. Similarly, the classification effect of the number 8 and other numbers
has the best classification accuracy among other groups, except one group (0, 8). In this
paper, the classification method of random forest is better than the method of SVM.
5 Conclusions
In this paper, we introduced the concept of persistent homology, related properties, and
the tools that use topology to describe: barcode, persistent diagram, betti curve. We
use the persistent homology to construct a series of cubical complexes to approximate
the image space, then obtain the geometric structure and the homology information of
the image. SVM and random forest are used to classify the images. The experiments
indicate that the combination of random forest and TDA technology can improve the
image classification effect in the MNIST data set.
There are many things for improvement in this work. We can use persistent homol-
ogy to classify other images, such as commodity images. It’s not just the MNIST data
set. TDA is well known for its robustness to noise, so different types of noise can be
studied. Since the persistence diagram is affected by the image scale, we can study image
classification at different scales.
References
1. Wang, B., Wei, G.-W.: Object-oriented persistent homology. J. Comput. Phys. 305, 276–299
(2016)
2. Zomorodian, A.: Topology for Computing. Cambridge University Press, Cambridge (2018)
3. Rosini, P., Landi, C.: Size theory as a topological tool for computer vision. Pattern Recogn.
Image Anal. 9(4), 596–603 (1999)
4. Carlsson, G., Ishkhanov, T., de Silva, V., Zomorodian, A.: On the local behavior of spaces of
natural images. Int. J. Comput. Vis. 76(1), 1–12 (2008)
5. Chung, M.K., Bubenik, P., Kim, P.T.: Persistence diagrams of cortical surface data. In: Prince,
J.L., Pham, D.L., Myers, K.J. (eds.) Information Processing in Medical Imaging, pp. 386–397.
Springer, Heidelberg (2009). https://doi.org/10.1007/978-3-642-02498-6_32
6. Bendich, P., Edelsbrunner, H., Kerber, M.: Computing robustness and persistence for images.
IEEE Trans. Vis. Comput. Graph. 16(6), 1251–1260 (2010)
7. Eda, K., Kawamura, K.: The surjectivity of the canonical homomorphism from singular
homology to Cech homology. Proc. Am. Math. Soc. 128(5), 1487–1495 (1999)
8. Duval, A.M., Klivans, C.J., Martin, J.L.: Critical groups of simplicial complexes. Ann. Comb.
17(1), 53–70 (2013)
9. Watanabe, S., Yamana, H.: Topological measurement of deep neural networks using persistent
homology. Ann. Math. Artif. Intell. (2021).https://doi.org/10.1007/s10472-021-09761-3
10. Chung, M.K., Ombao, H.: Lattice paths for persistent diagrams. In: Reyes, M., Hen-
riques Abreu, P., Cardoso, J., Hajij, M., Zamzmi, G., Rahul, P., Thakur, L. (eds.)
IMIMIC/TDA4MedicalData - 2021. LNCS, vol. 12929, pp. 77–86. Springer, Cham (2021).
https://doi.org/10.1007/978-3-030-87444-5_8
11. Curto, C., Paik, J., Rivin, I.: Betti curves of rank one symmetric matrices. In: Nielsen, F.,
Barbaresco, F. (eds.) GSI 2021. LNCS, vol. 12829, pp. 645–655. Springer, Cham (2021).
https://doi.org/10.1007/978-3-030-80209-7_69
12. Fajstrup, L.: Dipaths and dihomotopies in a cubical complex. Adv. Appl. Math. 35(2), 188–206
(2005)
Research on the Style Classification Method
of Clothing Commodity Images
1 Introduction
As e-commerce continues to expand and develop, the national e-commerce transaction
volume reached 34.81 trillion yuan in 2019 [1]. Figure 1 shows the market size and
growth of China’s apparel e-commerce industry in the past five years. The data shows
that the growth rate of the total scale of my country’s clothing e-commerce has been
unstable in the past five years. But it still keeps increasing year by year. It can be
seen that buying clothing products on the Internet has become a very popular way. The
contemporary public has a certain pursuit of personality and style in terms of clothing.
Therefore, the research on clothing style classification is very helpful for consumers to
retrieve their ideal style of clothing on the e-commerce platform.
In the clothing product display pictures of major e-commerce platforms, in order
to better display the effect, most of the clothing pictures are displayed on the models.
And there will be some exquisite backgrounds to enhance the overall visual effect. So,
some elements are not related to clothing. These elements will cause some errors when
classifying the style of the specific area of clothing, which affects the classification
accuracy. Therefore, this article pre-segmented the region of interest in the product
image to extract the target region. Secondly, the HOG feature of ROI region is extracted.
Then, the extracted data is put into the classifier as input data. Finally, the product
image style classification is completed. Experimental results show that the model can
improve the accuracy of clothing style classification results. And provide a reference in
the application of clothing style classification.
12000 29.83%
Fig. 1. A statistical chart of the market size and growth of China’s apparel e-commerce industry
in the past five years
when segmenting PET lung tumor images. When completing the segmentation task, the
traditional random walk algorithm must rely on manual methods to select seed points.
Therefore, the selection of seed points will have a certain impact on the segmentation
results. The position of each seed point and the total number of seed points will make the
segmentation effect different. But the traditional random walk method is very simple to
implement. When Liu Guocai et al. [7] studied the random walk algorithm, they improved
the disadvantage of manually selecting seed points. Using the method of region growth,
automatic seed point selection was realized, which greatly improved the segmentation
speed.
Compared with other feature description operators, the HOG descriptor has unique
advantages. It has good invariance in image geometric transformation and optical defor-
mation. Therefore, object detection, pedestrian detection, facial expression, and other
aspects have been widely used. There is also a smaller range of applications on clothing
product images.
Firmino M et al. [8] proposed a system for detecting and diagnosing nodules using CT
images. The characteristic values of pulmonary nodules were extracted and input into the
rule-based classifier and SVM classifier by combining the directional gradient histogram
algorithm with the watershed algorithm to complete the binary classification of benign
and malignant pulmonary nodules. Xiaofei Chao et al. [9] proposed using local binary
mode combined with HOG to retrieve clothing images. Sun Shuangchen [10] proposed a
retrieval model based on automatic learning of fashion tags to solve the problem between
user query intention and system retrieval results. It uses HOG features to extract the
underlying features. Finally, it is matched with SVM to complete the clothing fashion
retrieval system. Tian Xianxian et al. [10, 11] used Fisher features to improve the HOG
feature structure. Then, the image blocks generated by HOG are discarded or retained
through certain criteria to achieve the application purpose of MultiHOG feature. Sun
et al. [12] researched and proposed a method for multi-pedestrian tracking. The core
is to combine the enhanced online tracker and the HOG feature description operator
and the particle wave framework. And the results are good. After the improvement, the
tracking object information can be better described. When the tracking target is identified,
the above method is used, and the support vector machine is used. The output results
of dynamic fusion are input into the particle filter framework for processing to detect
similar objects. Pan Bin et al. [13] proposed using HOG to extract features to realize
automatic wine picking. Then, KSVM was used to classify the hop images. Finally,
a good effect was achieved and designed and implemented a clothing image retrieval
system. Madhogaria et al. [13, 14] proposed a novel vehicle recognition method that used
HOG features and MRF fusion in vehicle detection of aerial digital images. Alsahwa
et al. [15] used the method of extracting HOG features and putting them into SVM
classifier in the Marine biometric recognition system to solve the target recognition task
under this background.
404 Y. Zhang and R. Song
Support vector machines have been developed in many fields, such as the medical field,
vehicle transportation field, text recognition field, human body part recognition, human
behavior detection, and others. It has advantages in data sets with small sample size,
nonlinear and high-dimensional pattern recognition applications. There are also some
research achievements in the field of image classification.
Keping Wang et al. [16] improved the feature-weighted support vector machine and
proposed a new image semantic classification method. Since the correlation of each
feature is related to its degree of dispersion, a larger weight is assigned to the related
features and the unrelated features are eliminated. Finally, the SVM classifier is trained
with the weighted features to solve the image classification problem. When Gao Han [17]
used the SVM model for image classification, he proposed replacing the kernel function
with feature distance as a similarity measure. This new kernel function selection method
was adopted in the SVM model, which greatly improved its generalization ability. Ryu
et al. [18] researched the classification of handwritten documents, the core of which is to
apply a structured learning-based SVM to achieve classification. Use structured learning
methods to determine SVM parameters, construct and train relaxed structure SVM.
Finally, the optimal parameters can be estimated. Wang Kuiyang et al. [19] constructed
a braking intention SVM model based on support vector machines, which achieved high
accuracy in braking intention recognition. According to the problem of ship recognition
rate, Wu Yingzheng et al. [20] implemented ship image classification based on HOG
and SVM models. Dong Junjie [21] studied the image recognition technology of the
online clothing similarity retrieval system. The results of several model experiments
HOG feature
extraction
Output classification
results
were compared to find a more suitable one. A trainer framework combines HOG features
and SVM—designed and implemented a clothing image retrieval system.
To sum up, this paper firstly adopts the random walk algorithm to segment the apparel
merchandise images with complicated backgrounds. The clothing area is marked with
foreground seed points, and the image areas except clothing are marked with background
seed points. After segmentation, the target clothing image part is obtained. Secondly, the
HOG algorithm extracts the features of the image set that has not been segmented and
the image set that has been segmented. Finally, the extracted HOG features are put into
the support vector machine classifier to classify two styles of sports style and business
style. The flowchart is shown in Fig. 2.
In the formula, di is the degree of vertex vi , which means the sum of the weights
of all edges connected to vertex vi . The Laplace matrix constructed according to the
above-defined quantity is denoted as L. After the vertices in the graph are divided into
set VM and set VU , the Laplace matrix is expressed as
LM B
L= (2)
BT LU
The solution process of random walk can be transformed into solving the Dirichlet
problem [23]. The core of the problem is to determine the harmonic function u(x, y)
that meets the boundary conditions, so that the Dirichlet integral in Eq. (3) reaches the
minimum.
1 1 1 2
D[u] = (Au)T C(Au) = uT Lu = wij ui − uj (3)
2 2 2
eij ∈E
Among them, xis is the probability of vertex vi reaching the mark S for the first time.
The Histogram of Oriented Gradient (HOG) feature has been widely used since 2005.
It is a feature description operator used in image processing and computer vision. Dalal
and Triggs accurately defined its meaning at the CVPR conference [24]. The main idea
of the HOG algorithm is to use the direction and density distribution of the gradient to
describe the characterization and shape attributes of the local target in the image, which
focuses on the local area information of an image. The HOG method can be used to
obtain the edge information of the input image. An image will be divided into multiple
interconnected cell units. Each Cell unit generates a corresponding directional gradient
histogram, and all directional gradient histograms are combined. Normalization can be
used to describe the edge characteristics of the image.
The steps of the HOG method can be basically summarized as follows:
The mathematical model of the HOG algorithm first uses the gamma equalization
method to normalize and standardize the image. The specific operation is to take the
image with the index value of gamma, and then gamma equalizes each color channel.
The formula is shown in formula (5):
Through this method, when there is uneven illumination in the processed target
image recognition or target detection, the image can be enhanced under the condition
of weak illumination so that the robustness is improved. And the shadow caused by the
recognition can be reduced to a certain extent. Can also effectively avoid some noise
interference.
The first-order differential template gradient operator method is used to obtain the
gradient value of each pixel.
Gm (x, y) = f (x + 1, y) − f (x − 1, y) (6)
The formula (6) represents the gradient component vector of the image in the hor-
izontal direction. Normally, the gradient operator [−1, 0, 1] is selected to obtain the
image by convolution operation. The formula (7) represents the image in the vertical
direction. The gradient component vector is usually obtained by convolving the image
P with the gradient operator [1, 0, −1].
Formula (8) can calculate the gradient amplitude at the pixel point of the image.
2 2
M (x, y) = Gm (x, y) + Gn (x, y) ≈ |Gm (x, y)| + |Gn (x, y)| (8)
Formula (9) can calculate the gradient direction at the pixel point of the image.
Gm (x, y)
α(x, y) = arctan (9)
Gn (x, y)
Among them, M (x, y) in (8) represents the gradient amplitude, and α(x, y) in (9)
represents the gradient direction.
3.4 SVM
Corinna Cortes and Vapnik are equivalent to the support vector machines proposed in
1995 [25]. SVM can realize functions such as data analysis and pattern recognition in the
field of machine learning. Data classification and regression analysis are regarded as its
important advantages. The most basic understanding of SVM is a linear classifier with
the largest interval in the feature space, which is different from the perceptron. It can be
divided into linear classifiable support vector machines, non-linear classifiable support
vector machines [26], and other categories. This article will focus on linear classifiable
support vector machines [27], as shown in Fig. 3.
408 Y. Zhang and R. Song
L3
L1 L2
O
Fig. 3. Linear sortable graph
The ultimate goal of SVM is to find the optimal hyperplane, that is, to get the
maximum distance between each sample point and the hyperplane. It is therefore also
called the maximum interval hyperplane. Now suppose that there are a total of N sample
points to form the training set, and the data set is linearly separable. The feature vector
composed of these N points is denoted as xi . All the samples in the training set can
be divided into two types, denoted by si . Type one (denoted by s1 ) and category two
(denoted by s2 ). Formula (10) can be used to describe any hyperplane.
g(x) = 0 = wT x + w0 (10)
Among them, w = [w1 , w2 , w3 , ..., wi ]T represents the weight vector, and w0 repre-
sents the threshold or displacement term. Its size is the distance between the hyperplane
and the origin. The above linear equation describes that there are but not unique hyper-
planes that can divide the training samples into two categories. Figure 4 describes the
possible hyperplanes.
L3:g(x)=+1
L2:g(x)=0
L1 g(x)= -1
O
Fig. 4. Hyperplane
Research on the Style Classification Method 409
The distance from the sample point to the hyperplane is obtained by formula (11).
|g(x)|
d= (11)
w
Use ti as the class label. When w and w0 are normalized, the value of g(x) to the
nearest point in class s1 is t1 = 1. The value of the nearest point in class s2 is t2 = −1.
When the condition of formula (12) is satisfied, the minimization is as shown in formula
(13).
ti wT xi + w0 ≥ 1 i = 1, 2, . . . , N (12)
1
C(w) ≡ w2 (13)
2
At this time, the convex quadratic programming problem is obtained and it contains
inequality constraints. The Lagrangian objective function (14) is constructed. And the
Lagrangian multiplier method is used to solve this problem.
1 T N
L(w, w0 , λ) = w w− λi ti wT xi + w0 − 1 (14)
2
i=1
Among them, the Lagrange multiplier is represented by λi , λi ≥ 0 (i = 1,2,…,n).
Let θ (w) = max L(w, w0 , λ). When the sample point is not within the feasible
αi ≥0
solution range at g(x) < 1, the sample point does not meet the constraint conditions. If
λi is infinite, then θ (w) is infinite. g(x) ≥ 1 is in the feasible solution area, and θ (w) is
the original function at this time. Combine the above two situations to get a new objective
function (15).
C(w), x ∈ Feasible region
θ (w) = (15)
+∞, x ∈ Infeasible area
The problem is transformed into (16).
min θ (w) = min max L(w, w0 , λ) (16)
w,w0 w,w0 λi ≥0
According to the duality of the Grange function, the above formula is reduced to
(17).
max min L(w, w0 , λ) (17)
λi ≥0 w,w0
At this time, the problem is transformed into a dual problem including the minimiza-
tion problem. Solve the above problem, multiply all the non-zero support vectors and
add them up to get formula (18).
N
w= λi ti xi (18)
i=1
If you have the value of w, you can get b. And then you can get the hyperplane. The
SVM coincides with one of the two hyperplanes as shown in formula (19).
wT x + w0 = ±1 (19)
410 Y. Zhang and R. Song
Business
style
Sport
style
gradient, a small cell contains 128 values (8 * 8 * 2 = 128). Each block is obtained
by HOG through a sliding window, and the feature descriptor of the entire image can
be obtained by combining all blocks into one. The result of HOG feature extraction is
shown in Fig. 6.
SVM has certain advantages in solving the problem of small data scales. The weight
of the feature can show the importance of the feature, making the feature interpretable.
Another advantage of SVM is that it does not require a large number of adjustment
412 Y. Zhang and R. Song
parameters. Many experiments on clothing image retrieval by Dong Junjie [21] show
that the method of HOG+SVM is better than the method of LBP, GIST, and SIFT+BOW.
Therefore, this experiment uses the combination of HOG features and support vector
machines to solve the problem of clothing product image style classification.
This paper mainly studies the two-classification problem. The basic problem is to
study the classification of business style and sports style. According to the above method,
two sets of feature values of the original image and the segmented image are obtained
as the input value of the SVM classifier. First, the characteristic values of business style
and sports style clothing before segmentation are used as the experimental input data.
The specific experiment process is as follows:
Input: HOG extracts the gradient magnitude and gradient direction feature value
of the image pixels as input. Output: the label recognition result obtained after image
classification. The sports style is marked with “1” and the business style is marked with
“2”. When the segmented image features are used as input data, the steps are the same
as above.
After the classification result label is obtained, this experiment uses the following
two experimental evaluation indicators for evaluation:
From the results of the data shown in the table, it can be seen that the classification
accuracy of pictures that have not undergone random walk for background separation
processing is 57.5%. And the classification accuracy after separating the background
Research on the Style Classification Method 413
reaches 70%. It can be seen that the segmentation removes the interference factors of
the background. Improve the accuracy of the style classification of the main part of the
picture, and the time cost will be reduced.
5 Conclusion
The experimental results can be concluded that the accuracy of clothing style classifica-
tion after removing the background is better than that of unprocessed product image style
classification. Therefore, when categorizing clothing product styles, using appropriate
methods to remove the influence of irrelevant background factors is helpful for more
accurate classification. This article makes a certain contribution to the style classification
of clothing in e-commerce. The part that can be improved is to train the model with a
suitable public data set, perform style transfer, and use it in the self-built data set to
get the result. This article only extracts HOG edge features and then considers fusing
texture features and other features. In addition, the amount of data is not sufficient. In
the future, the method of data expansion can be considered to increase the data set to
improve the experimental results. Finally, more accurate segmentation can be applied in
segmentation mode and convenient automatic split mode.
Acknowledgments. Fund Project: Philosophy and Social Science Research Planning Project of
Heilongjiang Province (20GLE393).
References
1. 2019 China E-Commerce Report. http://dzsws.mofcom.gov.cn/article/ztxx/ndbg/202007/202
00702979478.shtml
2. Grady, L.: Random Walks for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intel.
28(11), 1768–1783 (2006)
3. Zhang, X., Wang, J., Kong, H., et al.: CT image lung tumor segmentation based on random
walk algorithm. J. Hebei Unive. (Nat. Sci. Edn.) 39(3), 311–322 (2019)
4. Wei, J., Xiang, D., Zhang, B., Wang, L., Kopriva, I., Chen, X.: Random walk and graph cut
for co-segmentation of lung tumor on PET-CT images. IEEE Trans. Image Proces. 24(12),
5854–5867 (2015)
5. Bagci, U., et al.: Joint segmentation of anatomical and functional images: applications in
quantification of lesions from PET, PET-CT, MRI-PET, and MRI-PET-CT images. Med.
Image Anal. 17(8), 929–945 (2013)
6. Dago, P.O., Ruan, S., Gardin, I., et al.: 3D random walk based segmentation for lung tumor
delineation in PET imaging. In: IEEE International Symposium on Biomedical Imaging, New
York (2012)
7. Liu, G., Hu, Z., Zhu, S., et al.: Random walk method for segmentation of head and neck tumor
PET images. J. Hunan Univ. Nat. Sci. Edn. 43(2), 141–149 (2016)
8. Firmino, M., Angelo, G., Morais, H.: Computer-aided detection (CADe) and diagnosis
(CADx) system for lung cancer with likelihood of malignancy. Bio Med. Eng. Line 15(1), 2
(2016)
9. Chao, X., Huiskes, M.J., Gritti, T., Ciuhu, C.: A framework for robust feature selection for
real-time fashion style recommendation. In: Proceedings of the 1st International Workshop
on Interactive Multimedia for Consumer Electronics, pp. 35–42. ACM (2009)
414 Y. Zhang and R. Song
10. Sun, S.: Design and implementation of a visual fashion product search engine based on hot
labeling. Sun Yat-sen University (2012)
11. Tian, X., Bao, H., Xu, C.: A pedestrian detection algorithm with improved HOG features.
Comput. Sci. 9, 320–324 (2014)
12. Sun, L., Liu, G., Liu, Y.: Multiple pedestrians tracking algorithm by incorporating histogram
of oriented gradient detections. IET Image Process. 7(7), 653–659 (2013)
13. Pan, B., Chen, W., Yao, Y.: Research on image classification algorithm in hop image
classification. Autom. Instrum. 2, 186–191 (2021)
14. Madhogaria, S., Baggenstoss, P., Schikora, M., Koch, W., Cremers, D.: Car detection by
fusion of HOG and causal MRF. IEEE Trans. Aerosp. Electron. Syst. 51(1), 575–590 (2015)
15. Alsahwa, B., Maussang, F., Garello, R.: Marine life airborne observation using HOG and
SVM classifier. In: Proceedings of OCEANS 2016 MTS/IEEE Monterey, pp. 1–5 (2016)
16. Wang, K., Wang, X., Zhong, Y.: A weighted feature support vector machines method for
semantic image classification. In: International Conference on Measuring Technology and
Mechatronics Automation, pp. 377–380 (2010)
17. Gao, H.: Image classification algorithm based on SVM. Jilin University (2019)
18. Ryu, J., Koo, H.I., Cho, N.I.: Word segmentation method for handwritten documents based
on structured learning. IEEE Sig. Process. Lett. 22(8), 1161–1165 (2015)
19. Wang, K., He, R.: Recognition method of braking intention based on support vector machine.
J. Jilin Univ. (Eng. Technol. Edn.). https://doi.org/10.13229/j.cnki.jdxbgxb20210187
20. Wu, Y., Yang, L.: Ship Image classification algorithm based on HOG and SVM. J. Shanghai
Inst. Ship Transp. Sci. 42(1), 58–64 (2019)
21. Dong, J.: Design and implementation of clothing image retrieval system based on HOG and
SVM. Sun Yat-sen University (2014)
22. Grady, L., Funka-Lea, G.: Multi-label image segmentation for medical applications based
on graph-theoretic electrical potentials. In: Sonka, M., Kakadiaris, I.A., Kybic, J. (eds.)
CVAMIA/MMBIA-2004. LNCS, vol. 3117, pp. 230–245. Springer, Heidelberg (2004).
https://doi.org/10.1007/978-3-540-27816-0_20
23. Jin, H., Bian, K.: Detection of moisture content in wheat flour by near infrared spectroscopy.
J. Chin. Cereals Oils Assoc. 25(8), 109–112 (2010)
24. Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: Proceedings
of IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR),
CA, USA, pp. 886–893 (2005)
25. Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20(3), 273–297 (1995)
26. Luo, K.: Research on CT image feature extraction and SVM classification of lung nodules.
Xihua University (2012)
27. Fan, X.: Research and application of support vector machine algorithm. Zhejiang University
(2003)
28. Zhang, X.: Research on facial expression recognition method based on deep learning.
Shanghai University of Engineering Technology (2020)
Topological Feature Analysis of RGB Image
Based on Persistent Homology
Abstract. In recent years, big data technology and its corresponding research
have become a research hotspot in various fields. The application of persistent
homology to analyze all kinds of big data is one of the research methods that have
attracted much attention. This paper mainly studies the application of continu-
ous homology to qualitative analysis of RGB three-channel image, extracts the
topological invariant features of the image, and writes them into the persistence
graph. The difference and similarity of “persistence map” of different images are
measured by Wasserstein distance. For similar images, there is a smaller distance,
while the distance between different images is larger. This method can better
distinguish images with similar topologies.
1 Introduction
Topological Data Analysis (TDA) [1] is data analysis, algebraic topology, computational
geometry, computer science, statistics, and other related fields. The main goal of TDA
is to study the qualitative characteristics of data by using the ideas of geometry and
topology. The persistent homology theory in TDA is an algebraic method to study the
topological properties of space. F frosting and Ferri [2] proposed this concept when they
studied the scaling theory in 1999. Nowadays, Persistent homology has been widely used
in computer science, medicine, social science, civil engineering and other fields. At the
same time, persistent homology has made great progress in image processing and target
recognition. Zhang Jingliang, Ju xianmeng [3] used the continuous coherence group cal-
culation methods and simple, complex row coherence group to classify and recognize
images. By calculating the homology of a simple complex, the corresponding barcode is
obtained. The topological features and the corresponding geometric structure informa-
tion of the image are obtained based on the barcode. In this paper, the color information
is further processed based on this method, and the result is good for distinguishing the
more complex and colorful images.
V = {vi , i = 0, 1, · · · , n} (1)
2.2 Filter
In the process of building complex filtration flow, it can be observed that the whole
two-dimensional point cloud model is constantly filled, at the same time, there are many
holes with simple complex as the boundary, and these holes continue to produce and
disappear with the change of the diameter of each sphere. In this paper, the change of the
sphere’s diameter is taken as the measurement standard. The length of the diameter when
each hole appears and disappears is recorded, which is used as the topological structure
of the two-dimensional model in the multi-scale range. By constructing a complex filter-
ing flow, the two-dimensional point cloud model appears obvious topological structure.
Some topological structures are filled after a short time, which is considered noise and
disturbance; In addition, some topological structures exist for a long time and are consid-
ered stable topological features. These stable topological features describe the intrinsic
characteristics of the two-dimensional model, and they are used as feature descriptors to
represent the two-dimensional model. This is the whole process of obtaining persistent
homology [5] by constructing a complex filter.
The Vietoris-Rips complex is defined as VR (x, r), the point set X = {x1 , x2 , · · · , xn },
satisfying the condition n = |X|, and r is the size of the filter value at a certain time. The
specific construction process [6] of the Vietoris-rips complex is as follows:
(1) Add points: for all point sets of x ∈ X, there is x ∈ VR0 (X, 0).
(2) Add a one-dimensional skeleton: one-dimensional simplex xi , xj exists in
VR1 (X , r), if and only if d(xi , xj ) holds.
(3) We define VR(X, r) to be the largest simplicial complex containing VR1 (X , r). In
other words, if a complex [x0 , x1 , · · · , xk ] is VR(X, r), if and only if all edges of
this complex exist on VR1 (X , r).
It can be seen that when the color difference between the two pixels is large, the
RGB itself will be too large, which will have a greater impact on the calculation of the
coherence. Therefore, we weigh the three components of RGB, and the weighting is
based on the image size. The weight formula is (4):
Where min(length, width) denotes the minimum length and width of the image. Since
the maximum value of RGB is 255, we set the RGB weight to be 1 when the minimum
value of the image length and width reaches 2550, which is ten times 255. From Fig. 2,
its size is 972 * 442, and the two-point five-dimensional data after weighting is (60,
303, 43.61, 43.80, 42.82), (631, 303, 7.85, 7.66, 8.64), and its distance is 65.09, For the
unweighted distance is reduced a lot. Although it is still much larger for the positional
distance alone, it will not select two points close to each other when selecting landmark
points. Therefore, the algorithm can ensure that certain color information is retained
under the premise of focusing on distance information.
Image complexity, regional complexity, and object complexity are three scales. It can
be seen that picture (a) has richer color information for butterflies, while picture (b) lacks
color information for scissors. It can also be said that picture (a) is more complicated, and
picture (b) is simpler. But the image information is more complicated. It is believed that
more data points are needed to represent the information of the original image. Therefore,
we introduce the image complexity [9] to measure the image complexity description can
be described separately from the overall perspective, the regional perspective, and the
target perspective, which corresponds to the entire image. This article mainly studies the
overall description of image complexity.
420 J. Ma et al.
Where N denotes the number of gray levels and ni is the number of each gray
level. We mainly consider the color complexity information, referring to the gray image
information entropy to obtain the color image information entropy. We only need to
change the gray level to the color level. However, grayscale images have 256 grayscale
values in total, while color images have 2563 color levels. This magnitude is too large,
which is not conducive to calculation. At the same time, it is almost impossible for an
image to contain all color levels, and in many cases, we do not need to divide the color
levels too finely. As shown in Fig. (2), although the black part of the scissors has slightly
different RGB components, it does not disproportionately impact the overall topology.
Therefore, similar colors can be regarded as a color level. In this paper, 64 adjacent
colors (four values of RGB are each combined) are regarded as a color level, so the
number of selected points for each image is H × 10.
When calculating Wasserstein distance using persistence graphs, due to the non-
negativity of distance, the number of birth-death points in each group of persistence
graphs will have a greater impact on the calculation. Generally, the distance with more
feature descriptors in two pictures will be larger than the distance with fewer feature
descriptors in the other two pictures. Still, the larger distance due to the gap of feature
descriptors does not mean that the gap in the topological structure is too large. If there
is too much difference in the number of feature descriptors between the two pictures,
the structural gap between the two is large. Therefore, we multiply the last calculated
distance by a coefficient max(i, j)/min(i, j), where i and j are the number of feature
descriptors in the group, respectively. When the two numbers are equal, the coefficient
is 1. The greater the difference between the two numbers, the greater the coefficient and
the greater the distance.
5 Numerical Experiment
5.1 Wasserstein Distance Introduction
Wasserstein distance [12] was first used to compare the difference between two his-
tograms and then describe the distance between two probability distributions. It can
be generally understood as the minimum transportation for moving many objects from
one location to another. The cost is defined as for formula (6), where X and Y are the
d-dimensional space sets satisfying the sum probability distribution, p represents the Lp
norm, and inf represents the lower bound. The value of p is 1 in the method in this paper.
1/
Wp (µ, v) = inf EX − Y p p , p ≥ 1 (6)
X ∼µ
Y ∼v
Firstly, the image sampling points are selected according to the complexity of the image,
and then continuous coherence is performed. Finally, the persistence map is obtained
according to the coherence information, and finally, the distance between the persistence
pictures is calculated according to the information of the persistence map. Figures 6 and
7 are two images and the resulting persistence maps:
Topological Feature Analysis of RGB Image 423
The image data used in this experiment are all from the Internet, and pure white
background images are used. Figure 8 shows 16 randomly selected pictures of different
sizes, including butterflies, scissors, rings, and dragonflies.
Tables 1, 2, and 3 are the data of the comparison experiment on the 8 images in
Fig. 10. Table 1 shows the Wasserstein distance between 4 butterfly pictures and 4
scissors pictures in Fig. 8 after introducing color complexity:
Table 2 shows the Wasserstein distance between butterfly and scissors without
introducing color complexity, and the sampling points for both are selected as 100.
If all four types of data are displayed, a 16×16 table is required, which is inconvenient
to display here. Table 3 and Table 4 respectively show the average distance between the
four types of images with and without complexity.
424 J. Ma et al.
As can be seen from Table 1 and 3, after introducing the weighted color space and
color complexity, it can be seen that the distance between similar images is smaller, and
the distance between different images is larger, which has a good effect for comparing
the differences and similarities between images. However, the data in Tables 2 and 4 are
Topological Feature Analysis of RGB Image 425
Table 3. Average distance between the four types of images with complexity.
Table 4. Average distance between the four types of images without complexity.
not very obvious for the distance between the same or different images, and the effect
is poor. The improved feature descriptor can better solve the problem of Wasserstein
distance repeated calculation.
426 J. Ma et al.
6 Conclusion
This paper processes the RGB space and introduces the color complexity to measure
the color richness of the image to define how much information the image contains.
The more complex the image, the richer the information it contains. Then the color
information is combined with the topological information, and the similarity and dif-
ference are measured by calculating the Wasserstein distance of the two images. The
distance between the two images is related to its topological features and related to
color information. Compared with traditional topological feature extraction methods,
this measurement method has a better effect on magnifying the differences of different
types of images.
References
1. Edelsbrunner, H., Harer, J.: Computational Homology: An Introduction, pp. 23–47. American
Mathematical Society (2010)
2. Frosini, P., Landi, C.: Size theory as a topological tool for computer vision. Pattern Recogn.
Image Anal. 9(4), 596–603 (1999)
3. Zhang, J., Ju, X.: Application of persistent homology in image classification and recognition.
J. Appl. Comput. Math. 31(04), 494–508 (2017)
4. Atienza, N., Gonzalez-Diaz, R., Rucco, M.: Persistent entropy for separating topological
features from noise in Vietoris-Rips complexes. J. Intel. Inf. Syst. 52(3), 637–655 (2019)
5. Carrière, M., Oudot, S.Y., Ovsjanikov, M.: Stable topological signatures for points on 3D
shapes. Comput. Graph. Forum 34(5), 1–12 (2015)
6. Zomorodian, A.: Fast construction of the Vietoris-Rips complex. Comput. Graph. 34(3),
263–271 (2019)
7. Liu, S., Chen, Y., Wang, H.: An improved hybrid encryption algorithm for RGB images. Int.
J. Adv. Sci. Technol. 95, 37–44 (2016)
8. De Silva, V., Carlsson, G.: Topological estimation using witness complexes. In: Proceedings
of the 1st Eurographics Conference on Point-Based Graphics, pp. 157–166. Eurographics
Association, Aire-la-Ville, Switzerland (2004)
9. Mario, I., Chacon, M., Alma, D., et al.: Image complexity measure: a human criterion free
approach. In: Proceedings of IEEE Annual Meeting of the North American Fuzzy Information
Processing Society, Detroit, MI, USA, pp. 241–246 (2005)
10. Gao, T., Li, F.: Using persistent homology to represent online social network graphs. In:
2017 IEEE 14th International Conference on Mobile Ad Hoc and Sensor Systems (MASS),
pp. 555–559. IEEE Computer Society (2017)
11. Adams, H., Chepushtanova, S., Emerson, T., et al.: Persistence images: a stable vector
representation of persistent homology. J. Mach. Learn. Res. 18, 1–35 (2015)
12. Panaretos, V., Zemel, Y.: Statistical aspects of Wasserstein distances. Ann. Rev. Stat. Appl.
6(1), 405–431 (2019)
Machine Learning and Deep Learning
and their Applications
A Novel Deep Image Matting Approach Based
on DIM Model
Abstract. The digital image matting task is an important research field of com-
puter vision, and the method of deep image matting is a new and efficient automatic
matting method. In the task of deep image matting, to solve the problem that the
details of edge in the feature images of the decoder are easy to lose, the layer-
skipping connection is introduced to concatenate the feature images, which have
the same size in the channel dimension between the encoder and the decoder, it
also realizes the fusion of shallow detailed information and deep semantic infor-
mation. To get deeper semantic information and wider receptive field, the encoder
uses the VGG19 network obtained by migration learning, and the decoder uses
the larger convolutional kernel of 9 × 9 accordingly. At the same time, in order to
solve the problems of slow speed in convergence and insufficient ability of refine-
ment in the refined network, four convolutional layers with the residual structure
are added in this network. Experimental results show that the improved network
has higher accuracy and richer information of shape. The ability of generalization
in this model is also stronger.
1 Introduction
The automatic matting task is an important direction in the field of image processing
and computer vision. It needs to extract foreground objects from images accurately, and
it pays more attention to the details of the edge in foreground objects. Unlike image
segmentation [1], which pays more attention to the division of image region, image
matting is based on users’ rough foreground and background information. Effectively
extracts edge details of foreground objects. This technology is widely used in image
synthesis, post-production of film, and other practical works.
Generally, in the matting work, the pixel color in an image is regarded as a linear
combination of foreground color and background color. The specific formula is:
In it, i is the pixel index of the image, and α i is the foreground opacity of the image
at the pixel point of i. I i is a known data.
The matting work can be equivalent to solving the values in the above equation, that
is, solving the α, which contains the foreground opacity of each pixel. However, for a
known image and its corresponding foreground and background images, they usually
contain three color channels: R, G and B, that is, the three above equations need to be
solved uniquely. At this time, the three equations contain three known quantities (the
values on three channels) and seven unknown quantities (the values on three channels
and unique values). Therefore, it is impossible to solve this problem by only a constraint
equation. So, users must provide a trimap with known foreground, background, and
unknown areas as auxiliary information formatting.
2 Related Work
At present, the common methods of image matting [2] can be roughly divided into
three categories: 1) sampling-based method; 2) the method based on propagation; 3) the
method based on deep learning.
The sampling-based matting method [3] generally selects candidate sampling points
for each unknown pixel in the known region on the trimap, then selects suitable fore-
ground and background sampling pairs in the candidate sampling points, and calculates
the values of sampling pairs as the values of unknown pixels. The main algorithms
include Robust Matting [4], Shared Matting [5], and Global Matting [6].
In the method based on propagation, the weighted sum of α values for similar pixels
in known areas is calculated in a small local range of pixels in unknown areas, and then
the α values of pixels in unknown areas are estimated. The main algorithms include
Poisson Matting [7], Closed-Form Matting [8], and KNN Matting [9].
The matting method based on deep learning regards the matting process as a super-
vised learning process. A linear or nonlinear model between α parameter and image
color is constructed through training and learning, and then the α value of unknown
area pixels is estimated. The main algorithms include DCNN Matting [10], Deep Image
Matting [11], Semantic Human Matting [12] and Deep Automatic Portrait Matting [13],
etc. The model in this paper is based on the Deep Image Matting algorithm (as shown
in Fig. 1), and then we have improved it.
The main work of this paper includes: 1) we use the layer-skipping connection in the
auto-encoder section of a DIM model to realize the fusion of the information at different
depths; 2) we use the VGG19 network as the main body to enhance the ability to extract
for deep semantic features; 3) in the refined part of this model, the residual structure of
convolution is adopted in corresponding positions, it enhances the refined ability of this
network and ensures the robustness of this model.
VGG19. In the first part of this model, the encoder is based on the VGG19 network
pretrained by the ImageNet dataset. The original full connection layer is removed, and
the four-channel kernels’ parameters in the first convolutional layer are initialized to
0. The VGG19 network is used as the backbone of the encoder. On the one hand, the
existing experimental results show that the VGG network has a stronger generalized
ability than other mainstream deep learning models, and it is usually the first choice for
transformation of the network; on the other hand, in the series networks of VGG, the
VGG19 network has stronger ability of extraction than other VGG networks, and it has
a larger receptive field on feature maps.
In VGG19, a convolutional block composed of four convolutional layers, whose
kernel size are 3 × 3, can be simulated as a convolutional layer, whose kernel size
is 3 × 3, so VGG19 can generate a larger receptive field and obtains more accurate
semantic information. Moreover, in the decoder part, we use the convolutional layers
whose kernel size is 9 × 9 on the corresponding position. The experimental results show
that the improved VGG19 network can extract more detailed information from images.
Therefore, our model uses layer-skipping connection in the auto-encoder part (as
shown in Fig. 4). It is embodied as follows: in the encoder, the output of each convo-
lutional block is concatenated with the input of each convolutional block at the corre-
sponding position in the decoder, and the concatenated feature maps are used as the input
of the next convolutional processing in the decoder. Experimental results show that the
shape of the original image is preserved by using the layer-skipping connection, and it
realizes the fusion of shallow detailed information and high-level semantic information.
A Novel Deep Image Matting Approach 433
Residual Structure. For a shallow network, because of the limited complexity, it is easy
to lead to the problem of insufficient extraction and insufficient ability of expression on
large datasets; however, if the network’s depth is increased blindly, on the one hand, it
may lead to the degradation of the network [17], and the dispersion of gradient [18], on
the other hand, it may have the over-fitting problem if don’t have massive datasets.
Therefore, in the refined part of our model, we use the residual structure (as shown
in Fig. 5), it is realized as follows: seven convolutional layers whose kernel size are 3 ×
3 are used for the refinement of α, and short-circuit layers are added to the second and
the fifth convolutional layers [18], that is, the input and the output of each short-circuited
convolutional block are directly added, to ensure that when the ability to express of our
model is insufficient, the short-circuit fails and the short-circuited convolutional layers
are added to our model; when the network degrades, the short-circuit layers work, and
the corresponding convolutional layers are short-circuited, it accelerates the speed of
convergence and reduces the risk of over-fitting.
The total loss function in this paper is mainly composed of two parts, and one is the loss
function of α prediction, the other is the loss function of α synthesis.
434 G. Yao and Z. Ma
The loss function of α prediction represents the error between the predicted values
of α and the true values of α, the specific formula is:
2
Lα =
i
αpi − αgi + ε2 . αpi , αgi ∈ [0, 1] (2)
In it, α ig means a certain pixel’s value on the real α, α ip is the pixel’s value at the
corresponding position on the predicted α, and ε is a certain minimum constant, it is
10–6 here; the loss function of α synthesis represents the error between the real RGB
images and the predicted RGB images. The predicted RGB images are synthesized of the
values of α prediction, foreground images and background images, the specific formula
is:
2
Lc =
i
cpi − cgi + ε2 . (3)
In it, cip means a certain pixel’s value on the predicted RGB image and cig is the pixel’s
value at the corresponding position on the real RGB image; the total loss function in this
paper is the weighted sum of these two loss functions, the specific formula is:
L = w × Lα + (1 − w) × Lc . (4)
In it, Lα means the loss function of α prediction, Lc means the loss function of α synthesis,
and w means a certain weight, it is 0.5 here. In the training process of the auto-encoder
network, we use the total loss function; but in the refined network and whole network,
we use the loss function of α prediction.
4 Experiment
4.1 Experimental Environment
The experiment in this paper is performed in Windows 10.0 system, and it is trained
and tested by building the framework of Tensorflow 2.4.1 and Keras 2.4.1. The CPU is
conFig.d with Inter (R) i7-10700KF, and the GPU is NVIDIA RTX2080-ti. This work
is supported by the Youth Innovation Talent Support Program of the Harbin University
of Commerce under Grant 2020CX39.
By default, our dataset uses the dataset in Xu’s paper. The batch size of our dataset
is set to 8, and the size of the input image in training dataset is set to 320 × 320. In the
stage of training, Adam optimizer is used in the auto-encoder network, the refinement
network and the whole network. The learning rates are set to 1.0 × 10–5 , 1.0 × 10–4 ,
and 1.0 × 10–5 , respectively. The training epoch is 20 by default, and the loss function is
the mean square error. In the stage of test, two general measurement indexes, the mean
square error (MSE) and the gradient error (Gradient), are set. They are used in many
matting algorisms, such as Fast Automatic Portrait Matting Based on Multitask Deep
Learning [19], Automatic Matting Algorithm for Human Foreground [20], etc. Their
formulas are listed as follows:
n 2
MSE(x, y) = xi − yi (5)
i=0
A Novel Deep Image Matting Approach 435
n
i
Gradient(x, y) = ∇x − ∇yi (6)
i=0
Among of them, xi is the predicted α of the image, yi is the real α, ∇ means the gradient
of the image.
The experiments are divided into two parts: the first part is the experiments between each
improved algorithm and the original DIM algorithm; The second part is the experiments
between the algorithm after adding all the improvements and other classical matting
algorithms. In the first part, we will compare the results which are gotten by different
structures after training, it includes whether there is a VGG19 network, whether there
is a layer-skipping connection, and whether there is a residual structure. The tested
dataset is the Composition-1k dataset consistently. In the comparison process, the batch
size, the loss function, the learning rate and other experimental factors are consistent,
and they are all the default values in the above experimental environment. The second
part compares the improved algorithm with the KNN Matting algorithm and the DCNN
Matting algorithm. The size of pictures in our tested dataset is reduced to 0.3 times (after
considering the speed of matting in the KNN algorithm). At the same time, we keep the
other experimental factors, such as the batch size, the loss function, and the DCNN
algorithm’s learning rate, which also belongs to the deep learning algorithm, consistent
with our algorithm. They are all the default values. The MSE and the Gradient are used
as the performance indexes of these experiments.
The first part of the experiment is a comparative experiment between each improved
algorithm with the original DIM algorithm: in the auto-encoder part of this network (as
shown in Table 1), the original algorithm uses VGG16 as the main body, after changing
to VGG-19 as the main body, the MSE drops significantly, it is mainly because the
deeper convolutional network has the ability of extraction for deeper dimension, and the
larger convolutional kernel can get the larger receptive field, but the Gradient doesn’t
have obvious change. After adding the layer-skipping connection, the Gradient begins
to drop significantly. The layer-skipping connection fuses the shallow features of shape
and the deep features of segmentation and keeps the richness and integrity of the output
information. The original model comprises four simple convolutional layers in the refined
part of this network (as shown in Table 2). After adding the convolutional layers with
residual structure, it not only makes the extraction of information richer but also avoids
the risk of over-fitting in our model. It makes these two performance indexes decrease
significantly. Finally, compared with the original model, the MSE decreased by 0.0020,
and the Gradient decreased by 0.0017. This fully demonstrates the effectiveness of our
improvements in the original algorithm (as shown in Fig. 6) (Table 3).
436 G. Yao and Z. Ma
The second part is the comparative experiment between our improved algorithm
and other classical matting algorithms. Because the KNN algorithm takes a long time
to calculate, we reduce the image size of Composition-1k dataset by 0.3 times as our
tested dataset. According to the test (as shown in Table 4), the KNN algorithm has a
good performance on smaller images, but the DIM algorithm and the improved DIM
algorithm have better performance, especially the improved DIM algorithm has the
best performance, which further shows the feasibility and the effectiveness of these
improvements.
The α results of several classical automatic matting methods on Composition-1k
(reduced by 0.3 times) are shown in Fig. 7: the KNN matting method gets a good α, but
it does not perform well in details, such as human’s hair and animal’s hair; the DCNN
matting method doesn’t get an ideal result, and a large misjudged area appears on the
output image, it may be related to the insufficient dataset and the small size of tested
images. The DIM algorithm has achieved a better α, but there are also some problems
of missing details; our improved DIM algorithm has gotten the ideal output, especially
in detail processing, it preserves much rich information of shapes, such as clothes’
corners and hair of human, the nose and hair of the animal, they are all visible in our α.
Experimental results show that our improved algorithm can obtain more accurate results
of α prediction, and the generated α has higher quality.
A Novel Deep Image Matting Approach 437
GT α
GT α
Fig. 6. Experimental results of DIM and Ours
DIM Ours GT α
DIM Ours GT α
Fig. 7. Experimental results of automatic matting
5 Conclusion
By analyzing the existing problems of the network in deep image matting, this paper
proposes to use the VGG19 network as the backbone of the encoder to enhance the ability
of extraction on feature maps; the auto-encoder uses the layer-skipping connection to
get a more detailed and rich shape information; in the part of refined network, we use a
deeper convolutional network with residual structure, it improves the refined ability and
the speed of convergence in this network. Experiments show that these improvements
significantly improve the accuracy of the synthetic dataset, and enhance the ability to
capture details for the generated α.
Although the improved model has obvious improvement in α prediction and extrac-
tion of shaped information, it still has shortcomings in real-time and portability, which
are mainly due to the problems that this model relies too much on hardware’s acceler-
ation and the amount of dataset is relatively insufficient. At the same time, it is also a
common problem that needs to be solved urgently in deep learning and matting. The
A Novel Deep Image Matting Approach 439
following research will be an important direction for us to reduce the parameters [21] in
our model without affecting the overall performance and to get more matting datasets
with lots of unique foreground pictures.
References
1. Huang, P., Zheng, Q., Liang, C.: Overview of image segmentation methods. J. Wuhan Univ.
(Sci. Edn.) 66(06), 519–531 (2020)
2. Liang, Y., Huang, H., Cai, Z., Hao, Z., Feng, F.: Summary of natural image matting technology.
Comput. Appl. Res. 38(05), 1294–1301 (2021)
3. Tang, J., Aksoy, Y., Oztireli, C., et al.: Learning-based sampling for natural image matting.
In: Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3055–3063(2019)
4. Wang, J., Cohen, M.: Optimized color sampling for robust matting. In: IEEE Conference on
Computer Vision & Pattern Recognition, pp. 1–8. IEEE Computer Society (2007)
5. Gastal, E., Oliveira, M.: Shared sampling for real-time alpha matting. Comput. Graph. Forum
29(2), 575–584 (2010)
6. He, K., Rhemann, C., Rother, C., et al.: A global sampling method for alpha matting. In: IEEE
Conference on Computer Vision & Pattern Recognition, pp. 2049–2056. IEEE (2011)
7. Sun, J., Jia, J., Tang, C., et al.: Poisson matting. ACM Trans. Graph. 23(3), 315–321 (2004)
8. Chen, Q., Li, D., Tang, C.: KNN matting. In: IEEE Conference on Computer Vision and
Pattern Recognition, vol. 35, no. 9, pp. 2175–2188. IEEE (2013)
9. Levin, A.: A closed form solution to natural image matting. IEEE Comput. Soc. 30(2), 228–
242 (2008)
10. Cho, D., Tai, Y.-W., Kweon, I.: Natural image matting using deep convolutional neural net-
works. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906,
pp. 626–643. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46475-6_39
11. Xu, N., Price, B., Cohen, S., Huang, T.: Deep image matting. In: IEEE Conference on
Computer Vision and Pattern Recognition, pp. 2970–2979 (2017)
12. Chen, Q., Ge, T., Xu, Y., Zhang, Z., Yang, X., Gai, K.: Semantic human matting. In:
Proceedings of the 26th ACM international conference on Multimedia (2018)
13. Shen, X., Tao, X., Gao, H., Zhou, C., Jia, J.: Deep automatic portrait matting. In: Leibe, B.,
Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 92–107. Springer,
Cham (2016). https://doi.org/10.1007/978-3-319-46448-0_6
14. Lutz, S., Amplianitis, K., Smolic, A.: AlphaGAN: generative adversarial networks for natural
image matting. In: British Machine Vision Conference (2018)
15. Badrinarayanan, V., Kendall, A., Cipolla, R.: SegNet: a deep convolutional en-coder-decoder
architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 1 (2017)
16. Shelhamer, E., Long, J., Darrell, T.: Fully convolutional networks for semantic segmentation.
In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3431–3440
(2015)
17. Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reduc-
ing internal covariate shift. In: International Conference on Machine Learning, pp. 448–456
(2015)
440 G. Yao and Z. Ma
18. He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition. In: IEEE
Conference on Computer Vision and Pattern Recognition (CVPR) (2016)
19. Xu, Z., Yang, Y.: Fast automatic portrait matting based on multitask deep learning. J. Wuhan
Univ. (Eng. Edn.) 53(08), 740–745+752 (2020)
20. Ran, Q., Feng, J.: Automatic matting algorithm for human foreground. J. Comput. Aided Des.
Graph. 32(02), 277–286 (2020)
21. Wang, R., Xu, S., Huang, J.: Image matting technology based on deep learning. JShanghai
Univ. (Nat. Sci. Edn.) 41(05), 1–9 (2021)
Application of Bloch Spherical Quantum
Genetic Algorithm in Fire-Fighting Route
Planning
Wei Zhao1,2 , Xuena Han1(B) , Hui Li1,2 , Zeming Li1 , and Xue Tan1
1 Harbin University of Commerce, Harbin 150028, China
2 Heilongjiang Provincial Key Laboratory of Electronic Commerce and Information Processing,
Abstract. Aiming at the fire-fighting path planning problem, the Bloch Quantum
Genetic Algorithm (BQGA) is studied and used. In this algorithm, first use Bloch
spherical coordinates to encode the quantum chromosome, construct the update
strategy of the quantum chromosome, establish the angle size and direction of the
quantum revolving door, and finally construct the phase formula in the mutation
operation. In this article, the three-dimensional (3D) point data obtained from the
three-dimensional building floors are converted into a 3D raster map; considering
the height gap of different areas, generating the feasible shortest path route can
directly generate a clear movement trajectory in the raster map to ensure the
driving process People’s safety in China. The experimental results show that the
algorithm does not need a priori knowledge of the map, has strong reliability and
practicability, and has less evolutionary algebra, fast convergence speed, and high
optimization efficiency.
1 Introduction
Quantum genetic algorithm combines the advantages of quantum computing and genetic
algorithm. A Quantum genetic algorithm is a brand-new evolutionary algorithm. It is
based on quantum principles and uses qubit encoding and quantum gates to update
the population to find the optimal global solution. Compared with traditional, quan-
tum evolutionary algorithms have a small population size, fast calculation speed, and
strong global optimization ability. Therefore, the quantum genetic algorithm has great
advantages and contains strong vitality, and has extremely high theoretical value and
application prospects [1].
Lewandowski et al. [14] used the Bloch sphere to describe the behavioral modeling
system in quantum reversible logic design. Later, Bloch spherical coordinates are often
used in quantum logic operations and quantum algorithms [1]. A quantum genetic algo-
rithm based on Bloch spherical coordinates was first proposed in China by Li Panchi
et al. [1]. The algorithm directly uses the Bloch spherical coordinates of quits to encode
quantum chromosomes. It uses quantum revolving gates to update qubits. A simple and
practical method is proposed to determine the direction of the turning angle of the revolv-
ing door. Mutation and a new operator based on the qubit Bloch spherical coordinate
are proposed [1]. At the same time, the three Bloch spherical coordinates of the quit
are all regarded as gene positions; therefore, each chromosome has three gene chains.
Compared with traditional optimization algorithms to solve complex nonlinear prob-
lems, the quantum genetic algorithm is produced under the guiding ideology, inherited
the advantages of the genetic algorithm itself, and improved some of the original short-
comings to some extent, such as the shortcomings of too large a population and too
long convergence time. Achieve better optimization effects than traditional algorithms.
The quantum genetic algorithm is applied in different disciplines to solve a variety of
practical problems. Facts have proved that its calculation results are often much better
than those obtained by traditional calculation methods, so they are widely used in differ-
ent optimization problems [1]. The simulation results of the extreme value optimization
of the fire path planning function show that the optimization performance of BQGA is
better than the ordinary ant colony algorithm [2].
This article focuses on the study of fire-fighting path planning for large-space building
fires. It is based on the existing literature and adopts a based on the Bloch spherical
coordinate algorithm. A quantum chromosome update strategy is constructed and its
population size can be adjusted freely, the convergence speed is fast, it has a strong
global optimization ability and rich population diversity and randomness [2].
On the Bloch sphere, a point p can be determined by two angles θ and ϕ, as shown in
the figure below:
As shown in Fig. 1, on the 3D Bloch sphere, any qubit corresponds to a point on the
Bloch sphere. Then the qubit can be expressed in Bloch spherical coordinates as:
In BQGA, the Bloch spherical coordinate coding of qubits is directly used. Let Pi
be the i chromosome in the population, and the BQGA coding method is:
cosϕi1 sinθi1 cosϕin sinθin
Pi = sin ϕi1 sin θi1 · · · sin ϕin sin θin (2)
cos θ cos θ
i1 in
Where 0 ≤ θ ≤ π, 0 ≤ ϕ ≤ 2π .
j 1
Xix = bj 1 + xij + aj 1 − xij
2
j 1
Xiy = bj 1 + yij + aj 1 − yij
2
j 1
Xiz = bj 1 + zij + aj 1 − zij (3)
2
where i = 1, 2, · · · , m and j = 1, 2, · · · n. m is the size of the population, n is
quantum digits.
3 Algorithm Flow
In the ideal algorithm, a person is regarded as a point with no size. The grid method [7]
is used to simplify a building as a whole to obtain an equal-sized 21 × 21 grid, and then
assign a value to each grid. The value of the grid represents the height, and then BQGA
is used for simulation. Proceed as follows:
Step 1 Initialize the population. Set the population size to popsize, the number of iter-
ations to maxgen, the corner step length to shiftstep, and the mutation probability to
Pm .
Application of Bloch Spherical Quantum Genetic Algorithm 445
As shown in Fig. 5 and 6, they respectively show the change trend of the best
individual fitness under different algorithms. The abscissa represents the number of
iterations, and the ordinate represents its corresponding fitness value. Figure 5 shows
the change trend of the best individual fitness of the BQGA algorithm, which drops from
the initial about 10000 to nearly 8100, which is a difference of about 1900. Figure 6
shows the change trend of the best individual fitness of the ACA algorithm, from the
initial 137 to nearly 115, the difference is about 22. In comparison, it is obvious that the
optimization efficiency of BQGA algorithm is higher than that of ACA.
Application of Bloch Spherical Quantum Genetic Algorithm 447
Fig. 5. The change trend of the best individual fitness of the BQGA model
448 W. Zhao et al.
Fig. 6. The changing trend of the best individual fitness of the ACA model
5 Conclusion
This paper uses a qubit-based Bloch spherical quantum genetic algorithm to solve the
three-dimensional fire path planning problem. Its coding method has three advantages:
First, it can avoid the random characteristics brought by the binary code generated by
measuring the qubit. Second, it can avoid the frequent binary numbers decoding process.
Third, it can expand the number of globally optimal solutions and increase the probability
of obtaining optimal global solutions. This method uses grid division to process the
3Dbuilding environment and then calculates the shortest path distance traveled by people
through the algorithm. A typical test example shows that the algorithm can always
achieve a smooth and safe path in a randomly distributed working environment with
various shapes. At the same time, the algorithm has all aspects of convergence speed,
iteration steps, execution time.
References
1. Li, S., Li, P.: Quantum Computation and Quantum Optimization Algorithms. Harbin Institute
of Technology Press, Harbin (2009)
2. Wang, J., Wang, H., Feng, L.: Building fire evacuation path planning based on quantum ant
colony algorithm. Comput. Meas. Control 28(07), 167–172 (2020)
3. Li, J., Han, K., Bao, T.: An improved bloch spherical quantum genetic algorithm and it’s
application. J. Railway Sci. Eng. 13(11), 2262–2269 (2016)
Application of Bloch Spherical Quantum Genetic Algorithm 449
4. Yan, F., Iliyasu, A.M., Liu, Z.-T., Salama, A.S., Dong, F., Hirota, K.: Bloch sphere-based
representation for quantum emotion space. J. Adv. Comput. Intel. Intel. Inf. 19(1), 134–142
(2015)
5. Yang, S., Xu, Y., Li, P.: Quantum-inspired artificial fish swarm algorithm based on the bloch
sphere search algorithm. Inf. Control 43(06), 647–653 (2014)
6. Chen, Y., Li, A., Huang, Y.: Quantum particle swarm optimization based on bloch coordinates
of qubits. J. Comput. Appl. 33(02), 316–318+322 (2013)
7. Yi, Z., He, R., Hou, K.: Quantum artificial bee colony optimization algorithm based on bloch
coordinates of quantum bit. J. Comput. Appl. 32(07), 1935–1938 (2012)
8. Yi, Z., Hou, K., He, R.: Adaptive quantum genetic algorithm based on bloch sphere. Comput.
Eng. Appl. 48(35), 57–61 (2012)
9. Li, P.: Quantum computation with applications to intelligent optimization and control. Harbin
Institute of Technology (2009)
10. Jafarizadeh, M.A., Karimi, N., Zahir, H.: Quantum discord for generalized bloch sphere states.
Eur. Phys. J. D 68(5), 1–9 (2014)
11. Yang, L., Zhang, A., Zhang, J., Song, J.: Real-time path planning of velocity potential for
robot in grid map environment. Comput. Eng. Appl., 1–8 (2021). http://kns.cnki.net/kcms/
detail/11.2127.tp.20210514.1007.008.html
12. Guo, L., Xiao, S., Wang, Z.: Geometric phase of two-level mixed state and bloch sphere
structure. Int. J. Theor. Phys. 52(9), 3132–3140 (2013)
13. Bartkiewicz, K., Miranowicz, A.: Optimal cloning of arbitrary mirror-symmetric distributions
on the bloch sphere: a proposal for practical photonic realization. Phys. Scr. T147, 014003
(2012)
14. Lewandowski, M., Ranganathan, N., Morrison, M.: Behavioral model of integrated qubit gates
for quantum reversible logic design. In: Conference: IEEE Computer Society International
Symposium on VLSI (2013). https://doi.org/10.1109/ISVLSI.2013.6654658
Chaotic Neural Network with Legendre
Function Self-feedback and Applications
Abstract. A new transient chaotic neuron model which introduces Legendre func-
tion into the self-feedback term is constructed. The dynamic characteristics of the
single neuron are analyzed by the time evolution graph and the inverted bifurca-
tion graph of the largest Lyapunov exponent. The network parameters are set for
settling the combinatorial optimization problem. The effectiveness of this model
has been verified by the simulation results in nonlinear function optimization,
traveling salesman problem (TSP).
1 Introduction
Transiently chaotic neural networks (TCNN) have been extensively researched [1–5],
which successfully and efficiently have solved TSP because of its complex nonlinear
dynamical behavior. By decreasing the self-feedback link weight exponentially, which
ensures, in chaotic search behavior, that can be astringed a point stably and behave
transiently and introducing a linear self-feedback, Chen-Aihara has proposed a CNN
with CSA [1]. Different self-feedback functions can generate different chaotic traversal
searches and different chaotic dynamics. Xu Nan, Xu Yaoqun, Sun Ming, Ye Yonggang
present a self-feedback term with trigonometric functions, nonlinear wavelet functions,
and Bessel functions to their models [6–10]. During the search process, this is why their
models have more complex chaotic search capabilities.
In this study, we put forward a Legendre function self-feedback CNN, which is unlike
the linear self-feedback term of Chen and Aihara’s network, because it is not simple and
nonlinear. Our presented chaotic neural network, with the suitable arguments, has a
stronger ability in terms of falling into a local minimum point and optimal performance.
These can be shown by simulations about two aspects, optimizing continuous function
and 10 cities TSP problem, and for practical questions, that is very complicated and
classical [11].
n
yi (t + 1) = kyi (t) + α[ wij xj + Ii ] − zi (t)g(xi (t) − I0 ) (2)
j=1,j=i
n
g(x) = λi P2i+1 (x) (3)
i=1
dn
Pn (x) = 1
2n ·n! · dxn (x
2 − 1)n n = 0, 1, 2, · · · (5)
The term zi (t)g(xi (t)−I0 ) is the nonlinear self-feedback term. In the usual stochastic
annealing process, the variable corresponds to the temperature. How does the chaotic
neural network introduce the chaotic mechanism? It asks this question that the connection
of the self-feedback weight, step by step, decreases as the value of zi (t) trails off. Then
the chaotic behavior in the chaotic neural networks is important to be a whole search
role. In order to anneal for the network, Eq. (4) is a program that adopts exponential
cooling. And the symbolic meaning is in Table 1.
Symbol Meaning
xi The output of neuron i
yi The inside of the state of neuron i
zi The self-feedback connection weight
Ii Input bias of neuron i
λi The parameter of the nonlinear function (0 ≤ λi ≤ 1)
g(x) The nonlinear function
wij The link weight, wij = wji
β The damping factor
ε Steepness parameter of the activation function (ε > 0)
(continued)
452 Y. Zhang et al.
Table 1. (continued)
Symbol Meaning
k Damping factor of nerve membrane (0 < k < 1)
Pn Legendre function
I0 Positive parameter
n
g(x) = λi P2i+1 (x) (8)
i=1
For the model, set i to 1 and 2, the chaotic neuron model has the next equations:
dy(t + 1) dg(x) dx
= k − z(t) · (11)
dy(t) dx dy(t)
dx dx(t) 1
= = x(t)(1 − x(t)) (12)
dy(t) dy(t) ε
dg(x) λ1 λ2
= (15x2 − 3) + (315x4 − 210x2 + 15) (13)
dx 2 8
The Lyapunov exponent can reflect the chaotic power of the model and then reflect the
strength of the global optimization ability of the network. In one-dimensional mapping,
Chaotic Neural Network with Legendre Function 453
λL > 0 indicates that the model is in a state of chaos. The higher the value, the stronger
the confusion. λL = 0 means that the model is in a stable boundary state. λL < 0
indicates that the motion is stable and the model does not have a chaotic state.
The self-feedback function in the model is a combination of Legendre functions of
λ1 = 1/3, n = 3 and λ2 = 1/100, n = 5, as shown in Fig. 1.
Then, we made some emulation by Matlab that the parameters are these:
ε = 0.018, λ1 = 1/3, λ2 = 1/100, y(1) = 0.1, z(1) = 0.98, k = 0.1, I0 = 0.85,
β = 0.001. And it takes our some time. Time evolution diagram and the inverted
bifurcation diagram of the maximum Lyapunov exponent of the neuron are shown as
Fig. 2.
Fig. 2. State bifurcation figure and the maximal Lyapunov exponents of the single neuron when
β = 0.001, λ1 = 1/3 and λ2 = 1/100
Fig. 3. State bifurcation figure and the maximal Lyapunov exponents of the single neuron when
β = 0.001, λ2 = 5/100 and λ1 = 1/3
Fig. 4. State bifurcation figure and the maximal Lyapunov exponents of the single neuron when
β = 0.001, λ2 = 1/10 and λ1 = 1/3
neurons exhibit transient chaotic dynamics, and the reverse bifurcation converges to a
stabilized equilibrium state.
Usually, the gradient descent method [12] is used to get command of the dynamic
behavior which belongs to the single neural unit, when the stable equilibrium state
begins. The length of the reversed bifurcation [13] can be impacted by β that is called
the simulated annealing parameter, which means that, as the value of β becomes small,
the reversed bifurcation behaves longer. Under the circumstance that the single neural
unit behaves alike as well as the Hopfield network, a stabilized equilibrium point is the
final state, which is the tendency of the network.
3 Application
Currently, chaotic neural networks are used in the following areas: continuous function
optimization problems, TSP problems, image encryption and communication. In this
paper, we focus on two aspects of continuous function optimization and the 10-city TSP
problem.
Minimum value of the function is 0, the minimum point of function is (0.7, 0.5),
local minimum point is (0.6, 0.4) and (0.6, 0.5). We set relevant parameters as follows:
ε = 0.018, k = 1, y1 (1) = y2 (1) = 0.1, z1 (1) = z2 (1) = 0.98, β = 0.001,
α = 0.05, I0 = 0.85, λ1 = 1/3, λ2 = 1/100.
Then, it is a new way, for chaotic neural network model which adopts Legendre
function self-feedback to deal with optimization function transiently. The development
of the energy function and the optimal solution are shown in Fig. 5.
Fig. 5. Time evolution figure of energy function and the optimal solution of X1 X2
Set λ2 = 3/100, other parameters remain unchanged, and study the influence of λ2
on the network, then time evolution figure of energy function and the optimal solution
is shown in Fig. 6.
Fig. 6. Time evolution figure of energy function and the optimal solution of X1 X2
many students and teachers to settle this problem more validly. TSP asks the following
question: There are many cities. We have to reach many cities only once, and the distance
between two cities came to light. Now, meeting the above conditions, What is the shortest
route? And find it.
A B
n n n
E= ( xij − 1)2 + ( xij − 1)2
2 2
j=1 j=1 i=1
(15)
n
n
n
+ D2 (xk,j+1 + xk,j−1 )xij dik
i=1 j=1 k=1
Where, xij refers to city i visited in order j, parameter A = B, dik is on behalf of the
distance between city i and city k, and the shortest effective path is equivalent to a global
minimum energy value. Then, we select normalized 10 city coordinates as follows Table
2.
i 1 2 3 4 5
X (i) 0.4 0.2439 0.1707 0.2293 0.5171
Y (i) 0.4439 0.1463 0.2293 0.716 0.9414
i 6 7 8 9 10
X (i) 0.8732 0.6878 0.8488 0.6683 0.6195
Y (i) 0.6536 0.5219 0.3609 0.2536 0.2634
In the literature, the shortest distance of the 10-city is 2.6776. Figure 7 shows the
optimal path of 10-city.
It is of significance for the neural networks to alter the arguments and energy
coefficients which is concerned to validity.
Chaotic Neural Network with Legendre Function 457
As can be seen from Table 3, when the network solves the 10-city TSP with the above
parameters, the highest proportion of the legal path is 99%, the average proportion of
the legal path is 97.9%, the average proportion of the optimal path is 96.2%, and the
maximum success rate of the optimal solution is 98%. On average, the difference between
the optimal ratio and the legal ratio is not significant, indicating the validity of this model.
Varying the function in the self-feedback with suitable parameters gives Table 3. it
can be seen that the nature of the self-feedback function has a degree of influence on the
effect of the optimal and legal ratios.
In the Chen’s chaotic neural network, set the parameters as follow: A = B = 1,
D = 2, α = 0.1, k = 1, I0 = 0.5, ε = 0.04, z(1) = 0.8, β = 0.002.
The program runs a total of 5 times, and 200 disparate original conditions of are
came into being at random in the interval [−0.1, 0.1].The consequences are collected in
Table 4.
The simulation experiment shows that when solving the TSP problem of 10 cities, the
solution ability of the neural network model with the Legendre function self-feedback
is better than that of the Chen’s neural network model.
Next, keeping other parameters the same, we study the influence of different λ on
the optimization of ability of TSP. The simulation consequences, with diverse λ1 and
λ2 , of 100 disparate internal conditions, are gathered in Table 5.
Under various conditions in the table, the program is run 20 times, and the simulation
results obtained are averaged. In addition, As seem from Table 5:
(1) The parameter λ1 and λ2 are very important in simulation results. When λ1 = 1/3,
λ2 = 0.016 and λ2 = 0.01, λ1 = 0.322, The chaotic neural network, in the TSP
question, which is equipped with Legendre function self-feedback has the greatest
performance.
(2) When λ1 = 1/3, 0.01 ≤ λ2 ≤ 0.02 and λ2 = 0.01, 0.32 ≤ λ1 ≤ 0.33, they also
have a better performance. When the parameters are farther and farther away from
the above ranges, it is obvious that the network optimization ability will decline.
Chaotic Neural Network with Legendre Function 459
4 Conclusion
In this study, the higher-order polynomial of Legendre basis function is recommended
into the self-feedback term of Chen’s transient chaotic neural network. The linear com-
bination of the higher-order polynomial of Legendre basis functions is recommended
used to form a new self-feedback function to construct a new transient chaotic neural
network. This model is employed to settle optimization functions, TSP problems, and
the influence of parameters on the network. The simulation consequences show that this
model has a strong power to optimize network performance.
References
1. Chen, L., Aihara, K.: Chaotic simulated annealing by a neural network model with transient
chaos. Neural Netw. 8(6), 915–930 (1995)
2. Xu, Y., Sun, M.: Wavelet chaotic neural network and its application in optimization calcula-
tion. In: 6th World Congress on Intelligent Control and Automation, WCICA 2006, Dalian,
pp. 3004–3009 (2006)
3. Zhao, L., Sun, M., Cheng, J., Xu, Y.: A novel chaotic neural network with the ability to
characterize local features and its application. IEEE Trans. Neural Netw. 20(4), 733–738
(2009)
4. Xu, Y., Sun, M.: Shannon wavelet chaotic neural network to solve TSP problems. Control
Theor. Appl. 25(3), 574–577 (2008)
5. Xu, Y., Yang, X.: Chaotic neural network with inverse trigonometric functions feedback and
its application. J. Harbin Univ. Commer. 26(3), 324–328 (2010)
6. Ye, Y.: Bessel function self-feedback chaotic neural network model and applications. Int. J.
Hybrid Inf. Technol. 7(4), 19–28 (2014)
7. Xu, Y., Xu, N., Qiu, Z.: Chaotic neural network with trigonometric function self-feedback.
In: Proceeding of the 39th Chinese Control Conference, Shenyang, pp. 7396–7401 (2020)
8. Sun, M., Xu, Y., Dai, X., Guo, Y.: Noise-tuning-based hysteretic noisy chaotic neural network
for broadcast scheduling problem in wireless multihop networks. IEEE Trans. Neural Netw.
Learn. Syst. 23(12), 1905–1908 (2012)
9. Xu, Y., Liu, J.: A novel chaotic neural network with self-feedback. In: 2016 Chinese Control
and Decision Conference, Yinchuan, pp. 3449–3454 (2016)
10. Sun, M.: Chaotic neural network and its application based on wavelet and hysteresis, pp. 39–
45. Harbin Engineering University (2010)
11. Ye, Y.: Simulation study on parameters of SLF chaotic neural network model. Int. J. Grid
Distrib. Comput. 8, 159–168 (2015)
12. Kwok, T., Smith, K.: Experimental analysis of chaotic neural network models for combina-
torial optimization under a unifying framework. Neural Netw. 13(7), 731–744 (2000)
13. Zhao, L., Sun, M., Cheng, J., Xu, Y.: A Novel chaotic neural network with the ability to
characterize local features and its application. IEEE Trans. Neural Netw. 20(4), 737–738
(2009)
460 Y. Zhang et al.
14. Wang, B., Liu, S., Qian, G.: Optimization method based on combination of wavelet neural
network and genetic algorithm. J. Aerosp. Power 23(11), 1953–1960 (2008)
15. Luo, Y., Li, X.: Adaptive synchronization of chaotic neural networks with time-delay based
on time-delay feedback control. J. Appl. Math. Comput. Math. 31(04), 443–453 (2017)
Deep Reinforcement Learning for Resource
Allocation in Multi-cell Cellular Networks
Ming Sun1(B) , Liangjin Hu1 , Yanchun Wang1 , Hui Zhang1 , and Jiaqiu Wang2
1 College of Computer and Control Engineering, Qiqihar University, Qiqihar 161006, China
2 Beijing China-Power Information Technology Co., Ltd., Beijing, China
1 Introduction
As a scarce mobile communications resource, the wireless spectrum has always been
an important research aspect in the mobile communication field [1, 2]. With the rise of
new concept networks, the new generation of wireless communication networks puts
forward higher and higher requirements for high rate, low energy consumption, and low
delay [3–9]. Therefore, it is important to manage spectrum resources efficiently in the
new generation of wireless communications.
At present, traditional methods, including heuristic algorithms, have been used for
the resource allocation of cellular networks. These methods affect the real-time per-
formance of the communication system and increase the computational burden of the
communication system [2], so that they cannot be applied to the new generation of
wireless communication networks.
Deep learning has the advantages of low computational complexity and low delay
[10–15], and has been used to solve the resource allocation of wireless mobile com-
munications. Some researchers have proposed supervised deep learning (DL) resource
allocation scheme [2], where the training data are generated by some heuristic algo-
rithms such as genetic algorithm, ant colony algorithm, simulated annealing algorithm,
etc. However, the generation of training data is often expensive and time-consuming,
so that the supervised DL approach is not suitable for large network systems. Some
researchers also use the reinforcement learning (RL) method [8, 12–15] to obtain the
optimal solution of the power control problem by interaction with the environment.
However, strategies in traditional RL are often stored in a table, which lacks flexibility
and is not feasible for large spaces of both the action and the state.
Due to the following three aspects, deep reinforcement learning (DRL) has become
a popular technology to solve complex control problems. Firstly, DRL combines the
perceptual ability of deep learning with the decision-making ability of reinforcement
learning. Secondly, DRL interacts with the environment continuously in the way of trial
and error. Thirdly, DRL obtains the best strategy by maximizing the cumulative reward.
In [15], the authors proposed a new online computing unloading method using DRL to
solve the mobile edge computing problem.
Inspired by the above, a deep Q network (DQN) is present in this paper to solve
resource allocation for full-frequency multiplexing cellular networks. In the proposed
DQN, the two aspects of resource allocation problems, i.e., channel allocation and power
control, are realized simultaneously to improve the energy efficiency of communication
systems. In the proposed DQN, the transition state is the power allocated for each channel.
In contrast, the transition action consists of the channel allocation scheme, the power
allocated for each channel, and the corresponding power adjustment amount. Besides,
the reward of the transition of the DQN is the total energy efficiency. The simulation
results show that, compared with the traditional artificial bee colony algorithm, the
greedy allocation algorithm, and the random allocation algorithm, the proposed DQN
can ensure the low computation delay and that the energy efficiency obtained by the
proposed DQN is larger than other methods as the number of channels increases.
2 System Model
As shown in Fig. 1, the downlink system of a cellular network with full-frequency mul-
tiplexing is composed of several OFDM cells. Assume that the cellular system contains
M cells, N orthogonal channel resources, and K users. Each cell has a base station in
the center, and users are randomly distributed in the cells. The channel resources are
reused with full frequency in the cells, and one single user only occupies each channel. In
addition, B and N 0 denote the bandwidth of each channel and the noise spectral density,
respectively. In this paper, the channel gain of the system is modeled as Eq. (1) [2]:
2
n
Hm,k = 10−(PLm,k +Xα )/10 hnm,k (1)
n is the channel gain as the base station m uses the channel n to communicate
where Hm,k
with the user k; PLm,k is the path loss model between the base station m and the user k;
X α represents the shaded fading which is a normal random variable with the mean of
zero and the standard deviation of α; hnm,k is the fast fading as the base station m uses
the channel n to communicate with the user k, which is an complex Gaussian random
variable with the mean of zero and the variance of one.
Deep Reinforcement Learning for Resource Allocation 463
Signal
Interference
signal
Base station
User
The total transmission rate and the energy efficiency of the system can be expressed
as [8]:
⎛ ⎞
M N K n pn H n
Dm,k
B log2 ⎝1 + ⎠
m,k m,k
R= (2)
m=1 n=1 k=1 N0 B + I m,k
n
D n pn H n
B log2 1 + m,k m,k m,k
M
N
K n
N0 B+Im,k
E= (3)
106 · pm,k
n
m=1 n=1 k=1
where R is the transmission rate with bit/(s·Hz); E is the energy efficiency of the system
n = 1 denotes that the base station m allocates the channel n to the
with Mbit/(s·W); Dm,k
user k which belongs to the m cell; Otherwise, Dm,kn = 0; pn is the power as the base
m,k
station m uses the channel n to communicate with the user k; = − ln(5BER)/1.6;
n is the interference as the base station m uses the
and BER is the bit error rate; and Im,k
channel n to communicate with the user k, which can be expressed as:
M
K
n
Im,k = n n
Di,j n
pi,j Hi,k (4)
i=1 j=1
i=m
Based on both the total transmission power constraints of a base station and the mini-
mum power constraints required for channel transmission, this paper models the resource
allocation problem of the above multi-cell cellular network with the full-frequency reuse
as the constrained optimization problem P1, shown as follows:
(P1) : max E
{D,p}
K
s.t. C1 : n
Dm,k = 1, Dm,k
n
∈ {0, 1}
k=1
C2 : n
pm,k ≥0
464 M. Sun et al.
C3 : pm
n
≥ pmin
C4 : N · pmin ≤ ptot,m ≤ pmax (5)
The basic model of DRL consists of action spaces A = {a1 , a2 , ..., an }, state spaces
S = {s1 , s2 , ..., sm }, reward signal R and the strategy π: S → A.
State space S: The state space is S = {s0 , s1 , ..., sT } where st = {p11 , p21 , ..., pM N}
n
represents the power set of each channel in each cell, and pm is the channel power as
the cell m uses the channel n to communicate with the user, and the dimension of st is
N × M.
Action space A: The action space is A = {a0 , a1 , ..., aT } where at is the t th action, and
at = {dkn , pm
n , p } where d n with the dimension N × M means the channel allocation
m k
scheme that the nth channel is assigned to user k, pm with the dimension M × 5
represents the power adjustment amount of base station M.
Deep Reinforcement Learning for Resource Allocation 465
Reward signal: The energy efficiency rt is used to act as the reward and punishment
value, shown below.
D n pn H n
log2 1 + m,k m,k m,k
M
N
K n
N0 B+Im,k
rt = n (6)
106 pm,k
m=1 n=1 k=1
where θ and θ are weights of the main network and the target network in DQN,
respectively, and η is the learning rate, and γ is the discount parameter.
p11 X1L1 1
W11 WL1 d11 p11 p1
p2
ReLU
ReLU
N
p1 d K1 p1N
1 1
pMN b 1 b L
d KN pMN pM
(N×K) (M×N) (M×5)
(Batch, M×N)
(Batch, N×K+M×N+M×5)
As shown in Fig. 2, the input of the proposed DQN is the state st = {p11 , p21 , ..., pM
N }.
After the hidden layer, the output layer at = {dk , pm , pm } includes the following three
n n
parts: the channel allocation dmn , the power allocation pm n , and the power adjustment
amount of the base station pm . According to the neuron output, the corresponding
maximum output value in each base station is select as the channel allocation scheme
dkn . In each base station, pm
n is the power selected by the DQN that needs to be adjusted
when entering the next state, and the adjustment amount pm of the base station is used
to adjust the select power pm n.
Step 1. Initialize parameters. Set the initial parameters of the DQN randomly. Initialize
discount factor γ and the learning rate η, and set the power adjustment amount pm
466 M. Sun et al.
as 5 levels, i.e., (−0.002, −0.001, 0, 0.001, 0.002), and set the size of experience pool,
and set the maximum number of episodes and the number of iterations in each episode.
Then, initialize the power pm n of the initial state s , which should satisfy the constraints
0
pm ≥ pmin and N · pmin ≤ ptot,m ≤ pmax .
n
Step 2. Store the state transition information in the replay memory. Based on the strategy
of exploration and exploitation, this paper uses ε-greedy policy to generate the action
at . That is, the action at is randomly selected or selected according to the output of the
deep neural network, i.e., at = arg maxa Q(s , a ; θ ). Reward r t is calculated according
to Eq. (6). The state information st is generated randomly by satisfying C2 and C3 of
Eq. (5). The next state st+1 is output by the DQN and is handled to stratify C2 and C3
of Eq. (5). Finally, the state transition information (sj , aj , rj , sj+1 ) value is stored in the
replay memory.
Step 3. Train the DQN. As the present capacity of the replay memory is reached, take a
small batch of state transition information randomly from the replay memory. And then
perform the gradient descent for the parameters θ of the DQN by using the loss function
[yj − Q(sj , aj ; θ )]2 , where
rj , if the episode ends at step j + 1
yj = (8)
rj + γ max
Q(s , a ; θ ), otherwise
a
In this section, the proposed DQN is evaluated by simulations. The simulation parameters
for the wireless network are shown in Table 1.
Number of base stations, mobile users Number of network layers and neurons
and channel resources
M, N, K M*N(Input)-180(FC-BN-ReLU)-120(FC-BN-ReLU)-
80(FC-BN-ReLU)-N*K+M*N+M*5(Output)
During the training for the DQN, the replay memory size is determined to be 64. A
total of 50 episodes of training are used in simulations, and the number of iterations for per
episode is set as 90. Figure 3 shows the energy efficiency in each episode as the number
of sub-channel is 16. During the training for the DQN, the state transition information in
the replay memory is randomly sampled in each episode, and the environment variable
of the channel gain H is initialized randomly.
Figure 4 shows that the resource allocation scheme based on the DQN proposed in
this paper is compared with the traditional artificial bee colony algorithm, greedy channel
allocation with WMMSE power control algorithm and random channel allocation with
WMMSE power control algorithm. In this figure, 50 groups of unseen channel gain H
are selected as external environment variables to make predictions, and the average value
of the energy efficiency of 50 groups is taken. The results show that as the number of
468 M. Sun et al.
channels is more than 12, the proposed DQN algorithm begins to outperform the artificial
bee colony algorithm. With the increase of the number of channels, the advantage of
the proposed DQN becomes more and more obvious. Meanwhile, it can be seen from
Figs. 3 and 4 that the energy efficiency predicted by the trained DQN in channel 14 is
obviously superior to other algorithms.
Fig. 4. Energy efficiency obtained by different algorithms under different number of channels.
As shown in Fig. 5, the proposed DQN has lower computational delay than other
algorithms. It indicates that the proposed DQN acts more quickly in terms of resource
allocation and can obtain higher energy efficiency at the larger number of channels.
Deep Reinforcement Learning for Resource Allocation 469
5 Conclusion
In order to improve the energy efficiency of next-generation cellular networks and reduce
the computing delay, this paper proposes a DQN to solve the resource allocation problem
in the cellular network with full-frequency multiplexing, which maximizes the energy
efficiency by the DQN for channel allocation and power control. Simulation results show
that the proposed resource allocation method based on the proposed DQN can achieve
larger energy efficiency than other algorithms under the premise of low computational
delay.
Acknowledgments. This work was supported in part by the Joint guiding project of Natural Sci-
ence Foundation of Heilongjiang Province under Grant LH2019F038, in part by the Young Inno-
vative Talents Program of Basic Business Special Project of Heilongjiang Provincial Education
Department under Grant 135309340.
References
1. He, X., Wang, K., Huang, H., et al.: Green resource allocation based on deep reinforcement
learning in content-centric IoT. IEEE Trans. Emerg. Top. Comput. 8(3), 781–796 (2020)
2. Ahmed, K., Tabassum, H., Hossain, E.: Deep learning for radio resource allocation in multi-
cell networks. IEEE Netw. 33(6), 188–195 (2019)
3. Liang, F., Shen, C., Yu, W., Wu, F.: Power control for interference management via ensembling
deep neural networks. In: 2019 IEEE/CIC International Conference on Communications in
China (ICCC), Changchun, pp. 237–242. IEEE (2019)
4. Lee, W., Kim, M., Cho, D.: Deep power control: transmit power control scheme based on
convolutional neural network. IEEE Commun. Lett. 22(6), 1276–1279 (2018)
5. Lee, W.: Resource allocation for multi-channel underlay cognitive radio network based on
deep neural network. IEEE Commun. Lett. 22(9), 1942–1945 (2018)
6. Lee, W., Kim, M., Cho, D.: Transmit power control using deep neural network for underlay
device-to-device communication. IEEE Commun. Lett. 8(1), 141–144 (2019)
7. Liu, D., Sun, C., Yang, C., Hanzo, L.: Optimizing wireless systems using unsupervised and
reinforced-unsupervised deep learning. IEEE Network 34, 270–277 (2020)
8. Liao, X., Yan, S., Shi, J., Tan, Z., Zhao, Z., Li, Z.: Deep reinforcement learning based resource
allocation algorithm in cellular networks. J. Commun. 40(2), 11–18 (2019)
9. Sun, M., Cao, W., Li, D., Ma, Z.: Strategy of maximizing capacity of OFDMA system for
ensuring fairness. Control Decis. 35(5), 1175–1182 (2020)
10. Challita, U., Dong, L., Saad, W.: Proactive resource management for LTE in unlicensed
spectrum: a deep learning perspective. IEEE Tran. Wirel. Commun. 17(7), 4674–4689 (2018)
11. Li, X., Fang, J., Cheng, W., et al.: Intelligent power control for spectrum sharing in cognitive
radios: a deep reinforcement learning approach. IEEE Access 6, 25463–25473 (2018)
12. Ye, H., Li, G., Juang, B.: Deep reinforcement learning based resource allocation for V2V
communications. arXiv arXiv:1805.07222v1 (2018)
470 M. Sun et al.
13. Abishi, C., Shital, A.R., Husnu, S.N.: DA-DRLS: drift adaptive deep reinforcement learning
based scheduling for iot resource management. J. Netw. Comput. Appl. 138, 51–65 (2019)
14. Tabrizi H., Farhadi G., Cioffi J.: Dynamic handoff decision in heterogeneous wireless sys-
tems: Q-learning approach. In: Proceedings of 2012 IEEE International Conference on
Communications (ICC), Ottawa, pp. 3217–3222. IEEE (2012)
15. Xiaoyu, Q., Luobin, L., Wuhui, C.: Online deep reinforcement learning for computation
offloading in block chain-empowered mobile edge computing. IEEE Trans. Veh. Technol.
68(8), 8050–8062 (2019)
Gauss Nonlinear Self-feedback Chaotic Neural
Network and Its Application
Abstract. The new network model with gauss nonlinear self-feedback is con-
structed. The appropriate initial value of each parameter is selected, making the
network reflect the chaotic dynamic characteristics. The comparison of simulation
figures illustrates the feasibility of the application of the model, and the network’s
sensitivity to chaotic parameters is also reflected. It is proved that the width coef-
ficient increase or decrease in different structure-function influences the output to
change differently by comparing Gaussian and the contrary multiquadric function.
The new model is used in the combinatorial optimization problem. The experi-
mental data show that the new network can effectively escape from the minimum
range of chaotic characteristics and achieve a high convergence effect to a station-
ary point if appropriate parameters are adopted. We also study the influence of the
ability of optimization belonging to main parameters in the network.
1 Introduction
At present, the nonlinear chaotic neural network has been studied primarily. Such as
Xu Yaoqun et al. constructed a new chaotic neural network (CNN) with trigonometric
function as self-feedback connection term [1], Yang Xueling et al. constructed a class
of CNN with ant-trigonometric function as self-feedback [2]. The feasibility of CNN
with nonlinear self-feedback has been obtained [3]. All these have laid a theoretical
foundation for further research on it. This paper continues the direction of research
based on Chen’s CNN [4], which uses a kind of radial basis function. Gauss function
constitutes a nonlinear self-feedback connection item, and it shows a new characteristic
different from linear self-feedback. The output graph and Lyapunov exponent of the
new model are analyzed. The classic problem “traveling salesman shortest path (TSP)”
is a kind of Non-deterministic Polynomial (NP) problem. It is solved using this new
network, and the ability to find the shortest path is studied.
self-feedback function. Gauss function is a kind of radial basis function, which has a
strong function approximation ability. It is adopted as the self-feedback term of the new
network. This makes the feedback term of the new network show a nonlinear trend. The
chaotic behavior of neurons can investigate the chaotic ergodic ability of the new neural
network, so the model of a single neuron is studied first [5]. The single neuron model of
the new network is constructed as follows:
The new Gauss nonlinear self-feedback single neuron model shows as:
1 − u22
f (u) = e δ (4)
c
In this model, x(t) is output; y(t) is used to examine the trend of independent variables
over time; z(t) is self-feedback item; f(u) is one of radial basis function named Gauss
which and z(t) is multiplied by the nonlinear self-feedback connection in this model [6];
k is damping factor and its value between 0 and 1; δ is extension constant of Gauss; c
also a parameter of Gauss.
The following simulation experiment is configured for Lenovo G450 Notebook PC,
the operating system is Windows XP, Intel Core 2 DuoT6600CPU, Hard disk capacity
is 320 GB, Graphics chip is NVIDIA GeForce G210M, Program running environment
is Matlab7.0, Computer memory is 2 GB.
The setting of initial value is very important for the embodiment of chaos. This is
because only a small part of the numerical interval of a model meets the requirements of
chaos. And the chaotic trajectory changes with time. Lyapunov exponent is the quantity
to quantitatively describe this variation. Lyapunov exponent is often used to measure
the chaotic characteristics of systems. When the initial values of parameters are very
close, it will lead to the emergence of adjacent orbits in space, then whether the change
trend between orbits converges or diverges with the passage of time, and the speed
of change can be directly reflected by Lyapunov exponent. If the maximum Lyapunov
exponent is greater than zero in a time period, it means that the system has entered a
chaotic state. When the Lyapunov exponent is positive, it indicates that even if the initial
values are very similar, the difference will change over time, and the exponential rate
will increase, so it cannot be predicted. This is chaos. “Bifurcation” in dynamics refers
to the continuous change of a parameter of the dynamic system, but it causes the sudden
and sharp response of the system behavior. Until a certain point, the system suddenly
changes from law to irregularity, which is the state of stepping into chaos.
Therefore, the dynamic properties of the model are analyzed by using the bifurcation
diagram of the neuron and the trend chart of maximum Lyapunov exponent over time.
Select parameters: ε0 = 1.25, y(1) = 0.283, z(1) = 0.5, k = 0.1, I0 = 0.85, δ = 0.25,
c = 0.0001 fixed invariant. The bifurcation diagram of the neuron and the trend chart
Gauss Nonlinear Self-feedback CNN 473
of maximum Lyapunov exponent when β = 0.0025 and β = 0.003 are shown as Fig. 1,
Fig. 2, Fig. 3 and Fig. 4.
Conclusions can be drawn from the above 4 chart: When the parameters are properly
selected, the neuron model has the characteristics of transient chaotic dynamics and the
chaotic behavior can fully reflect the sensitivity of the simulated annealing parameters
β. With the increase of β, the convergence speed will accelerate. With the decrease of
β, the more the chaotic search process is. From the numerical point of view, the two
simulate test β value difference of 0.0005, it is far apart. But the balance point one about
t = 2750 s, the other about the t = 2250 s, the time difference of 500 s. It shows that the
small change of β will lead to a larger difference in the search process of chaos.
From another point of view, we can easily see through the formula (3): Self-feedback
connection z(t) is a process of constant attenuation and the speed of the reduction depends
on the value of β. The greater the β value, the faster the z(t) changes and the more
rapidly the simulated annealing temperature is reduced, then the network can reach a
stable equilibrium point. Not the proposed beta is too large so what the chaotic behavior
cannot be fully reflected and not easy to obtain the optimal solution of traversing the
global network; The smaller the value of β is, z(t) changes more slowly and network
continue to have higher simulated annealing temperature. Although the network can get
rich chaotic dynamical behavior but slow convergence speed and seeking the optimum
speed, the value of beta should not be too small.
After learning that the network has the characteristics of transient chaos, we study
the influence of the parameters of Gauss function on the model. First of all, to understand
the Gauss function curve as shown in Fig. 5. When the value of u is the same, the bigger
the width δ value is and the bigger the output f(u) value is. Compared with another kind
of radial basis function named contrary multiquadric function as shown in Fig. 6: When
the value of u value is the same, the bigger the width δ value is AMD the smaller the
output f(u) value is. It is proved that the width parameter has different influence on the
output result in different radial basis function structure.
Next, the influence of the width parameter on the network is studied. Select param-
eters: ε0 = 1.25, y(1) = 0.283, z(1) = 0.5, k = 0.1, I0 = 0.85, β = 0.0025, c = 0.0001
fixed invariant. The bifurcation diagram of the neuron and the trend chart of maximum
Lyapunov exponent when δ = 0.2 are shown as Fig. 7 and Fig. 8.
Figure 7 and Fig. 8 are compared with Fig. 1 and Fig. 2, it shows that the size of the
width parameter affects the network’s chaotic search process under the condition that
other parameters are fixed. Convergence point is located at t = 2450 s when δ = 0.2 for
this model. Convergence point is located at t = 2750 s when δ = 0.25. The shorter the δ,
the shorter the chaotic search process, easy to fall into local minimum points. The larger
the δ value is, the more capable of fully reflecting the chaotic behavior, can be separated
from the small point limit, so as to find the global optimal solution.
Gauss Nonlinear Self-feedback CNN 475
Fig. 6. The figure of Contrary Multiquadric function when δ = 0.3 and δ = 0.9
Fig. 10. The trend chart of maximum Lyapunov exponent when c = 0.0003
Figure 9 and Fig. 10 are compared with Fig. 1 and Fig. 2, it shows that the single
neuron’s chaotic search ability is very sensitive to the parameters c. Convergence point is
located at t = 2750 s when c = 0.0001. Convergence point is located at t = 2300 s when
c = 0.0003. The difference between the c value is only 0.0002 but the convergence point
is far away. Therefore, the value of c should be appropriate to ensure that the network
has chaotic characteristics and has a strong ability to find the best.
Next, the influence of the parameter I0 of Gauss function on the dynamic behavior
of neurons is investigated. Select parameters: ε0 = 1.25, y(1) = 0.283, z(1) = 0.5, k =
0.1, β = 0.0025, δ = 0.25, c = 0.0003 fixed invariant. The bifurcation diagram of the
neuron and the trend chart of maximum Lyapunov exponent when I0 = 0.75 as shown
as Fig. 11 and Fig. 12.
Figure 11 and Fig. 12 are compared with Fig. 9 and Fig. 10, it shows that parameter
I0 is also very sensitive to the chaotic search. See from formula (2) and formula (4): In
Gauss Nonlinear Self-feedback CNN 477
Fig. 12. The trend chart of maximum Lyapunov exponent when I0 = 0.75.
the nonlinear self-feedback model, I0 greatly affects the output of Gauss function. So
it is very important in the convergence of the network (corresponding to the inverted
bifurcation and the equilibrium point in the image). By numerical results we can see
that the difference of I0 is only 0.1, but the convergence of the network is a difference of
300 s. The larger the value of I0 , the shorter the time of chaotic search. If the I0 value is
too large, it is not easy to generate chaos. The smaller the value of I0 can be fully ergodic
chaotic region, it is conducive to find the global optimal solution. However, if the value
of I0 is too small, the search takes too long and does not meet the actual requirements.
Therefore, the range of I0 should be adjusted according to the situation.
1 − u22
f (u) = e δ (8)
c
In the formula: γ is the positive scale parameter for input; Ii is the input deviation of
the neuron; Wij is the connection weights from neuron i to neuron j, and Wij = Wji , Wii
= 0.
The dynamic characteristic of the CNN model is very sensitive to the value of γ
which represents the influence of energy function on the change of internal state [8].
If the value of γ is too large, the energy function will be too strong and the network
convergence too fast. The network can’t get the transient chaos phenomenon, and it is
very easy to linger in the local range and cannot be traversed globally. If γ is too small,
the energy function will be too weak. Although transient chaos can be obtained, it is
possible not to converge to the optimal solution in a certain time range.
478 N. Xu et al.
A B D
n n n n n n n
E= ( Vxi − 1)2 + ( Vxi − 1)2 + dxy Vxi Vy,i+1 (9)
2 2 2
x=1 i=1 i=1 x=1 x=1 y=1 i=1
Gauss Nonlinear Self-feedback CNN 479
In order to verify the effectiveness of the new network for this problem, the original
data is the classical normalized 10 city coordinates: (0.4, 0.4439); (0.2439, 0.1463);
(0.1707, 0.2293); (0.2293, 0.716); (0.5171, 0.9414); (0.8732, 0.6536); (0.6878, 0.5219);
(0.8488, 0.3609); (0.6683, 0.2536); (0.6195, 0.2634). The shortest path of the 10 cities
is 2.6776 (see Fig. 13).
The Gauss nonlinear self-feedback CNN is applied to solving the 10 city TSP.
Research on the function of the network solution ability and the main parameters.
Select parameters: ε0 = 0.8, z(1) = 0.5, k = 1, I0 = 0.7, β = 0.001, A = 8.3, D =
3.7, γ = 0.9, δ = 1 fixed invariant. The influence of different c values on the solution of
TSP is studied. The following Table 1 gives the results of the initial value of 200 random
assignment in this case.
Seen from the Table 1: Regardless of the c value, the proportion of the legal path [9]
is 100%, but its value greatly influences the proportion of the optimal path. The optimal
ratio is more than 70% when c = 1; The optimal path proportion is gradually decreasing
and the whole is more than 60%. When 0.17 ≤ c < 1; The ability to find the optimal
network greatly decreased when c ≤ 0.15. The value of the parameter c can be explained
to the Gauss nonlinear self-feedback chaotic neural network to solve the TSP. c value
should be properly selected to ensure the ability of the network to obtain the global
optimal solution [10].
Select parameters: z(1) = 0.5, k = 1, I0 = 0.7, β = 0.001, A = 8.3, D = 3.7, γ =
0.9, δ = 1, c = 5 fixed invariant. The influence of different ε0 values on the solution of
TSP is studied. The following Table 2 gives the simulation results of the initial value of
200 random assignment in this case.
See from the Table 2: Influence of the steepness parameter ε0 value on the network is
more obvious. The optimal path ratio is less than 70% when ε0 < 0.5. With the decrease
of ε0 , network optimization ability decreased [11] gradually. Not only reduce the number
of optimal path to and decrease the number of legal paths when ε0 is too small such as
ε0 = 0.1 in the Table 2. This shows that the network is easy to fall into local minimum
point and can’t make full use of the chaotic characteristics to find the global optimal
solution; The optimal ratio is lower than 70% when ε0 > 1.2. The bigger the ε0 value,
the smaller the optimal rate, but the legal path ratio remains unchanged at 100%; The
optimal ratio keeps at 70% when 0.5 ≤ ε0 ≤ 1.2. This shows that the network has a better
ability to solve the TSP and has a better ability to find the best. Therefore, under the
same condition, the range of steepness parameters suggested that the control at 0.5–1.2.
Gauss Nonlinear Self-feedback CNN 481
4 Conclusion
A new nonlinear self-feedback chaotic neural network model is constructed in this paper.
The new model has the characteristics of chaotic dynamics by analyzing the inverted
bifurcation and Lyapunov exponent [12]. The new network is applied to solve the TSP.
The simulation experiment shows web searches that the global minimum point is sen-
sitive to Gaussian function parameters c and the steepness parameter ε0 . The network
can have a good ability to find the best solution if the model parameters are selected
appropriately.
References
1. Xu, Y.Q., He, S.P.: Self-feedback chaotic neural networks with trigonometric functions and its
applications. In: Proceedings of the 2009 China Conference on Control and Decision Making,
Guilin, Guangxi (2009)
2. Xu, Y.Q., Yang, X.L.: Chaotic neural network with anti-trigonometric function self-feedback
and its application. J. Harbin Univ. Commer. (Nat. Sci. Edn.) 26(3), 324–328 (2010)
3. Tong, X., Li, J.B.: Robust encryption face recognition algorithm based on composite chaos
in cloud environment. Mod. Electron. Technol. 42(06), 166–169 (2019)
4. Chen, L., Aihara, K.: Chaotic simulated annealing by a neural network model with transient
chaos. Neural Netw. 8(6), 915–930 (1995)
5. Xu, Y., Yu, Q.: Optimizing logistic distribution routing problem based on wavelet chaotic neu-
ral network. In: Proceedings of the 2017 Chinese Control and Decision Conference, CCDC,
pp. 3460–3464 (2017)
6. Hopfield, J.: Neural networks and physical systems with emergent collective computational
abilities. Proc. Natl. Acad. Sci. 79, 2554–2558 (1982)
7. Wang, X.-Y., Bao, X.-M.: A novel image block cryptosystem based on a spatiotemporal
chaotic system and a chaotic neural network. Chin. Phys. B 22(5), 050508 (2013)
8. Abdoun, N., El Assad, S., Hoang, T.M., Deforges, O., Assaf, R., Khalil, M.: Designing two
secure keyed hash functions based on sponge construction and the chaotic neural network.
Entropy 22(9), 1012 (2020)
9. Ye, Y.G., Xu, Y.Q.: SLF chaotic neural network and its application. Comput. Eng. Appl.
(2015)
10. Xiu, C.B., Liu, C., Guo, F.H., Cheng, Y., Luo, J.: Research on control strategy and application
of delayed chaotic neuron/network. ActaPhys. Sin. (2015)
11. Hu, Z.Q., Li, W.J., Qiao, J.F.: Frequency conversion sinusoidal chaotic neural network based
on adaptive simulated annealing. J. Electron. 47(3), 613–622 (2019)
12. Hu, N.W., Qi, M.L.: Molecular dynamics simulation of displacement cascades in Ni-Mo alloy.
Nucl. Sci. Tech. 26(6), 060603 (2015)
Guidance Prediction of Coupling Loop Based
on Variable Universe Fuzzy Controller
Ming Zhao1,2 , Yang Liu1 , Hui Li1,2 , Yun Cao1(B) , Yuru Zhang1 , and Hao Jin1
1 Harbin University of Commerce, Harbin 150028, China
2 Heilongjiang Provincial Key Laboratory of Electronic Commerce and Information Processing,
1 Introduction
Because of the significance of decision-making, researchers have brought the LSO guid-
ance prediction approach into focus [1–4]. A discrete-time series of LSO is applied to
model the pilot’s final actions [5]. “Instruction sending-operation responding-deviation
correcting” process is realized based on LSO guidance instructions [6]. Ming presented
a model of LSO for the digital simulation of a pilot-carrier system [7]. Four LSO grade
technology should be proposed: effect capability integration of multiple flight states
attributes; effect capability integration at multiple reference positions; effect capabil-
ity quantification and evaluation on the final approach [8]. However, current research
mainly focuses on the dynamic simulation of LSO guidance decision-making, and there
is a lack of analysis on the characteristics as a person of LSO.
The rest of this paper is structured as follows: in the next section, we first analyze
the influence factors of the landing process. Section 3 researches the landing guidance
system characteristics. The control approach of variable universe fuzzy logic will be
introduced in Sect. 4. The guidance prediction system of LSO has been spread to prove
the feasibility of the improved algorithm in Sect. 5.
In the longitudinal loop, two factors impact safety: glideslope deviation and sink rate
deviation.
Sink Rate. The ideal landing longitudinal sink rate is constant. However, there is pitch,
yaw and roll motion during the landing process, and the pilot sometimes changes the
attitude by altering the sink rate or throttle.
In the lateral loop, the centering deviation and drift rate are the two important factors
for landing safety.
Centering Deviation. It is the lateral factor that is the most concerned with LSO. The
centering means that the carrier-based aircraft must be correctly aligned with the center-
line of the landing operation area on the deck. Otherwise, lateral collision accidents may
occur during landing. The centering quality of aircraft based on the “centring scaleplate”
on the ramp should be judged by LSO.
As shown in Fig. 1, when the plane is centered on the left, it may crash into the sea.
Else if the aircraft is on the right side, an aircraft crash accident may happen.
Drift Rate. The lateral deviation is compensated by changing the drift rate, however,
when a great right drift rate has occurred, a large position deviation should exist. The
aircraft body is likely to strike the carrier’s starboard. Therefore, the pilot should change
the drift rate according to the horizontal instruction provided by LSO in real-time to
achieve the purpose of lateral centering.
484 M. Zhao et al.
Very Right
Right
A Little Right
OK
Cables A Little Left
Left
Desired touchdown
Very Left
Ramp
Carrier
Inner-loop Cockpit
Outer-loop Inner-loop controller
command
control control Airframe
Outer-loop technique technique Outer-loop
Engine state
command
Pilot vision sensitivity
APCS
Position Attitude
For a Multiple Input Single Output (MISO) fuzzy system, the controller f : xk ∈ U ⊂
Rn → yk+1 ∈ V ⊂ R based on the fuzzy rule shown in (1) could be expressed as [15]:
m
n
yj μAij (xik )
j=1 i=1
yk+1 = f (xk ) = (5)
m
n
μAij (xik )
j=1 i=1
Where μAij (xik ) is the membership function of input fuzzy sets Airi , and yj is the center
of output fuzzy set Bro .
Step 1: At k = 1 moment, the input of controller is x1 = (x11 , x21 , ..., xn1 ), and x1ij =
α(xi1 )x0ij , y1j = β(y1 )y0j , then the output of controller is:
m
n
y1j μA1 (xi1 )
ij
j=1 i=1
y2 = f (x1 ) = (7)
m
n
μA1 (xi1 )
ij
j=1 i=1
Step k: At k moment, the input of controller is xk = (x1k , x2k , ..., xnk ), and xkij = α(xik )x0ij ,
ykj = β(yk )y0j , then the output of controller is:
m
n
ykj μAk (xik )
ij
j=1 i=1
yk+1 = f (xk ) = (8)
m
n
μAk (xik )
ij
j=1 i=1
The universe Uik shrinks with the decrease of xik , and because of the invariance of
input universe Uik division at k moment, the fuzzy sets Akij still shrink. The neighbourhood
of expected control point xik = 0 is divided particularly, which is equivalent to increase
the number of rules.
cause severe coupling of pilot control, and the risk coefficient of landing would be
increased simultaneously. Therefore, an important working principle of LSO is to send
as few instructions to pilots as possible on the premise of ensuring the safety of pilot’s
operation [6]. Based on it, in the actual control process, the priority order of LSO com-
mand information needs to accomplish, and the most critical guidance instructions for
pilots are sent first.
The design principles of the LSO coupling fuzzy guidance predictin system are as
follows:
(1) The “Climbing” instruction has a higher priority than “Left/Right” instruction of
the same level.
(2) The “Left/Right” instruction has a higher priority than “Falling” instruction of the
same level.
(3) The “Left/Right” instruction has a higher priority than “Climbing” instruction of
the next level.
Longitudinal Lateral
PB1 PM1 PS1 AZ1 NS1 NM1 NB1
PB2 PB2 PB2 PB2 PB2 PB2 PB2 PB2
PM2 PB1 PM2 PM2 PM2 PM2 PM2 NB1
PS2 PB1 PM1 PS2 PS2 PS2 NM1 NB1
AZ2 PB1 PM1 PS1 AZ2 NS1 NM1 NB1
NS2 PB1 PM1 PS1 NS2 NS1 NM1 NB1
NM2 PB1 PM1 NM2 NM2 NM2 NM1 NB1
NB2 PB1 NB2 NB2 NB2 NB2 NB2 NB1
Since the instructions sent by LSO to the pilots are discrete signals, it is necessary to
first discrete the output curves of the longitudinal and lateral fuzzy controllers of LSO.
The discrete instruction bar chart is shown in Fig. 3.
As shown in Fig. 3, the instruction priority would be automatically determined by
LSO based on the current flight states. One-dimensional instruction is selected from
the 14 longitudinal and lateral instructions to send to the pilot. The pilot can clarify
the current most major flight deviation and perform maneuver operations to ensure the
safety of the landing process.
488 M. Zhao et al.
Very High
High
A Little High
OK
A Little Low
Low
Very Low
0 1 2 3 4 5 6 7 8 9 10
t/s
Very Right
Righ
t
A Little Right
OK
A Little Left
Left
Very Left
0 1 2 3 4 5 6 7 8 9 10
t/s
Very High
High
A Little High
OK
A Little Low
Low
Very Low
0 1 2 3 4 5 6 7 8 9 10
t/s
6 Conclusions
In the landing process of carrier-based aircraft, the characteristics of LSO deviation
instructions, the glideslope deviation, and sink rate deviation of the longitudinal loop,
and the centering deviation and drift rate deviation of the lateral loop are the influencing
factors of carrier-based aircraft landing safety. The deviations above are input into the
landing guidance model of LSO and an intelligent guidance prediction system fuzzy
intelligent guidance system based on variable universe fuzzy logic. Simulation results
show that the guidance prediction model established by fuzzy logic conforms to the
operating characteristics of the true LSO, and the output results accord with the correction
of deviation effect under the real environment. In the meantime, the establishment of
the system model also provides an effective solution for the objects with uncertainty,
nonlinearity, and task environment complexity, especially for the object is an individual.
References
1. Zhu, Q., Yang, Z.: Dynamic recurrent fuzzy neural network-based adaptive sliding control
for longitudinal automatic carrier landing system. J. Intell. Fuzzy Syst. 37(1), 53–62 (2019)
2. Zuo, Z., Wang, L., Liu, H., Wang, Y.: Similarity for Simulating automatic carrier landing
process of full-scale aircraft with scaled-model. Acta Aeronautica et Astronautica Sinica
40(12), 123005 (2019)
3. Zhou, J., Jiang, J., Yu, C., Xiao, D.: Carrier aircraft dynamic inversion landing control based
on improved neural network. J. Harbin Eng. Univ. 39(10), 1649–1654 (2018)
4. Hess, R.: Simplified approach for modeling pilot pursuit control behaviour in multi-loop flight
control task. Inst. Mech. Eng. 220(2), 85–102 (2006)
5. Wang, L., Zhu, Q., Zhang, Z., Dong, R.: Modeling pilot behaviors based on discrete-time
series during carrier-based aircraft landing. J. Aircr. 53(6), 1922–1931 (2016)
6. Li, H.: Modeling landing signal officer instruction associated with operation guide system.
Int. J. Control Autom. 8(2), 373–382 (2016)
7. Shi, M., Cui, H., Qu, X.: Modeling landing signal officer for carrier approach. J. Beijing Univ.
Aeronaut. Astronaut. 32(2), 135–138 (2016)
8. Li, H.: Integrated evaluation technology of landing signal officer for carrier-based aircraft.
Int. J. Multimedia Ubiquit. Eng. 11(1), 169–178 (2016)
9. Li, H.: The essence of fuzzy control and a kind of fine fuzzy controller. Control Theory Appl.
14(6), 868–871 (1997)
10. Shi, P., Xu, Z., Wang, S.: Variable universe adaptive fuzzy pid control of active supension.
Mech. Sci. Technol. Aerosp. Eng. 38(5), 713–720 (2019)
11. Du, E., Wang, S., Chang, L.: Variable universe fuzzy controller design of missile mark
trajectory with feed-forward compensation. J. Acad. Armored Force Eng. 31(2), 84–89 (2017)
12. Li, H.: Interpolation mechanism of fuzzy control. Sci. China (Series E) 28(3), 259–267 (1998)
13. Yang, Z., Wang, H.: Maximum power point tracking for photovoltaic power system based on
asymmetric fuzzy control. Mech. Autom. 41(2), 153–156 (2012)
490 M. Zhao et al.
14. Liu, J., Zhang, Y.: Variable universe fuzzy pid control method in piezoelectric ceramic
precision displacement system. Autom. Instrum. 32(2), 45–49 (2017)
15. Li, D., Shi, Z., Li, Y.: Sufficient and necessary nonditions for boolean fuzzy systems as
universal approximators. Inf. Sci. 178(2), 14–24 (2008)
16. Li, H.: Adaptive fuzzy control of four-stage inverted pendulum based on variable universe.
Sci. China (Series E) 32(1), 65–75 (2002)
17. Chen, G.: On approaching precisions of standard fuzzy systems with different basic functions.
Acta Autom. Sinica 34(7), 823–827 (2008)
Heart Disease Recognition Algorithm Based
on Improved Random Forest
1 Introduction
Predicting whether heart disease occurs has always been a research hotspot in the modern
medical field. According to the “China Cardiovascular Health and Disease Report 2019”
issued by the National Medical Center, the number of people suffering from heart disease
in my country each year is gradually increasing compared with previous years, and the
death rate of cardiovascular disease is the first. The use of machine learning classification
algorithms for heart disease diagnosis is significant for establishing a stable and reliable
predictive model for heart disease. Scholars at home and abroad have conducted a lot of
research. In 2014, Dai et al. used classification models such as SVM, logistic regression,
and naive Bayes to predict heart disease and obtained a prediction accuracy of 82% [1].
Kedar et al. used the K-nearest neighbor algorithm to predict and classify heart disease
in 2016 and obtained an accuracy of 75% [2]. Chen Jianghong and others used SVM to
diagnose heart disease in 2017 with missing data [3]. Luo Benjie designed a heart disease
recognition platform based on random forest in 2017 [4]. Wen Bowen used improved
grid search technology to optimize the parameters of random forest in 2018 [5]. Ding
Weijie compared the recognition accuracy of seven classification algorithms on the heart
disease data set in 2019 [6]. Zhao Jinchao et al. used an optimized random forest for
heart disease prediction in 2021 [7].
In traditional diagnosis, the doctor judges whether a person has heart disease based
on clinical experience. There will inevitably be a risk of misjudgment and cause med-
ical accidents. Combining medical diagnosis with machine learning, data mining and
other technologies can achieve the dual protection of man-machine integration, greatly
reducing the risk of medical misdiagnosis and improving medical efficiency. The opti-
mized random forest proposed in this paper has a high accuracy rate for diagnosing heart
disease. It is more suitable for the diagnosis of heart disease than some other machine
learning algorithms.
2 Related Theories
As the name suggests, random forest uses a random way to build a forest composed of
multiple trees. It is an integrated algorithm based on a decision tree as a learner. The
random forest algorithm integrates many independent and identically distributed decision
trees into a forest, and together they are used to predict the final result. Compared with a
single classification algorithm, the integration of decision trees has a better classification
effect, lower overfitting, and stronger generalization ability. It is essentially an integrated
improvement of decision trees.
The two most critical parts in the construction of a decision tree are node splitting and
pruning operations. Choosing the best split node is to choose the feature node with the
highest purity and how to prune to make the decision tree fit the least, which determines
the quality of the classification effect. Random forest introduces randomness theory to
avoid the limitations of decision trees.
trees. It is an integrated algorithm of decision trees. Several decision trees first determine
each sample of the input, and then the results of each tree are summarized to determine
the sample type. For the judgment of each sample, only when more than half of the
decision trees are judged incorrectly will there be a misjudgment, so the accuracy rate
is greatly improved.
The randomness of random forest is mainly reflected in two aspects: one is to ran-
domly select data from the data set, and the self-service method introduces higher diver-
sity to the training subset of each predictor so that the correlation between the base
predictors is lower; The second is to select a feature subset from the features randomly,
and randomly select a certain size subset every time a node is split, and then select the
best feature from the feature subset to split according to the importance of the feature.
The algorithm execution steps are as follows:
(1) Use the Bootstrap sampling method from the original data set D to extract D training
sample sets of the same size as the original data set used to build the model.
(2) Randomly select some features each time a node is split and select the best split
node from the subset of features.
(3) Each decision tree determines the class of the sample.
(4) Voting to get the final category of the sample, the voting formula can be expressed
by formula (1).
k
H(x) = argmax I (hi (x) = Y ) (1)
Y i=1
Compared with other single classifiers, the random forest has a better classification
effect. The generalization ability of the model is strong. Because of its randomness, it
does not need to perform feature selection when making classification predictions.
3 Data Preprocessing
This experiment uses discrete data one-hot encoding as the experimental data. Data
preprocessing was completed through missing value processing on this data set, hot
editing of discrete data code, and normalization processing.
There are two commonly used methods for processing missing values in data mining:
filling method and the deletion method. The filling method mainly includes meaning
filling, mode filling, and using machine learning algorithm modeling prediction filling,
etc. It is mainly suitable for scenarios with relatively large feature data missing and
relatively important features. In this experiment, the number of missing data is small, so
the deletion method is adopted to process the missing values. The experimental results
show that the deletion of a few missing values will not reduce the prediction effect of
the model but may increase the accuracy of the prediction.
494 H. Xin and H. Yu
The purpose of one-hot encoding is to convert sub-type variables into numerical values
for later modeling and prediction. This article maps all sub-type variables to integer
values and represents them as binary vectors, except that the integer index is 1, and the
rest are 0. The corresponding category matrix is generated according to the number of
categories of each classification feature, the category corresponding to the sample is 1,
and the rest are 0.
First, convert the values of all classification features into corresponding textual mean-
ings and then perform one-hot encoding. After deletion of missing values and one-hot
encoding, the data set has changed from 1328 samples and 14 feature columns to 1315
samples and 26 feature columns.
Before modeling, a heat map is built to analyze the correlation of data features, as shown
in Fig. 1.
The horizontal and vertical coordinates of the heat map respectively represent the
feature columns of the data set, and the right side represents the correlation coefficient
between the features. The darker the color, the stronger the correlation between the two
features. It can be seen from the above Fig. that the diagonal color is the darkest, the
correlation coefficient is 1, and the colors of the other positions are lighter. It can be seen
that the correlation between the features of the data set is low, and there is no redundancy
between the features.
The experiment is based on Python language and performs random forest modeling and
prediction in the Anaconda environment. Use learning curve and grid search technology
to optimize the design of random forest parameters. According to the importance of
the influence of each parameter on the prediction result, five important parameters of
Heart Disease Recognition Algorithm 495
adjustment method, which not only reduces the algorithm optimization time, but also
accurately knows the prediction of each parameter to the model The effect of the effect.
The final optimal combination of parameter settings is: n_estimators: 81, max_depth:
12, min_samples_split: 3, min_samples_leaf: 1, max_features: by default, the algorithm
has the highest accuracy rate. The flow chart of the experimental model is shown in
Fig. 2.
The ROC curve refers to constantly moving the prediction “threshold” of the classifier
to generate a set of key points (FPR, TPR) on the curve. From a qualitative point of view,
the closer the ROC curve is to the upper left corner, the better the classification effect
of the model. From a quantitative point of view, the larger the AUC value, the better the
classification ability of the model.
random forest algorithm proposed in this paper has a better predictive effect on heart
disease. The experimental results are shown in Table 3.
It can be seen from Table 3 that the heart disease data set after data preprocessing has a
good predictive effect on the random forest model itself, indicating that the random forest
algorithm itself is suitable for the diagnosis of heart disease. The model’s accuracy used
in this experiment is 94.9%, and the AUC value is as high as 98.5%, which is significantly
improved compared to decision trees and KNN-RF.
In order to better display the AUC value, the ROC curve of each algorithm is drawn.
Figure 4 is a schematic diagram of the ROC curve of a decision tree, Fig. 5 is a schematic
diagram of the ROC curve of a random forest without tuning parameters, and Fig. 6 is
a schematic diagram of the ROC curve of KNN-RF. Figure 7 is a schematic diagram of
the ROC curve of the model used in the experiment.
Heart Disease Recognition Algorithm 499
Fig. 4. Decision tree ROC curve. Fig. 5. Random forest ROC curve.
Fig. 6. KNN-RF ROC curve. Fig. 7. Optimized random forest ROC curve.
6 Conclusion
Through preprocessing of the Kaggle heart disease data set and the heart disease data set
provided by the Cleveland Medical Center, an optimized random forest algorithm was
proposed. In this diagnosis and recognition system, feature correlation analysis, adjust-
ment of model hyperparameters, and the comparison of related studies were adopted to
obtain a better accuracy rate and prediction effect. Although the random forest algorithm
has a high degree of fit and good generalization ability, due to the small heart disease
data set, the predictive effect of the model on a large data set is unknown. to prove the
predictive effect of the model on a large data set. The algorithm needs to be further
optimized in conjunction with feature engineering.
References
1. Dai, W.Y., Brisimi, T.S., Adams, W.G.: Prediction of hospitalization due to heart disease by
supervised learning methods. Int. J. Med. Inform. 84(3), 189–197 (2015)
2. Kedar, S., Bormane, D.S., Nair, V.: Computational Intelligence in Data Mining, vol. 1, pp. 69–
56. Springer, New Delhi, India (2016)
500 H. Xin and H. Yu
3. Chen, J.H., Wan, H.J., Zhang, Q.H.: diagnosis of heart disease based on support vector machine
with missing data. Math. Pract. Knowl. 47(02), 130–135 (2017)
4. Luo, B.J.: Design and Implementation of A Heart Disease Prediction Platform Based on
Random Forest. University of Posts and Telecommunications, Beijing (2018)
5. Wen, B.W., Dong, W.H., Xie, W.J., Ma, J.: parameter optimization of random forest based on
improved grid search algorithm. Comput. Eng. Appl. 54(10), 154–157 (2018)
6. Ding, W.J.: Research on Classification Algorithm in Prediagnosis of Heart Disease. Xidian
University, China (2019)
7. Zhao, J.C., Li, Y., Wang, D., Zhang, J.H.: An optimized random forest heart disease prediction
algorithm. J. Qingdao Univ. Sci. Technol. (Nat. Sci. Edition) 42(02), 112–118 (2021)
8. Wang, Y.S., Xia, S.T.: Overview of random forest algorithms for ensemble learning. Inf.
Commun. Technol. 12(01), 49–55 (2015)
Improved YOLOv3 Road Multi-target Detection
Method
1 Introduction
Object detection has always been an important research direction in the field of computer
vision. Traditional methods, such as histogram of oriented gradient (HOG) [1], scale-
invariant feature transform (SIFT) [2], deformable parts model (DPM) [3], mainly rely
on artificial extraction of the object feature. Meanwhile, only describing the appearance
and shape of an object cannot show its deep characteristic information, and the number
of candidate object boxes is more, leading to poor generalization ability of the model.
Besides, the practical application of the model is limited by a large amount of calculation
in the complex background and its weak robustness [4].
With the development of computing technologies, many researchers have focused on
deep learning methods. Currently, the detectors can be divided into two main categories.
The first category is region-based (two-stage framework) that performs object detection
into two steps, such as the R-CNN (R-CNN, Fast-RCNN, Faster-RCNN) [5–7] series
model. The other category is unified pipeline-based (one-stage framework) that uses
direct regression, such as SSD (SSD, DSSD, DSOD) series model and YOLO (YOLOv1,
YOLOv2, YOLOv3) [8–11] series model.
2 Related Work
YOLOv3 is mainly composed of two parts: darknet-53 feature extraction network and
prediction network. The whole neural network consists of 252 layers, among which
darknet-53 has 185 layers. It borrows the idea of a residual network [12] and uses a
large number of jump connections to avoid the problem of gradient dispersion. Besides,
YOLOv3 uses the feature pyramid network (FPN) [13] for feature extraction. In the
pyramid, the lower the feature layer, the less semantic information, but the more precise
location information; by contrast, the higher the feature layer, the more the seman-
tic information, and the fuzzier the location information. The feature fusion pyramid
obtains more abundant information through sampling on the fusion of low-level fea-
tures. Darknetconv2d_BN_Leaky (DBL) structure is the basic component of YOLOv3,
and it consists of a convolution layer, a batch normalization layer, and a Leaky-Relu
activation function. The network structure is shown in Fig. 1.
KITTI training set contains 7481 pictures with marked information. In road target
detection, the algorithm only needs to detect pedestrians and vehicles. So, this study
merges Van, Truck, and Tram into Car. Also, it merges Pedestrian (sitting), Cyclist, and
Pedestrian into Person and deletes misc.
Improved YOLOv3 Road Multi-target Detection Method 503
The number of pedestrian samples in the KITTI data set is less than that of vehicle
samples, which may cause problems such as overfitting. Therefore, this study incor-
porates self-made data set HCU-data into the KITTI data set. HCU-data contains 531
pictures, which are mainly mobile phone camera photos and road traffic surveillance
photos. In most of the pictures, people and cars are overlapped. Therefore, pictures with
non-target obstructions such as leaves and railings are added to improve the detection
ability of the model. Some of the pictures in the dataset are shown in Fig. 2.
The width and height of the anchor box in the original YOLOv3 algorithm are obtained
by clustering the COCO data set, but they do not necessarily apply to KITTI data sets.
Besides, this paper only focuses on the detection of vehicles and pedestrians, so it is
necessary to re-cluster the KITTI data set to select a more appropriate prior box. The
original YOLOv3 algorithm adopts a K-means [14] algorithm to select the anchor box.
However, this algorithm is sensitive to the selection of initial points. As more clustering
number increases, only locally optimal results may be obtained. Therefore, this paper
adopts the K-means++ [15] algorithm and define the distance by GIoU:
The point represents the coordinates of sample points, and the centroid represents
the center of the cluster. The algorithm steps are as follows:
3. Use formula 3 to calculate the probability of each sample being selected as the next
cluster center:
D(x)2
P= (3)
D(x)2
x∈X
The width and height of the anchor box obtained by the above algorithm on the
mixed dataset are listed in Table 1.
After the anchor box is obtained by the two algorithms, the detection result on the
mixed data set is listed in Table 2.
As can be seen from Table 2, compared with the original algorithm, the anchor box
obtained by the K-means++ clustering algorithm contributes to 0.81% higher accuracy
and 0.99 f/s faster detection speed on the mixed data sets.
In Fig. 3, the three superposition methods have equal IoU. Meanwhile, the left one
has the best regression effect and the right one has the worst regression effect.
Since IoU is only a ratio and is not sensitive to the size of the object, Rezatofighi
[16] et al. put forward GIoU in CVPR2019, and they proposed that IoU can be set as the
regression loss, as shown in Formula 5 [16].
|Ac − U |
GIoU = IoU − (5)
|Ac |
However, when the two contained boxes are in the horizontal and vertical directions,
the GIoU loss almost degenerates into the IoU loss. Zheng [17] et al. proposed DIoU to
handle this problem. The calculation of DIoU is shown in Formula 6 [17]:
ρ 2 b, bgt
DIoU = IoU − (6)
c2
However, considering the aspect ratio of the three factors of bounding box regression,
Zheng et al. proposed CIoU based on DIoU (see Formula 7).
ρ 2 b, bgt
CIoU = IoU − − αν (7)
c2
Where α represents the weight function and v is used to measure the similarity of
aspect ratio. The definition of v is shown in Formula 8 [17]:
2
4 ωgt ω
ν = 2 arctan gt − arctan (8)
π h h
Where ω and h are the length and the width.
The derivative of ν to ω is shown in Formula 9:
∂ν 8 wgt w h
= 2 arctan gt − arctan × 2 (9)
∂ω π h h w + h2
The derivative of ν to h is shown in Formula 10:
∂ν 8 wgt w w
= − 2 arctan gt − arctan × 2 (10)
∂h π h h w + h2
When both the length and width fall into the range of [0,1], ω2 + h2 is usually small,
1
and a gradient explosion appears, Therefore, ω2 +h 2 is replaced by 1 in the paper.
506 J. Fan et al.
This paper uses CIoUloss instead of the bounding box loss function in YOLOv3, and
the definition of CIoUloss is shown in Formula 11 [17]:
s
B 2
obj j∗ j
loss = Ii,j (1 − CIoU ) ∗ 2 − ωi hi
i=0 j=0
s2
B
j
j
obj j j
− Ii,j C i log Ci + 1 − C i log 1 − Ci
i=0 j=0
(12)
s
B 2
j
j
noobj j j
−λnoobj Ii,j Ci log Ci + 1 − C i log 1 − Ci
i=0 j=0
s2
obj
j
j
j
j
− Ii,j P i log Pi + 1 − P i log 1 − Pi
i=0 c∈classes
Table 3 shows that the average accuracy of IoU and GIoU is not much different, and
that of DIoU and CIoU is similar. CIoU contributes to the highest average accuracy of
92.18%, with a detection speed of 42.85 f/s.
3 Experiment
3.1 Experimental Environment and Parameter Setting
The experiment in this paper is performed by renting an instance on Featurize. The
instance runs Ubuntu 18.04 system, and the GPU configuration is manually set. The
software used in the experiment includes TensorFlow-GPU 1.13.2, kearas 2.1.5, and
opencv_python 4.1.2. The hardware platform is equipped with Intel (R) Core (TM)
I5-9400F CPU @ 2.90 GHz (6 cores), 218 GB disk space, and GeForce RTX 2080
Improved YOLOv3 Road Multi-target Detection Method 507
Ti graphics card (11 GB video memory, 31 GB main memory, with 26.90 TFLOPS
single-precision and 13.45 TFLOPS double-precision computing power).
The model uses batch gradient descent, and the idea of transfer learning was used to
load the pre-training weight. The epoch was set to 50000; the initial learning rate was
set to 0.001; the batch size was set to 64; the momentum was set to 0.9, and the weight
attenuation coefficient was set to 0.0005. The first 10000 training sessions were frozen.
When 40000 and 45,000 iterations were performed, the learning rate was respectively
decreased to 10% and 1% to accelerate the convergence of the model.
As for the detection of the first group of pictures, the original algorithm failed to
detect the two cars because they are only partially exposed and some of them are blocked
by leaves. In comparison, the improved algorithm successfully detected the two cars.
As for the detection of the second group of pictures, the original algorithm failed
to detect the black car at the school gate because it is far away from the camera and is
blocked by railings. In comparison, the improved algorithm successfully detected the
car.
508 J. Fan et al.
For the detection of the third group of pictures, the original algorithm detected the
pedestrians but failed to detect the black car behind due to the occlusion and small object.
In comparison, the improved algorithm can detect this object.
The results before and after the algorithm improvement are listed in Table 4.
Since there are only two categories of car and person in the data set, this paper uses
mAP, FPS, and F1-score to evaluate the performance of object detection models. The
calculation formula of mAP is shown in Formula 13:
APcar + APperson
mAP = (13)
2
The calculation formula of F1-SCORE is shown in Formula 14:
pre × rec
F1 = 2 × (14)
pre + rec
The proposed model is compared with YOLOv2, original YOLOv3, Faster-RCNN,
SSD, and the algorithms of other papers. The comparison result is shown in Table 5.
Table 5 shows that the average accuracy of the improved algorithm in this paper
reaches 92.79%, and the detection speed reaches 42.88 f/s. Meanwhile, the improved
algorithm achieves the highest f1-score, which is significantly better than that of other
models.
Improved YOLOv3 Road Multi-target Detection Method 509
4 Conclusion
Aiming at the problem that the yolov3 algorithm fails to detect small or blocked objects,
this paper uses the K-means++ algorithm to re-cluster on the mixed data set and selects
a priori frame more suitable for road objects to improve the average accuracy and speed
of detection. Then, the loss of boundary box is replaced with CIoU loss to improve the
positioning accuracy and average accuracy. The trained model is compared with other
models. The results show that the improved yolov3 algorithm performs significantly
better than other models, with an average accuracy of 92.79%, a detection speed of
42.88 f/s, and an F1 score of 91.62%.
There are still many deficiencies in the YOLOv3 algorithm. For example, the multi-
scale feature fusion method has poor portability, so it is not suitable for migration learn-
ing. Also, the detection of different scenes needs to select the appropriate feature fusion
structure to change the scene [19]. This paper also tried to add a 104 × 104 scale in
addition to the three detection scales. However, the experimental results on mixed data
sets are not as good as those of the original algorithm. With the development of computer
hardware, the YOLOv4 model or YOLOv5 model can be further optimized in the future.
References
1. Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: Proceedings
of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 886–893.
IEEE (2005)
2. Luo, J., Oubong, G.: A comparison of SIFT PCA-SIFT and surf. Int. J. Image Process. 3(4),
143–152 (2009)
3. Felzenszwalb, P.F., Girshick, R.B., Mcallester, D.: Object detection with discriminatively
trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 32(9), 1627–1645 (2009)
4. Li, Z., Huang, M.H.: Real-time vehicle detection based on YOLO_v2 Model. China Mech.
Eng. 29(15), 1869–1874 (2018)
510 J. Fan et al.
5. Girshick, R., Donahue, J., Darrell, T.: Rich feature hierarchies for accurate object detection
and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition (CVPR), pp. 580–587. IEEE (2014)
6. Girshick, R: Fast R-CC. In: Proceedings of the 2015 IEEE Conference on Computer Vision
and Pattern Recognition (CVPR), pp. 1440–1448. IEEE (2015)
7. Ren, S., He, K., Girshick, R.: Faster R-CNN: towards real-time object detection with region
proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137–1149 (2016)
8. Liu, W.: SSD: single shot multibox detector. In: European Conference on Computer Vision,
pp. 21–37 (2016)
9. Redmon, J.: You only look once: unified, real-time object detection. In: Proceedings of the
2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779–788.
IEEE (2016)
10. Redmon, J.: YOLO9000: better, faster, stronger. In: Proceedings of the 2017 IEEE Conference
on Computer Vision and Pattern Recognition (CVPR), pp. 7263–7271. IEEE (2017)
11. Redmon, J.: Yolov3: an incremental improvement. In: Computer Vision and Pattern
Recognition, arXiv preprint arXiv:1804.02767 (2018)
12. Lin, T Y.: Feature pyramid networks for object detection. In: Proceedings of the 2017 IEEE
Conference on Computer vision and Pattern Recognition (CVPR), pp. 2117–2125. IEEE
(2017)
13. He, K M.: Deep residual learning for image recognition. In: Proceedings of the 2016 IEEE
Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778. IEEE (2016)
14. Kong, F.F., Song, B.B.: Improved panoramic traffic monitoring object detection of YOLOv3.
Comput. Eng. Appl. 56(8), 20–25 (2020)
15. Arthur, D.: K-means++: the advantages of careful seeding. In: Eighteenth Acmsiam Sympo-
sium on Discrete Algorithms, New Orleans: Society for Industrial and Applied Mathematics,
pp. 1027–1035 (2007)
16. Rezatofighi, H.: Generalized intersection over union: a metric and a loss for bounding box
regression. In: Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern
Recognition (CVPR), pp. 658–666. IEEE (2019)
17. Zheng, Z H.: Distance-IoU loss: faster and better learning for bounding box regression. In:
Association for the Advance of Artificial Intelligence (AAAI), pp. 12993–13000 (2020)
18. Li, Y.P., Hou, L.Y.: Motion object detection in automatic driving based on YOLOv3 [J].
Comput. Eng. Des. 40(4), 1139–1144 (2019)
19. Liang, H., Wang, Q.W.: Review of small object detection technology. Comput. Eng. Appl.
1(13), 12–20 (2020)
Research on Ctrip Customer Churn Prediction
Model Based on Random Forest
Zhijie Zhao1,2 , Wanting Zhou1,2(B) , Zeguo Qiu1,2 , Ang Li1,2 , and Jiaying Wang1,2
1 School of Computer and Information Engineering,
Harbin University of Commerce, Harbin 150028, Heilongjiang, China
2 Heilongjiang Provincial Key Laboratory of Electronic Commerce and Information Processing,
Abstract. With the rapid development of the tertiary industry, the online reser-
vation market has great potential. The research on the customer churn factor of
customer-centric hotels is of great significance to the development of Ctrip. In
order to qualitatively analyze the causes of Ctrip’s scheduled customer churn, this
paper uses the current value and potential value in the customer value system to
determine the influencing factors of Ctrip’s customer churn. It introduces a random
forest algorithm to construct the Ctrip customer churn prediction model. Finally,
the confusion matrix and ROC curve are used to evaluate the performance of the
model. The results show that the random forest algorithm can better solve the
two-classification problem of customer churn prediction, and the accuracy of the
prediction model reaches 94%. In the analysis of influencing factors, the random
forest algorithm effectively avoids the collinearity interference between factors in
the traditional research methods and sets the objective weight, The weight of its
influencing factors can provide a scientific basis for Ctrip to formulate targeted
retention strategies.
1 Introduction
In 2020, China will build a well-off society all-around way, and the national GDP will
increase year by year. The people’s pursuit of a better life is not limited to material satis-
faction. The rapid economic development makes people more pursue spiritual pleasure,
and China has entered a new era of mass tourism. The general construction of network
infrastructure in China has caused a significant impact on the traditional offline tourism
industry. Multi-level and diversified tourism demand and consumer market patterns have
been formed. The Online Travel Agency (OTG) has developed ahead of schedule and
tourism consumption has become the rigid demand of the public. According to the 47th
Statistical Report on China’s Internet Development released by China Internet Network
Information Center, as of December 2020, the number of online travel booking users
in China had reached 342 million, accounting for 34.6% of the total Internet users [1].
According to the report on China’s online tourism industry in 2020, the tourism market
presents a pattern of “one super, multiple powers” for the OTA platform. The mainstream
online tourism booking platforms Meituan, Qunar tourism, Ctrip Travel, Tongcheng
Yilong, etc., are highly competitive. Among them, Ctrip Trave occupies 40.7% of the
market share, far ahead of other giants. However, due to the rapid development of the
network, the homogenization of products is serious. The competition among industries
has become increasingly fierce, and the amount of customer churn has increased sharply.
Therefore, to maintain the industry’s leading position, it is important to study customer
churn for Ctrip booking.
According to the customer life cycle theory, customer churn is inevitable. Research
has found that, on the one hand, a company’s customer churn will generate opportunity
costs due to reduced sales, and on the other hand, it will reduce attracting new customers.
For enterprises, the cost of redeveloping a new customer is about 300 to 600 dollars,
which is 5,6 times the cost of retaining an old customer [2]. Still, enterprises can extend
the customer life cycle to the greatest extent through reasonable marketing strategies, and
effectively retain the original customers has become the primary concern of enterprises.
This paper uses the hotel reservation data released by the OTA giant “Ctrip Travel”
as the basic data set to study the customer churn prediction of Ctrip hotel reservations.
Referring to the current value and potential value of customer lifetime value in the
customer value system to determine various variables to construct the characteristic
dimensions of Ctrip’s scheduled customer churn prediction, determine the prediction
model of customer churn based on random forest algorithm, and finally determine the
influencing factors of customer churn. The model can be used to analyze the factors
of Ctrip’s customer churn. For Ctrip, it is helpful to retain customers, reduce customer
churn and save costs.
2 Related Research
Customer value has become the center and focus of customer relationship management.
The “Two-Eight Law” indicates that 80% of the profits of enterprises come from 20%
of important customers. Identifying customer value has become the research focus of
enterprises with customer competitiveness at this stage. Verhoef et al. (2001) [3] proposed
for the first time that customer value is composed of two parts: current customer value
and potential customer value. Current value refers to the value created by customers
for the enterprise based on existing cooperation, and potential value refers to the total
profit contributions that customers can bring to the enterprise in the future. Liu Xiao
[4] used K-means and rough domain sets to classify aviation customers’ value and
established aviation customers’ value evaluation index system from the dual perspectives
of customers’ current and potential value. Tian Bo built a prediction model for Ctrip
customer churn based on the XGBoost algorithm and constructed a customer lifetime
evaluation system using current value and potential value [5].
Research on Ctrip Customer Churn Prediction 513
forest algorithm to build a Ctrip customer churn prediction model, and use the evaluation
index to compare and evaluate the model, and finally measures to retain customers.
According to the data processing sequence, the model is mainly divided into data pre-
processing, model construction and evaluation, churn reason analysis, and retention
measures. The research framework of this article is shown in Fig. 1.
Fig. 1. Flow chart of Ctrip customer churn prediction based on random forest.
defines customer churn as interaction with online hotel web pages but ultimately does
not purchase behavior. There are 500588 churn customers with browsing records and
have not placed an order and 189357 customers who have not to churn in the dataset.
Feature Selection Based On Customer Value. Customers with different values have
different contributions to the enterprise. This paper refers to the customer value eval-
uation system proposed by Shen Ziyao et al. [11], divides the customer value into the
customer lifetime value improvement system of current value and potential value, and
finally constructs a customer lifetime value evaluation system with 16 characteristic
dimensions in combination with the existing data indicators in the data set, As shown in
Table 1.
Data Outlier and Missing Value Processing. Due to the lack of multi-dimensional
data in the original data set, the lack of situation of each variable is shown in Fig. 2.
Most of the missing values of the 16 influencing factors based on the current value and
potential value are concentrated at about 30%. For attributes with missing values greater
than 30%, they still belong to the case of high missing values. In order to reduce the
impact of filling values on the original data, the high-dimensional extreme value – 999,
which has no practical significance, is selected to fill in the missing value.
The three influencing factors with less than 30% missing value, such as login duration
within 24 h, last visit duration within a year, and annual visit times, are filled in different
ways according to their distribution patterns. The same distribution pattern of influencing
factors is shown in Fig. 3. For the features with the same skew distribution, the missing
values are filled with the median.
User price preference_24 h most viewed hotel price, user price preference_24 h
browse average hotel price, customer value, and customer value_last one year. Generally
speaking, the four influencing factors related to value are positive, and the negative values
in the data are regarded as outliers. The median replaces the outliers related to the price
of user preferences, and the mode replaces the outliers related to the customer value.
Among them, H (X ) represents the final result of the model, I(•) represents the
indicative function, n represents the number of decision trees in the forest, hi (X ) is each
classification tree, and Y is the output variable. Random forest, as a nonlinear modeling
tool, can avoid overfitting. In the process of Ctrip’s customer churn data processing, the
random forest algorithm is used to establish a prediction model, which does not require
data pre-processing such as normalization and standardization of sample data, and the
weight of characteristic factors can be simply obtained to determine the importance of
affecting customer churn.
Where TP means true positive, that is, the actual value is 1, and the predicted result
is also 1; FP means false positive, that is, the actual value is 0, and the predicted result
is 1; TN is true negative, that is, the actual value is 0, and the predicted result is also 0;
FN means false negative, that is, the actual value is 1, and the prediction result is 0. Five
518 Z. Zhao et al.
indicators are used to evaluate the model based on these four values: accuracy, precision,
recall, F1-score, and AUC.
Accuracy is used to calculate the proportion of the number of correctly predicted
samples in the total samples, see formula 2; Accuracy is used to calculate how many
customers are churn in the total number of customers predicted to be churn, see formula
3; Recall rate refers to how many customers churn are predicted to be churn successfully,
see formula 4; F1 score is a comprehensive index of accuracy and recall, see formula 5.
It can be seen that the higher F1 score is, the higher the accuracy and recall will be; AUC
can measure the quality of a prediction model and it is determined by the area between
the ROC curve and coordinate axis, which is between 0.1 and 1. The larger the value,
the better the performance will be.
Accuracy = TP+TN
TP+TN +FP+FN (2)
Precision = TP
TP+FP (3)
Recall = TP
TP+FN (4)
The ROC curve and AUC can reflect the global performance index of the prediction
model. The ROC curve is composed of the horizontal axis false positive rate (FPR) and
the vertical axis true positive rate (TPR). True positive rate is the percentage of customers
who are actually churn and predicted to be churn, which is equivalent to the recall rate,
as shown in formula 6. False positive rate is the ratio of the number of customers who
are actually not churn but misjudged as churn, accounting for the total number of actual
non-churn customers, see formula 7.
TPR = TP
TP+FN (6)
FPR = FP
TN +FP (7)
The more the ROC curve deviates to the upper left corner, the better the performance
of the model, the greater the true positive rate and the smaller the false positive rate. The
AUC value is the area below the ROC curve. The larger the AUC area, the better the
model performance of the prediction classifier.
Model Performance Evaluation and Result Analysis. In order to further verify the
accuracy of the prediction model based on random forest, eliminate the error caused by
the inconsistency of the original data set to each prediction classification model. Based
on the same pre-processed Ctrip customer churn data, this paper uses four classification
prediction methods: logical regression, naive Bayes, support vector machine, and deci-
sion tree to compare and analyze the prediction with the random forest algorithm model.
The five index results of the model evaluation are shown in Table 3.
Research on Ctrip Customer Churn Prediction 519
It can be seen from Table 3 that for the five indicators of accuracy, precision, recall,
F1-score, and AUC, the values of random forest are higher than the other four prediction
models. The accuracy of random forest is 94%, the precision is 96%, and the recall rate
is 79%, with the highest AUC of 0.97. Similarly, it can be seen from the ROC curves of
the five models in Fig. 4 that the ROC curve of the random forest (rfc) is the uppermost
curve among the five prediction models. The more to the upper left corner, the higher
the true positive rate and the smaller the false positive rate. Its model performance is the
best, with the largest area of 97%. The prediction accuracy of customer churn is high,
the misjudgment rate is low, and the AUC is 97%. The AUC areas of logistic regression,
naive Bayes, and support vector machine are close, all about 60%. It can be seen that
the random forest model is better than other models for the two-classification problem
of customer churn and performs well in model generalization and accuracy testing.
520 Z. Zhao et al.
Table 4. Loss factor weight values of the optimal random forest prediction model.
According to Table 4, the random forest algorithm based on all factors of the value
system eliminates the number of annual user orders based on customer loyalty indicators
in the potential value of Ctrip reservations. For the online hotel reservation industry,
homogenization is serious. Compared with the annual visits of users under the same
customer loyalty index, the annual orders of users have little impact on the loyalty of the
online hotel reservation industry. Compared with offline physical stores, users” annual
orders reflect users’ loyalty. Only when the merchants are satisfied, the number of annual
orders will increase. For online, in the Internet era, traffic is king, and the satisfaction of
merchants is reflected in the number of visits. For customers with more extensive visits
and fewer orders, such customers are still high-value customers, maintaining customer
stickiness and loyalty to merchants. On the contrary, customers with fewer visits and
more orders may only be interested in the activities launched by the merchant at this
Research on Ctrip Customer Churn Prediction 521
time and make a purchase, but they have not made a purchase reservation the rest of
the time. Such customers are more likely to lose due to external factors. Therefore,
the number of annual user orders for Ctrip’s scheduled customer churn factor has been
eliminated. In addition, all influencing factors in current value and potential value are
directly proportional to the loss of Ctrip’s scheduled customers.
Analyzing the experimental results, the factor that has the greatest impact on Ctrip’s
scheduled customer churn is the value of customers in the past year, with a weight of
0.133476, followed by customer value with a weight of 0.124215; and the third is the
number of annual visits, with a weight of 0.083267. For Ctrip, the loss of a customer
is more related to its value and the number of annual visits. For high-value customers,
Ctrip has developed a membership system to retain customers, which increases the cost of
customer transfer and makes it easier to retain customers; Customers with a high number
of annual visits indicate that customers’ loyalty satisfaction with Ctrip is higher than that
of other companies and that such customers are not easy to lose. Therefore, for Ctrip, the
most important factors affecting customer churn are the customer’s value in the past year,
customer value, and the number of annual visits. From the weight of the experimental
results, it can be seen that the biggest impact on the loss of Ctrip’s scheduled customers
is the current value of customers. It shows that the most important thing that affects
the loss of Ctrip’s scheduled customers is improving the current value of customers,
keeping customers in a high-value stage, and avoiding churn. On the contrary, for Ctrip,
the three factors that negatively impact customer churn are user cancellation rate within
one year, 24-h login duration, and price sensitivity index. Their weights are 0.017746,
0.016944, and 0.007949 respectively. In particular, the price sensitivity index has the
smallest weight. For the Ctrip reservation system, customers pay more attention to factors
related to the hotel’s measures such as check-in experience, environment, sanitation, and
transportation, and pay less attention to the price. Moreover, most customers who book
rooms are business and tourism customers, and they do’ not pay much attention to the
price sensitivity index.
According to the customer value system, the factors affecting Ctrip’s scheduled cus-
tomer churn are the revenue value in the customer’s current value, the growth value in
the potential value, and the customer loyalty in the potential value. Given the analysis
of customer churn factors, in terms of retaining Ctrip customers, enterprises should pay
attention to customers and competitors, understand the current value of customers and
competitors’ products, understand customer needs, and develop more satisfactory ser-
vices and housing for customers to improve their overall competitiveness. The greater the
customer’s contribution to the enterprise, the higher the customer’s value. Ctrip should
attach more importance to high-value customers, face up to customer value, provide per-
sonalized services to high-value customers, and increase customer orders. The website
is the most crucial face for online enterprises. Optimize the website’s design, increase
Ctrip’s own promotion, and design different website links and jumps for customers of
all ages to increase the traffic of users to improve customer satisfaction and loyalty.
522 Z. Zhao et al.
5 Conclusion
In the online travel market, where market competition continues to intensify, customer
competition is indispensable. Compared with the massive cost of attracting a new cus-
tomer, most enterprises have begun implementing effective customer management on
existing customers. Customer management in the tourism industry provides a new way
of thinking. This paper uses the current value and potential value in the customer value
system to construct an index system for Ctrip’s customer churn factors. It uses the ran-
dom forest algorithm to construct a Ctrip customer churn prediction model, objectively
obtains the weights of factors that affect Ctrip’s reservations and provides Ctrip with
measures to retain lost customers and cause analysis. It is hoped that Ctrip can extend the
life cycle of customers and obtain greater profits. The research results show that: (1) The
prediction accuracy evaluation of four machine learning algorithms: logistic regression,
naive Bayes, support vector machine, and decision tree through the confusion matrix
and ROC curve shows that the accuracy and AUC value of the random forest algorithm
is the largest. It proves that the random forest model has higher usability in classifica-
tion and prediction problems. (2) The random forest algorithm is used to analyze the
influencing factors of customer churn, which effectively overcomes the interference of
the weight setting and the complex linear relationship between factors in the traditional
research on the influence of customer churn. It has more advantages over traditional
classical methods in prediction feasibility and accuracy. (3) According to the analysis
results of influencing factors, it can be seen that the most weighted is the revenue value
of the current value of the customer, indicating that the most significant factor affecting
Ctrip’s customer loss is the revenue value. From the analysis results, Ctrip has provided
effective measures to retain customers to obtain greater profits.
Acknowledgment. The Heilongjiang Provincial Science Fund Project supported this work (No.
LH2019F044).
References
1. China Internet Network Information Center.: The 47th Statistical Report on China’s Internet
Development, vol. 2 (2020)
2. Xiao, J., He, X., Teng, G., Xie, L.: Research on cost-sensitive semi-supervised integrated
model of customer churn Prediction. Syst. Eng. Theory Pract. 41(01), 188–199 (2021)
3. Donkers, V.B.: Predicting customer potential value an application in the insurance industry.
In: Decision Support Systems (2001)
4. Liu, X., Wang, X.: Research on aviation customer value classification based on K-means and
neighborhood rough set. Oper. Res. Manage. 30(03), 104–111 (2021)
5. Tian, B., Jiang, S., Gui, B.: Research on customer churn prediction based on customer
segmentation. Commun. World 27(06), 183–184 (2020)
6. Amin, A., et al.: Customer churn prediction in telecommunication industry using data
certainty. J. Bus. Res. 94, 290–301 (2019)
7. Kwon, H., et al.: Lifelog data-based prediction model of digital health care app customer
churn: retrospective observational study. J. Med. Internet Res. 23(1), e22184–e22184 (2021)
Research on Ctrip Customer Churn Prediction 523
8. Alexandros, D., Charalampos, A.: Designing a real-time data-driven customer churn risk
indicator for subscription commerce. Int. J. Inf. Eng. Electron. Bus. 12(4), 1–14 (2021)
9. Feng, Y.: Research on management methods of customer churn in tourism industry. Bus. Res.
08, 138–140 (2008)
10. Park, E., Kang, J., Choi, D., Han, J.: Understanding customers’ hotel revisiting behaviour: a
sentiment analysis of online feedback reviews. Curr. Issues Tourism 23(5), 605–611 (2020)
11. Shen, Z., Yuan, X.: integrated energy service customer identification based on parallelized
K-means. Power Eng. Technol. 40(02), 107–113 (2021)
12. Breiman, L.: Random forests. Mach. Learn 45(1), 5–32 (2001)
13. Kang, Y., Chen, Y., Gu, S., Yao, X., Huan, Q., Tang, Y.: Evaluation of sustainable utilization
of regional water resources based on random forest. Hydropower Energy Sci. 32(03), 34–38
(2014)
14. Zhang, R., Zhang, Y., Zhao, H., Ding, Z.: Quantitative comparison of the effects of bank
credit evaluation models. Financ. Supervision Res. 01, 66–85 (2021)
Business Intelligence
and Communications
An Evolutionary Game Analysis of Product
Crowdfunding Opportunistic Behavior
Considering Price Acquisition Model
1 Introduction
In the context of unstable development and the poor survival rate of small and medium-
sized enterprises, the problem of “difficult and expensive financing” has been widely
concerned. In particular, COVID-19 has once again caused a major impact on them.
At this time, the advantage of crowdfunding model becomes more and more distinct
among the financing methods of Internet finance. Product crowdfunding has a high
participation rate, which is a friendly boost for the development of small and medium-
sized enterprises. The reason is that product crowdfunding adopts the form of “pre-sale
+ group purchase”.
On the one hand, it can obtain start-up funds and promotion, and on the other hand,
it can carry out reasonable planning and coordination for the production of product
quantity. However, there is a serious information asymmetry between the platform,
sponsors, and investors. With the support of the third-party agent operating company,
the sponsors have prominent problems of opportunistic behavior, such as false publicity
is not in line with the real thing, exaggerated product quality is not in line with the
price, hiding negative information and other moral hazard problems. Therefore, the
opportunistic behavior of product crowdfunding sponsors is regarded as a key academic
research object.
However, the existing relevant research also has some regrets. First, the feasibility
of decision theory analysis from a single perspective in actual operation remains to be
discussed. In the research process, each subject’s mutual influence of interest relation-
ship and decision-making behavior should be considered. That is, the value acquisition
mode and the dynamic evolution of behavioral decision-making should be considered.
The value acquisition model is designed to solve these three problems: “who to charge”,
“what to charge” and “how much to charge” [1]. At the same time, the complete infor-
mation assumption of the complete information static game model is inconsistent with
the actual conditions. The decision of the platform group and the initiator group is a
dynamic evolution process, which the evolutionary game model should analyze. Sec-
ond, platform type incentive reward type in both positive practice condition, according
to current rules, through the horse race mechanism to allocate resources has grown up
in the form of platform, basically can achieve a fair and just, a short period of time
can’t change, and for the evaluation standard of self-discipline behavior, it is difficult
to determine. Accordingly, most of the temporary deposits for the later project quality
assessment compensates for investors’ complaints, etc. Therefore, the negative incentive
of punishment can be discussed and strengthened. Third, most of the relevant literature
studies using evolutionary game theory models use global stability analysis instability
analysis, and the equilibrium points obtained are no more than (0,0) and (1,1). Consid-
ering local equilibrium, that is, stability analysis with added constraints can draw more
effective conclusions.
Therefore, this paper will focus on the dynamic evolution process of the initiator’s
autonomous behavior and opportunistic behavior decision-making and the platform’s
active and negative regulatory decision-making in product crowdfunding. Considering
the interesting relationship between crowdfunding subjects, namely the platform value
acquisition model, the finite rational evolutionary game model is used to analyze the local
stability. A stable equilibrium point is explored and simulated using MATLAB. Through
the analysis of the evolution process, the platform mechanism design is expected to
provide optimization suggestions, maintain the high development momentum of product
crowdfunding.
Hypothesis 2. The set of strategies adopted by the initiator group is {S1 self-discipline
behavior, S2 opportunism behavior}, and the set of strategies adopted by the platform
group is {P1 active regulation, P2 negative regulation}. “Self-disciplined behavior [2]”
means that the sponsors provide products that meet the standards in full accordance
with the platform crowdfunding product rules; “Opportunistic behavior” refers to the
initiator’s use of more information in his/her possession to exaggerate the quality, launch
products that are not newly developed, sell them on other platforms, hide adverse news
and other behaviors. “Active supervision” refers to the strict examination and full man-
agement of enterprise qualification, trademark and patent, product quality and page
design in the early stage. If the examination fails, the project will not be allowed to go
online and the violation will be fined. “Negative supervision” refers to the relaxation or
lowering of audit standards in the early stage, and also includes the basic supervision to
deal with investors’ rights protection in the later stage of the project to ensure the normal
operation of the platform’s reputation.
Platform
Positive regulation P1 Negative regulation P2
Originator Self-discipline behavior W1 − C1 − R,R − D1 W1 − C1 − R,R − D2
S1
Opportunistic behavior (1 − m)W2 − C2 − C3 − W2 − C2 − C3 − R −
S2 (1 − m)R − nC4 − (m + nC4 , R − D2 − D3
n)F,(1 − m)R − D1 − D3
+ (m + n)F
According to the above analysis, the five local equilibrium points of system evolution
are (0, 0), (0, 1), (1, 0), (1, 1), (x0, y0), where x = x0, y = y0, 0 The stability of the local
equilibrium points of the system can be judged by checking the sign of the determinant
and trace of the Jacobi matrix of the dynamic system. When the sign is opposite, the
stability is obtained, and the evolution-stable strategy ESS is obtained. The Jacobi matrix
A of the system is as follows:
A11 A12
A=
A21 A22
Based on the replication dynamic analysis of the initiator’s group behavior decision
and the platform group supervision decision above, the local equilibrium analysis is
discussed in the following 9 cases.
It can be clearly seen from Table 2 above that the local equilibrium point is (1, 0) in
case 1, 2 and 3; (0,0) in case 4, 5, 7 and 8; (0,1) in case 6; and there is no local equilibrium
point in case 9. The following conclusions can be drawn:
Conclusion 1. The condition for the initiator group to evolve into self-disciplined behav-
ior is that the utility of opportunistic behavior is less than that of self-disciplined behavior,
that is, W2 − C2 − C3 − nC4 < W1 − C1 . The evolution process can be accelerated by
increasing W1 , decreasing W2 , and increasing n.
According to the local equilibrium point in case 1, 2 and 3 is (1, 0), it can be seen that
the initiator group stably chooses the self-disciplined behavior, and W2 − C2 − C3 − nC4
< W1 − C1 , where, the cost of the initiators of self-disciplined behavior C1 , the cost of the
initiators of opportunistic behavior C2 , the loss caused by opportunistic behavior to the
initiators C3 and the compensation cost of the complained opportunist initiators C4 will
not change in a short time. If the initiators want to be guided to choose the self-disciplined
behavior, then in the mechanism design, it is necessary to increase the revenue W1 of the
sponsors of self-disciplined behaviors through platform crowdfunding, reduce the rev-
enue W2 of the sponsors of opportunistic behaviors through platform crowdfunding, and
improve the probability of the sponsors of opportunistic behaviors being complained n.
An Evolutionary Game Analysis 533
Conclusion 2. The conditions for the final evolution equilibrium of the platform to be
actively regulated are as follows: the sponsor’s group stably chooses the opportunistic
behavior decision, the platform charging service fee is less than the fine and the regulatory
cost difference is less than the difference between the fine and the service fee, that is
(1 − m)W2 − C2 − C3 + mR − nC4 − (m + n)F > W1 − C1 ; mR − (m + n)F < 0;
D1 − D2 < (m + n)F − mR The evolution process can be accelerated by increasing D2 ,
increasing F, decreasing R, and increasing n.
In the 9 cases of partial equilibrium analysis, only in case 6, the platform evolves
the active regulation into a stable strategy. It can be seen from one side that it is difficult
for the platform to adopt the stable active regulation strategy. The supervision cost is
one reason, and the value acquisition mode of the platform is another reason. In order to
make the platform choose active supervision, D1 − D2 < (m + n)F − mR, except the
probability m of not being approved by the platform and the cost D1 of active supervision
of the platform will not change in a short time, when the mechanism is designed, It can
increase the cost of the platform’s negative supervision (D2 ), increase the punishment
of opportunism by active supervision (F), reduce the fees charged by the platform to the
sponsors (R), and improve the probability of the sponsors being complained (n).
Here, case 2 is selected as the data graph of the control experiment. At this time, W1
= 8, W2 = 10, C1 = 6, C2 = 5, C3 = 2, C4 = 3, D1 = 6, D2 = 5, m = 0.4, R = 2,
n = 0.5, F = 0.5. When other conditions remain unchanged, only one place of data is
modified for comparison. When W1 = 10 is increased, 1-(b) is obtained; when W2 = 9
is decreased, 1-(c) is obtained; when n = 0.8 is increased, 1-(d) is obtained, as shown
in Fig. 1 below. It is obvious that Fig. B, C and D all accelerate the evolution process,
that is, increasing W1 , decreasing W2 and increasing N make the initiator group reach
the stable evolution strategy of self-regulation earlier.
Fig. 1. The dynamic evolution process of sponsor group with time in case 2
Prove conclusion 2: In order to pay attention to the platform group and actively
monitor the evolution process speed, condition 6 meets the research conditions, at which
time W1 = 8, W2 = 15, C1 = 6, C2 = 5, C3 = 2, C4 = 3, D1 = 6, D2 = 5, m = 0.1,
R = 2, n = 0.5, F = 3. When other conditions remain unchanged, only one data is
modified for comparison. When D1 = 5.5 is increased, 2-(b) is obtained; when F = 3.5
is increased, 2-(c) is obtained; when R = 1 is decreased, 2-(d) is obtained; when n =
0.8 is increased, 2-(e) is obtained, as shown in Fig. 2 below. It can be seen that Fig. B,
C, D and E all accelerate the evolution process, that is, the increase of D2 , the increase
of F, the decrease of R and the increase of N make the platform group reach the stable
evolution strategy of active supervision in advance.
Fig. 2. The dynamic evolution process of the platform group with time in case 6
An Evolutionary Game Analysis 535
Prove conclusion 3: Opportunism behaviors for the sponsors and platform, the regu-
lation of the negative regulatory steps to simplify the proof, the originator of the strategy
of local stability selection opportunism behavior and platform using negative regulatory
situation, the originator and platforms can be observed at the same time, the dynamic
evolutionary process of 4, 5, 7, 8 can be as a condition of research, here, case 4 is selected
as the data graph of the control test. At this time, W1 = 8, W2 = 15, C1 = 6, C2 = 5,
C3 = 2, C4 = 3, D1 = 6, D2 = 5, m = 0.1, R = 2, n = 0.5, F = 1. Under the condi-
tion that other conditions remain unchanged, only one data is modified for comparison.
When W1 = 10 is increased, the result is 3-(b); when W2 = 13 is decreased, the result
is 3-(c); when D2 = 5.5 is increased, the result is 3-(d) (the maximum value of t on
the horizontal axis in Fig. D is 50; the rest Fig. is 20); when F = 1.5 is increased, the
result is 3-(e); when R = 1 is decreased, the result is 3-(F). When n = 0.8 is increased,
4-(g) is obtained, as shown in Fig. 3 below. Can see from Fig. b−g group of sponsors
and platforms have slowed the evolution process, but the platform group slow amplitude
is relatively obvious, the originator group only slow and small, namely, increasing W1
and W2 , D2 , increase, reduce the R, F improve n can allow the originator group delay
to reach the steady evolution of opportunistic behavior strategy, Moreover, the stable
evolution strategy of passive supervision is delayed for the platform group, which has a
larger effect on the platform group and a smaller effect on the initiator group.
5 Conclusion
First, the only way to solve the initiator’s opportunistic behavior is to make the utility of
opportunistic behavior smaller than that of self-disciplined behavior. Effective informa-
tion integration and high sharing among partners are the key factors for the success of
the alliance [5]. Therefore, the mechanism design of contemporary crowdfunding mode
should focus on how to improve the effectiveness of self-disciplined behavior and how to
reduce the effectiveness of opportunistic behavior, rather than reward and punishment.
536 G. Yang et al.
Acknowledgment. This research was supported by the 2020 Ideological and Political Work
Research Project of Harbin University of Commerce (2020SZY002) and Humanities and Social
Science Research Planning Foundation of Ministry of Education of China (21YJAZH099).
References
1. Chen, Y.L.: Research on Business Model of Platform Enterprises in Two-Sided Market. Wuhan
University, Wuhan (2014)
2. Wang, W.Y., Zhang, N.: An evolutionary game analysis of the opportunistic behavior of P2P
online lending platform’s illegal operation and government supervision. Enterp. Econ. 37(10),
163–172 (2018)
3. Chen, J.H., Wang, H., Zhang, Y.Q.: An evolutionary game analysis of stakeholder value co-
creation in service ecosystem environment. Oper. Res. Manage. Res. 28(11), 44–53 (2019)
4. Smith, J.M.: Evolution and The Theory of Games. Cambridge University Press, Cambridge
(1982)
5. Zhang, S., Hu, X.: Game analysis on logistics cloud service discovery and combination. Int. J.
Serv. Sci. Technol. 8(10), 193–202 (2015)
6. Zhao, J., Li, Y., Ding, Y., Liu, C.: The value of leading customers in a crowdfunding-based
marketing pattern. PLoS ONE 14(4), 1–18 (2019)
7. Yuan, Y., Guo, H.H.: Building innovative platform to incubate trend design – dialogue with
Gao Zheng, head of JD crowdfunding. Decoration 01, 31–36 (2017)
Research on the Model of Word-of-Mouth
Communication in Social Networks Based
on Dynamic Simulation
1 Introduction
In the context of the Internet era, Internet Word of Mouth has profoundly affected con-
sumers’ willingness to purchase products and purchase decisions by virtue of faster
transmission speed and wider coverage. When consumers make purchase decisions for
new and never-experienced products, they tend to actively search for product-related
information through the Internet to obtain product reviews and suggestions from expe-
rienced consumers [1]. Among various Internet communication platforms, social media
has not only become a daily communication tool for more and more consumers, but
word-of-mouth recommendations from the social circle of acquaintances are generally
more likely to be adopted by consumers because of their high credibility.
Word-of-mouth communication is the informal exchange of information about prod-
ucts, services, or brands between non-commercial consumers. Compared with traditional
2 Related Work
The principles of network information dissemination and disease transmission are very
similar [10]. First of all, both transmission mechanisms are transmitted through infectious
channels. The spread of disease means that healthy people are infected after they come
into contact with patients. In the online network, people who don’t know the topic
information are infected after contact with the disseminator, so that the information can
be further disseminated [11]. Second, the traditional infectious disease model divides
the population into healthy people, infected people, and cured people. Similarly, in
a complete online information dissemination network, users can also be divided into
groups of people who do not know the information, people who know that they have not
yet chosen to disseminate, people who know and choose to disseminate behavior, and
people who are immune to the information [12]. Therefore, this article will study the
information dissemination mechanism in online social networks based on the infectious
disease model.
Hypothesis: The total number of people is fixed during the process of information dis-
semination in social networks. The number of friends and relationships on Weibo remains
unchanged. After disseminating the information, the user will not contact the same infor-
mation again. The user’s level of interest in the information will not remain unchanged
and will be affected by the surrounding friends. And the trust relationship between friends
will also affect the user’s credibility of the information. The main body of information
dissemination is mainly composed of the disseminators and receivers of the information.
In information dissemination, the information disseminator and information receiver are
generally regarded as a node respectively. In Weibo, all users have a certain probability
of seeing this information, and there is a certain probability to forward, comment, or be
directly immune. Therefore, based on this dissemination mechanism and users’ potential
behavioral responses, comprehensively considering the microblog information dissem-
ination process, the nodes in the microblogging network are divided into the following
five forms:
1) S-node: The initial participating group refers to the initial group that has the
ability to obtain information, can see word-of-mouth information, and may conduct
communication behaviors.
2) Node E: A group of lurkers, a group of users who have been exposed to word-of-mouth
information but have not yet made a dissemination decision.
3) Node I: Communicator group refers to the user group that has browsed word-of-mouth
information in social networks and made clear decisions on dissemination.
4) R node: immune group refers to the user group that has completed or refused to spread
after obtaining word-of-mouth information.
5) S’ node: Participate in the group again. It means that after the immune group exits the
communication, there is a certain probability that µ will become the initial participant
540 Z. Fan et al.
group again and participate in the word-of-mouth information dissemination after re-
acquiring the word-of-mouth information or continue to refuse the dissemination.
According to the expressions of the above 5 types of nodes, construct the SEIRS
theoretical model, as shown in Fig. 1:
Use N to represent the number of all nodes in the social network. It is assumed that
the number of nodes in a certain period of time is constant, that is, the total number of
users at different nodes remains N . S(t) represents the number of unknown nodes at time
t, E(t) represents the number of latent morphological nodes at time t, I (t) represents
the number of nodes in the propagation node state at time t, and R(t) represents weak
immunity at time t The number of state nodes, O(t) refers to the number of nodes exiting
at time t. Based on the analysis of the above theoretical model, a dynamic differential
equation is established:
dS(t)
= −µ1 S(t)I (t) + µR(t)
dt
dE(t)
= µ1 S(t)I (t) − µ2 E(t)
dt
dI (t)
= µ2 E(t) − µ3 I (t)
dt
dR(t)
= µ3 E(t) − µR(t)
dt
This article uses MATLAB software to simulate the proposed word-of-mouth commu-
nication model. µ2 is the probability that a latent person turns into a communicator.
Therefore, it has a greater impact on the number of word-of-mouth communicators. We
first simulate and analyze the changes of µ2. By changing different µ2 sizes, observe
the changes in the number of latent groups and word-of-mouth communicators groups.
The simulation result is shown in Fig. 2.
From Fig. 2, we can see that with the continuous increase of µ2 from 0.1 to 0.7, the
balance point of the number of lurking nodes first increases and then decreases. In other
words, when the probability of turning a lurker node into a communicator node gradually
increases, the extreme value of the lurker node first increases and then decreases, instead
542 Z. Fan et al.
of continuously increasing. Therefore, for the promotion and exposure of specific word-
of-mouth topics in actual social networks, the higher the exposure, the better. Instead,
the best value should be grasped in advance, and the communication efficiency should
be maximized while spending the appropriate publicity price. As the probability of µ2
increases, the extreme value of the communicator group is also increasing, but as time
goes by, the number of communicator groups will also decline rapidly. This is also in line
with the law of the actual word-of-mouth dissemination process. However, many infor-
mation recipients are affected by the trust relationship in social networks and become
information disseminators, disseminating word-of-mouth information. However, as an
ordinary communication group, the word-of-mouth information is usually forwarded
and spread only once, and few people will spread the word-of-mouth information that is
forwarded multiple times. Therefore, as time goes by, many communicators will trans-
form into an immune group and will not spread word-of-mouth information again in a
short period of time. Therefore, if you want to make word-of-mouth topics more popular
quickly in real word-of-mouth communication, you need to increase efforts to convert
more lurkers into communicators.
In order to verify the impact of the re-transformation of the immune group into the
initial group on word-of-mouth communication, we verified the changes of each group
in the process of word-of-mouth communication when the return probability µ = 0
and µ = 0.01. The experimental results are shown in Fig. 3. From Fig. 3, we can see
that no matter whether the return probability is 0 or not, the number of communicators
will increase rapidly and reach a peak. Over time, the number of communicator groups
will also rapidly decline. However, the difference is that if the return probability is 0,
the number of communicators will drop rapidly and gradually approach 0. It can be
seen from Fig. 3 that when the return probability is 0.01, the number of communicators
will slow down and will gradually stabilize while maintaining a certain number of com-
municators. In actual word-of-mouth communication, if the communicator group only
spreads word-of-mouth information once, the popularity of word-of-mouth information
Research on the Model of Word-of-Mouth 543
will drop rapidly after reaching its peak, and it will be replaced by the heat of other
information.
In order to further verify the impact of return probability µ on word-of-mouth com-
munication, we simulated the situation where µ increased from 0.01 to 0.05. As shown
in Fig. 4, Fig. 4a shows the relationship between the initial population number and µ.
Figure 4b shows the relationship between the number of latent groups and µ. Figure 4c
shows the relationship between the number of communicators and µ. Figure 4d shows
the relationship between the number of exiting communicators and µ. From the figure,
we can see that the greater the probability of return, the more positive impact on word-
of-mouth communication will be. The number of word-of-mouth communicators is at a
higher level, which is more conducive to the spread of word-of-mouth.
number of final communicators. It shows that the greater the number of people who
become latent, the more conducive it is to become communicators in the end. In the
actual word-of-mouth dissemination process, it is necessary to increase the exposure
of the information as much as possible so that more people can receive word-of-mouth
information, thereby increasing the number of lurkers. µ2 is the probability that the
latent group transforms into the communicator group. It can be seen from Fig. 5b that
the larger µ2 is, the greater the number of communicators will be, which is the same as
the actual situation. But through simulation, we can find that although with the increase
of µ2, the number of communicators will increase rapidly. However, as time goes by,
the number of communicators will rapidly decay, and eventually, the number will stabi-
lize. When the µ2 probability is small, although the number of communicators grows
slowly, the number of communicators stays longer. After the final number stabilizes,
the number of communicators is not much different from when µ2 is high. This also
shows that in actual word-of-mouth communication, while pursuing a higher transfer
probability from lurkers to communicators, we should also find ways to increase the
duration of the number of high communicators. If the number of communicators can be
maintained at a high level for a long time, it will bring more word-of-mouth exposure,
which is equivalent to the spread of word-of-mouth. µ3 is the transformation probability
of the immune population. From Fig. 5c, it can be seen that this probability has a greater
impact on the number of communicators. In actual word-of-mouth communication, if
the communicator quickly loses interest in word-of-mouth information, the number of
communicators will rapidly decay.
For the return probability µ introduced in this article, it can be seen from Fig. 5d that
it has a very large impact on the number of communicators. When µ keeps increasing
from 0.1 to 0.7, it has little effect on the time for the number of communicators to reach
the maximum. The impact on the maximum number of communicators is also small.
But it has a very large impact on the stable number of communicators. This also shows
that in the process of word-of-mouth communication, if a part of the immune group
gets word-of-mouth information again to become a group of word-of-mouth commu-
nicators after they withdraw from the communication, it will stabilize the number of
word-of-mouth communicators with a higher number. This will achieve better word-
of-mouth communication effects. Therefore, in the actual word-of-mouth communica-
tion process, after a period of time of word-of-mouth communication, certain measures
should be taken to increase the exposure of word-of-mouth information again. Some
of the groups who have withdrawn from word-of-mouth communicators can be trans-
formed into potential communicators again through certain incentive measures. In this
way, word-of-mouth communicators can be stabilized at a higher value, and a better
word-of-mouth communication effect can be achieved.
Through the above simulation analysis, it can be seen that in the process of word-
of-mouth communication in social networks, various elements in users, information
and media will have different effects on the efficiency and speed of word-of-mouth
communication.
Research on the Model of Word-of-Mouth 545
5 Conclusion
This article is based on the model of infectious disease dynamics; combined with the
characteristics of actual word-of-mouth communication, new user status nodes are added,
and a new social network information dissemination model is constructed. The influence
of the transition probability of each group in word-of-mouth communication on word-
of-mouth communication is analyzed by simulation. The experimental results show
that the model constructed in this paper can truly reflect the law of social network
information dissemination. The communication model constructed in this article can
provide theoretical support for enterprises in formulating word-of-mouth communication
strategies to achieve the best word-of-mouth communication effect.
Acknowledgment. This research was supported by the Heilongjiang philosophy and Social
Science Fund Project (21GLC186).
References
1. Guerreiro, J., Pacheco, M.: How green trust, consumer brand engagement and green word-
of-mouth mediate purchasing intentions. Sustainability. 13, 7877 (2021)
546 Z. Fan et al.
2. Cheng, X.S.: The paradox of word-of-mouth in social commerce: exploring the juxtaposed
impacts of source credibility and information quality on SWOM spreading. Inf. Manage.
58(7), 103505 (2021)
3. Hussain, S., Ahmed, W., Jafar, R.M.S.: eWOM source credibility, perceived risk and food
product customer’s information adoption. Comput. Hum. Behav. 66, 96–102 (2017)
4. Cai, S.Q., Wang, W., Zhou, P.: Research on the dissemination of negative word-of-mouth
information in network communities based on multi-agents. Comput Sci. 43(4), 70–75 (2016)
5. Cai, S.Q., Yuan, Q., Zhou, P.: Research on the linear threshold propagation model of negative
word of mouth under corporate response. J. Syst. Eng. 32(2), 145–155 (2017)
6. Deng, W.H., Yi, M.: Research on the tree communication of online community word-of-
mouth information based on the information diffusion cascade theory. Chinese J. Manage.
14(2), 254–260 (2017)
7. Li, P., Yang, X., Yang, L.X.: The modeling and analysis of the word-of-mouth marketing.
Phys. A Statist. Mech. Appl. 493, 1–16 (2018)
8. Edmond, A., Michael, A., Susan, L.A.: An approach for combining ethical principles with
public opinion to guide public policy. Artif. Intell. 287, 303–349 (2010)
9. Li, Z., Xu, Y., Li, K.: The influence factors of collective intelligence emergence in knowledge
communities based on social network analysis. Int. J. Intell. Syst. 9(1), 23–43 (2019)
10. Yu, W., Shanshan, C., Xinchu, F.U.: Review and prospect of propagation dynamics models.
Commun. Appl. Math. Comput. 32(2), 267–294 (2018)
11. Thomas, W.C.: Equivalent probability density moments determine equivalent epidemics in a
sirs model with temporary immunity. Theor. Popul. Biol. 113, 1–9 (2017)
12. Pu, C., Li, S., Yang, X.X.: Traffic-driven SIR epidemic spreading in networks. Phys. A Statist.
Mech. Appl. 446, 29–137 (2015)
Spatial Correlation Analysis of Green Finance
Development Between Provinces
and Non-provinces Along the Belt and Road
in China
Abstract. Based on the panel data of China’s green finance development from
2008 to 2018, this paper constructs the green finance linkage network with the
gravity model. It uses the social network analysis method to analyze the network
structure characteristics and the green finance linkage network differences between
The Belt and Road and non-B&R provinces in China. The conclusions are as
follows: the spatial acceptance effect of the provinces along The Belt and Road is
relatively stronger, and the intermediary role in the green financial development
network is relatively stronger. Among the sections that divide the green finance-
related network, the provinces along The Belt and Road are more in the main
beneficiary section and two-way spillover board, promoting the development and
communication of green finance among the provinces. The paper puts forward
some policy suggestions, such as strengthening the spatial connection of green
finance in different provinces and implementing financial policies with regional
differentiation.
Keywords: Green finance · The Belt and Road · Gravitation model · Social
network analysis
1 Introduction
With the conflict between the traditional energy industry, manufacturing industry, and
environment becoming more intense, Green finance has developed rapidly. Boao Forum
for Asia 2021 has put one theme of “The Belt and Road Initiative”. In this context, it is
of great significance to conduct in-depth research on green finance from the perspective
of The Belt and Road. There are more and more studies related to it. In studying the
impact of green finance on the economy and environment, most scholars start from the
micro-level. For example, taking commercial banks as the research object, some scholars
analyze the impact of green credit on the risk-taking and competitiveness of commercial
banks by using the method of difference of differences and regression analysis (Shao
Chuanlin and Yan Yongsheng 2020; Gao Xiaoyan and Gao Ge 2018). From a macro
perspective, a few scholars have discussed the role of green finance in the construction
of green technology innovation systems and the process of industrial transformation and
efficiency improvement in China (Yan Jinqiang and Yang Xiaoyong 2018; Gu Beibei
et al. 2021). Some scholars used the mediating effect model to discuss the dual impact of
green credit on the economy and environment (Wang Yanli et al. 2021). Some scholars
used the propensity matching score method to prove the green finance experimental
area (Huang Haifeng and Zhang Jing 2021). In the research on the development status
of green finance, some scholars choose to study it from the perspective of financial
instruments. Some scholars chose the model of maximizing benefits to analyze the
impact of China’s green bond environment and its external characteristics on both sides
of the bond transaction and local government (Ba Shusong et al. 2019). Some studies use
the factor model to verify green incentives in China’s stock market, and green resources
are mainly allocated to the top green enterprises (Liu Yong and Bai Xiaoying 2020).
Some scholars also focus on the regional analysis. For example, some studies selected
the fixed effect space Dubin model, took the green finance in Guangdong Province as
the research object, and got its development characteristics and influencing factors (Yu
Fengjian and Xu Feng 2019). In the research on opening strategies related to green
finance, Some scholars explored the development status of green finance in The Belt
and Road strategy and provided suggestions for the development of green finance under
the strategy (Cao Mingdi and Dong Ximiao 2019; Yu Hongyuan and Wang Wanfa 2021).
From the above research, previous literature reflects the increasing practice of green
finance in China. However, there are several limitations: First, there are still few regional
analyses and studies, and the spatial characteristics of green finance revealed are only
limited to local areas. Second, when analyzing the regional characteristics of green
finance, it ignores the role of national strategy in developing regional green finance. Given
the above shortcomings, this paper makes the following improvements: First, the social
network analysis method is adopted to comprehensively analyze the changing trend of
the whole network and each node. Second, this paper combines the development strategy
of The Belt and Road when analyzing the correlation network. From the perspective of
The Belt and Road strategy, this paper reveals the spatial correlation difference of green
finance development between provinces along the Belt and One Road and provinces
outside the Belt and One Road.
2 Model Construction
2.1 The Gravity Model
This paper chooses the gravity model to construct a regional social network. In order to
increase the applicability of the gravity model to the social network analysis method,
this paper revises it. The model is as follows:
√
3
√2
Gi Ii Ci 3 Gj Ij Cj G i Ii
Rij = Kij Kij = √ √ (1)
Dij2 2
Gi Ii + 2 GJ IJ
In Eq. (1), Rij represents the correlation strength of green finance development level
between I and J province. Gi and Ii jointly represent the development level of green
Spatial Correlation Analysis 549
finance in I province. Considering the availability and scientific of data, Gi and Ii are
expressed by the green credit ratio and total investment in environmental pollution control
of a province (Dong Xiaohong and Fu Yong 2018). Ci is an indicator of the spatial
connection of the development level of green finance. The carbon emissions of a province
are closely related to the local economy and environment, so this paper uses the carbon
emissions of province I to represent Ci . Dij represents the distance between province I
and province j, and Kij is the gravitational parameter.
According to the formula (1) model, this paper calculates the gravity matrix of the
intensity of inter-provincial green finance development level correlation. It conducts
binarization processing on the obtained matrix. The threshold value selected for pro-
cessing in this paper is the mean value of matrix columns. The calculation results higher
than the threshold value are denoted as 1, and the results lower than the threshold value
are denoted as 0.
In Eq. (3), (4) and (5), the meanings of N and L are consistent with that of Eq. (2).
In Eq. (4), gjk represents the number of shortest paths between province j and k; gjk(i)
represents the number of shortest paths between province j and k through I. In formula
(5), Dij represents the shortest path of the province i and province j.
This paper uses Ucinet software to analyze the overall network characteristics of green
finance development level (as shown in Table 1).
As shown in Table 1, the overall density level of China’s green finance development
network has little change. Although the variation trend of network density was not stable
in the sample period, the overall network density remained in the range of 0.21–0.23, with
a small fluctuation range. In the sample investigation period, the maximum number of
network relationships is 200, far less than the maximum number of network relationships
Spatial Correlation Analysis 551
870. This shows that there is still great room for progress in promoting inter-provincial
green finance links.
In this paper, NetDraw, a visualization tool of Ucinet, is used to draw the spatial
association network diagram of green finance development level between provinces and
non-provinces along The Belt and Road in China (as shown in Fig. 1).
Note: ■ means The Belt and Road province ■ means non-B&R provinces
There are obvious regional green finance-related small groups in the overall network
of the provinces along The Belt and Road, such as the three eastern provinces, Gansu
and Ningxia in the northwest, and the southern marginal provinces such as Shanghai,
Zhejiang from Fig. 1. The Belt and Road strategy enhances the relevance of the green
financial network. It can be seen from Table 1 that the network correlation degree from
2008 to 2018 is 1. It shows that there is a connection between the development of green
finance in all provinces of China.
This paper mainly shows the influence of each province on the green finance development
network in China through three indicators: degree, closeness and betweenness. Table 2
is obtained using data from 2018.
The Difference in Degree. The degree mainly reflects the province’s degree in the data
core position in the national linked data network. According to Table 2, there are 17
provinces exceeded the national average. These provinces are more associated with
other provinces in the degree network. The last five-degree in China are Jilin, Shanghai,
552 C. Zheng and X. Dong
Heilongjiang, Qinghai, and Hainan. This fully shows that these provinces are less corre-
lated with other key provinces in the degree network. The reason may be that these places
are remote or their carbon emissions are too high, which affects their development of
green finance. According to Table 2, Shaanxi, Xinjiang, Guangdong, Liaoning, Ningxia
have a higher indegree. They all belong to the provinces along The Belt and Road route.
Drived by the Belt and Road development strategy, their connection to green finance
development has also been enhanced.
The Difference in Closeness. Closeness represents the extent to which other regions
within the same network do not control a province. The 12 provinces exceeded the
national average. It shows that these provinces play the role of central actors in the
network. The bottom five provinces are Shandong, Henan, shaanxi, Hubei, and Anhui.
Due to their economic and location limitation, these provinces play the role of marginal
actors in the network. It can be found that the top five provinces are all along The Belt
and Road. Under the influence of The Belt and Road strategy, the efficiency of economic
connectivity between these provinces and other provinces has been enhanced.
Plate 1 2 3 4 N P A C
1 51.000 17.000 7.000 0.000 10.000 31.034 68.000 Br
2 3.000 22.000 0.000 5.000 6.000 17.241 73.333 M
3 6.000 2.000 41.000 14.000 8.000 24.138 65.079 Br
4 0.000 5.000 4.000 23.000 6.000 17.241 71.875 Bi
Note: 1 stands for first plate; 2 stands for second plate; 3 stands for third plate; 4 stands for
fourth plate; N is the abbreviation of Number of members; P is the abbreviation of Percentage of
expected internal relationships (%); A is the abbreviation of Actual Internal Relationship Ratio
(%); C is the abbreviation of Characteristics of plate; Br is the abbreviation of Broker board; M is
the abbreviation of Main beneficiary block; Bi is the abbreviation of Bidirectional overflow plate
The first plate is a brokered board according to the results in Table 4, which both
accepts external contacts and sends contacts to other boards. It plays an important role
of “intermediary” and “bridge” in China’s development network of green finance. The
second plate is the main beneficiary block. The major provinces of this plate are all in
the western region, where the economy is underdeveloped, and the environment is under
great pressure. Its green finance development is greatly affected by other sectors. The
third plate is a brokered board. It also acts as an “intermediary” in the green finance
development network, but its spillover relationship is less than that of the first sector.
The fourth plate is a two-way overflow plate, which sends out the contact and receives
the contact of other plates.
It can be seen that the provinces along The Belt and Road are mainly distributed in the
main beneficiary plate and the two-way overflow plate, according to Table 3. Most of the
provinces along The Belt and Road are located in marginal areas. They are remote, and
the conflict between the natural environment and economic development is serious. The
implementation of The Belt and Road strategy has enhanced economic exchanges and
cooperation between provinces along The Belt and Road and other Chinese provinces
or countries along The Belt and Road.
Based on 30 provinces’ data from 2008 to 2018, this paper constructs the spatial cor-
relation network of green finance development in China and investigates its network
structure characteristics. The results are as follows:
(1) The development of green finance in China constitutes an overall network. But
overall network density was relatively low and generally remained stable during the
sample period. It shows that there is still much room for progress in the connection
of green finance development in China’s provinces.
Spatial Correlation Analysis 555
(2) Provinces along The Belt and Road have a higher degree and closeness. This means
that the spatial acceptance effect of provinces along The Belt and Road is relatively
stronger, and the intermediary role is relatively stronger in the network. Their close-
ness is relatively low, indicating that the degree of control over other provinces in
the green finance network is relatively low.
(3) The Belt and Road provinces are mainly distributed in the main beneficiary plate
and the two-way overflow plate. The main members of these two plates are the
western region and the southern marginal provinces. Their economic exchanges
and cooperation with other regions have been strengthened due to The Belt and
Road.
References
Chuanlin, S., Yongsheng, Y.: Is green finance a “double-edged sword” for commercial banks to
take risks? J. Guizhou Univ. Finance Econ. 1, 68–77 (2020). (in Chinese)
556 C. Zheng and X. Dong
Xiaoyan, G., Ge, G.: Research on the relationship between green credit scale and commercial
banks’ competitiveness. Econ. Issues 7, 15–21 (2018)
Jinqiang, Y., Xiaoyong, Y.: Promoting the construction of green technology innovation system
with green finance. Fujian Forum (Humanit. Soc. Sci. Edition) 3, 41–47 (2018)
Gu, B., Chen, F., Zhang, K.: The policy effect of green finance in promoting industrial transforma-
tion and upgrading efficiency in China: analysis from the perspective of government regulation
and public environmental demands. Environ. Sci. Pollut. Res. 28(34), 1–18 (2021)
Wang, Y., et al.: The dual impacts of green credit on economy and environment: evidence from
China. Sustainability 13(8), 4574 (2021)
Huang, H., Zhang, J.: Research on the environmental effect of green finance policy based on the
analysis of pilot zones for green finance reform and innovations. Sustainability 13(7), 3754
(2021)
Shusong, B., Yujia, C., Weihao, Z.: Green bond theory and China’s market development analysis
[J]. J. Hangzhou Normal Univ. (Soc. Sci. Edition) 1, 91–106 (2019)
Yong, L., Xiaoying, B.: Green incentives in China’s stock market: the perspective of sustainable
development. Econ. Manage. 12(1), 155–173 (2020)
Fengjian, Y., Feng, X.: The development of green finance and its influencing factors in Guangdong
Province from the spatial perspective: an empirical study based on the fixed-effect spatial Durbin
model. Sci. Technol. Manage. Res. 39(15), 63–70 (2019)
Zhang, D., Mohsin, M., Rasheed, A.K., Chang, Y., Taghizadeh-Hesary, F.: Public spending and
green economic growth in BRI region: mediating role of green finance. Energy Policy 153
(2021)
Mingdi, C., Ximiao, D.: Green finance and The Belt and Road initiative: evaluation and prospect.
J. Renmin Univ. China 33(04), 2–9 (2019)
Hongyuan, Y., Wanfa, W.: Green The Belt and Road construction: progress, challenges and
deepening paths. Stud. Int. Stud. 02, 114–129 (2021)
Xiaohong, D., Yong, F.: Analysis on the spatial dynamics of green finance and green economy
coupling development. Ind. Technol. Econ. 37(12), 94–101 (2018)
Study on Decision-Making Behavior of Effective
Distribution of Fresh Agricultural Products
After COVID-19 Epidemic
1 Introduction
As fresh and green agricultural and sideline products, fresh agricultural products are
related to farmers’ income and to residents’ three meals a day, which is closely related to
people’s quality of life and modern agricultural development [1]. Although the work of
agriculture, rural areas, and farmers has entered a new development stage, ensuring stable
agricultural production and supply and increasing farmers’ income is still the focus, and
the effective circulation of fresh agricultural products has undoubtedly become the focus
of extensive attention [2].
Under the background of the new crown epidemic, as the speed of virus transmission
is fast and the scope of the spread is large, it is necessary to strictly control the traffic
personnel and vehicles. For the fresh agricultural products with high requirements such
as freshness and aging, the longer the transportation time, the greater the consumption
[3]; Secondly, the continuous innovation of production technology makes the output of
agricultural products reach new highs, resulting in a large number of unsalable fresh
agricultural products. Due to the suspension of work in some areas, the distribution
efficiency of fresh agricultural products has been greatly reduced [4].
To sum up, to ensure the effective distribution of fresh agricultural products during
the epidemic, the decision-making behavior of both parties is studied with the leading
logistics enterprises and local governments as the game players. In addition, the existing
research mainly discusses the impact of the epidemic on agricultural production and the
countermeasures. It does not consider the main body’s strategic choice in the effective
distribution process of fresh agricultural products under the background of COVID-19.
On this basis, this paper proposes an evolutionary game model for the effective distri-
bution of fresh agricultural products after COVID-19 and studies the strategic choice of
participants and the conditions for achieving stable conditions. Then, the measures to
improve the distribution efficiency of fresh agricultural products under COVID-19 are
discussed. Finally, the corresponding suggestions and solutions are put forward.
In the evolutionary game theory, the two groups of logistics companies and local gov-
ernments will take certain actions in order to survive better. When all individuals in a
group choose the same behavior, the mutants will either be. The system is eliminated, or
the strategy is changed to adapt to the system environment [5]. This requires participants
to constantly adjust their strategies to adapt to the system in the dynamic changes of
the system, which is exactly in line with the research on the decision-making behavior
of logistics enterprises and local governments. When a logistics company chooses to
“actively invest”, it will have a strong social awareness and be able to actively connect
with areas where agricultural products are unsellable, continue to invest in the logistics
resources needed, and fight the epidemic together with the people of the whole coun-
try; if it chooses “passive investment”, the logistics company for their benefit, they will
not be willing to participate in it, or perfunctory. When the local government chooses
“coordination”, it will adopt a certain way to encourage logistics enterprises to invest in
logistics resources in slow-sale areas during a special period; if it chooses “uncoordi-
nated”, the government will adopt other methods to solve the transportation problem of
agricultural products.
From the perspective of input costs and benefits of both sides of the game, in order
to better calculate the parameters and rationalize the evolutionary game model, the
following assumptions are put forward:
Study on Decision-Making Behavior 559
– H1 : Logistics enterprises and local governments that maintain limited rationality will
make decisions according to the actual situation.
– H2 : Because logistics enterprises know their investment level but do not know whether
the local government will encourage and how much the incentive level is; However,
local governments do not know whether logistics enterprises will actively invest in
logistics resources, which makes the information obtained by both parties asymmetric.
– H3 : Based on the dynamic nature that both sides of the game achieve equilibrium
through continuous trial and error, it is considered that the decision-making behavior
of logistics enterprises and local governments is a dynamic process.
– H4 : The strategies adopted by logistics enterprises are “positive investment” and
“negative investment”, while the strategies adopted by local governments are
“coordination” and “disharmony”.
– H5 : In the game process, the proportion of individuals that logistics enterprises choose
to actively invest in the population at time t is x, while the proportion of individuals
that local governments choose coordination strategies at time t is y.
In order to facilitate the analysis of different strategies and the evolution paths of both
sides of the game under different circumstances, the relevant parameters of the model
are set as shown in Table 1.
Definition Symbol
Basic income of logistics enterprises without active investment A > 0 A
The basic income when local government is uncoordinated B > 0 B
The cost of logistics enterprises’ active investment C1 > 0 C1
The cost of local government’s coordination strategy C2 > 0 C2
Total revenue of local government coordination D > B > 0 D
The local government that adopts the coordination strategy will give some rewards E
to the logistics enterprises that actively invest E > 0
When the local government adopts the coordination strategy, it will punish the F
logistics enterprises with negative investment F > 0
The cost-benefit conversion coefficient of local government facing the problem of α
agricultural products distribution when enterprises put in negative investment
0<α<1
Amplification factor of value added when logistics enterprises actively invest β > 1 β
Based on the above analysis and assumptions, the income matrix of the game between
logistics enterprises and local governments is analyzed and calculated [6], as shown in
Table 2.
560 Y. Yang et al.
Local government
Coordinate Uncoordinated
Logistics enterprises Active engagement βA + E − C1 , D − C2 βA + C1 , B
Negative input A − F, αD − C2 +F A, αB
The local government chooses the coordination strategy at time t to obtain the
following benefits:
U21 = x(D − C2 ) + (1 − x)(αD + F − C2 )
At t time, the benefits of choosing the uncoordinated strategy are as follows:
U22 = xB + (1 − x)αB
At time t, the total income of the local government population is as follows:
U2 = yU21 + (1 − y)U22
Similarly, the dynamic equation of local government replication is as follows:
ẏ = y(U21 − U2 ) = y(1 − y){x[(1−α)(B + D) − F] + α(D − B) − C2 + F}
enterprises will generally be higher than the passive input of the enterprises, otherwise,
the coordination strategy of the local governments will not have any significance; If the
local government adopts an uncoordinated strategy to the logistics enterprise, the income
from the positive investment of the logistics enterprise will be less than the income from
the negative investment of the enterprise. This highlights the impact of local govern-
ment coordination on the behavior choice of logistics enterprises to a certain extent.
Similarly, suppose the logistics enterprises adopt the strategy of active investment. In
that case, the local government’s revenue from adopting the coordination strategy will
be greater than that from adopting the uncoordinated strategy, which improves the local
government’s enthusiasm to coordinate the logistics enterprises to a certain extent; This
highlights the impact of local government coordination on the behavior choice of logis-
tics enterprises to a certain extent. Similarly, suppose the logistics enterprises adopt the
strategy of active investment. In that case, the local government’s revenue from adopt-
ing the coordination strategy will be greater than that from adopting the uncoordinated
strategy, which improves the local government’s enthusiasm to coordinate the logistics
enterprises to a certain extent. To sum up, the constraints of the equilibrium solution of
logistics enterprises and local governments in the game system are as follows:
βA + E − C1 > A − F, βA − C1 < A, D − C2 > B, αD − C2 + F < αB
Because the simultaneous replication of dynamic equations can better reflect the
speed and direction of their evolution [8], there are five local equilibrium points in the
system by solving the dynamic replication equations:
E1 (0, 0), E2 (0, 1), E3 (1, 0), E4 (1, 1), E5 (x∗ , y∗ )
Since the equilibrium point obtained from the replicated dynamic equation is only
the local asymptotically stable point in the evolution process, the Jacobian matrix in the
system is analyzed by using the method proposed by Friedman [9], and the stability in
the evolutionary equilibrium is obtained. If the trace of the Jacobian matrix is less than
zero and the determinant is greater than zero, it is an evolutionary stability strategy [10].
After calculation, the symbols of determinant and trace are shown in Table 3.
Because the broken lines of stable, unstable, and saddle points are the boundary lines
of game evolution under different strategies, the trend chart of interaction between local
government and logistics enterprises can be obtained.
562 Y. Yang et al.
By analyzing the parameters that affect the evolution of the system one by one, and
drawing the corresponding trend diagram. The linkage reaction of parameters is not
considered here. For the linkage reaction of parameters, the influence of main parameters
on the system evolution results is observed when the conditions are satisfied.
Parameter E, F. When E increases, E5 moves vertically downward (see Fig. 1(a)), the
area of region E5 E2 E1 E3 decreases and the area of region E5 E2 E4 E3 increases, and the
probability of the system converging to the ideal mode increases, which is conducive to
the benign evolution of the system. When F increases, E5 moves to the left and down(see
Fig. 1(c)), the area of region E5 E2 E1 E3 decreases, while the area of region E5 E2 E4 E3
increases, and the system converges to the ideal mode faster.To sum up, the promotion
of the reward value of the enterprise’s positive investment and the increase of the penalty
for the enterprise’s negative investment make the logistics enterprises more and more
tend to the positive investment strategy.
Parameter C1 , C2 . C1 refers to the cost that logistics enterprises pay when they actively
invest. When C1 increases, E5 moves vertically upward (see Fig. 1(b)), the area of region
E5 E2 E1 E3 increases. Both sides of the game gradually tend to the most unsatisfactory
state, which is not conducive to the benign evolution of the system.C2 refers to the
cost of local government coordination. When C2 increases, E5 level moves to the right
(see Fig. 1(d)), the area of region E5 E2 E1 E3 increases and the area of region E5 E2 E4 E3
decreases, and the probability of the system converging to the ideal mode decreases.
From the above analysis, it can be seen that if the cost is increasing, the strategy choice
of both sides of the game is not conducive to the benign evolution of the system.
In order to verify the stability of the equilibrium point, the simulation value is set
according to the relevant constraints, and then the evolution path of logistics enter-
prises and local governments under the influence of the major epidemic is simulated.
564 Y. Yang et al.
logistics enterprises choose to actively invest, and the government chooses to coordinate.
The government and logistics enterprises will jointly solve the effective distribution of
fresh agricultural products during the epidemic period, which is the ideal situation, as
shown in Fig. 3(c).
The simulation results show that the model is feasible. Accordingly, the develop-
ment suggestions of the participants are given respectively. For logistics enterprises,
the unnecessary waste of some resources can be reduced by increasing the batch of
single transportation, to reduce the transportation cost invested by enterprises; Or by
reducing unnecessary links in transportation and selecting direct transportation mode if
possible, to reduce the cost of loading, unloading and handling of agricultural products.
In addition, by increasing the rendering of humanistic feelings against epidemic dis-
ease, enterprises can realize that the investment in logistics resources in unsalable areas
of agricultural products will produce greater social benefits, far exceeding the benefits
created by enterprises themselves, and stimulate the initiative of enterprises to actively
invest while deepening the cultural heritage of enterprises [11]. For local governments,
we can build an information service platform for the circulation of agricultural prod-
ucts. Relevant departments can publish the current situation of unsalable agricultural
products, circulation progress, resource investment of various logistics enterprises and
rewards obtained after investment here, and ensure the openness, transparency and shar-
ing of information. The government can improve the enthusiasm of resource investment
of logistics enterprises by comparing the resource investment of various enterprises dur-
ing the epidemic period and using the power of public opinion and the psychology of
comparison among enterprises.
– the game players must actively participate in the effective transportation process
of fresh agricultural products, jointly seek development and actively adjust the
development strategy.
– From the perspective of the whole system, it is necessary to introduce a third party
to supervise this to promote the cooperation between the two sides of the game and
ensure the efficient distribution of fresh agricultural products.
This study aims to effectively solve the distribution problem of fresh agricultural
products during the COVID-19 period, which can provide a theoretical basis for the
prevention and control of other public health events in the future.
566 Y. Yang et al.
Acknowledgments. This work is partly supported by the project supported by the National Social
Science Foundation (16BJY125), Heilongjiang philosophy and social sciences research planning
project (19JYB026), Key topics in 2020 of the 13th five year plan of Educational Science in
Heilongjiang Province (GJB1320276), Project supported by undergraduate teaching leading talent
training program of Harbin University of Commerce (201907), Key project of teaching reform and
teaching research of Harbin University of Commerce in 2020 (HSDJY202005(Z)), Innovation and
entrepreneurship project for college students of Harbin University of Commerce (202010240059),
School level scientific research project of Heilongjiang Oriental University (HDFKY-200202), Key
entrusted projects of higher education teaching reform in 2020 (SJGZ202 00138).
References
1. Run, H., Ye, Q.: The enlightenment of epidemic situation on the circulation of fresh
agricultural products. Macroecon. Manage. 09, 34–35 (2020)
2. Sun, W., Xu, M.: Mechanism and empirical study on the impact of agricultural product
circulation system on farmers’ income. Commercial Econ. Res. 11, 126–129 (2021)
3. Pan, J., Guo, M.: COVID-19’s influence on the circulation of agricultural products. Bus. Econ.
Res. 11, 155–157 (2020)
4. Zhang, X.: Novel coronavirus pneumonia impact on agricultural product supply chain and
coping mechanism. Agric. Econ. Manage. 04, 45–51 (2020)
5. Smith, J.M., Price, G.R.: The logic of animal conflict. IMA J. Appl. Math. 246(5427), 15–18
(1973)
6. Liu, Z.: The “double subjects” in the regional innovation system in the transition period:
based on the evolutionary game between the government and entrepreneurs. Econ. Issues 05,
113–122 (2020)
7. Huang, M.: Research on cooperative product development and cooperation mechanism of
supply chain based on evolutionary game. Chin. Manage. Sci. 6(18), 155–162 (2010)
8. Liu, X.: Research on the evolutionary game between government and enterprise under the
fiscal and tax incentive policies – taking small and medium-sized science and technology
enterprises as an example. Tech. Econ. Manage. Res. 10, 16–21 (2019)
9. Friedman, D.: Evolutionary game in economics. Econometrica 59(3), 637–666 (1991)
10. Zhang, J., Xi, X.: Evolutionary game analysis on nitrogen emission reduction of local govern-
ments and enterprises under the trading of river basin emission rights. Chin. Manage., 1–12
(2020)
11. Nie, L., Zhang, L.: Evolutionary game analysis and simulation of government and sewage
enterprises in green technology innovation. Econ. Issues 10, 79–86 (2020)
Study on the 2-Mode Network Characteristics
of the Types and Issuing Places of Chinese
Provincial Green Bonds
Abstract. Although China’s green bonds started considerably late, they have
become an important part of international green bonds. According to the develop-
ment status of Green bonds in China. In this article, 2-mode network analysis is
used to construct the network of the types and issuing places of China’s provincial
green bonds. The UCINET software is used to analyze and study the network
centrality and singular value decomposition. This paper studies the functions and
roles of different issuing places in China’s inter-provincial green bond market from
the relationship between the green bond types and issuing places. It is concluded
that all issuing places have generally established connections with the green bond
market, and the eastern region has obvious advantages in the types and quantity
of green bond issuance. There are two specific market clusters in the green bond
market: multi-instrument + central and eastern cluster and green corporate bonds
+ central and western cluster. Some measures are put forward, such as strength-
ening the regional cooperation capacity of green bonds among regions, delegating
power to the lower levels, and stimulating the vitality of green bond issuers.
1 Introduction
In April 2020, the International Conference on Green Finance Leadership webinar was
successfully held. The international community has more and more recognized the con-
cept of green development. Under this background, green finance emerges at the historic
moment. In order to promote the rapid development of green bonds, the government has
introduced many relevant policies. In 2017, for example, the “Guidelines on the Assess-
ment and Certification of Green Bonds (Interim)” was issued, which plays an important
role in effectively reducing the assessment risks of green bonds. The development status
and characteristics of the green bond market are analyzed, and its development trend
is forecast (Zhang J.W. 2019). By studying whether domestic green bonds are labeled,
this paper analyzes the level of “real green” bonds and puts forward suggestions and
measures such as improving the enthusiasm of “real green” bond issuers, building a
good investment environment for “real green” bonds, and increasing the number of “real
green” bond investors (Liao Y. et al. 2021). Reboredo J.C. (2018) studied co-movement
between the green bond and financial markets, finding that the green bond market cou-
pled with corporate and treasury bond markets weakly co-moves with stock and energy
commodity markets. GARCH model is used to analyze the yield rate of the green bond
market and compare the difference with the yield rate of the traditional bond market, and
the conclusion is drawn that they are negatively correlated (Xu X. and Li Y. 2018). Based
on the vector autoregression model’s forecast error variance decomposition method, Gao
Y. and Li C.Y. (2021) studied the risk spillover effect among the Chinese green bond
market and the traditional fixed income markets, stock markets, and foreign exchange
markets. Various types of financial market and the risk spillover effects from green bonds
market is stronger than that of other markets. In addition, the risk spillover between the
green bond market and the traditional fixed income market is characterized by great
uncertainty. In the comparative analysis of Chinese and foreign green bond markets, the
literature mostly studies the advantages of foreign green bond markets and the disad-
vantages of domestic green bond markets. Theoretical analysis was used to study the
differences between domestic and foreign green bond markets in bond varieties and
issuing subjects (Gao Q. et al. 2020; Shang W.X. et al. 2017; Shi R. and Chisi S. 2020).
Wang J.Z. et al. (2020) examined the market reaction to China’s green bond issuance.
They found that Enterprise participation in sustainable financing practice increases firm
value in the long run and thus is favored by shareholders. Pham L. and Huynh T.L.D.
(2020) analyzed the link between investor attention and the green bond market perfor-
mance. Their analysis revealed that appropriate information and attention for directing
financial flows towards sustainable investment is important.
The relationship between the types of green bonds and issuing places can be regarded
as a 2-mode network. This paper constructs the green bond types and issuing places
affiliation network, analyzes the characteristics of the 2-mode network between Chinese
provincial green bond types and issuing places, and reveals the relationship between
green bond types and issuing places to provide the basis for the cluster of the green
bond market. Issuing places and green bond types have the characteristics of “duality”,
and the relationship between green bond types and issuing places is regarded as a 2-
mode network. The 2-mode network can integrate two sets of issuing places and green
bond types into the same network structure and analyze the green bond market from the
relationship between the two sets and within the sets.
The relevant data in this paper are mainly from China Financial Information Network
Green Bond Database. In order to eliminate the interference to the effective analysis of
the data, 31 provinces (except Hong Kong, Macao, and Taiwan) in 2019 were selected,
and 6 provinces (including Inner Mongolia, Liaoning, Jilin, Heilongjiang, Hainan, and
Ningxia) whose issuance amount of green bonds were 0 were excluded. Five types of
Study on the 2-Mode Network Characteristics 569
green bonds and 25 places of green bonds are taken as the research objects. Therefore,
the two sets of a 2-mode network of green bond types and issuing places established in
this paper are respectively 5 sets of green bond types and 25 sets of issuing places, and
the network relationship is represented by the number of green bonds issued by issuing
places, then, the 25 × 5 2-mode matrix Xij of the issuing places and the green bond types
is obtained. 0 and 1 represent the affiliation relationship, and binarization was used to
process the relational data. First of all, a 2-mode network Xij1 based on the proportion
of green bond issuance amount of different types in each issuing place, the total green
bond issuance amount of the issuing place should be established. And then calculate
the average of all the values of Xij1 . Bigger than the average is equal to 1, and less than
the average is equal to 0. Thus, a 2-mode network Xij2 is obtained. This represents the
affiliation network between each issuing place and the type of green bonds. The 2-mode
network of Xij2 is shown in Table 1.
Province F C E D A Province F C E D A
Beijing 0 1 0 0 0 Hunan 0 1 0 0 0
Tianjin 0 1 0 1 1 Guangdong 0 0 1 0 1
Hebei 0 0 1 0 0 Guangxi 1 0 0 0 0
Shanxi 0 0 1 0 0 Chongqing 0 1 1 0 0
Shanghai 1 0 0 0 1 Sichuan 0 1 0 0 0
Jiangsu 0 1 1 0 1 Guizhou 1 1 0 0 0
Zhejiang 1 1 0 0 0 Yunnan 0 0 1 0 0
Anhui 0 1 0 1 0 Xizang 0 0 1 0 0
Fujian 1 0 0 0 0 Gansu 1 0 0 1 0
Jiangxi 1 0 0 0 0 Qinghai 1 0 0 0 0
Shandong 1 1 1 0 0 Ningxia 0 0 0 0 1
Henan 1 0 0 0 0 Xinjiang 1 0 0 1 0
Hubei 0 0 1 0 0
Note: F, C, E, D and A are abbreviations for finance bonds, corporatebonds, enterprise bond, Debt
financing instrument and asset-backedsecurity.
Closeness centrality. For the issuing place, closeness centrality is the shortest distance
between the type of green bonds it issues and other types of green bonds and the issuing
place. Its closeness centrality Cc (ni ) is expressed as follows:
g+h −1
i=1 mink d(k, j)
Cc (ni ) = 1 + (1)
g+h−1
In the expression, Cc (ni ) denotes the closeness centrality of the issuing place ni ; g
denotes the number of issuing places; h denotes the number of types of green bonds; k
denotes the type of green bonds issued by issuing place ni ; j denotes other issuing places
or types of green bonds; mink d(k,j) denotes the shortest network distance between k
and j.
For green bond types, closeness centrality is a function of the shortest network
distance from the issuing place it radiates to other issuing place and the green bond type.
Cc (mk ) of green bond type is expressed as follows:
⎡ g+h ⎤−1
mini d(i, j)
Cc (mk ) = ⎣1 + ⎦
j=1
(2)
g+h−1
In the expression, CB (ni ) denotes the betweenness centrality of the place of issue;
The issuing place nj denotes another issuing place that shares the green bond type mk
with the issuing place ni ; XijN is the number of the same green bond class shared by ni
and nj . Similarly, if a green bond type has the market of only one issuing place, the green
bond type can get (g + h + 2) “points” of betweenness; For all the provinces radiated
by a green bond type, the green bond type can obtain 1/Xkl M betweenness centrality, m is
l
another type of green bond that shares the issuing place with mk , and XijN is the number
of issuing places that mk shares with ml .
Singular value decomposition (SVD) can reduce the number of dimensions of relational
data to find out the common factors behind it and divide the structure at a certain level.
Study on the 2-Mode Network Characteristics 571
For the 2-mode network of green bond types and issuing places, this technology can
be used to find the specific market cluster of green bond types and their corresponding
issuing places to help issue places form a cooperative market community and accelerate
the development of the green economy. The calculation results of the matrix Xij1 using
social network analysis software are shown in Table 2. The contribution rates of the
two singular values in Table 2 are all over 20%, and the cumulative contribution rate is
53.1%, indicating that the index has strong explanatory power.
Table 2. The types of green bonds are affiliation network SVD analysis results
3.1 Analysis of Network Characteristics of the Types of Green Bonds and Issuing
Places
The centrality analysis of matrix Xij2 results in the following (Table 3).
According to Table 2, from the network centrality, Jiangsu, Shandong, and other
eastern regions have a high network centrality. They are at the core of the green bond
type affiliation network structure, while Qinghai and other central and western provinces
Table 3. Measurement results of green bond issuance network centrality of some provinces
Table 3. (continued)
are at the end of the network centrality. This shows that the green bonds of Jiangsu
and other provinces have developed rapidly, which has a spillover effect on the green
bond market of other provinces. The eastern region used to adopt an extensive mode of
economic development, causing serious ecological damage. Moreover, due to the high
level of financial development, it is easier to promote the issuance of green bonds.
On the contrary, the economic development in central and western regions is rela-
tively backward, and the demand for green financing is small. In addition, the level of
financial development is lower than that in the eastern region, and the basic financial
facilities are backward, and the promotion of green bonds is slower. In terms of the
types of green bonds issued, green financial bonds obtained the largest centrality index.
Green financial bond is the main body of green bond issuance. The results of closeness
centrality and betweenness centrality are also consistent.
In general, green financial bonds, green corporate bonds, green enterprise bonds and
other types of green bonds have obtained a large index of centrality. The issuance of
green bonds has established extensive affiliation relationship with the issuing places.
As the Fig. 1 shows, the correlation number of green financial bonds in the affiliation
Study on the 2-Mode Network Characteristics 573
network of green bond types and issuing places is the largest, and the indexes such as
centrality are the largest. Therefore, the structural advantage of green financial bonds is
the most obvious, occupying the core and absolute advantage in the affiliation network
of green bond types.
On the one hand, these bonds have fewer restrictions on market issuers, and national
policies support green bonds. For example, on March 6, 2019, the National Development
and Reform Commission and other relevant departments jointly issued a notice on issuing
the “Green Industry Guidance Catalogue (2019 Edition)”. The central and eastern regions
have developed economies and more green financial institutions. In particular, there are
many green bond issuers in the eastern region, which makes them more dependent on
the central and eastern provinces.
Green Enterprise Bond + Central and Western Cluster. According to Table 4 and
Table 5. The type of green bonds of this cluster is mainly green enterprise bonds. Domes-
tic provinces mainly include Hebei, Shanxi, Chongqing, Xizang, and other central and
western provinces. Green enterprise bonds are the fourth-largest type of green bonds in
China in 2019, with great growth potential. Green corporate bonds are mainly issued by
574 Y. Xiao and X. Dong
Table 5. The load value of the types of green bonds on each dimension
central government agencies, wholly state-owned enterprises, and state holding enter-
prises and are subject to strict issuance conditions. The number of non-state-owned green
bond issuers in the central and western regions is small, indirectly promoting the ratio
of green enterprise bond issuance to the total green bond issuance.
Study on the 2-Mode Network Characteristics 575
Based on a 2-mode network analysis, this paper constructs the affiliation network of
green bond types and issuing places and analyzes its network centrality and singular
value decomposition. The main conclusions are as follows:
On the whole, green bonds are characterized by a market with the central and east-
ern regions as the core and the western regions as the edge. Jiangsu and Shandong
provinces, economically developed and rich in financial resources, show strong advan-
tages in spillover effect. The green bond cooperation among the issuing places needs to
be conducted around specific provinces.
Green financial bonds and green corporate bonds are the types of green bonds with
large issuance. Their centrality is high, indicating that they play a key role in the affiliation
network of green bond types and issue places. They become factors to be considered
when establishing effective cooperative relations between different issuing places in the
green bond market.
According to the results of singular value decomposition, there are two clusters of
issuing places and green bonds with different characteristics: multi-instrument + central
and eastern cluster and green corporate bond + central and western cluster.
4.2 Measures
Increase opportunities for collaborative green bond issuance between issuing places.
Enhancing inter-provincial cooperation in green bond issuance and related activities
will be conducive to increasing the types of green bonds and promoting the develop-
ment of green finance in China. The government should build a unified green bond
platform, increase the number of intermediary services for green bonds, help backward
provinces develop green bonds, and improve the innovation capacity of green bond
financial instruments.
Focus on developing green bond types that are dominant in each issuing place.
According to the centrality and singular value decomposition results, the position of each
issuing place in the green bond market is different, so the positioning of each issuing place
should be clear. According to their respective economic development level, financial
resources, and location advantages, the green bond types with outstanding advantages
should be developed.
The government should streamline administration and delegate powers, stimulate
green bond issuers’ vitality, and promote the development of green bonds in western
provinces. Due to various factors such as location and economy, there are differences in
the number and vitality of green bond issuers between the central and the west. In the cen-
tral and eastern regions, green financial bonds and green corporate bonds are dominant.
The government should give more autonomy to the issuers of green corporate bonds and
simplify the issuance procedures of green corporate bonds to advance the development
of the green bond in the western region and promote the balanced development of the
market of China’s green bond.
576 Y. Xiao and X. Dong
References
Zhang, J.W.: The current situation, characteristics and trend of China’s green bond market. Banker
212(06), 109–111 (2019). (in Chinese)
Liao, Y., An, G.J., Shang, J., Ma, Y.: A report on the operation of China’s real green bond market
in 2020. Bonds 02, 49–55 (2021). (in Chinese)
Reboredo, J.C.: Green bond and financial markets: co-movement, diversification and price
spillover effects. Energy Econ. 74, 38–50 (2018)
Xu, X., Li, Y.: Study on yield spillover effect of green bond market. Time Finance 717(35),
170–171 (2018). (in Chinese)
Gao, Y., Li, C.Y.: The risk spillover effect between China’s green bond market and financial
market. Financ. Forum 01, 59–69 (2021). (in Chinese)
Gao, Q.X., Jiang, L.Y.: Development status of green bonds in foreign countries and its implications
for China. Environ. Sustain. Dev. 044(006), 114–119 (2019). (in Chinese)
Shan, X.W., Li, Y.H., Wang, Y.X.: Analysis on the development status of green bond market at
home and abroad. Mod. Econ. Inf. 000(013), 288–289, 292 (2017). (in Chinese)
Shi, R., Chisi, S.: International development experience of green bonds and its implications for
China. Rural Econ. Sci. Technol. 31(498), 113–114 (2020). (in Chinese)
Liu, F.J., Chen, D.D., Zhu, J.H., Qian, Z., Li, B.B.: The structure and interaction pattern of
China’s inter-provincial inbound tourism source market: based on 2-mode network analysis.
Prog. Geogr. 8(8), 932–940 (2016). (in Chinese)
Wang, J., et al.: The market reaction to green bond issuance: evidence from China. Pacific-Basin
Finance J. 60, 101294 (2020)
Pham, L., Toan, L.: How does investor attention influence the green bond market? Finance Res.
Lett. 35, 101533 (2020)
Study on the Influence of Internet Payment
on the Velocity of Money Circulation in China
1 Introduction
Internet payment is a new payment method that relies on network terminals for pay-
ment and settlement, that is, the currency payment or fund flow on the Internet. With
the emergence of Internet payment, new third-party payment products such as Alipay
and Tenpay have been developed, which have brought great convenience to people in
daily life but hit the traditional payment methods to a certain extent. According to the
statistics from iResearch, the Internet payment transactions in China totaled only 365
billion yuan in the first quarter of 2011, soaring to about 6.3 trillion yuan in the second
quarter of 2019. Fluctuations in the velocity of money circulation will affect China’s
macroeconomic policies, thus reflecting the problems in the country’s overall economic
operation. Electronic money is a form of Internet payment, and its development will
affect the supply and demand of money, thereby weakening the effectiveness of macro
policies [1]. Therefore, the stability of the velocity of money circulation has become a
focus in academic research.
2 Literature Review
Foreign scholars have conducted researches on Internet payment early, but most of them
focus on the influence of the development of electronic money on national monetary
policy. Freedmen (2000) [3] pointed out that electronic money will substitute the money
issued by the central bank, thereby affecting the monetary policy formulated by the
central bank. Sullivan (2002) [4] holds the view that both the money multiplier and
monetary base will be affected by electronic money. On the one hand, the ability of the
central bank to supply money will be hit. On the other hand, the central bank’s monetary
policy’s effectiveness will be weakened. Williamson (2003) [5] believes that electronic
money transactions will reduce the utilization rate of cash, and the central bank will
also impact the settlement of electronic money. Al-Laham et al. (2009) [6] reported that
electronic money increases the velocity of money circulation and changes the monetary
multiplier. Hiroshi et al. (2014) [7]. stated that electronic money has the advantage of
lower transaction cost compared with paper money. From a fintech perspective, Lee et al.
(2018) [8] argued that mobile payments could narrow the income gap.
Most domestic scholars have explored the influence of Internet payment on the
velocity of money circulation from two perspectives: first, from the perspective of elec-
tronic money, network money, or digital cash, Dong Xin (2001) [9] found that network
money can be a substitute of the money in circulation, hence accelerating the circulation
of money. However, Zhou Guangyou (2006) [10] believes that despite the substitution
effect of electronic money on traditional money, the velocity of money circulation has
not been inhibited from increasing by electronic money. Some scholars also pointed out
the complexity of the influence of Internet payment on the velocity of money circulation.
Yin Long (2000) [11] and Chen Yulu (2002) [12] believe that it is difficult to identify the
impact of electronic money on the velocity of money circulation in the short term, which
increases the difficulty in forecasting the velocity of money circulation. Pu Chengyi
(2002) [13] found that digital cash affects the supply of money since it can be regarded
as a substitute for the cash in circulation, thereby slowing down the velocity of money
circulation in the early stage but accelerating it in the later stage, and the velocity of
money circulation changes in a V-shaped pattern. Second, from the perspective of third-
party payment, Fang Yiqiang (2009) [14] believes that the development of third-party
payment will positively affect the velocity of money circulation, and the continuous
Study on the Influence of Internet Payment on the Velocity 579
improvement of the payment systems will accelerate money circulation. Li Nan et al.
(2014) [15]. found that third-party Internet payment impacts the capital market and fund
circulation between the real economy and will promote the velocity of fund circulation.
Li Shujin et al. (2015) [16] reported that third-party payment positively affects the veloc-
ity of money circulation, and its accelerating effect on V0 and V1 is greater than that
on V2. Fang Xing et al. (2017) [17] discovered that the use of electronic money and the
size of third-party payments would positively affect the velocity of money circulation.
The literature review found that the velocity of money circulation has been exten-
sively studied at home and abroad, but there are still controversies. Although the research
on the velocity of money circulation from third-party payments has attracted much atten-
tion, some econometric models, especially the different adjustments in the error correc-
tion models, are insufficiently reasonable, and the short-term and long-term effects on
the velocity of money circulation have not been made clear. In this paper, based on the
previous studies, the long-and short-term influence of Internet payment on the velocity
of money circulation was analyzed.
3 Theoretical Analysis
3.1 Analysis of Factors Affecting the Velocity of Money Circulation
Most foreign research on the factors influencing the velocity of money circulation is
based on Fisher’s equation of money quantity, MV = PY. Fisher pointed out that, among
the influencing factors of the velocity of money circulation, financial development and
personal psychology have a prominent impact. Baumol&Tobin believes that financial
innovation and actual income are the primary factors affecting the velocity of money
circulation. Mitchell believes that the level of economic monetization is the main factor
influencing the velocity of money circulation, which coincides with the view of Gold-
smith, who also believes that the level of financial development is the factor affecting
the velocity of money circulation. Among the domestic scholars, Hou Ying et al. (2012)
[18] variables such as GDP, monetization degree, and inflation rate to investigate the
influencing factors of the velocity of money circulation and found that these influencing
factors maintain a certain equilibrium and stable relationship with the velocity of money
circulation. In Xia Deren’s viewpoint, the degree of economic monetization is the reason
for the decline in the velocity of money circulation [19]. Ai Hongde et al. (2002) [2]
used variables such as economic monetization, the level of financial development, and
savings rate to study the velocity of money circulation and obtained the conclusion that
both the economic monetization and savings rate will accelerate money circulation, but
the level of financial development plays an inhibitory role. Zhou Guangyou (2006) [10]
concluded that among various factors affecting the velocity of money circulation, the
degree of financial digitization, monetary digitization, and cash ratio positively affect the
velocity of money circulation. Zhang Shuming et al. (2007) [20] stated that the degree
of economic monetization and the level of interest rates negatively correlate with the
velocity of money circulation. In contrast, the level of financial modernization and the
savings rate is significantly positively correlated with the velocity of money circulation.
Li Shujin et al. (2015) [16]. studied variables such as the level of economic monetiza-
tion and the degree of financial modernization, discovering that third-party payments
580 X. Liu and Q. Liu
produce different effects on the acceleration of money circulation at various levels. Liu
Da (2017) [21] included financial electronization and savings rate as control variables
affecting the velocity of money circulation. The study showed that third-party payments
have a long-term co-integration relationship with the degree of financial electronization
and savings rate. They play a joint role in the velocity of money circulation.
It can be easily found from the domestic and foreign studies on the factors influ-
encing the velocity of money circulation. These studies mostly used the variables such
as the degree of financial electronization, economic monetization, and savings rate as
the main factors affecting the velocity of money circulation. At the same time, Internet
payment, as a new payment method, can be considered a financial innovation to a certain
extent. Financial innovation will inevitably affect the velocity of money circulation in
China [22]. Therefore, Internet payment will impact the degree of financial electroniza-
tion, the degree of economic monetization, and the savings rate. The level of economic
monetization is generally measured by the ratio of the broad money quantity to the gross
national product, which has a reciprocal relationship with the dependent variable V 2
selected below [23]. Therefore, the degree of financial electronization and intention to
save were selected as the control variables in this paper.
Internet payment is essentially the transaction of electronic money between buyers and
sellers. Its low conversion cost can reduce the liquidity between the quantity of money
at various levels, thus blurring the boundaries among M 0 , M 1 , and M 2 . On the one hand,
Internet payment has accelerated the quantity of money at all levels, and on the other
hand, it functions as a substitute for traditional money. Under the combined effects of
these two aspects, Internet payment will inevitably have varying degrees of impact on
the velocity of money circulation at all levels. Particularly, the quantity of money at low-
level M 0 will be replaced by electronic money, and M 0 will be partially converted to the
quantity of money at a higher level. From this perspective, the demand for a low-level
quantity of money M 0 is reduced. According to Fisher’s equation, the velocity of money
circulation V 0 will rise. The changes in the second-level quantity of money M 1 depend
on the quantity of money converted from the lower-level to M 1 on the one hand, and also
hinge on the quantity of money partially converted to the higher-level quantity of money
M 2 . At the same time, part of M 1 is replaced by electronic money. Thus these three parts
will result in the change in the velocity of money circulation V 1 . The changes in the high-
level quantity of money M 2 depend on the part of the quantity of money converted from
M 0 and M 1 on the one hand, and on the other hand, it is partially replaced by electronic
money. Therefore, the changes in the high-level quantity of money M 2 depend on the
effect of these two aspects, and then affect the velocity of money circulation V 2 . The
mechanism of the influence on the velocity of money circulation is shown in Fig. 1.
Study on the Influence of Internet Payment on the Velocity 581
Electronic Money
Partially replaced by
Partially converted toPartially converted to
V0 V1 V2
Vi (i=0,1,2)
+ -
+
IPF FE SAVING
4 Model Construction
4.1 Selection of Variables
In this paper, the velocity of money circulation at three different levels V i (i = 0,1,2) was
selected as the explained variable, the Internet payment frequency as the explanatory
variable, and the degree of financial electronization and the intention for saving as the
control variables to establish an ECM model to study the influence of Internet payment
on the velocity of money circulation.
Mt Vt + Me Ve = PY (3)
Where t represents traditional money, and e is electronic money. The increase in IPF
will produce two kinds of effects: either accelerates or converts the velocity of money
circulation. From the above analysis, it can be seen that these two kinds of effects will
increase the velocity of money circulation. The report of iResearch reveals that third-
party Internet payment institutions have the function of realizing high-speed money
circulation in different regions. Based on this, hypothesis 1 was proposed: there is a
positive relationship between IPF and the velocity of money circulation in China.
Saving. The intention for saving refers to people’s comprehensive consideration of the
liquidity of the money they hold and the risk. When the intention for saving increases,
people hold cash with strong liquidity or the intention for current deposits decreases,
and the velocity of money circulation drops accordingly. Since the intention for saving
represents people’s subjective factors, it is difficult to be calculated and measured with
specific data, and appropriate variable from the outside can be used for measurement.
Based on this, “the proportion of residents who prefer saving” was adopted in this paper to
replace the intention for saving. The intention for saving is an individual’s psychological
factor. When people face uncertainties in their future income, they will be “forced to
save”. As a result, people will reduce their consumption behavior, which will lead to
the declined velocity of money circulation. Thus, hypothesis 3 was proposed: SAVING
negatively correlates with the velocity of money circulation in China.
5 Empirical Analysis
5.1 Stationarity Test
The data from the first quarter of 2007 to the third quarter of 2019 were selected. To
eliminate the influence of seasonal factors, the X-12 seasonal adjustment method was
adopted, and SA was used o express the adjusted variables. Meanwhile, the logarithmic
form of each variable was used to enhance the stationarity of the time series. Considering
the selected data of time series will show spurious regression due to non-stationary
characteristics, the ADF unit root stationarity test was performed on each variable.
Before the ADF unit root test, it is necessary to check whether the variable has a drift
item or a time trend item. It can be found from Fig. 2 that the drift items and trend
items should be added to LNV 0 SA, LNV 2 SA, LNIPFSA, LNFESA, LNSAVINGSA, and
drift item should be added to LNV 1 SA.
The AIC information criteria were employed to determine the lag order of each
variable. The results are provided in Table 2.
Study on the Influence of Internet Payment on the Velocity 585
It can be seen from Table 2 that the minimum values of AIC and SBIC are on the
same lag order. Thus the optimal lag order can be determined as 1. The results of ADF
unit root test are displayed in Table 3. After the first-order difference, all variables at the
1% significance level show no unit root, and all variables are stationary time series.
All variables are stationary after the first-order difference, and these variables are inte-
grated of the first-order. Thus, it can be considered that there is a long-term stable
relationship among them, which means that there is a possible co-integration relation-
ship. A test was performed to check this possibility. The main methods of co-integration
test are EG-ADF and Johnsen test method. The EG-ADF test method cannot deal with
586 X. Liu and Q. Liu
the situation where there are multiple co-integration relationships. Besides, this method
consists of two steps and will produce large errors. Therefore, the Johnsen test method
was selected in this paper for the test. The test results are shown in Table 4. When the
critical level of all variables is 5%, the trace statistic of the null hypothesis “there is no
more than one co-integration relationship” is greater than 5% critical level. Hence the
null hypothesis is rejected, and the null hypothesis that “there are no more than two co-
integration relationships” is accepted, indicating that “the velocity of money circulation
at each level has a co-integration relationship with IPF, FE, and SAVING.”
Table 4. (continued)
Note: the numbers in the brackets represent t statistics, and *, **, *** represent 10%,
5%, and 1% significance levels, respectively.
The elasticity coefficient of the first-order lag of error correction term ECM is sig-
nificant at the 1% level, indicating that IPF, FE, and SAVING all significantly affect the
velocity of money circulation at all levels. When the velocity of money circulation at each
level in the short-term deviates from the long-term equilibrium relationship, the existing
adjustments of −1.0848, −0.5664, and −1.1076 will restore this long-term equilibrium
state. From the error correction model, it can be seen that IPF is significantly positively
correlated with the velocity of money circulation at all levels at the 10% significance
level, which is consistent with Hypothesis 1. The elasticity coefficient of IPF shows that
its accelerating effect on V 0 is stronger than that on V 1 and V 2 . This is mainly because,
in the short-term, Internet payment reduces the demand for cash held by people so that
it has a greater positive, stimulating effect on V 0 than it has on V 1 and V 2 .
FE shows a positive promotion effect on V 0 and V 1 , but it has a negative inhibitory
effect on V 2 , not consistent with the Hypothesis 2. There may be the following rea-
sons. With the increasingly strengthened financial electronization in China, residents
or companies have a positive expectation on the financial products with preventive and
speculative motives covered by money supply M 2 , such as financial bonds, commercial
paper, and transferable certificates of deposit, so they tend to accept these financial prod-
ucts. From this perspective, the holding of these financial products is relatively stable,
and it may lead to a decrease in the velocity of money circulation V 2 .
The estimation results of the model reveal that the residents’ intention for saving has
a significant negative correlation with the velocity of money circulation at all levels at
the 5% level, which is consistent with Hypothesis 3. Since the error correction model
is not applicable to the analysis of the changes in each variable, the impulse response
function was used to analyze the dynamic fluctuations in each variable.
It can be seen from Fig. 3 that in the impulse analysis with lag periods of 20, the impact of
V 0 on itself in the current period is 11.2%, and then drops rapidly to −1.8% in the second
period, after which the value fluctuates and remains at 1%. Regarding other variables
Study on the Influence of Internet Payment on the Velocity 589
.12
.10
.08
.06
.04
.02
.00
-.02
-.04
-.06
2 4 6 8 10 12 14 16 18 20
LNV0SA LNIPFSA
LNFESA LNSAVINGSA
influencing V 0 , the impact of IPF, FE, and SAVING on V 0 in the current period is zero.
The impact of IPF on V 0 in the second period is −3.7%, and reaches the peak of 1.9%
in the fourth period, then gradually fluctuates until the 18th period, stabilizing at the
0.6% level; the impact of FE reaches the bottom of −2.5% in the second period, then
fluctuates to a peak of 1.2% in the third period, and finally remains at −0.4% in the 17th
period; the impact SAVING in the fourth period reaches the bottom of −4.1%, which
fluctuates until the 16th period when it finally stabilizes at −0.18%. Specifically, the
impact of IPF on V 0 increased from −3.7% in the first quarter of 2008 to 1.9% in the
fourth quarter of 2009. The impact of FE reached the minimum value of −2.5% in the
first quarter of 2008 and the maximum value of the impact was 1.2% in the first quarter
of 2009.The minimum value of SAVING, which gradually showed its impact after the
second quarter of 2009, was −4.1%.
.10
.08
.06
.04
.02
.00
-.02
2 4 6 8 10 12 14 16 18 20
LNV1SA LNIPFSA
LNFESA LNSAVINGSA
In the impulse analysis of V 1 , lag 20 was also sent for analysis. It can be seen from
Fig. 4 that the impact of V 1 on itself in the current period is 9.1%. Then it drops to the
bottom of 3.6% in the second period and remains fluctuating until the 12th period when it
finally stabilizes at 4.2%. Similarly, the influence of other variables on V 1 was analyzed,
590 X. Liu and Q. Liu
and it was found that the impact of these factors on V 1 in the current period is 0. The
impact of IPF on V 1 reaches a maximum of 0.7% in the second period, after which it
fluctuates until the 11th period, maintaining a 0.1% level; FE has a 2.2% impact on V 1
in the third period, and this Fig. drops to the minimum of 0.2% in the fourth period,
finally stabilizing at 0.8% in the 13th period; the impact SAVING in the third period has
reached the maximum of 2.3%, after which the value reduces until the 15th period when
it finally remains at 1%. Specifically, the impact of IPF on V 1 reached a peak of 0.7%
in the first quarter of 2008. From the fourth quarter of 2008 to the third quarter of 2009,
the impact of FE decreased from 2.2% to 0.2%. SAVING declined from a peak of 2.3%
in the fourth quarter of 2008 until stabilizing at 1% around the fourth quarter of 2017.
.10
.08
.06
.04
.02
.00
-.02
-.04
5 10 15 20 25 30 35 40 45 50
LNV2SA LNIPFSA
LNFESA LNSAVINGSA
In order to better show the dynamic changes of the impulse response of each variable,
in the impulse response analysis of V 2 , lag 50 was selected. From Fig. 5, it can be
found that the impact of V 2 on itself is 0.9% in the current period, after which it has
been fluctuating until it gradually stabilizes at 3.3% in the 46th period. Regarding other
variables affecting V 2 , the impact of IPF, FE, and SAVING in the current period is 0.
The impact of IPF reaches a peak of 1.9% in the fourth period, then remains fluctuating
and gradually decreases until it finally stabilizes at 1.3% in the 38th period; the impact
of FE drops to the bottom of −3.2% in the second period, and then remains fluctuating
until the 44th period when it finally maintains the level of −1.7%; the impact SAVING
in the third period reaches a maximum of 1.7%, which falls to −0.8% in the 4th period,
and reaches a steady state in the 40th period, remaining at the level of 0.4%. Specifically,
the impact of IPF on V 2 peaked at 1.9% in the fourth quarter of 2007; The impact of
FE in the second quarter of 2007 was the least −1.7%; The impact of SAVING peaked
at 1.7% in the third quarter of 2007.
for saving (SAVING) were thoroughly studied and analyzed through the co-integration
test, ECM model and impulse response analysis. On the whole, Internet payment fre-
quency and financial electronization have a positive correlation with the velocity of
money circulation in China, while the intention for saving shows a negative and inhibitory
effect on the velocity of money circulation in China. In the long term, the velocity of
money circulation at all levels in China maintains a stable relationship with the Internet
payment frequency, financial electronization and the intention for saving. By establish-
ing an ECM model and impulse response analysis, it can be concluded that the Internet
payment frequency, financial electronization and the intention for saving will produce
varying degrees of influence on the velocity of money circulation in China in the short
term. With the expansion of the scale of Internet payment, the early simple attribution
of Mackinnon’s Chinese financial conundrum to the decline in the velocity of money
circulation is no longer applicable. There are complex factors behind the velocity of
money circulation.
In the long run, there is a stable relationship between the activity of Internet payment
and the velocity of money circulation at all levels in China. In the short run, Internet
payment indeed will promote V 0 , V 1 and V 2 . Therefore, it can be said that Internet
payment has accelerated money circulation in China. However, the accelerating effect
of IPF on V 0 is stronger than that on V 1 and V 2 , which can be ascribed to two aspects.
First, the scale of Internet payment transactions is constantly expanding; second, Internet
payment leads to the conversion of low-level money to high-level money, which gradually
blurs the boundary between the quantity of money at various levels. When the velocity
of money circulation at all levels in the short-term deviates from the equilibrium and
stable relations with various factors, greater efforts are required to bring the velocity of
money circulation at all levels back to this equilibrium state, which means the increase
in the instability of the velocity of money circulation, thereby affecting the formulation
of monetary policies. The following suggestions are proposed:
First, the amount of money at all levels can be subdivided. The rapid development of
Internet payment and the increasing scale of transactions will affect the division of the
amount of money at various levels, thereby influencing the effectiveness of monetary
policies, and will bring certain challenges to the supervision of Internet payment. With
the expansion of the scale of Internet payment and the strengthening of financial elec-
tronization, the monetary authority should further optimize the division of the amount of
money at all levels, consider Internet payment, and use it as a reference for formulating
policies.
Second, it is necessary to understand the impact of the velocity of money circulation
on monetary policies. Since the velocity of money circulation has an important influence
on the effectiveness of monetary policies, the supply of money should be considered.
Various factors affecting the velocity of money circulation need to be considered in the
formulation of monetary policies. In addition, attention should also be paid to the impact
of changes in the velocity of money circulation on the economy. When the economic
operation is under pressure, such as inflation, the central bank needs to respond to the
velocity of money circulation and make corresponding expectations on the changes in
the velocity of money circulation to accurately regulate the money supply and mitigate
economic pressure such as inflation.
592 X. Liu and Q. Liu
Third, laws and regulations on Internet payment should be improved. China’s central
bank needs to fulfill its supervisory duties to ensure the proper development of Internet
payment. It cannot excessively suppress the development of Internet payment, nor allow
the uncontrollable development of Internet payment. Characterized by convenient use
and low transaction costs, Internet payment itself plays a certain role in promoting
the sound development of China’s economy. Therefore, China’s central bank should
encourage, support, and correctly guide the development of Internet payment in China
and take preventive measures to develop relevant policies to solve the problems in the
development of Internet payment.
References
1. Xie, P., Liu, H.: ICT, mobile payments and electronic money. J. Financ. Res. 10, 1–14 (2013)
2. Hongde, A., Nan, F.: An empirical analysis of monetary velocity in China. J. World Economy.
(08), 53–59 (2002)
3. Freedmen, C.: Monetary policy implementation: past, present and future will electronic money
lead to the eventual demise of central banking? Int. Financ. 3(2), 221–227 (2000)
4. Sullivan, S.M.: Electronic money and its impact on central banking and monetary policy.
Hamilton College Working Paper (02/1) (2002)
5. Williamson, S.D.: Payment systems and monetary policy. J. Monetary Econ. 2, 475–495
(2003)
6. Al-Laham, M., Al-Tarwneh, H., Abdallatn.: Development of electronic money and its impact
on the central bank role and monetary policy. Issues Inf. Sci. Inf. Technol. (6), 339–349 (2009)
7. Fujiki, H., Tanaka, M.: Currency demand, new technology, and the adoption of electronic
money: micro evidence from Japan. Econ. Lett. (1), 5–8 (2014)
8. Lee, H., Han, K.: A study on mobile easy payment service based on fintech to reduce smart
divide and income gap. Int. J. Adv. Sci. Technol. 116, 35–48 (2018)
9. Dong, X., Z, H.: Challenge of networks money to the central bank. Econ. Theory Bus. Manage.
(07), 21–25 (2001)
10. Zhou, G.Y.: The impact on money circulation velocity of electronic money: an empirical
study with cointegration method. China Econ. Q. 03, 1219–1234 (2006)
11. Yin, L.: The effect of e-money on the central bank. J. Financ. Res. 04, 34–41 (2000)
12. Chen, Y., B, W.: The development of e-currency and the analysis on the risk faced by the
central bank. Stud. Int. Financ. (01), 53–58 (2002)
13. Pu, C.: The influence of digital cash on currency supply and currency speed. J. Financ. Res.
05, 81–89 (2002)
14. Fang, Y.Q.: The development of payment system and the change of velocity of currency
circulation. Shanghai Financ. 09, 27–31 (2009)
15. Li, N., H, X., and X, E.: The impact of payment system innovation on chinese monetary
system. Finan. Forum. 19(11), 29–34 (2014)
16. Li, S., Zhang, X.: The influence of the third-party on the velocity of chinese monetary
circulation. Finan. Forum. 20(12), 25–33 (2015)
17. Fang, X., Guo, Z.R.: Third-party internet payment, currency circulation speed and monetary
policy effectiveness-research based on TVP-VAR model. Inquiry Econ. Issues. 03, 183–190
(2017)
18. Hou, Y., Chen, J.N.: Research on correction of velocity of money and influence factors: based
on empirical analysis of VEC. Financ. Theory Pract. 02, 107–111 (2012)
19. Xia, D.R.: Changes of monetary velocity of economic monetization process. J. Central Univ.
Finan. Econ. 04, 24–28 (1991)
Study on the Influence of Internet Payment on the Velocity 593
20. Zhang, S.M., Du, W.L.: A quantitative study on the velocity of money in China. J. Shandong
Univ. (Philosophy and Social Sciences) 02, 93–98 (2007)
21. Liu, D.: Does the “internet payment” speed up the monetary circulation velocity? J. Central
Univ. Financ. Econ. 02, 32–42 (2017)
22. Liang, D.P., Qi, Z.Y.: Research on the impact of financial innovation on currency circulation
speed. Econ. Sci. 02, 27–34 (2004)
23. Liu, M.Z.: China’s M2 /GDP (1980–2000): trend. Level and Determinants. Econ. Res. J. 02,
3–12 (2001)
System Ordering Process Based on Uni-, Bi-
and Multidirectionality – Theory and First
Examples
1 Introduction
A theory related to whole-part relations is the holon theory by Arthur Koestler [1], where
he first defined the word holon, which stands for the hole and the part in one word. In the
later development, especially with production and computational research, the holon was
also related to the then-emerging decentral production. Some approaches like the holon
theory and similar ones like bionic and fractal manufacturing are related primarily to
flexible production’s decentralizing problem [2]. The theory of self-organization on the
other side is related to nonlinear phenomena, especially the chaos theoretic approaches,
e.g., by Prigogine [3], Lorenz, van der Pool, the theory of laser, and the synergetics
of Herman Haken [4, 5] and interdisciplinary approaches [6]). This theory can com-
bine natural sciences and social sciences through increasingly nested relations. On the
other hand, chaos theory deals with deterministic approaches in a rather mathematical
form, and this may be a first step in understanding nature more deeply. The interdis-
ciplinary methods inherited from chaos theory, self-organizational, and systems theory
also use Einstein’s idea [7] to relate different applications or disciplines with the same
mathematical expressions.
An example is logic calculus - beginning, e.g., with predicate logic of first-order [8]
- allowing us to combine very different areas of knowledge in the language. The origin
or orgiton theory framework is in the development process [14, 15], and consists of
origins, orgitons or cybernetical units of mass, energy, and information. The hierarchic
ordering there is quite similar to the holon theory [1]. However, the focus is more on
lively functional elements, although, in the generalization, it is universally applicable
and can be best understood by a dynamic graph theory approach, where knots and edges
are interchangeable and can be regarded by this as phase changes.
This paper aims to argue for the plausibility of uni-, bi- and multidirectional order
increase - as a proposition of orgiton theory - as a systemic property that can be motivated
by the no-intersection theorem in chaos-theory. We then further want to answer how to
use this in the epistemology of science as a sustainability framework and in logistics and
production modeling. The method used is a system theoretic argumentation introducing
and using basic axioms and principles.
The paper begins in Sect. 2 with the definition of the higher-order gaining principles.
In Sect. 3, we then give an application concerning an epistemology of science. Next, in
Sect. 4, we provide the application in production and logistics, with an orgiton model
in principle, Witness modeling, and some first considerations towards a further general-
ization. Finally, in Sect. 5, we conclude and give a short outlook on the extension of this
approach towards more connected graphs.
The no-intersection theorem in chaos theory states, according to [5; p. 76]: “Two distinct
state-space trajectories cannot intersect (in a finite period of time). Nor can a single
trajectory cross itself at a later time.”
This theorem is especially closely connected to differential equation systems of a
certain order and the Lipschitz condition, to be continuous and differentiable to a certain
degree. This restricts the application to differentiable functions and the assumption of
being continuous, which means related or having a “connection”.
Relating now the state space of the Lorenz equations according to Fig. 1, we can
draw in the lower two-dimensional diagrams a two-directional mapping, which indicates
a snapshot of bidirectionality. In this case of a bi-bi directionality as the Lorenz-attractor
has two fix-points or nodes. The seeming intersection in the lower dimensions expresses
the coupling, which is invariant in all the views but changes with perspective, as the
information is incomplete in the lower dimensions. Due to the no-intersection theo-
rem, we gain the complete and integrated information only by integration, hence the
unidirectional trajectory in the state-space.
Still, in the meta-view, there occurs a unifying process towards what we understand
as one system. The important fact is that we see by this mechanism that a system can
be coupled along more than one dimension, meaning interdependence, bidirectionally.
Altogether, this connectivity can be observed on lower dimensions with different per-
spectives as the synchronicity of a specific order. The views may then change the relative
directionality but not the orgitonal order of unidirectionality (), bidirectionality (),
or multidirectionality (), meaning the possible multiplicity with at least two ele-
ments. More precisely, in addition to the change of direction, also at least one repetition
in the same direction. The further increase in production efficiency can then be under-
stood in general as the osmotic principle [9–11], which denotes a multidirectional path
596 B. Heiden et al.
in, e.g. logistics and production. In biology or the biotechnological, it leads in the long
term to a production maximisation in diverse, or as we would say multidirectional, eco-
logical niches [12]. Here, the complementary principle of nature also applies, which has
recently been rediscovered by Grossberg [13] in the neuro-computational modeling of
our brain.
Fig. 1. Lorenz attractor - solutions for r = 25, p = 10, b = 8/3, x2(0) = 0, y2(0) = 1, z2(0) =
20, x2, y2, z2… State-space coordinates, t1… Time variable (0..15) x2 ˙ = p*(y2-x2); y2 ˙ = r*x2-
y2-z2*x2; z2=˙ x*y-b*z2; Lorenz equation according to [5]. The arrows depicted here denote the
directional state of a state-trajectory or its at-the-moment dynamic movement. The above-shown
simulation is done with Mathcad 15, analogously to [5]. Additionally, to demonstrate the meaning
of directionality in state-space trajectories, there are here depicted arrows () that indicate the
directionality about time, as each, with one arrow connected, point pair in the picture corresponds
to a time sequence in points. Hence each arrow represents onedirectionality () in time as well as
in the state-space-room. We can also see that a higher dimensional direction may appear different in
the lower dimensions by simultaneously conserving the bidirectional coupling or bidirectionality
(). The Mathcad 15 is provided in [26].
In orgiton theory [14, 15], which is a self-organization theory, that is based on cyber-
netic units, which can be understood as a cyclic interconnection between mass, energy,
and information as potential variables, there is the notion of uni- and bi-directionality,
which indicate the potential increase in order by increasing orgitonal evolution, or
indices. Relating this to the production and logistics area with Axiom 1, the following
can be stated:
With relation to the Lorenz equations, which may be valid more widely in nonlinear
applications, Axiom 2 can be formulated:
Its coupling (of the Lorenz equation) shows in the “synchronicity” or “state Axiom 2
space parallelity” of the autocorrelation function.
Figure 1 shows that the two points depicted there correlate with each other after
specific time intervals, as they repeat their trajectory approximately. The same is anal-
ogously valid for the parallelity in state-space. When looking at the same points from
different perspectives as done in Fig. 1, we see that those points’ properties may change
their movement concerning other observation parameters. However, they stay coupled
to one “single line”, which in effect is the meaning of the no-intersection theorem. In
the higher dimension, in this case, three-dimensional, the order gets “unidirectional”,
although, in two dimensions, it brings different shapes of directions. The coupling is
then indicated in the higher-order unification.
This means that the bifurcation of directions, and their repetitive approximative
behavior in lower dimensions, indicates order, self-order, or autocorrelation in time and
above the state space.
An interesting note may be given on the autocorrelation function, which can be
regarded as bi-directionality, related to a self-cycle, meaning that the trajectory is return-
ing approximately to itself (cf. [27, 28] where this is related to order increase). In cyber-
netics, this is also called a feedback cycle. Hence we see, that a feedback cycle is a
bidirectional entity and can also be regarded as a holon or orgiton. Rudolf E. Kálmán,
in his famous paper [16], has related this to his important terms of observability (one-
directional) and controllability (bi-directional). A similar notion is given in management
theory by Otto Scharmer in theory U, where he denotes a cycle of a creative act using
first observing and then prototyping, which leads to effective, innovative products [17,
18]. These products then mean a higher-order step.
So simultaneously, at the same time, two variables are coupled with their direction
or opposite direction when there occurs a cycle. The overall cyclic behavior leads then
to the abstraction of the two-directional ().
As seen in the Lorenz attractor, different states can simultaneously indicate even
higher order, each achieved by bifurcation. Thus, there are two node attractors in this
case, leading to two coupled cyclic behaviors in one meta-coupled trajectory.
This can be understood as a bifurcated bidirectional trajectory or a bi-bi directional
one.
Due to Axiom 3:
In general hence
The coupling links the “two directions”, employing the higher order, the sum-
function, integral, etc. The higher-order gives the system more momentum (see also
Sect. 3) and hence increases its Echtzeit (real-time), or range of Raumzeit (room-
time) validity. This system’s growth in the room-time range is also evidence of the
solidification-fluidization theorem [19], which links the growth of an open system with
a larger scaling range of controllability.
3 Bidirectionality in Science
According to Fig. 2, the knowledge-gaining process is divided by the time arrow. We can
imagine dual parameters A and A’ in the imaginary dimension, like ideas, innovations,
applications, and natural science or basic research. A and A’ are complementary, anal-
ogously to the imaginary axes in the complex Gaussian numbers plane. The parameters
and cones are connected in the present - or presencing - point of the continuous-time
arrow. So this duality of antipodes is a matter of coupled bidirectionality. Figure 2 can
also be denoted as a Sanduhr (sandglass), indicating real-time processes with different
cone angles for many influencing parameters.
System Ordering Process Based on Uni-, Bi- and Multidirectionality 599
This part describes the elementary production or logistic scalable orgiton model for
studying their energy usage, simulated in the simulation software Witness. The basic
model consists of three nodes (Fig. 3a). The advanced model consists of four and virtual
three nodes (Fig. 3b) and four nodes (Fig. 3c). The idea is to provide a model that
describes the central and decentral dichotomy or problem as a holon or an orgiton. For
this, it has the properties of uni- (), bi- () and multidirectional (), especially
when nested of lower-order forms (cf. also Fig. 3a-c).
The orgiton model, modelled in the production and simulation software Witness, is based
on the primary facility location problem (cf. e.g. [24]) and consists of (1) producer, (2)
supplier, and (3) customer in three possible arrangement combinations (see Fig. 3).
In Fig. 3, centralised production facilities (“Producer”) are shown. The raw materials
are delivered to the production from the supplier (“Supplier”). A production process runs
in one single location, and finished products are delivered afterwards to the customer
(“Customer”). The above-mentioned “transportation processes” are performed using
trucks (“Truck1”, “Truck2”) through the routes (“R1”, “R2”). Since the return shipments
are not taken into consideration for this special case, it yields us to consider the relations
“Supplier”-“Producer”, and “Producer”-“Customer” as unidirectional. Anyhow the back
travelling is included by the factor 2 in the “PC” equation.
System Ordering Process Based on Uni-, Bi- and Multidirectionality 601
Fig. 3. Elementary scalable production or logistic orgiton model in Witness: a) with centralised
production facilities (); b) with decentralised coupled production facilities (); c) with decen-
tralised coupled and “meta”-uncoupled production facilities (). The Witness 14 files are
provided in [26].
In Fig. 3b, decentralised coupled production facilities are shown. In this example, the
raw materials are delivered from the supplier (“Supplier”) to the first production facil-
ity (“Producer1”), where the first production phase is performed (e.g. pre-assembling).
Afterwards, the semi-finished products are transported to the second production facility
(“Producer2”), where the final product is performed. Afterwards, they are transported
to the customer (“Customer”) similar to the in Fig. 3a shown example. The transport
processes are performed with three trucks (“Truck1”, “Truck2”, “Truck3”). In this exam-
ple, routes “R1”, “R3” can, similar to the in Fig. 3a shown example, be considered as
unidirectional. However, we consider a route “R2” as bidirectional since both facilities
are considered as one company, and hence, are coupled. It refers to Axiom 4, described
above. According to it, bidirectionality is both coupling and reduction since in cou-
pled facilities, the interaction level grows (cf. Principle 1), leading to more cooperation.
Practically this means, e.g. the dynamic allocation of production volumes to reduce peak
utilisation.
In Fig. 3c, decentralised coupled and “meta”-uncoupled production facilities ()
are shown. The two production facilities are decentral and cooperative (), which we
also call conficient. The meta-relation from customer to producer is meta-uncoupled
602 B. Heiden et al.
() once triggered, and there is a meta-uncoupling () inside the company as central
control. Here, the raw materials are delivered with a truck (“Truck1”) from the supplier
(“Supplier”) to the two production facilities (“Producer1, Producer2”), which are inde-
pendent of each other, where the whole production cycle is performed. Afterwards, the
final products are dispatched to the customer (“Customer”), using trucks “Truck2” and
“Truck3”. This example can be considered as a “scaled model” of the in Fig. 3a, shown
example, where all routes are unidirectional. In general, in the case of an uncoupled ()
and united () model, the order of the whole system increases, according to Axiom 1,
with the increasing number of routes (edges), which leads to increasing complexity and
hence a potential higher order orgiton or holon, as it is more diverse.
4.3 Results
The first results of the considerations, especially concerning Fig. 3, are that the order
according to subfigure or case “a, b, c” is that the energy consumption is increasing
linearly due to the chosen scale of comparable length and the same production-logistics
sequence according to the summarising sentence that “a(A) product is delivered to the
customer via supply and production”, which comprises 3, 4(3), 4 knots from “a–c”.
Hence the complexity is in “b, c” increased compared to “a”, but “c” can be regarded
as a unidirectional and bidirectional case versus “b”, a bidirectional one, as it is unity
on a higher level. There are two sorts of bifurcations: One towards cooperation and
bidirectional communication (), the other towards non-cooperation and unidirectional
() communication. In their nesting, they get multidirectional ().
5 Conclusion
First, in Sect. 2, we have developed four axioms and two principles related to the princi-
ple of uni-, bi- and multidirectional, with recourse to the chaos-theoretic no-intersection
theorem. According to Sect. 3, from epistemic knowledge gaining to social mass related
processes, increasingly fluidization occurs through coupling and an overall inertialisa-
tion, by gaining knowledge momentum, leading to an increasingly higher order, of a
system that is increasingly real-time connected, multidirectional and conficient. In [25],
we defined conficient as the property of being cooperative and efficient. According to
Sect. 4, it seems that, as a first result, cooperation is leading to more communication pro-
cesses that are subdividing the internal process of the manufacturer, and hence, allowing
for a flexible adaption according to the geometry. This has the advantage that the order
increases, and with it the possibilities room. Compared to the higher level no-cooperation
case “c”, case “b” is relatively less conficient or less efficient and cooperative. Com-
pared to the less complex case “a,” it is – as a possibility - most efficient, understanding
here efficiency as the ability to cope with complexity. That means that there is a range
of optimality of cooperation and its “dual” concerning its meaning in condensed form:
conficient or cooperative and efficient systems.
Overall for both investigated cases, as for the general case, the no-intersection the-
orem predicts unification of higher-order systems. That means that the unification is
hence of less order, as it is unidirectional, but it has the potential to increase in order,
System Ordering Process Based on Uni-, Bi- and Multidirectionality 603
References
1. Koestler, A.: The Ghost in the Machine. Anchor Press Ltd. (1967)
2. Tharumarajah, A., Wells, A.J., Nemes, L.: Comparison of the bionic, fractal and holonic
manufacturing system concepts. Int. J. Comput. Integr. Manuf. 9, 217–226 (1996)
3. Glansdorff, P., Prigogine, I.: Thermodynamic Theory of Structure, Stability and Fluctuations.
Wiley Interscience a division of John Wiley & Sons Ltd, Brüssel (1971)
4. Dürr, H.P., et al.: Selbstorganisation. Die Entstehung von Ordnung in Natur und Gesellschaft.
Piper Verlag, München (1986)
5. Hilborn, R.C.: Chaos and Nonlinear Dynamics - An Introduction for Scientists and Engineers.
Oxford University Press, New York (1994)
6. Götschl, J.: Self-organization: new foundations for a more uniform understanding of reality
(Original in German: ‘Selbstorganisation: Neue Grundlagen zu einem einheitlicheren Real-
itätsverständnis’). In: Vec, M., Hütt, M.T., Freund, A. (eds.) Self-organization - A system of
thought for nature and society (Original in German: ‘Selbstorganisation - Ein Denksystem
für Natur und Gesellschaft’), pp. 35–65. Böhlau Verlag, Köln (2006)
7. Einstein, A., von Smoluchowski, M.: Untersuchungen über die Theorie der Brownschen
Bewegung - Abhandlung über die Brownsche Bewegung und verwandte Erscheinungen.
Verlag Harri Deutsch, Frankfurt am Main (2001)
8. Mates, B.: Elementare Logik - Prädikatenlogik der ersten Stufe. Vandenhoeck und Ruprecht
in Göttingen (1978)
9. Heiden, B., Tonino-Heiden, B., Alieksieiev, V.: Artificial Life - Investigations about a Uni-
versal Osmotic Paradigm (UOP). In: Arai, K. (ed.) Intelligent Computing, LNNS, vol. 285,
pp. 595–605. Springer Nature (2021). https://doi.org/10.1007/978-3-030-80129-8_42
10. Villari, M., Fazio, M., Dustdar, S., Rana, O., Ranjan, R.: Osmotic computing: a new paradigm
for edge/cloud integration. IEEE Cloud Comput. 3, 76-83–76-83 (2016)
604 B. Heiden et al.
11. Heiden, B., Volk, M., Alieksieiev, V., Tonino-Heiden, B.: Framing Artificial Intelligence (AI)
Additive Manufacturing (AM). Procedia Computer Science, vol. 186, pp. 387–394. Elsevier
B.V. (2020). https://doi.org/10.1016/j.procs.2021.04.161
12. Moser, A.: Bioprozeßtechnik. Springer Vienna (1981)
13. Grossberg, S.: Conscious Mind. How Each Brain Makes a Mind. OXFORD UNIV PR,
Resonant Brain (2021)
14. Heiden, B., Tonino-Heiden, B., Wissounig, W., Nicolay, P., Roth, M., Walder, S., Mingxing,
X., Maat, W.: Orgiton Theory (unpublished) (2019)
15. Heiden, B., Tonino-Heiden, B.: Philosophical Studies - Special Orgiton The-
ory/Philosophische Untersuchungen - Spezielle Orgitontheorie (English and German Edition)
(unpublished) (2021)
16. Kálmán, R.E.: Contributions to the Theory of Optimal Control. Boletin de la Sociedad
Matematica Mexicana 5, 102–119 (1960)
17. Scharmer, O., Käufer, K.: Leading from the Emerging Future - From Ego-System To Eco-
System Economies - Applying Theory U to Transforming Business, Society, and Self. Berrett-
Koehler Publishers Inc., San Francisco (2013)
18. Heiden, B.: Wirtschaftliche Industrie 4.0 Entscheidungen - mit Beispielen - Praxis der
Wertschöpfung. AV Akademiker Verlag (2016)
19. Heiden, B., Tonino-Heiden, B.: Emergence and Solidification-Fluidisation. In: Arai, K. (ed.)
LNNS, vol. 296, pp. 845–855. Springer Nature Switzerland AG (2022). https://doi.org/10.
1007/978-3-030-82199-9_57
20. Götschl, J.: Zur Epistemologie der Selbstorganisation: Von Konvergenzen zu Korrelationen
zwischen Systemwissenschaften der Natur und Systemwissenschaften vom Menschen. In:
Fabisch, H., Fabisch, K., Kapfhammer, H.-P. (eds.) Die Welten von Psyche und Soma - Zur
Verbindung von Psychoanalyse und Neuropsychiatrie - Gedenksymposion für Frau Univer-
sitätsprofessor Dr.in med. Dr.-in phil. Margarete Minauf im Meerscheinschlössl in Graz am
25. November 2017, pp. 85–104. Verlag Dr. Kovac (2020)
21. Tarski, A.: The Semantic Conception of Truth: and the Foundations of Semantics. Res. 4,
341–376 (1944)
22. Quine, W.v.O.: Wort und Gegenstand (Word and Object). Reclam (1993)
23. Luhmann, N.: Soziale Systeme. Suhrkamp Verlag AG (2018)
24. Farahani, R.Z., Abedian, M., Sharahi, S.: Dynamic Facility Location Problem. In: Zanjirani
Farahani, R., Hekmatfar, M. (eds.) Facility Location. Contributions to Management Science,
pp. 347–372. Physica, Heidelberg (2009). https://doi.org/10.1007/978-3-7908-2151-2_15
25. Heiden, B., Walder, S., Winterling, J., Perez, V., Alieksieiev, V., Tonino-Heiden, B.: Uni-
versal Language Artificial Intelligence (ULAI). In: Schulz, F. (ed.) Advances in Artificial
Intelligence Research. Nova Science Publishers, Incorporated (2020)
26. Heiden, B.: Support Material for the Paper ‘System Ordering Process Based on Uni-, Bi- and
Multidirectionality – Theory and First Examples’ (2022, this paper). https://github.com/Ber
nhardHeiden/Directionality-Principle. Accessed 3 Dec 2021
27. Heiden, B., Tonino-Heiden, B.: Key to artificial intelligence (AI). In: Arai, K., Kapoor, S.,
Bhatia, R. (eds.) IntelliSys 2020. AISC, vol. 1252, pp. 647–656. Springer, Cham (2021).
https://doi.org/10.1007/978-3-030-55190-2_49
28. Heiden, B., Leitner, U.: Additive manufacturing – a system theoretic approach. In: Drstvenšek,
I. (ed.) ICAT 2018, Maribor, pp. 136–139. Interesansa - zavod, Ljubljana (2018)
The Impact of Intellectual Property Protection
on China’s Import of Computer and Information
Service Trade: Empirical Research Based
on Panel Data of 34 Countries or Regions
Abstract. The 14th Five-Year Plan indicates the need to improve the scientific
and technological innovation system and mechanism based on a sound intellec-
tual property protection and application system. Based on the panel data of 34
countries or regions from 2007 to 2019, the article uses the extended trade gravity
model to empirically study the impact of China’s intellectual property protection
on the import of computer and information services trade. The results show that
the improvement of China’s intellectual property protection level has effectively
restrained the imitation of domestic member-consumers and has a significant pro-
motion effect on the growth of computer and information service trade imports.
The market expansion effect of intellectual property protection is greater than the
market power effect.
1 Introduction
The “14th Five-Year Plan” clearly pointed out the need to improve the intellectual prop-
erty protection and application system, better protect and encourage high-value patents,
and cultivate patent-intensive industries. Under the “14th Five-Year Plan” background,
the importance of intellectual property protection in China is still self-evident. The intel-
lectual property rights involved in the “TRIPS” agreement include industrial designs,
patents, and integrated circuit layout designs. This kind of power is the right to self-
apply, grant, and be protected for high-tech products. As the share of high-tech products
and services in international trade continues to increase, it becomes more and more
important to identify the relationship between intellectual property protection and these
non-traditional trades. The trade-in computer and information services is recognized by
the “EBOPS (Expanded Balance of Payments Service Classification)” as a type of trade
Fund Project: This article is a phased research result of the National Social Science Fund Project
“China-Eurasian Economic Union FTA Creation Research from the Perspective of the ’Silk Road
Economic Belt’” (project number: 18JL094).
in services. With the continuous improvement of residents’ wages and the continuous
expansion of middle-income groups, the demand for modernized high-tech services is
also increasing. From 2007 to 2019, China’s computer and information service trade
imports increased from 2.208 billion dollars to 26.86 billion dollars each year. At the
same time, China’s computer and information service trade imports accounted for the
proportion of total service trade imports from 1.7% to 5.36% year by year. However,
under the epidemic’s impact, China’s service trade deficit will decrease in 2020, including
a decrease in imports of computer and information services. However, the great domestic
demand for such modern high-tech services is more in line with the development trend
of the world’s service trade structure.
China’s intellectual property protection system has been strengthened and improved
year by year. Because of its technology-intensive product trade characteristics, the com-
puter and information service trade is destined to have an inseparable relationship with
intellectual property protection. In 2020, under the background that China’s computer
and information service trade imports will decrease in 2020, resulting in a larger gap
in this supply service, can we continue to strengthen China’s intellectual property pro-
tection to promote the import of computer and information services, to make up for the
impact of the new crown epidemic on imports? This is worthy of our in-depth study.
However, the similar growth trends of the two do not indicate that intellectual prop-
erty protection has a significant role in promoting the import of computer and infor-
mation services. First of all, as a major imitating country, service providers provide
software processing, system integration, repair, and maintenance services across bor-
ders. This process is easy to imitate, leading to patent and even integrated circuit layout
design infringements, although intellectual property rights are protected. The system
is gradually being improved, but whether the protection of intellectual property rights
can eliminate the doubt that exporting countries will face the threat of imitation is still
unknown. Second, after the protection of intellectual property rights reaches a certain
level, will it increase the effect of market expansion and expand the scale of imports? It
remains to be verified whether the exporting country has created a market power effect by
raising services to suppress imports. Based on the above considerations, this article aims
to provide suggestions for improving China’s intellectual property protection system
and optimizing the import structure of service trade by studying the impact of China’s
intellectual property protection on the import of computer and information services.
2 Empirical Research
In 2019, OECD countries and APEC countries or regions (excluding countries that over-
lap with OECD organizations) accounted for 88.19% of the world’s exports to China in
this industry. They have been stable since 2015. Around this data, it can be seen that
these developed countries have a large proportion of China’s exports to China in the
The Impact of Intellectual Property Protection 607
computer and information service industries. Therefore, this paper selects more repre-
sentative panel data of 34 countries or regions1 in 31 OECD countries and 3 additional
APEC countries or regions from 2007 to 2019, and establishes an extended trade gravity
model to conduct empirical research on the research content of this paper.
Given the research in this article, we assume that the intellectual property protection
of the importing country has the expected impact on the import trade. So the core
explanatory variables that need to be examined in this article are included in intellectual
property protection (IPR), as well as China’s use of foreign direct investment (FDI) and
the level of technological innovation (TI) of exporting countries that have the expected
impact on China’s computer and information service trade imports, taking into account
the factor of time t, that is, the model equation setting becomes:
1 These 34 countries include 31 OECD countries and 3 APEC countries or regions. OECD
countries include Australia, Austria, Belgium, Canada, Czech Republic, Denmark, Estonia,
Finland, France, Germany, Greece, Hungary, Italy, Ireland, Israel, Japan, South Korea, Lux-
embourg, the Netherlands, Norway, New Zealand, Poland, Portugal, Slovenia, Slovakia, Spain,
Sweden, Switzerland, Turkey, the United States and the United Kingdom. APEC countries or
regions include Hong Kong, Russia and Singapore.
608 H. Yang and S. Pang
IMijt represents the import trade volume of country i from country j in the computer and
information service industry in year t; GDPit GDPjt represents the economic development
level and market capacity of country i and country j in year t, measured by the product of
the two gross domestic products; Distanceij represents the distance between China and
the capital of the exporting country; IPRit represents the level of intellectual property
protection in China in year t; FDIijt represents China’s use of foreign direct investment in
year t; TIjt represents the technological innovation capability of the exporting country in
year t, although the number of patent applications can be used to measure technological
innovation capabilities by eliminating the influence of government patent agencies and
other human factors (Chen Lijing, Gu Guoda, 2011) [1], but the amount of patent grants
can better reflect the quality of technological innovation, Dai Ming, Chen Xiao, and
Jiang Han (2017) also believe that China’s technological level should be measured by
the amount of China’s foreign patents granted [2], so this article uses the number of
patents granted by each exporting country based on computer technology to measure the
technological innovation capability of the exporting country according to the selected
research objects.
Computer and service trade import data comes from OECD Statistics; each country’s
GDP product data comes from the World Bank’s statistical database, and GDP data is
calculated in 2010 constant US dollars; the distance between China’s capital and its
trading partners comes from CEPII-Database; the calculation source of the level of
intellectual property protection is based on the GP index researched by Ginarte and
Park (1997) [3] to measure the level of intellectual property protection in China. Since
the GP index itself is an indicator of the intensity of legislation and does not take
into account the factors of the law enforcement of intellectual property protection, we
adjusted the GP index accordingly. Refer to Han Yuxiong and Li Huaizu (2005) [4]
to introduce “law enforcement” on the basis of the GP index, the research method of
“intensity” considers the influencing factors of law enforcement intensity, constructs an
index system of China’s intellectual property protection law enforcement intensity, then
calculates the level of intellectual property protection law enforcement intensity, and
finally calculates China’s actual intellectual property protection level based on the GP
index. The law enforcement system includes four indicators, including legal influence
factors, economic development level, social public awareness and international society
influence factors. Han Yuxiong and Li Huaizu (2005) believe that different indicators
have the same impact weight on law enforcement, so the four indicators are calculated
using an average weighting method. The arithmetic average of the four indicators is
the level of law enforcement of intellectual property protection over the years, and it
expressed by F(t); the intensity of intellectual property protection legislation over the
years is expressed by G(t), and it is calculated through the expansion of the GP index
researched by Ginarte and Park (1997); IPR(t) is used to express the actual level of
intellectual property protection over the years, then the actual intellectual protection
level can be expressed as:
According to the above formula, the actual protection level of intellectual property
in China from 2007 to 2019 can be calculated; China’s use of foreign direct investment
(FDI) data comes from the National Bureau of Statistics; the number of patents granted
by computer technology in various countries comes from the WIPO database.
Since this article uses panel data from the sample countries, the Hausmann test and
LM test must be passed before estimating the model’s parameters to compare fixed
effects, random effects, or mixed regression, which is better. But the distance between
the two countries’ capitals does not change with time, so the variable distance cannot
be identified in the fixed-effects model. Therefore, this article conducts an LM test to
determine which is better than random-effects model or mixed regression. After the LM
test, the result shows Prob > chibar2 = 0.0000. We want to reject the null hypothesis
that the model adopts mixed regression, indicating that the use of random effects is
better than mixed regression. The following table shows the stepwise regression esti-
mation results based on the random effects model. The specific conditions are shown in
Table 2.
610 H. Yang and S. Pang
domestic consumers, and China has strengthened intellectual property protection on the
import of computer and information services trade. The effect of market expansion is
greater than the effect of market power, which has promoted the import scale of computer
and information service trade.
Fourth, the regression coefficient β4 of China’s use of foreign direct investment is
significant at the 5% confidence level, and the sign is positive, indicating that the influ-
ence of FDI on import trade is becoming more and more obvious. Foreign companies
have brought in a large amount of advanced technology, machinery, equipment, informa-
tion and management experience directly related to production, sales and consumption to
directly invest in China, which will also drive domestic demand for computers and infor-
mation services, and increase China’s actual purchasing power to promote the growth
of import trade in this service industry.
Fifth, the regression coefficient β5 of the exporting country’s technological innova-
tion capability is significant at the 5% confidence level, and the sign is positive. On the
one hand, if the technology gap between the two countries is large, China, as a major
imitating country, will further strengthen the company’s own imitating ability. If this
happens, it will inhibit the transfer of advanced technology to the country and constitute
a trade barrier, which will reduce China’s imports. On the other hand, under the govern-
ment’s intervention and other external factors, Chinese companies will choose countries
with a large technological gap to import products and services that are much higher
than their domestic technology levels, allowing companies to conditionally apply and
promote technologies from these countries. Therefore, the possible explanation for this
is that the application and promotion of these technological innovations has a greater
impact on imports than the increase in domestic imitation ability that inhibits imports.
At the same time, the import of high-tech-intensive products also drives the import of
corresponding high-tech services, that is to say, the improvement of the technological
innovation capacity of exporting countries is not an obstacle to China’s computer and
information service imports, but a facilitating factor.
3 Conclusions
This article uses data from 2007 to 2019 in China, OECD countries and APEC countries
or regions (excluding countries that overlap with OECD organizations). The extended
trade gravity model analyzes the impact of intellectual property protection on the import
of computer and information services trade from a quantitative and qualitative perspec-
tive. The main conclusions are as follows: first, while importing computer and informa-
tion services in China, the improvement of China’s intellectual property protection level
has effectively restrained the imitation behavior of domestic member consumers and
has a significant promotion effect on the growth of computer and information service
trade imports, and the protection of intellectual property rights on the effect of market
expansion is greater than the effect of market power. Secondly, there is still a lot of
room for improvement in intellectual property protection in China. Before the exporter
forms the market power effect of intellectual property protection and restrains Chinese
imports, enhancing intellectual property protection can further promote trade imports of
computer and information services.
612 H. Yang and S. Pang
References
1. Chen, L., Gu, G.: The impact of technological innovation and intellectual property protection
on the structure of china’s imported goods: an empirical analysis based on time series data
from 1986 to 2007. Int. Trade Issues. 05, 14–21 (2011)
2. Dai, M., Chen, X., Jiang, H.: The interactive effect of exporting country’s technological level
and importing country’s intellectual property protection on export trade of exporting coun-
tries——based on the empirical study of China’s export trade. Tech. Econ. 36(05), 103–109
(2017)
3. Ginarte, J., Park, W.: Determinants of patent rights: a cross-national study. Res. Policy 26(3),
283–301 (1997)
4. Han, Y., Li, H.: Quantitative analysis on the level of protection of intellectual property rights
in China. Res. Sci. Sci. 03, 377–382 (2005)
5. Liu, C., Li, X., Xu, L.: The influence of regional intellectual capital on regional economic
development-evidence from China. Int. J. u - e Serv. Sci. Technol. 8(6), 89–102 (2015)
The Influence Mechanism of Business
Environment on the Allocation
of Entrepreneurship
1 Introduction
China has achieved long-term, rapid, and stable economic growth in the early stage of
reform and opening up, relying on demographic dividends and resource advantages.
However, the disappearance of these dividends and the pressure of aging, the short-
age of resources, and the backwardness of technology have made Chinese economic
development unbalanced and insufficient. Traditional economic development methods
do not affect current economic development challenges. When the development con-
ditions change, Chinese economic development and the adjustment of economic struc-
tures can only rely on innovation. Because innovation can break through the economic
development bottleneck, crack economic development challenges, find new develop-
ment advantages and guide economic development [1]. Innovation will become the first
driving force to lead Chinese economic development. However, the implementation of
innovation is inseparable from policy support, market subject participation, and transfor-
mation of thinking. In addition, the quality of the business environment largely affects the
behavior of market participants represented by entrepreneurs and the allocation direction
of production factors, which in turn affects the economic efficiency of the entire society.
Special Project of Philosophy and Social Science in Heilongjiang Province (20JYH078).
Therefore, this article will analyze the influence mechanism of the business environment
on the allocation of entrepreneurship in economic development. And then, it will provide
solutions for the problems encountered in the process of economic operation.
2 Literature Review
2.1 Research on the Business Environment
The market environment reflects the process and results of marketization in a country
or region. Its quality is related to the direction of entrepreneurship, and it is a key factor
affecting the development of entrepreneurship.
The Impact of Competition. The competition mechanism is the core of the market
economy and an important means for market entities to realize their interests. A fair
and efficient market competition environment adjusts product prices through supply and
demand, rather than excessive human intervention to distort market prices and cause
resource mismatch. Fair competition promotes the efficient operation of the market,
effectively stimulates the vitality and creativity of market players, and actively attracts
them to actively participate in economic construction. In an environment of fair compe-
tition, entrepreneurs rely on their talents to configure production factors by market rules,
reduce unnecessary non-productive activities that consume social wealth, and devote
more energy to more creative and productive activities. During the activities, the effi-
ciency of economic operation will be improved, and the economy will achieve Pareto
optimality.
The Impact of Integrity. In the process of the market economy becoming increasingly
mature and market scale gradually expanding, the spirit of contract has gradually pene-
trated all aspects of economic life, regulating and guiding the behavior of market entities.
The quality of integrity management derived from the spirit of contract is a beneficial
weapon to help enterprises and entrepreneurs establish a good image and enhance their
competitiveness. An enterprise’s operation and sustainable development require that its
managers have the quality of integrity and products within the scope of laws and regula-
tions. The honest market environment reduces the financing barriers of enterprises to a
certain extent, enables the rational allocation of funds, promotes the flow of production
factors from surplus departments to shortage departments, improves resource alloca-
tion efficiency, reduces enterprise production costs and transaction costs, and promotes
healthy economic development.
and service” and a source of motivation for stimulating market players’ vitality and cre-
ativity. It can regulate market players’ behavior and resolve law enforcement conflicts and
help establish good government-enterprise relationships. In addition, regulatory agencies
should also optimize and innovate regulatory models, expand the scope of supervision of
market entities and strengthen supervision by cooperating with other agencies to achieve
comprehensive supervision of cross-regional markets. At the same time, the regulatory
authorities should advocate to create a social atmosphere for honest law enforcement,
introduce relevant policies and standard systems to regulate entrepreneurial behavior,
and improve the efficiency of supervision and management.
The construction of a legalized business environment can regulate and restrict the behav-
ior of market participants and attract more foreign investment and promote regional
economic development.
The Impact of Investor Protection. Investors provide financial support for the estab-
lishment and continuous operation of enterprises. While protecting enterprises and their
operators, the law must also protect investors’ legitimate rights and interests and main-
tain market fairness and justice. The financial support of small and medium investors
broadens the financing channels for small and micro enterprises and helps solve the
financing difficulties of enterprises. Improving the legal protection of small and medium
investors and enhancing the investment confidence of capital providers are an incentive to
continue to provide financial support for small and medium investors. At the same time,
standardizing corporate capital allocation behavior and improving capital use efficiency
can constrain corporate management behavior.
The Impact of Law Enforcement and Justice. Law enforcement can regulate and
restrict the behavior of market entities. Justice is the last line of defense to safeguard social
fairness and justice. The efficiency of law enforcement and judicial transparency is cru-
cial to constructing a legalized business environment and the allocation of entrepreneur-
ship. Judicial transparency is the key to realizing government credibility and reflects
whether social fairness and justice are maintained and respected. The open judicial
system fundamentally solves the problem of human rights protection, which not only
maintains the authority of the national judicial department but also effectively resolves
the conflicts between government and enterprises [4]. For enterprises and entrepreneurs,
improving the efficiency of law enforcement and enhancing judicial transparency reduce
the cost of litigation for enterprises and protect enterprises’ legal rights and interests.
It can also regulate and restrict entrepreneurs’ behavior and guide entrepreneurship to
flow into the productive activities that create social value.
The Impact of Openness. Since the reform and opening up, the transnational flow of
labor, capital, technology, and other production factors has greatly promoted China’s
economic growth. The influx of high-quality production factors enables China to enjoy
the technology spillover effect and promotes the development and innovation of China’s
technology. The improvement of technological level has improved the quality of Chinese
products, improved Chinese industrial structure, and made Chinese position in the global
value chain gradually rise. In addition to the free flow of production factors that will
618 J. Li and T. Zhang
The Impact of the Level of Economic Development. The level of foreign direct
investment and national income can intuitively reflect the economic development of
a country. And the level of a country’s economic development will directly affect the
allocation direction of entrepreneurship. When the level of national income increases,
people will save after satisfying their basic living needs. The conversion of savings
to investment will drive the improvement of a country’s investment level. To a certain
extent, the increase of idle capital will provide financial support for entrepreneurs and
reduce the difficulty of financing small and medium-sized enterprises [6]. In addition, it
can also reduce the possibility of rent-seeking for entrepreneurs due to insufficient funds
and optimize the allocation direction of entrepreneurship.
The Impact of Social Security Services. A sound social security system is conducive
to narrowing the income gap and alleviating social conflicts. First of all, a sound med-
ical insurance system can protect the health of the labor force and ensure the driving
force needed for sustained economic development. Secondly, in the context of aging,
the improvement of the old-age insurance system helps to ensure the normal replace-
ment of labor and promote the rationalization of the employment structure. Finally, the
improvement of the housing provident fund system has effectively solved the problem
that hinders the free flow of labor factors and ensures the flow of labor factors to regions
with higher production efficiency.
The Impact of Government Services. The government service environment runs
through the entire development process of the enterprise. The digital government service
platform can greatly simplify the government approval process, lower the market access
threshold and reduce the waste of social resources caused by vicious competition. It can
also promote the free flow and aggregation of production factors and provide market
participants with a fairer business environment. While streamlining administration and
delegating power, the government must supervise the market strictly to maintain market
order and regulate the behavior of market entities by establishing a credit supervision
system, which ensures that market participants operate with integrity within the legal
boundaries [7].
The Impact of Financial Services. The financial service system plays an irreplace-
able role in mobilizing savings, allocating idle funds, and diversifying investment risks.
There are significant differences in the financing difficulties faced by companies of differ-
ent sizes. Compared with large companies, small enterprises have difficulties obtaining
financial support from large banks because of lacking reliable financial information and
mortgageable fixed assets. Therefore, small enterprises can only obtain financial sup-
port by relying on “soft information” such as the manager’s ability, personal qualities,
industry development prospects, and corporate profitability, forming a good expectation
for the capital supplier. It can be seen that only when entrepreneurs are committed to
productive activities can they solve the problem of lack of funds.
Traditional
Culture Affect the Comprehens
ively
Thoughts
Social Influence
Education and
Environment Method the
Behaviors of Direction of
Social Entrepreneu
Public Groups
Opinion rship
reform of “decentralization, regulation, and service” and further optimize the business
environment.
4.1 Establish a Market Environment for Fair Competition and Honest Operation
In order to establish a good market environment, the government and the market must
clarify the boundaries and perform their respective duties. Firstly, based on implement-
ing a negative list system for market access, the regulatory authorities should curb illegal
competition among market players, ensuring that market players participate in market
competition fairly and obtain equal access to production factors. Secondly, the gov-
ernment must establish a sound integrity mechanism to restrict entrepreneurs’ illegal
production behavior and reduce entrepreneurs’ non-productive activities. Finally, the
government must regulate the behavior of government and enterprise personnel and
save social resources.
4.2 Build a Legal Environment that Protects the Legitimate Rights of All Parties
The protection of the legitimate rights and interests of both investors and entrepreneurs
is the key to promoting the Chinese legalized business environment. Firstly, relevant
departments should formulate laws to provide institutional guarantees for investors and
entrepreneurs. Secondly, the supervisory department must standardize law enforcement
and improve the credibility of. Thirdly, the judiciary must uphold fair justice, improve
its adjudication capabilities, and fully protect investors and entrepreneurs’ legitimate
rights and interests.
The way of education and the atmosphere of public opinion affect the thoughts and
behaviors of social groups. In order to guide the deployment of entrepreneurship to
productive activities, firstly, China should change the existing education methods and
education concepts and focus on cultivating students’ exploration and innovation spirit.
Secondly, China should abandon the traditional education method and increase the flex-
ibility of education. Thirdly, the whole society should establish a good fault-tolerant
mechanism to reduce the psychological burden of entrepreneurs.
5 Conclusion
Acknowledgment. In writing the thesis, my supervisor Juan Li helped me sort out the context of
the article, find logical loopholes and make the whole thesis more and more perfect. My parents
and friends have always encouraged me to go forward bravely and silently support me on my way
to school. I am very grateful for their love and tolerance for me, and I hope they can be happy
forever.
References
1. Zhuang, Z.: Southern imitation, entrepreneurship and long-term growth. Econ. Res. 62–70 +
94 (2003)
2. Schumpeter, J.A.: The Theory of Economic Development (1934)
3. Baumol, W.J.: Entrepreneurship: productive, unproductive, and destructive. J. Politics 98, 893–
921 (1990)
4. Zheng, F., Wang, Z., Wei, H.: Business rule of law environment index: evaluation system and
guangdong empirical. Guangdong Soc. Sci. (05), 214–223, 256 (2019)
5. Li, H., Li, X., Yao, X., Zhang, H., Zhang, J.: The impact of entrepreneurial and innovative
spirit on China’s economic growth. Econ. Res. 44(10), 99–108 (2009)
6. Zhou, C., Liu, X., Gu, Z.: Business environment and China’s foreign direct investment: from
the perspective of investment motivation. Int. Trade Issues (10), 143–152 (2017)
7. Liu, J., Guan, L.: Optimization of business environment, government functions and new growth
momentum of enterprise TFP- “Window lighting” or “Pro-upper and clearer”. Soft Sci. 34(04)
(2020)
The Transmission and Preventive Measures
of Internet Financial Risk
Heilongjiang, China
1 Introduction
Internet finance is an inevitable trend of economic informationization and a concrete
manifestation of financial modernization. Internet finance refers to integrating Internet
technology (such as computer technology, communication technology, network technol-
ogy, etc.) with financial services to realize the automation and electronation of financial
services, optimize the configuration of financial services, and improve the efficiency of
financial services.
The impact of Internet finance on financial development is two-sided. On the one
hand, Internet finance can promote economic development, promote financial develop-
ment, promote economic globalization, and improve the efficiency of financial services.
On the other hand, the development of Internet finance makes electronic technology
integrated into the financial industry, and the era of big data leads to a large amount of
information, which exacerbates financial risks. However, the development of Internet
finance has become an inevitable trend of economic development and a new engine for
the practice and development of the digital economy.
The first stage is the initial stage of Internet finance (before 2005). At this stage,
Internet finance has not been established, which is mainly reflected in the attempt of
traditional financial institutions to connect with the Internet and try to move the traditional
financial business to the online operation. The emergence of Taobao in May 2003 has
provided a broader space for the development of Internet finance.
The second stage is the embryonic stage of Internet finance (2005–2012). At this
stage, traditional financial institutions continue to deepen their business cooperation
with the Internet. The emergence of third-party payment platforms such as Alipay and
P2P has made Internet finance enter a new stage of development. During this stage,
the People’s Bank of China began to issue third-party payment licenses, which laid a
foundation for the rapid development of the Internet.
The third stage is the rapid development of Internet finance (2013–2014). The year
2013 is known as the “first year of Internet finance”. In this stage, the combination of
the Internet and finance was further deepened. Many financial and structured financial
products appeared on the network platform on a large scale, laying a foundation for the
rapid development of the Internet. At the same time, in June 2013, Alipay and Tianhong
Fund jointly launched “Yu ’ebao” product, which opened up a new situation of Internet
finance and opened up a new market situation of low-risk financial products with higher
returns than deposits. At the same time, the rapid development of third-party payment,
P2P, crowdfunding platforms, etc., also makes the Internet finance industry ushered in
a boom of development.
The Transmission and Preventive Measures 625
The fourth stage is the regulatory development of Internet finance (from 2015 to
now). With the gradual expansion of the Internet financial business, P2P platforms also
have frequent thunderclap times. Since July 2015, the state and local governments have
successively issued regulatory regulations for the Internet financial industry, which pro-
vides a legal basis for the development of Internet finance and puts an end to the chaotic
era of Internet finance. It opens a period of evidence-based development of Internet
finance and protects investors’ legitimate rights and interests.
First, in the P2P trading platform, there exists between the borrower and net credit
platform. The information asymmetry between lenders and borrowers, financial markets
as part of the overall economic activity, also because of market failure caused by asym-
metric information, in this case, with the advantage of the information of one party to
pursue the maximization of self-interest, to adverse selection and moral, Thus, it will
bring the risk of a run on the financial market and cause the risk of Internet finance.
Second, in P2P platforms, borrowers default and fail to repay loans on time. This credit
risk will lead to insufficient information in the whole industry, thus causing Internet
financial risks. Third, due to the poor operation of some P2P platforms, the money lent
by the platforms cannot be recovered. Investors cannot meet their demands when they
want to withdraw money, and violent incidents occur on the platforms, bringing risks to
Internet finance.
While the rest of the investors remain patient, delay consumption, choose to spend at T
= 2, we call the second type of investors.
This expression indicates the conditions that should be satisfied when the equilibrium
is reached and is the Pareto equilibrium solution. Since it is a quadratic continuously
The Transmission and Preventive Measures 629
From Table 1 can see there is two Nash equilibrium strategy (not withdrawal, not
withdrawal) and (withdrawal, withdrawal), the corresponding return is (R, R) and (C1 ,
C1 ), because the R > C1 , it can be seen that before a Nash equilibrium (not withdrawal,
not withdrawal) is superior to the latter equilibrium (withdrawal, withdrawal), also is
the best policy. Both investors will choose not to withdraw money in this state, and there
will be no run on the bank. But because there is an information asymmetry phenomenon,
both investors may choose to withdraw. However, this equilibrium, policymakers do not
want to happen, but due to the random appear when the two equilibrium, once appear,
the equilibrium that market will appear in the run, if the Internet agency risk, about the
effects of risk, is not big. Still, if this risk is transmitted to other peers or the financial
industry, it will lead to the collapse of the entire financial market [5].
Firstly, improve risk prevention and control capabilities and strengthen network security
supervision measures. Network security is the premise of the development of Internet
finance, but also an important support force for the development of inclusive finance.
Only by improving network security can we gain people’s confidence and make people
accept Internet financial services from their hearts. Only by improving customer demand
can we continuously strengthen the process of Internet finance, and only when there is
demand can there be progress. Customers will put forward some practical suggestions
while accepting Internet finance. Financial practitioners can promote the development of
financial informatization according to customers’ needs, so as to promote the continuous
progress of Internet finance. Therefore, the development of Internet finance cannot be
separated from the care of network security, and the continuous improvement of network
security can promote the further development of inclusive finance.
630 N. Zhao and F. Yao
Secondly, Improve relevant laws and regulations. First of all, to stabilize Internet
finance development, the state has promulgated relevant laws. The publication of Guid-
ance on Promoting the Healthy Development of Internet Finance in 2015 has certain
constraints on the development of Internet finance. It has accurately classified the finan-
cial forms such as third-party payment, crowdfunding, and P2P. But the “guidance”,
“recommended”, “method”, etc. There are disadvantages, which will be a loophole of
Internet finance. At present, China doesn’t have perfect laws and regulations policy, the
lack of implementation details, the supporting measures and the measures, and China’s
financial regulator lacks independence, not an independent agency of the Internet finan-
cial regulation. Internet finance is a double-edged sword, as well as low cost, high
efficiency, the advantages of the portable performance, is good, universality. There are
legal issues, such as the common phenomenon in the social activities such as buying
and selling citizens personal information, telecom fraud, etc., so the refining industry
sector of the telecommunications network fraud obligation, it is imperative to perfect
the relevant punishment mechanism of telecom network fraud.
Thirdly, we will further improve the social credit system. Internet financial trading is
based on network, so the credit standards for consumers demand is higher. Still, the social
credit system in China is compared to the rapid development of Internet financial lag,
third-party payment platform, if the low barriers to entry, payment verification method are
relatively simple, not only will bring some risk for subsequent management, At the same
time, it will also bring opportunities for illegal elements. So set up a credit rating system,
perfect the social credit system, reduce Internet financial risk, and effectively evaluate
the credit risks of the Internet financial. Increasing Internet finance’s criminal and illegal
costs and strengthening the sharing of credit information among the government, public
security departments, and the market are conducive to creating a green and safe operating
environment for Internet finance.
6 Conclusion
According to the characteristics of Internet finance, this paper deduces the risk trans-
mission mechanism of Internet finance by using the bank run model and systematically
expounds the path and principle of risk transmission of Internet finance. Based on the
analysis of the development of Internet finance, the risk transmission path of the influence
of the three forms of Internet finance, such as third-party payment, P2P and crowdfund-
ing, on Internet finance is proposed, and the return of the multi-equilibrium model is
constructed by game theory. According to the characteristics of Internet risk transmis-
sion, this paper puts forward how to use it for risk management to improve the risk
governance ability and risk management efficiency of Internet finance.
Acknowledgments. This work was supported by the grants of 2017 National Social Science
Foundation of China (No. 17BJY119).
The Transmission and Preventive Measures 631
References
1. Liao, Y.: The supervision research on China’s internet financial development and risk——
based on P2P Platform, Yu’ebao and the third-party payment as an example. Econ. Manage.
29, 51–57 (2015)
2. Grisse, C., Nitschka, T.: On financial risk and the safe haven characteristics of Swiss franc
exchange rates. J. Empirical Financ. 32, 153–164 (2015)
3. Chan, S.: Financial repression and financial risk: the case of China. Post-Communist Econ.
33(1), 119–132 (2021)
4. Chuanmin, M.I., Runjie, X.U., Jing, T.: Analysis of internet financial spatial aggregation and
systematic risk prevention——based on t-SNE machine learning model. Collected Essays
Finance and Econ. (2019)
5. Rongda, C., Lean, Y., Chenglu, J.: The development process, development mode and future
challenges of internet Finance in China. J. Quant. Tech. Econ. 37(01), 3–22 (2020)
Emotion Analysis System Based on SKEP Model
Abstract. In recent years, with the rise and development of e-commerce and
social networking, appeared on the Internet a lot of subjective texts, and emo-
tional analysis task is defined with the text analysis, the subjectivity of the emo-
tional color processing, summing up. Reasoning process, so the sentiment analysis
technology is designed to deal with subjective text. This paper proposes an emo-
tional analysis system based on the SKEP model based on network neologies,
ambiguous sentences, irony, or implicit emotional expressions in subjective texts.
In addition, traditional sentiment analysis methods often have disadvantages such
as low accuracy, weak semantic representation, incomplete feature extraction, and
low contextual correlation. Through text pre-training of large-scale unsupervised
e-commerce, fine-tuning is carried out on small-scale supervised data. Based on
the same data set, compared with THE SEKP model, the F1 value of the opinion
extraction task and the accuracy of the sentence-level sentiment analysis task were
significantly improved. Still, the accuracy of the evaluation object-level sentiment
analysis did not show good results.
1 Introduction
With the emergence of subjective text, the sense of participation of Internet users has
been enhanced, and the user stickiness and system retention of APP. The task of sen-
timent analysis is defined as analyzing, processing, inducing, and reasoning subjective
texts with emotional color. That is, corresponding to the input text, output description
entity, attribute, emotion, opinion holder, time five dimensions. At present, most emotion
analysis technologies mainly solve the extraction of the three dimensions of description
entity, attribute, and emotion, corresponding to the opinion holder, while the two dimen-
sions of time mainly use some existing information extraction technologies directly.
According to the different emphases of the techniques, we will specifically divide the
emotion analysis techniques into three categories: opinion extraction, sentence-level
emotion classification, and more fine-grained evaluation object-level emotion classifi-
cation. Opinion extraction and sentence-level sentiment classification are both extracted
from a single emotion dimension, and the only difference between them is the granularity
of the input text. One is the granularity of words, and the other is the granularity of sen-
tences or texts. However, the evaluation object-level emotion classification will extract
information from multiple dimensions such as entity, attribute and emotion, which is
difficult to realize technically.
Hopes to use in electric business review sentiment analysis technology to automati-
cally extract the user’s point of view and view of polarity, through automated analysis of
the electricity business review user’s point of view, including logistics soon, authenticity,
cost-effective, easy to use, these views, on the one hand, the characteristics of can help
the user to quickly understand the product, auxiliary for the consumer decision-making,
on the other hand, can also help our businesses to be timely find the advantages and
disadvantages of its products, to improve their products.
2 Related Work
Because opinion extraction techniques are usually fairly basic, we can eventually get
the sentiment analysis results for each word. We can then use the results of opinion
extraction as a lexicon for other sentiment categorization results. Hence the term many
people use for opinion extraction techniques to build sentiment dictionaries. Common
methods to construct an emotion dictionary include manual annotation and automatic
method [1]. The advantage of manual annotation is that the accuracy rate is very high, but
the disadvantage is that many labor costs are needed. Of course, there are some partial
automatic methods. These are mainly based on the manual annotation of prominent words
and then automatically emotional annotation of other words. Constructing an emotion
dictionary is an important link in the task of emotion analysis. Yang Shuxin, Zhang Nan
et al. [2] believed that the emotion dictionary could reflect the unstructured features
of texts and adopted an automatic partial method to screen words in sentences using
the emotion dictionary. Knowi et al. [3] proposed that using an emotion dictionary for
corpus emotion analysis is an important work content and significance of the dictionary.
Through in-depth research on the emotion dictionary, Zargari Hamed et al. [4] proposed
an emotion dictionary based on the analysis of emotion components such as emotion
words, negative words, and intensifier words, which makes up for the heterogeneous
influence of the phenomenon of multiple languages of emotion compound in the emotion
dictionary.
As for sentence-level emotion analysis aims to output the overall emotional tendency
of the whole sentence and discourse. It is generally accepted that this study started with
Pang [5] and Turney [6]. Generally speaking, sentiment analysis at the sentence or
discourse level is treated as a typical text classification problem. In terms of specific
methods, it mainly experienced three stages: rule-based method, traditional machine
learning, and deep learning method [7]. The rule-based method requires experts to set
rules and then requires a lot of labor costs.
Moreover, this rule-based method has no generalization ability, and the model per-
formance is relatively weak [8]. After a long period of time, with the development of
traditional machine learning technology, methods based on traditional machine learning
gradually occupy the mainstream position [9]. Traditional machine learning methods
require the manual design of some features, and the quality of feature design will greatly
634 Z. Yanrong et al.
affect the model’s effectiveness. With the emergence of deep learning methods in recent
years, a great advantage is that there is no need to define some features manually [10].
The evaluation object-level emotion analysis task, or the objective level emotion
analysis, is a more fine-grained emotion analysis ability, aiming at the whole sentence
and chapter, output the emotion analysis results for the specific evaluation object or
dimension. It is specifically divided into the emotion analysis task of the given evaluation
object and the joint extraction of the evaluation object and emotion [11]. The emotion
analysis task of the given evaluation object first needs to give the target entity to be
analyzed, and at the same time needs the user to input the current input text to be
analyzed, and the model needs to give a judgment according to the two input results.
However, the joint extraction task of evaluation object and emotion only requires the
user to input the text, and the model can automatically identify the description object
and its corresponding emotion [12].
Sentence or discourse level emotion analysis aims to output the overall emotional ten-
dency of the whole sentence and discourse. For example: “What China has achieved is
really incredible”, we need SKEP model to understand the overall emotion expressed
in this sentence, and then output the corresponding result of sentiment classification, as
well as the positive probability and negative probability.
Generally speaking, sentiment analysis at sentence or discourse level is treated as a
typical text classification problem. In terms of specific methods, it mainly experienced
three stages: rule-based method, traditional machine learning and deep learning method.
The rule-based method requires experts to set rules, which not only consumes a lot of
labor costs, but also has weak generalization ability. After a long period of time, with the
development of traditional machine learning technology, methods based on traditional
machine learning gradually occupy the mainstream position. Firstly, the input discrete
text is continuously processed, which is the process of feature extraction and feature
selection [16].
“This store has great facilities. I’m very satisfied.” First, we manually define the
features that we think can be used to express the emotion of a sentence. Perhaps the
result of the word breaking is the emotional feature that we find useful. Then, we use
the feature name combined with the feature value to represent the result of the word
breaking. The other is a fragment of two words, which extracts N-element grammar as
a feature that is useful for emotional classification. As shown in Table 2.
Similarly, emotion words and degree words can be used to construct features. In this
way, after all the features are constructed, the discrete input text can be expressed as
the vector form of real numbers, and then the real number vector can be applied to the
classification model. Common models include SVM, maximum entropy, Naive Bayes,
Logistic regression, etc. [17].
In addition, in addition to feature extraction and feature selection, we also need to
define a loss function to guide model training. When the number of training rounds of
the model is enough, and the effect is good, we can deploy the model and predict the
emotion of the new samples. See Fig. 1.
Traditional machine learning methods require the manual design of some features,
and the quality of feature design will greatly affect the effectiveness of the model. With
the emergence of deep learning methods in recent years, compared with traditional
machine learning methods, a great advantage is that there is no need to define some
features manually. The input text only needs to be represented by embedding encoding,
and then the model can automatically learn and extract features [18].
The most classical model for emotion analysis task of a given evaluation object is the
LSTM model based on Attention mechanism [20], as shown in Fig. 2.
This model first represents the input word as a word vector, and then inputs it into
the LSTM model according to the sequence of statements, so that the implicit vector of
each position can be obtained. The model introduces a new Embedding vector Va for the
current given evaluation object, and this vector is automatically learned in model training.
For example, “Mi phones” is represented as Va, and then Va and the implicit vector of
each position of the input text (H1,..Hn) uses Attention to calculate correlation and get
a weight of Attention. Attention model can be understood as a method of automatic
correlation calculation. Attention can be used to automatically learn [21] which words
in the input text are important for us to classify emotions, and then the Attention weight
vector will give higher weight to these words. After we get the Attiention weight vector,
our model will perform a weighted sum of H at all positions based on the Attiention
weight. So we get the semantic representation r of the current evaluation object, and
then we take that r to do affective classification. The advantage of the model is to learn
a semantic representation related to each evaluation object by introducing the vector
Va and Attiention mechanism of the evaluation object. Therefore, it is better to use this
semantic representation for emotion classification.
In view of the evaluation objects, emotion combined extraction task we usually
practice it as a typical problem of information extract, two-way LSTM + CRF model
is the most classic [22], you first need to put every word for embedding embedded said,
then will put each embedding vector according to the sequence of input into our LSTM
model, then we could get the LSTM implicit vector model outputs, CRF directly in the
implicit vector based by one layer, let the CRF sequence annotations for each position. In
this way, according to the results of CRF annotation, we can directly output the extraction
638 Z. Yanrong et al.
results of the model. The advantage of the model is that such dimension and emotion
are extracted together by introducing a sequence annotation task [23].
The method of this model is to first train the pre-training model continuously with
a large amount of subjective text data, such as “This product came really fast, and I
appreiated it” commentary text, and then remove the words that are considered to be
related to emotional analysis from these sentences so that the model can be automatically
restored. For example, to block the word “fast” and let the model predict in the pre-
training stage, the original position should be an emotional word, and then to block the
word “appreiated”, and then let the model predict what word is in the pre-training stage,
or even whether the model predicts a positive or a negative emotional word before the
position. Through these simple and effective methods, the SKEP model has a better
perception of multiple types of emotional knowledge, and then it will achieve better
results on the downstream emotional analysis tasks.
Objective Optimization Function. The objective optimization function L(Sentiment
Pre-training Objectives) of the emotional pre-training model SKEP consists of three
Emotion Analysis System Based on SKEP Model 639
objective optimization functions. They are sentient word objectives (Lsw ), word polarity
objectives (Lwp ) and aspect sentient pair objectives(Lap ),so L = Lsw + Lwp + Lap . The
calculation method of objective function loss in emotional words is shown in formulas
(1) and (2):
i=n
Lsw = − mi × yi log ŷi (2)
i=1
Where x̃i is the output vector of the encoder in the transformer, and ŷi is the probability
distribution obtained after x̃i passes through the output layer and then through softmax.
However, after reasoning to get the token of each location, the loss of each token will
not be calculated, but only the loss of the location of emotional words, and the location
of non emotional words will not participate in the calculation. This is also the role of,
which will screen which words are emotional words and which are not.
The calculation method of the objective function Lwp of emotional polarity words is
similar to that of Lsw . The difference is that Lsw calculates token and loss and polarity
loss. In fact, polarity can be understood as another type of token here. The difference is
that there are only two classes: positive and negative.
The calculation method of attribute word emotion word pair objective function Lap
is shown in formulas (3) and (4):
ŷa = sigmoid x̃1 Wap + bap (3)
a=A
Lap = − ya log ŷa (4)
a=1
Here, x1 is the output vector of the [CLS] position. ya is an aspect sentimental pair,
that is, an attribute word emotion word pair. ŷa is the probability evaluation value of ya .
It should be noted that we already have a dictionary library of aspect sentimental pairs,
that is, each pair has a corresponding ID representation.
A total of 7 data sets were collected and sorted for opinion extraction task, sen-
tence level emotion classification task and evaluation object level emotion classification
task, including CoTE-BD, Cote-MFW and Cote-DP of Chinese Academy of Sciences,
ChnSentiCorp of Chinese Academy of Sciences, NLPCC14-SC of Sooksu University,
SE-ABSA16_PHNS and SE-ABSA16_CAME of Harbin Institute of Technology.
640 Z. Yanrong et al.
SKEP model of data taken from June 30, 2020, last updated data, in SKEP model,
on the basis of through small-scale supervised data of fine-tuning, changing the batch
Emotion Analysis System Based on SKEP Model 641
size (batch_size) and training rounds (epochs) and so on in order to achieve the best
training effect, in which the training model in Chinese using skep_ernie_1. 0 _large_ch,
this model is in pre training ernie_1 SKEP model. 0 _large_ch continues to advance in
massive Chinese data based training by the training model in Chinese.From Table 3 the
result of the experiment shows that extraction task overall improved values of F1, of
which COTE - DP value increased by 4%, the F1 sentences to emotion classification
task accuracy level range is not very big, ChnSentiCorp increased 0.25%, NLPCC14 -
SC increases by 0.72%, but the accuracy evaluation object level emotion classification
task optimization effect is far inferior to SKEP model.
5 Conclusion
In recent years, a large number of studies have shown that pretraining models based on
large corpora can learn common language representations, which is beneficial to down-
stream NLP tasks and can avoid training models from scratch. With the development of
computing power, the emergence of deep model Transformer and the enhancement of
training skills make the pre-training model of large corpus develop continuously from
shallow to deep.SKEP model uses an unsupervised method to mine the emotional knowl-
edge automatically, and then uses the emotional knowledge to construct pre-training
targets so that the machine can learn to understand the emotional semantics.
References
1. Zhong, J., Liu, W., Wang, S., Yang, H.: Text emotional analysis methods and application
overview. Data Anal. Knowl. Discov. 5(06), 1–13 (2021)
2. Yang, S., Zhang, N.: Text emotional analysis. Comput. Appl. 1–6 (2021)
3. Li, M., Cui, X.: Usefulness evaluation of user-generated content based on domain emotional
dictionary--take douban reading as an example. Intell. Theor. Pract. 1–15 (2021)
4. Zargari, H., Zahedi, M., Rahimi, M.: GINS: a global intensifier-based N-Gram sentiment
dictionary. J. Intell. Fuzzy Syst. 40(6), 11763–11776 (2021)
5. Pang, B., Lee, L., Vaithyanathan, S.: Thumbs up: sentiment classification using machine
learning techniques. Empirical Methods Nat. Lang. Process. 79–86 (2002)
6. Turney, P.: Thumbs up or thumbs down semantic orientation applied to unsupervised classi-
fication of reviews. In: Proceedings of Annual Meeting of the Association for Computational
Linguistics, pp. 417–424 (2002)
7. Wang, T., Yang, W.: A review of text affective analysis methods. Comput. Eng. Appl. 57(12),
11–24 (2021)
8. Meng, J., Lv, Pin., Yu, Y., Zheng, Z.: Cross-disciplinary emotional analysis based on CNN.
Comput. Eng. Appl. 1–11 (2021)
9. Tang, L., Xiong, C., Wang, W., Zhou, Y., Zhao, Z.: An overview of short-text emotional
tendency analysis based on in-depth learning. Comput. Sci. Explor. 15(05), 794–811 (2021)
10. Liu, T., Feng, M.: Affective analysis based on in-depth learning. Inf. Commun. 01, 88–89
(2020)
642 Z. Yanrong et al.
11. Shen, Y., Zhao, X.: An overview of emotional analysis at different levels based on in-depth
learning. Inf. Technol. Stand. (Z1), 50–53+58 (2020)
12. Chen, P., Feng, L.: Overview of aspects extraction in emotional analysis. Comput. Appl.
38(S2), 84–88+96 (2018)
13. Wang, J., Zhang, Z.: Emotional analysis of short text on Weibo based on improved subject
model [J]. Information and Computer (theoretical version), (06): 134–135+141 (2019).
14. Wang, Q., Wu, Z.: Research progress of view mining technology for public opinion
monitoring. New Industrialization 9(06), 74–77 (2019)
15. Zhang, X., Wang, D.L., et al.: Boosting contextual information for deep neural network
based voice activity detection. In: IEEE/ACM transactions on audio, speech, and language
processing (2016)
16. Kim, Y.: Convolutional Neural Networks for Sentence Classification. Eprint Arxiv, 2014.
17. Qian, Q., Huang, M., Lei, J., et al.: Linguistically regularized LSTMs for sentiment classi-
fication. In: Proceedings of the 55th Annual Meeting of the Association for Computational
Linguistics (Volume 1: Long Papers) (2017)
18. Teng, Z., Vo D, T., Yue, Z.: Context-sensitive lexicon features for neural sentiment analysis. In:
Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing
(2016)
19. Li, X., Bing, L., Li, P., et al.: A unified model for opinion target extraction and target sentiment
prediction (2018)
20. Wang, Y., Huang, M., Zhu, X., et al.: Attention-based LSTM for aspect-level sentiment
classification. In: Proceedings of the 2016 Conference on Empirical Methods in Natural
Language Processing (2016)
21. Liu, F., Xu, M., Deng, X.: An emotional analysis combining attention mechanism and sentence
ordering. Comput. Eng. Appl. 56(13), 12–19 (2020)
22. Guan, P., Li, B., Lv, X., Zhou, J.: Bidirectional LSTM emotional analysis with enhanced
attention. Chin. J. Inf. 33(02), 105–111 (2019)
23. Tang, D., Qin, B., Feng, X., et al.: Effective LSTMs for target-dependent sentiment
classification. Comput. Sci. (2015)
24. Devlin, J., Chang, M W., Lee, K., et al.: BERT: pre-training of deep bidirectional transformers
for language understanding (2018)
25. Tian, H., Gao, C., Xiao, X., et al.: SKEP: sentiment knowledge enhanced pre-training for
sentiment analysis. In: Proceedings of the 58th Annual Meeting of the Association for
Computational Linguistics (2020)
Evaluation of Enterprise Development
Efficiency Based on AHP-DEA
Abstract. With the rapid development of China’s economy and the acceler-
ated expansion of enterprise-scale, it is necessary to conduct objective evalua-
tion through quantitative evaluation method to determine whether the investment
of enterprises can obtain the best efficiency and enhance the competitiveness of
enterprises. To overcome the lack of comparability of the traditional absolute eval-
uation method and better evaluate the relative development status of enterprises,
the relativity evaluation of input-output efficiency should be carried out among
different enterprises. A comprehensive development performance evaluation sys-
tem is constructed based on the AHP analytic hierarchy process (AHP) model and
the DEA data envelopment analysis model. First, the AHP method is used to deter-
mine the weight of the evaluation index at the same level and relative to the target
level. The indexes with larger weight values relative to the target level are selected
as the index of the DEA evaluation model among all the indexes, thus avoiding the
randomness of DEA index selection. Based on the annual report data of tourism
companies between 2017 and 2019, the AHP-DEA model was used to evaluate the
static efficiency of enterprise development. Then the DEA-Malmquist model was
used to analyze competitiveness with three years of panel data for dynamic anal-
ysis. At last, the effective and ineffective development enterprises were pointed
out, and the reasons for the invalid development were analyzed, aiming to provide
useful evidence for the manager to make decisions.
1 Introduction
In order to obtain the largest market share, the scale of enterprises continues to expand, the
product system is increasingly improved, and the market order is continuously optimized
[1]. And with the development of enterprises, the competition among enterprises is
also intensifying. Since enterprise competitiveness is closely related to multiple factors
such as total assets, the number of products, operating income, marketing expenses,
profits, etc., simply expanding the scale to have more market share is not necessarily
beneficial for enterprise development. Therefore, effective evaluation of competitiveness
in enterprise development has great application value and can be widely used in other
industries.
Existing research mainly evaluates the enterprise competitiveness from two aspects,
qualitative and quantitative. The qualitative analysis is mainly conducted from enter-
prise culture and the connotation of enterprise competitiveness [2]. Quantitative research
mainly establishes evaluation index systems from internal and external aspects such as
the external environment in which the enterprise is located and the enterprise’s devel-
opment. Quantitative analysis uses hierarchical analysis, principal component analy-
sis, neural network, fuzzy comprehensive evaluation method, data envelopment analy-
sis (DEA), comprehensive index evaluation method, and survey questionnaire method.
Many methods involve selecting evaluation indexes with subjectivity, arbitrariness, and
not scientific objectivity, and there is no uniform standard in selecting enterprises. In
order to overcome the subjectivity and arbitrariness of evaluation, a comprehensive and
objective comprehensive evaluation model of enterprise competitiveness evaluation is
constructed using the widely used multi-criteria methods AHP hierarchical analysis and
DEA data envelopment analysis. The main contributions of this paper include.
(1) Using AHP to establish the hierarchy structure of the enterprise, the elements of
each level are described quantitatively so that the peer weight and relative target
level weight of the elements of each level can be calculated and the weight order
of each index can be determined.
(2) The DEA evaluation index is determined, and the DEA-BCC model is used to make
a static analysis on the development performance of 10 listed tourism enterprises.
(3) This paper uses DEA Malmquist model to dynamically analyze the three-year
development of 10 listed tourism enterprises from 2017 to 2019.
inputs and multiple index outputs, with unit invariance and function uncertainty, that
is, changes in the numerical units of the input-output index variables of the decision
units do not affect the final efficiency value results, and the units of each index can be
inconsistent. For the production function, there is no need for No prior assumptions to be
required for the production function, which simplifies the complex production process, so
DEA can overcome the influence of subjective factors and simplify the algorithm while
improving the accuracy. Since there are many models in DEA with different conditions
and emphasis, the BCC model and the Malmquist model were chosen to analyze the
company’s competitiveness from both static and dynamic quantitative aspects [9].
BCC Model. Relative to the DEA-CCR model based on constant payoff size first pro-
posed by A. Charnes W.W. Copper and E. Rhodes in 1978, the BCC model [10] adds
a condition that the sum of the weight coefficients Lambda of the reference bench-
mark is 1. The analysis results obtained not only consider the pure scale efficiency. The
BCC model focuses on the analysis of static efficiency values of multiple decision units
simultaneously. If we want to measurethe efficiency
n of a set of n decision making units
(DMUs) of the same kind, denoted as DMUj j=1 , each DMU decision making unit has
m input index, denoted as {xi }m
i=1 , and s output index, denoted as {yr }r=1 . A constant is
s
introduced to denote the non-Archimedean infinitesimal, and sr+ and sr− are the slack
variables for output and input, respectively. When the firm reaches the Pareto optimal
state, that is, θ = 1 and sr+ = 0 and sr− = 0. This is the decision unit with stronger
validity, and the other is when θ = 1, without considering the slack variables, which is
the decision unit with weaker validity.
Malmquist Index Model. The Malmquist index model was proposed to address the
limitations of the BCC model with same period data, i.e., cross-sectional data, which
is a dynamic efficiency value analysis model for multiple decision units with a time
horizon. RolfFare et al., in 1994 [11] combined the Malmquist index. They proposed the
DEA total factor productivity (TFP, which refers to the ratio of the value of the output of
an economic system to the value of all inputs, reflecting the combined productivity level
of all factors of production that are input to production) model. This model decomposes
the Malmquist index into the change in integrated technical efficiency (EC) and the
change in technological progress (TC), which can also be divided into the change in the
value of pure technical efficiency (PEC) and the change in the value of scale efficiency
(SEC).
Hierarchical Element Selection for AHP. First of all, the competitiveness of enter-
prise development includes the growth status of enterprise competitiveness, and the
evaluation of enterprise competitiveness can be measured from three aspects: enterprise
economic scale situation, enterprise profitability, and enterprise solvency.
646 W. Geng and M. Gao
1) The enterprise’s economic scale can determine the enterprise’s position in the com-
petition of the same industry, as well as the enterprise’s ability to resist risks, and
the corresponding indexes are selected as follows.
Total assets: The total number of assets owned by the enterprise, including current
assets, long-term assets, fixed assets, intangible and deferred assets, and other long-term
assets, etc.
Total Products Produced: The type and number of products.
Owner’s equity: It is the interest of the enterprise’s investors in the enterprise’s net
assets and contains the profit that the owner can make according to the number of assets
provided.
2) Profitability of the enterprise is the driving force for the development of the enter-
prise, mainly including operating income, operating costs, marketing expenses and
net profit.
Operating income: reflects the enterprise’s income, which is the main support to
maintain the development of an enterprise. The size of the marketing amount can observe
the operation of the enterprise and the occupation of the industry market.
Operating cost: It is the cost of business against the turnover. The lower the cost, the
more likely the enterprise will be profitable. Therefore, operating costs can effectively
reflect the efficiency of business management.
Marketing costs are the various expenses related to marketing activities. In order
to attract new customers and increase the market share, enterprises need to do various
promotional and promotional activities or advertising investments. Marketing expenses
directly affect the profit of the enterprise. The enterprise should also control the marketing
amount.
Net profit: The amount of total profit minus income tax of the enterprise, which can
reflect the important index of business efficiency of the enterprise.
The current ratio of current assets to current liabilities is a simple estimate of short-
term debt servicing capacity.
The gearing ratio measures an enterprise’s ability to use creditors’ funds to carry out
business activities. It is an important index of the level of indebtedness and the degree
of risk.
In summary, the overall framework of the AHP model is shown in Fig. 1.
Evaluation of Enterprise Development Efficiency Based on AHP-DEA 647
Table 1. The weight of each index at the same level and relative to the target level
First level index Second level index Weight of same level Weight for target level
Competitive Enterprise scale 0.2790 0.2790
Profitability 0.6491 0.6491
Solvency 0.0719 0.0719
Enterprise scale Total assets 0.0497 0.1782
Total products 0.0196 0.0704
Owner’s equity 0.2096 0.7514
Profitability Operating income 0.1608 0.2477
Net profit 0.3595 0.5538
Operating cost 0.0817 0.1259
Operating expenses 0.0472 0.0727
Solvency Current ratio 0.0629 0.8750
Asset-liability ratio 0.0090 0.1250
648 W. Geng and M. Gao
It can be seen from Table 1 that among the primary evaluation indexes, the lowest
weight is given to the solvency of the enterprise, while the highest weight value is given
to the profitability. Among the secondary ‘indexes, the weighting of the impact on the
economic scale of the enterprise is Owner’s Equity > Total Assets > Total Product
Production. The weight value of the impact on profitability is net profit > operating
income > operating cost > marketing expense.
To determine the input and output indexs of the DEA model, the weight values of each
evaluation index can be calculated according to AHP, and the evaluation indexes affecting
the competitiveness of enterprises are mainly selected from the economic scale and
profitability of enterprises, and the indexs with larger weight values are selected, which
can comprehensively and accurately show the operating conditions and development
trends of these enterprises. Among them, “operating revenue (billion)”, “owner’s equity
(billion)” and “net profit (billion)” are used as three output indexes, “operating Costs
(billion)” and “Operating expenses (billion)” are used as 2 input indexes x.
Based on the list of top-ranked tourism enterprises selected from public data, a total of 10
listed tourism enterprises, including Ctrip, China National Travel Service, China Youth
Travel Service, Tempus International, Zongxin Travel, Xi’an Tourism, Tibet Tourism,
Huangshan Tourism, Qujiang Cultural Tourism, and Yunnan Tourism Co. were selected
from the Shenzhen and Shanghai markets, respectively, as research subjects. These enter-
prises were selected considering factors such as tourism market ranking, tourism char-
acteristics, and e-commerce applications, and representative enterprises were selected
to represent the current development status of the tourism industry.
Because the size of the tourism market is still expanding continuously, the current market
is always in a state of expansion, so the BCC model based on the output perspective is
selected, i.e., the inputs are kept constant, making the output maximized, and the results
of the DEAP2.1 software used to calculate the specific efficiency of the 10 enterprises
from 2017–2019 are obtained. As shown in Fig. 2:
Evaluation of Enterprise Development Efficiency Based on AHP-DEA 649
Fig. 2. The efficiency chart of variable scale return and constant scale return of 10 tourism
enterprises from 2017 to 2019
From Fig. 2, we can see that if constant returns to scale (CRS) are assumed, each
firm’s overall efficiency of development is low, and most firms are inefficient and far
from the efficient curve1. Thus variable returns to scale (VRS) should be chosen as a
premise.
Analysis of the efficiency of each enterprise from 2017 to 2019: 8 enterprises had
effective, comprehensive technical efficiency and two enterprises had ineffective effi-
ciency in 2017. Only four enterprises developed in 2018 were effective; three were
ineffective. In 2019, only four pure scale efficiency effective enterprises. There are eight
pure technical efficiency effective enterprises. Overall, the overall technical efficiency
value, subject to variable payoffs for scale, decreases from 0.891 in 2017 to 0.878 in
2019, indicating that the firm’s overall efficiency is decreasing. The pure technical effi-
ciency is also decreasing from 0.977, 0.962 to 0.954, while the pure scale efficiency of
the firm is effective overall from 0.912, 0.915 to 0.922, although the scale efficiency is
increasing less. However, the decreasing pure technical efficiency of travel firms indi-
cates that the technical input efficiency of travel firms is low. From the Fig., it can be seen
that Ctrip network is a typical example, which has been in an effective state in 3 years.
It is a typical representative enterprise of tourism e-commerce, focusing on the applica-
tion and development of advanced technology, and has been in the leading position in
terms of technical efficiency; although the scale gain remains the same, the enterprise
is still ineffective development. Thus the enterprise development mainly depends on
technological progress.
Fig. 3. Chart of total factor productivity of 10 tourism enterprises from 2017 to 2019
(1) Figure 3 shows the dynamic trends of integrated efficiency change effch, tech-
nological change techch, pure technological change pech, scale change sech, and
total factor productivity i.e., productivity index tfpch for the three years 2017–
2019, where total factor productivity change is determined by integrated technical
efficiency change (i.e., resource allocation efficiency) and technological progress,
while purely technical changes and scale changes determine integrated technical
efficiency change. Where the productivity index is greater than 1, firm productivity
rises, and if the productivity index is less than 1, firm productivity is falling.
(2) Stage-by-stage trend analysis of tourism enterprises’ development efficiency during
the three years. Through the Malmquist model, it is also possible to obtain efficiency
changes among different years for all enterprises, as shown in Table 2.
The table shows that relative to 2017, the production index for 2018, i.e., total factor
productivity, is greater than 1 and shows an upward trend, above the average. This is
mainly due to the technological progress index of 1.235, which is 23.5% higher than the
efficiency value in 2017. The decline in ure technical efficiency and the increase in scale
efficiency led to a combined efficiency value of 0.992, which is 0.8% lower than in 2017.
One of the main factors is that 2018 is the opening year of the integrated development
of culture and tourism. People’s enthusiasm for tourism is high, the business volume of
enterprises has increased. Therefore, each tourism enterprise has increased its investment
Evaluation of Enterprise Development Efficiency Based on AHP-DEA 651
in the research and development of new tourism products and expansion of enterprise-
scale, which increased from 5.40 trillion in tourism revenue in 2017 to 5.97 trillion in
2018. The main source of revenue is the first-tier cities in the north. Still, the second-tier
and new first-tier cities also become the focus of people’s outbound travel; thus, the
national tourism form is better. The efficiency of the tourism scale is improved.
5 Conclusion
References
1. Zhang, J.T., Wang, Y., Liu, L.G.: Model system construction of intelligent tourism application
in the context of big data. Enterp. Econ. 36(5), 116–123 (2017)
2. Jiang, A.Y., Li, Z.R.: Research on credit evaluation of science and technology-based SMEs.
Dev. Res. (6), 60–64 (2014)
3. Tian, H.D., Shen, W.H., T, Y.B., Gan, G.J.: Research on the development path of forest
recreation based on hierarchical analysis--take Guangxi Cat’s Hill National Nature Reserve
as an example. Forestry Econ. 1–12 (2020)
4. Li, Y.: Research on risk evaluation of PPP projects based on F-AHP evaluation method. J.
Hunan Acad. Arts Sci. (Nat. Sci. Ed.) 32(04), 69–74 (2020)
5. Li, J.Y.: Exploring the decision making problem of airport cab drivers based on hierarchical
analysis. Sci. Technol. Innov. (34), 14–15 (2020)
652 W. Geng and M. Gao
Harbin Vocational College of Science and Technology, Harbin 150300, Heilongjiang, China
quy@hrbcu.edu.cn
1 Introduction
J.A. Sehumpete, an Austrian American economist, first put forward the concept of
“innovation”. He regards “innovation” as a process of establishing a new production
function. He believes that innovation introduces a group of new combinations of pro-
duction factors and conditions into the production system. Most scholars believe that
although the innovation efficiency of China’s electronic and communication equipment
manufacturing industry has experienced a series of fluctuations, it is generally on the rise.
Sun (2014) used the DEA-Malmquist index to measure the operating efficiency of ten
LED manufacturing enterprises in Taiwan from 2003 to 2009. The innovation efficiency
of both upstream and downstream enterprises has improved accordingly [1]. Although
the research on innovation efficiency of high-tech industry and manufacturing industry
is very rich, the research on electronic and communication equipment manufacturing
industry is still very scarce. Henisz and Zelner (2001) analyzed the impact of the insti-
tutional environment on telecom industry investment by using cross-border panel data
of 147 countries from 1960 to 1994. The results show that the stronger a country’s gov-
ernment control, the faster the growth rate of Telecom investment [2]. Hemmert (2004)
analyzed the impact of institutional factors on the innovation capability of the high-tech
industry. This paper divides institutional factors into four aspects, a total of 27 indicators,
and empirically tests Germany and Japan’s pharmaceutical and semiconductor industries
[3]. Andonova (2006) found that the change of institutional environment has a greater
impact on wired information technology than wireless communication technology by
using cross-border panel data from 1960 to 2002 [4].
In terms of research methods, the DEA method is used in most efficiency measure-
ments. Wang and Huang (2007) used DEA method to evaluate the efficiency level of
the Research and Development (R&D) process in various countries. The results show
that more than 50% of countries are not fully effective in R&D activities, and more than
two thirds of countries are in the stage of increasing returns to scale [5]. Sharma and
Thomas (2008) used the DEA method to evaluate the innovation efficiency of more than
20 countries, and found that the innovation resources of developing countries can be
used reasonably [6]. Lee et al. (2009) used the DEA method to measure and compare
national R&D plans’ performance and provided policy suggestions for the government
to effectively formulate and implement national R & R&D plans [7]. Guan and Chen
(2012) used a two-stage DEA model to find that the overall efficiency of national inno-
vation mainly depends on the downstream commercialization efficiency [8]. But DEA
can only measure the static change of DMU efficiency, Malmquist index can measure
the dynamic change. Therefore, this paper selects the Global- Malmquist (GM) index
model to analyze the EECE of 20 provinces in China. The innovations are as follows:
Firstly, the research subject of this paper is the electronic and communication equip-
ment manufacturing industry, and there is little research on this industry at home and
abroad, so the research subject is novel. Secondly, based on the common frontier and
regional frontier, this paper uses the DDF-GM model to evaluate the EECE of the three
regions, and through a series of spatial analyses, puts forward effective policy measures
to improve EECE.
Innovation Efficiency of Electronic and Communication Equipment Industry 655
QG (x) = Q1 a1 ∪ Q2 a2 ∪ . . . ∪ QT aT (5)
The Global-Malmquist index can be defined as formula 6:
ScoreG(at ,bt )
t
GMt−1 = (6)
( t−1 t−1 )
ScoreG a ,b
Referring to the existing literature, this paper constructs the input-output index system
as follows.
This paper selects high-tech output R&D internal expenditure, high-tech output R&D
personnel full-time equivalent, new product development expenditure as input variables.
They represent the human investment, science and technology financial investment,
and production stage investment respectively. New product sales revenue and patent
applications are output variables. They represent the economic benefit output and R&D
output. The above indicators are all from China Statistics Yearbook on High Technology
Industry.
According to the availability of existing data, 20 provinces in China are selected
as research samples, which are Guangdong, Guangxi, Guizhou, Hebei, Anhui, Beijing,
Fujian, Gansu, Henan, Hubei, Hunan, Jiangsu, Sichuan, Tianjin, Zhejiang, Chongqing,
Jiangxi, Shandong, Shaanxi and Shanghai. They are divided into three regions: the east-
ern region (Hebei, Shandong, Jiangsu, Beijing, Fujian, Guangdong, Zhejiang, Shanghai,
Tianjin), the central region (Hunan, Jiangxi, Anhui, Henan, Hubei), and the western
region (Shaanxi, Sichuan, Chongqing, Gansu, Guangxi, Guizhou).
3 Empirical Research
As can be seen from Fig. 1, under the common frontier, the average EECE in the past
ten years has increased by 6.47%. In 2012, the value of EECE (1.2097) was the highest,
greater than 1 was positive growth, while in 2014, the value of EECE (0.9520) was the
lowest, less than 1 was negative growth. Most of the years are positive growth, only 2014
(0.9520) and 2015 (0.9692) are less than 1. Since the “Western Development” in 2016,
technical support and preferential policies have narrowed the gap between the western
region and advanced regions. In terms of different regions, the values of EECE of 2010
(0.9237), 2013 (0.9461) and 2015 (0.8750) in the eastern region were less than 1, only
2014 (0.8843) in the central region was less than 1, and 2011 (0.9497), 2013 (0.9932)
and 2014 (0.8032) in the western region were less than 1. In terms of decomposition
efficiency, the contribution of TC at the national level is greater than that of EC. In terms
of regions, the growth of EECE in eastern region and western region is mainly supported
by TC. It shows that the eastern and western regions attach importance to the cultivation
of independent innovation ability. However, the growth of EECE in the central region
depends on the support of EC. It shows that the central region pays more attention to the
management.
Innovation Efficiency of Electronic and Communication Equipment Industry 657
Fig. 1. China’s EECE and its decomposition index under meta frontier.
As can be seen from Fig. 2, under the regional frontier, the average EECE in the
past ten years has increased by 3.66%. In 2016, the value of EECE (1.1952) was the
highest, greater than 1 was positive growth, while in 2014, the value of EECE (0.9442)
was the lowest, less than 1 was negative growth. Most years are positive growth, only
2010 (0.9822) and 2017 (0.9992) are less than 1. In terms of different regions, most of
the years in the eastern region are negative growth, but the overall average growth rate
in this decade is 1.94%, only in 2011 (1.0192), 2012 (1.1791), 2014 (1.0826) and 2016
(1.1895) the EECE is greater than 1. In the central region, only 2014 (0.8400) was less
than 1, while in the western region, 2010 (0.9660) and 2014 (0.8235) were less than 1.
From the perspective of decomposition efficiency, the contribution of TC is also greater
than that of EC at the national level. In terms of regions, the growth of EECE in the three
regions is mainly supported by TC.
Generally speaking, whether meta-frontier or regional frontier, the contribution of
technology is greater than that of efficiency. Under the meta-frontier, the EC and TC
were 1.0290 and 1.0445, and that were 1.0073 and 1.0408 under the regional frontier.
Fig. 2. China’s EECE and its decomposition index under regional frontier.
658 Y. Song et al.
As shown in Fig. 3, the highest EECE in Hubei (1.1467) and Jiangxi (1.1668) is at
the common frontier, both of which are located in the central region, which makes
the overall EECE of the central region higher than the other two regions. The values
of EECE of Gansu, Shaanxi, Beijing, Zhejiang, Fujian, Guangdong and Guizhou are
relatively low. EECE in three regions is growing positively. The eastern and western
regions are technology oriented innovation, while the central region is efficiency oriented
innovation progress. The EECE and its decomposition indexes in eastern region were
1.0386, 1.0065 and 1.0360 respectively, those in central region were 1.1212, 1.0787
and 1.0497 respectively, and those in western region were 1.0569, 1.0214 and 1.0529
respectively. The EECE and its decomposition index of the three regions are all greater
than 1, showing a positive growth, indicating that the development of EECE in China is
good at present and has great potential in the future.
Fig. 3. Average EECE in different province and regions under the meta-frontier.
As shown in Fig. 4, the highest EECE in Hubei (1.1609) and Jiangxi (1.1408) is
also at the regional frontier. Both provinces are located in the central region, which
makes the overall EECE of the central region higher than the other two regions. The
values of EECE in Sichuan, Guangxi and Shanghai are relatively low. All three regions
are technological oriented innovation and progress. The EECE and its decomposition
indexes in the eastern region were 1.0194, 1.0106 and 1.0173, respectively, those in the
central region were 1.0816, 1.0119 and 1.0772 respectively, and those in the western
region were 1.0248, 0.9985 and 1.0458 respectively. The EC of the western region is
less than 1, and the TC of the western region is less than that of the central region, which
leads to the lower overall EECE of the western region.
Innovation Efficiency of Electronic and Communication Equipment Industry 659
Fig. 4. Average EECE in different province and regions under the regional frontier.
The overall growth of EECE has a positive growth. Whether in the common frontier
or regional frontier, Hubei and Jiangxi have the highest EECE, and the EECE in the
central region is higher than that in the eastern and western regions. The central region is
located in the middle of China. Its geographical position is superior. It can not only rely
on the advanced technical level of the eastern region, but also rely on the rich human
and geographical resources of the western region. Therefore, its development has more
potential. It is necessary to strengthen regional cooperation and attach importance to
the construction of R&D team. Strengthen cooperation in R&D, market cooperation,
cooperative production, etc.
Based on the above analysis, we conclude that: (1) From 2009 to 2018, the EECE showed
positive growth in most years, and the overall growth is positive. The EECE growth rate
was 6.47% in the common frontier, and 3.66% in the regional frontier. (2) The EECE
in 2009–2018 was the highest in central region, followed by the western region and
lowest in the eastern region. Under the two frontiers, the EECE in the eastern region
increased by 3.86% and 1.94% respectively, while that in the central region increased
by 12.12% and 8.16% respectively, while that in the western region was 5.69% and
2.48% respectively. (3) The decomposition index of EECE is driven by efficiency and
technology, and the contribution of technology is greater than that of efficiency.
In order to improve the overall innovation efficiency, we need to take some measures
to adjust. First, we should strengthen the government’s financial, human and material
investment in innovation and R&D, specify incentive policies, and increase investment
in innovation resources. We should also attach importance to technology introduction
and absorption. At the same time, we should promote the transformation of science and
technology into productivity. Second, it is necessary to strengthen technical cooperation
and exchanges among regions in the electronic and communication equipment manu-
facturing industry. Technological innovation and development have a synergistic effect,
especially the dissemination of advanced technology and ideas, which can accelerate
the division of labor and cooperation among regions and improve the competitiveness
of regions. The third is to enrich the innovation efficiency evaluation system of the
660 Y. Song et al.
Acknowledgments. We thank the funding sponsored by the General Project of Teaching Reform
of Higher Vocational Education in Heilongjiang (SJGZY2019172) and Key Entrusted Project of
Teaching Reform of Higher Vocational Education in Heilongjiang (SJGZZ2019023).
References
1. Sun, C.C.: Assessing the relative efficiency and productivity growth of the Taiwan LED
industry: DEA and Malmquist index application. Math. Probl. Eng. 4, 1–13 (2014)
2. Henisz, W.J., Zelner, B.A.: The institutional environment for telecommunications investment.
J. Econ. Manag. Strategy 10(1), 123–147 (2001)
3. Hemmert, M.: The influence of institutional factors on the technology acquisition performance
of high-tech firms: survey results from Germany and Japan. Res. Policy 33, 1019–1039 (2004)
4. Andonova, V.: Mobile phones, the internet and the institutional environment. Telecommuni-
cations Policy 30, 29–45 (2006)
5. Wang, E.C., Huang, W.: Relative efficiency Of R&D activities: a cross-country study
accounting for environmental factors in the DEA approach. Res. Policy 36(2), 260–273 (2007)
6. Sharma, S., Thomas, V.J.: Inter-country R&D efficiency analysis: an application of data
envelopment analysis. Scientometrics 76(3), 483–501 (2008)
7. Lee, H., Park, Y., Choi, H.: Comparative evaluation of performance of national R&D programs
with heterogeneous objectives: a DEA approach. Eur. J. Oper. Res. 196(3), 847–855 (2009)
8. Guan, J., Chen, K.: Modeling the relative efficiency of national innovation systems. Res.
Policy 41(1), 102–115 (2012)
9. O’Donnell, C.J., Rao, D.S., Battese, G.E.: Metafrontier frameworks for the study of firm-level
efficiencies and technology ratios. Empirical Econ. 34, 231–255 (2008)
10. Wang, Q.W., Zhao, Z.Y., Zhou, P., Zhou, D.Q.: Energy efficiency and production technology
heterogeneity in China: a meta-frontier DEA approach. Econ. Model. 35, 283–289 (2013)
Internet Financial Regulation Based
on Evolutionary Game
1 Introduction
With the development of Internet technology in China, the integration of the Internet
and traditional industries is accelerating. At present, Internet finance has been a new
financial operation mode formed by the deep integration of the Internet and finance. It
has become an important part of China’s financial system [1]. By the end of 2020, the
market size of Internet financial management has reached 25.86 trillion yuan, and it has
generated 993.25 billion yuan of profits for investors. In 2020, the Internet index of the
financial industry was about 30.2, which is much higher than other industries. At present,
the development of the Internet financial shows a diversified competition phenomenon,
and the transaction scale continues to expand. Due to the rapid development of the
financial Internet, the corresponding internet regulatory system is not comprehensive.
There are many problems, such as fuzzy supervisory responsibility attribution, lack of
market access criteria, backward regulatory means, and lagging regulatory laws. The
scale of Internet finance continues to expand. Since 2016, the People’s Bank of China,
China Banking Regulatory Commission, China Securities Regulatory Commission, and
China Insurance Regulatory Commission have decided to jointly regulate the Internet
finance industry and issued a series of regulations. However, these regulations are less
effective because they do not form a complete regulatory system. Based on the current
situation of China’s Internet financial regulatory agency, this paper uses the dynamic
evolutionary game model to analyze the evolution of Internet financial regulation and
tries to provide practical suggestions for the formulation and improvement of relevant
policies.
the players can adopt strategies according to the form of regulation by the regulatory
authorities. If Internet finance companies strictly abide by the regulatory rules, they will
lose the extra benefits of law violations. Conversely, if companies do not comply with
regulations, they may be subject to high penalties while obtaining high benefits.
It is assumed that the Internet financial management platform can obtain normal
returns (α) through legal operation and excess returns (β) through normal financial
innovation. Additional income (θ ) could be obtained through illegal means. If the reg-
ulatory agency regulates, the illegal enterprises will pay fines(M). As far as regulators
664 S. Jiang and Y. Huang
are concerned, the cost of supervision of financial enterprises is C, the overall benefit of
society is H, and the negative impact and externality caused by the illegal behavior of
the Internet financial market is u˛ . Based on the above assumptions, this paper presents
the payoff matrix of both sides in the game, as shown in Table 2.
E 1a = y(H − C + M) + (1 − y)(H − C) = My + H − C
E 1b = −˛uy + 0(1 − y) = −˛uy
E 1 = xE 1a + (1 − x)E 1b = xy(M + u˛ ) − u˛ y + x(H − C) [6].
The above equation shows that no matter what decisions the regulatory authority
makes, its expected revenue can be expressed as a function of x, then we could construct
the evolutionary equation of the regulatory agency’s behavior:
Similarly, the expected returns (E 2j and E2g ) and mean values (E 2 ) of financial enterprises
under different decisions can also be obtained.
E 2j = x(α + θ − M) + (1 − x)(α + θ )
E2g = x(α + β) + (1 − x)(α + β) = α + β
E 2 = yE 2j + (1 − y)E2g = −xyM + y(θ − β) + α + β
Assuming that the financial enterprises to the project information truthfully informed
the truth degree (D), the degree of deception (F) of the financial enterprise and its own
ability (H), the cost is a function of the degree of truthful disclosure (D), deception
degree of financial enterprises (F) and the degree of consumer supervision (K), thus its
profit function can be expressed as G(D, F, H) − Z(D, F, K). If the financial enterprise
tells the true information truthfully, it can temporarily ignore the consumer choice, and
the return function is G(D) − Z(D). If the financial enterprise deceives the consumer, the
666 S. Jiang and Y. Huang
consumer will supervise it, then the income function is G(F) − Z(F) − PF 2 , where P
is the punishment intensity of the fraud; If the financial enterprise cheats the consumer
and they doesn’t supervise, the income function is G(D) − Z(D).
Let’s start to analyse the state of the consumer. If consumers give up supervision,
whether Internet financial enterprises choose to tell real information or not, their return
fuction is G(ω) (ω is the depth of consumer supervision). If the consumer gives up
supervision and the firm tells the truth, the consumer still gets G(ω). If consumers
choose regulation and financial firms commit fraud, the consumer income is G(ω) −
Z(K) + PF 2 . If the consumer chooses to supervise and the financial enterprise tell the
true information of the project, the consumer benefit is G(ω) − Z(K). We can get the
following matrix (Table 4):
The probability of financial enterprises truthfully inform the true information of the
project is x, and the probability of consumers choosing supervision behavior is y. Thus,
the earnings of financial enterprises under different strategies (E1x and E11−x ), and the
mean value is AE 1 .
F(x) = dx/dt = x(E1x − AE1 ) = x(1 − x)(G(D) − Z(D) − G(F) + Z(F) + yPF 2 )
(3)
– Factors influencing the strategic choice of regulatory agencies. The first factor is
the regulatory willingness of regulators. The probability of regulation is positively
correlated with the negative social effects of the industry’s indulgence of bad behav-
iors. The regulation can speed up the prosperity of the Internet financial industry and
reduce the negative impact of enterprises’ illegal activities. Then, the supervisory
authority’s willingness to supervise will decrease, and the systemic risks in the indus-
try will increase. The second factor is the comprehensive social benefits brought by
regulation. The greater the effect of the regulatory authority after disposing of illegal
enterprises, the greater the possibility of the regulatory agency to carry out regulatory
work.
– Factors Influencing Internet Financial Enterprises and Consumers’ Choice
Strategies
The first factor is illegal earnings. If other factors are determined, the benefits brought
by business illegal (cheating) behavior will exceed the costs. The companies will still
consider violations of the law as the best choice even if the regulatory authorities and
consumers conduct strict supervision.
The second factor is the cost of deception. When companies choose whether to cheat
or not, they must consider the cost of the cheating. A company’s willingness to break
the law is inversely proportional to its costs. A single consumer’s ability to participate
668 S. Jiang and Y. Huang
in the game of the Internet finance industry is very limited, and the supervision from
consumers must rely on the entire investment group. Therefore, consumers should pay
more attention to their screening ability and cooperate with government departments to
conduct effective public supervision.
5 Conclusion
Based on evolutionary game theory, this paper constructs a dynamic evolutionary game
model among regulators, consumers, and internet financial companies. As far as business
Internet Financial Regulation Based on Evolutionary Game 669
and regulatory authorities are concerned, the initial game state has an influence on the
stable equilibrium point of repeated games in the process of asymmetric replicative
dynamic evolutionary games. According to the replicating equation of regulators and
financial enterprises, when the violation of the law can bring high profits, the enterprise
will not comply with the law to obtain additional profits, whereas financial enterprises
will comply with the law provisions [10]. Similarly, it will bring great social benefits
when the market without supervision, the regulatory subjects will give up the supervision
of the market. On the contrary, they will carry out a series of supervision activities. As
far as financial firms are concerned, the tendency to tell the true information of the
project increases only with the greater the loss caused by the fraud. The more serious
the phenomenon of corporate fraud and concealment, the higher the willingness of
consumers to supervise. Both enterprises’ and consumers’ decisions need a long-term
evolutionary game to reach a stable state.
References
1. Li, Z.: Thoughts on internet finance. Manag. World 4(7), 1–7 (2015)
2. Cao, D., Cao W., Wu, J.: A game study on financial innovation and regulation in the internet
era. J. Southeast Univ. (Philos. Soc. Sci. Ed.) 16(4), 59–64+135 (2014)
3. Smith, J.: Evolution and Game Theory. Shanghai Fudan University Press (2008)
4. Liu, Y., Xu, C., Yu, P.: Internet finance: origin, risk, and regulation. Soc. Sci. Res. 3, 28–33
(2014)
5. Huang, Y., Xu, H.: On protection of rights and interests of financial consumers in P2P online
lending. Hebei Law 34(9), 16–27 (2016)
6. Zhang, C., Liu, J.: Dynamic evolutionary game between internet financial innovation and
financial regulation. Guizhou Soc. Sci. 1, 151–159 (2020)
7. Wang, T., Qin, J.: Study on optimization of local financial regulatory framework in China: an
analysis based on dynamic evolutionary game model. Shanghai Econ. Res. 4, 14–22 (2016)
8. Song, Y., Xu, Y., Zhang, Z.: Soc. Sci. Res. 4(4), 25–31 (2018)
9. Su, Y., Rui, Z.: Development of internet finance and the emergence of government regulatory
system: based on the perspective of evolutionary game. Finance Account. Bull. 11, 19–22+129
(2015)
10. Liu, H.: On the dilemma of government regulation of internet finance and the way to solve
it. Legal Bus. Stud. 35(5), 58–69 (2018)
Research on Task Scheduling Method of Mobile
Delivery Cloud Computing Based on HPSO
Algorithm
1 Introduction
Particle swarm optimization transforms these massless particles into moving birds with
two properties: speed, how fast or slow they are moving, and position, which is where
they are moving. Every particle has their own behavior, in a certain range of space alone
to find the optimal solution, and record the optimal solution for individual an extreme
value, and to be Shared with other particles, and the extreme value is taken as the overall
optimal solution, and other particles adjust their current state according to this value.
Its advantages such as strong optimization ability and easy implementation have been
used to solve many practical problems [1]. It is a search algorithm in a population, using
individuals in the population to search a certain space of value. In this particular case,
the population can be called a swarm, and the individuals can be regarded as particles.
Other particles adjust their current state based on this value and speed and record the
best position during the adjustment.
With the continuous progress of contemporary society, more and more people buy
the goods they need through online shopping, which brings the number of orders is
increasing, so how to carry out distribution also has It has become a problem that needs
to be solved at present. For now, the vast majority of logistics in our country’s road
transportation, or a few special cases with the method of rail and air transport, however,
in this process, the number of trucks that can be put into use has a certain limit, so how to
use the limited conditions to achieve the lowest cost and the fastest delivery of logistics
is one of the goals that all logistics companies need to achieve. Therefore, this article
takes an in-depth look at cloud computing task scheduling. The main contents of this
article are as follows:
(1) Firstly, it summarizes many algorithms of the swarm intelligence algorithm and
related research on cloud task scheduling and finds that task assignment is
inseparable from the swarm intelligence algorithm.
(2) The task scheduling problem of cloud computing was studied. Under time con-
straints, road gap, road accessibility, and driving cost, the task assignment model
was established with the optimal path as the optimization goal.
(3) Integrating the relevant PSO in swarm intelligence, an HPSO task scheduling
method is proposed to find the most appropriate path under relevant constraints
(4) Through relevant simulation experiments to ensure the feasibility of task scheduling
method, the leadership of HPSO is obtained from the algorithm fitness and the
shortest path
2 Related Work
Cloud task scheduling is mainly divided into single cloud task scheduling and cross-cloud
task scheduling, optimized through different methods.
In terms of task allocation, the research results of relevant scholars are as follows:
Entisar S et al. proposed a new resource allocation model for task scheduling based on
multi-objective optimization (MOO) and PSO algorithm because of the complexity of
task scheduling and the complexity of clients’ demands on execution time and through-
put. On this basis, a new multi-objective PSO is proposed. Execution time and wait time
are shortened, and throughput is improved [2]. Awad AI et al. Proposed the task model
of lmbpso under the constraints of reliability and time, saving the cost and obtaining a
feasible plan [3]. Fahimeh ramezani et al. Established a multi-objective comprehensive
model of task optimization with the ultimate goal of minimum time [4]. Masdari M
et al. analyzed the cloud environment task and workflow scheduling scheme. They pro-
posed the particle swarm optimization model, which improved the efficiency and solved
the appropriate task and workflow scheduling [5]. According to the most difficult job
scheduling problems in parallel distributed computing environments such as clusters,
grids, and clouds, Kaur N et al. used meta-heuristic algorithms such as genetic algorithm
and ant colony algorithm to solve the approximate optimal solution of job scheduling
problems and effectively utilized computing resources [6].
As for the swarm intelligence algorithm, related scholars’ research results are as
follows: Yousif s et al. Came up with a new method. Based on the BSA algorithm,
swarm intelligence algorithms such as the bat algorithm solve an optimization module
of variables with different relationships in different ranges [7]. Xu Z et al. The effects
672 J. Li et al.
3 Model Construction
Based on the assumption that many logistics enterprises take road transportation as the
condition, M delivery vehicles will deliver N pieces of goods to the designated place.
Assuming all the restrictions are met, all the designated locations have been visited,
and the delivery vehicles eventually return to its point of departure. For the multi task
allocation of logistics, the following assumptions are made:
(1) Starting from the departure of the delivery vehicles and ending of the task, the task
will not be changed during the process. That is to say, the customer will not cancel
or modify the logistics information.
Research on Task Scheduling Method 673
(2) When carrying out logistics distribution, all delivery vehicles should have no
ongoing or pending distribution tasks.
(3) All trucks will depart from the logistics company and return to the company after
the task returns.
Under the condition of the above hypothesis, the logistics company received logistics
task, the task to the cloud platform, cloud platform to classify different tasks, will be
neat, the same type of task and then distributes tasks to delivery vehicles drivers, delivery
vehicles drivers perform tasks and logistics information on to the cloud platform, the
platform to record the number of delivery vehicles driver to perform a task, And feed
back the information uploaded by the delivery vehicles driver to the logistics company.
See the figure below (Fig. 1):
The optimization of the current logistics distribution problem is mainly the process
of optimizing by shortening the time and distance. In real life may appear crowded
roads, road construction, road restrictions such as special cases, by adding the time
constraints, road space constraints, the smooth general characteristic, road and driving
cost constraints of the four different constraint conditions, improve the accuracy of the
logistics distribution route planning in the process, find the shortest path distribution
path, The optimization objective function in the process of task scheduling is presented
as follows:
D = min( t
Xab dab ) (1)
a∈ZM b∈ZM
674 J. Li et al.
Constraints include:
⎧
⎪
⎪ T ≤ Tc
⎨ t
Xab tab > tc
(2)
⎪ t > zfc
⎪ Xab zfab
⎩
C=k (1 + gab )Xab
t d <f
ab max
Among them,at that time,dab ∈ M judge whether the current node is on the path,
if so, then let Xtab = 1, not so Xtab = 0; zfij ∈ (0, 1), zfij the closer to 1, the higher
the driving comfort coefficient zfc indicates minimum comfort during the driving of the
delivery vehicles. In the interval, when the comfort level is lower than this value, the
delivery vehicles driver refuses to drive; tab ∈ (0, 1), tab the closer to 1, the lower the
degree of traffic jam; tc represents the minimum acceptable congestion threshold. When
this value is 0, indicates that this part is not allowed to pass (Table 1).
Symbol Explain
M The set of paths that a point of departure contains
dab how far are a and b
t
Xab Judge the location of the path
ZM The set of network nodes for the path
T Total travel time of all transport vehicles
Tc Total maximum time consumed by all logistics vehicles
zfab The comfort factor for driving between nodes a and b
tab The level of traffic jams on the roads
C Total cost in logistics distribution process
k Average cost of fuel per kilometer
g Percentage increase in fuel consumption under different driving conditions
fmax Maximum allowable cost of vehicles during distribution
Vab The velocity from a to b
tc Minimum acceptable congestion threshold
Research on Task Scheduling Method 675
The standard PSO algorithm is mainly applied to the optimization of continuous solution
space problems. Its mathematical description is as follows:
It is assumed that the problem solved is carried out in n-dimensional space and the
number of particle swarm is M. With the occurrence of iteration, the position of particle
will also change correspondingly. It is suppose that the position where particle I reaches
after t iterations is expressed as Xi (t) = {vi1 (t), vi2 (t), ..., vin (t)}, i = 1, 2, ..., M , And
the corresponding velocity of the particle is denoted as vi (t) = {vi1 (t), vi2 (t), ..., vin (t)},
Then the JTH (j = 1,2… n) is expressed in the following formula:
The position update of particle i at time t can be calculated by the following formula
In the formula, w is the inertia weight, c1 and c2 are the acceleration factors, r1 and
r2 are the random numbers in [0,1], and gj is the optimal value in the j dimension.
Hybrid particle swarm algorithm flow chart as shown, first on the particle swarm ini-
tialization, set the number of initial population, and then the fitness calculation, it is
concluded that fitness curve, and then according to the fitness value of particles are
updated, find out the optimal particle and group optimal individual particles, individual
respectively and the optimal particle and crossover operation for the optimal particle
groups and groups, Each produces a new particle after the intersection, and this particle
after the mutation will produce another new particle (Fig. 2).
676 J. Li et al.
Where, n is the number of cities, and patha,b is the distance between two cities
ab.
3. Cross operation
Cross the individual extremum and the whole extremum at a certain position to form
a new individual. The operation is as follows:
[5 1 4 2 3] cross
−→[5 2 1 2 3]
[5 4 1 3 2]
Because of the special case of the crossover operation, if the repeated particle
exists in the new individual generated after the crossover operation, then the repeated
Research on Task Scheduling Method 677
4. Variation operation
5 Simulation Experiment
5.1 Experimental Environment and Parameter Setting
The hardware environment of the experiment is: Intel (R) Core (TM) I5-6200U CPU @
2.30 ghz, 12.00 GB memory, Windows10 system, Matlab R2016b simulation platform.
Simulation task C = 31, maximum number of iterations = 500, Ca = 200, p = 1, the
parameters are set as follows (Table 2).
(1) Fitness changes of particle swarm optimization algorithm (PSO) and mixed PSO
The maximum iteration of the algorithm is 500 times. The fitness change curves of
PSO and hybrid PSO algorithms are shown in Fig. 3.
The position of each city is plotted according to the coordinate points of each city,
as shown in Fig. 4. The shortest distance is obtained by the hybrid particle swarm
optimization algorithm:15601.9195. There may be many shortest paths, but the shortest
distance is only one. According to Fig. 6, you can see the change of distance. It is not
difficult to find that both the shortest distance and the average distance decrease in a
certain interval and then plateau in the later period (Fig. 5).
Fig. 4. Schematic diagram of city location Fig. 5. Optimize the shortest path diagram
Research on Task Scheduling Method 679
Fig. 6. Comparison of the shortest distance and average distance in each generation
6 Conclusion
With the optimization target, this paper to realize the shortest path problem with time
constraints, road space constraints, the road smooth general characteristic and driving
cost constraints, four constraint conditions of cloud computing task scheduling model
is established, based on cloud computing task scheduling model, the collection related
features of swarm intelligence algorithm, hybrid particle swarm algorithm in cloud com-
puting task scheduling method, Each city is simulated as a point on the coordinate axis,
and the optimal path is selected. Through the simulation experiment, the approximate
distribution map of the city is drawn, and the optimal path is obtained through the hybrid
particle swarm optimization algorithm, which greatly reduces the transportation distance
and cost, and improves efficiency.
Acknowledgment. This work is partly supported by the project supported by the National Social
Science Foundation (16BJY125), Heilongjiang philosophy and social sciences research planning
project (19JYB026), Key topics in 2020 of the 13th five year plan of Educational Science in Hei-
longjiang Province (GJB1320276), Project supported by under-graduate teaching leading talent
training program of Harbin University of Commerce (201907), Key project of teaching reform and
teaching research of Harbin University of Commerce in 2020 (HSDJY202005(Z)), Innovation and
entrepreneurship project for college students of Harbin University of Commerce (202010240059),
School level scientific research project of Heilongjiang Oriental University (HDFKY200202), Key
entrusted projects of higher education teaching reform in 2020 (SJGZ20200138).
References
1. Wang, R.: A novel hybrid particle swarm optimization using adaptive strategy[J]. Inf Sci 579,
231–250 (2021)
2. Alkayal E, Jennings R, Abulkhair F (2017) Efficient task scheduling multi-objective particle
swarm optimization in cloud computing[C]. Local Computer Networks Workshops. IEEE
680 J. Li et al.
3. Awad, I., El-Hefnawy, A., Abdel, K.M.: Enhanced particle swarm optimization for task
scheduling in cloud computing environments[J]. Procedia Comput Sci 65(1), 920–929 (2015)
4. Ramezani F, Jie L, Hussain F (2013) Task scheduling optimization in cloud comput-
ing applying multi-objective particle swarm optimization[C]. International Conference on
Service-Oriented Computing. Springer, Berlin, Heidelberg
5. Masdari, M., Salehi, F., Jalali, M.: A survey of PSO-based scheduling algorithms in cloud
computing[J]. J Netw Syst Manag 25(1), 122–158 (2017)
6. Kaur, N.: Comparative analysis of job scheduling algorithms in parallel and distributed
computing environments[J]. Int J Adv Comput Res 8(3), 948–956 (2019)
7. Yousif, S., Saka, M.P.: Enhanced beetle antenna search: a swarm intelligence algorithm. Asian
J Civ Eng 22(6), 1185–1219 (2021). https://doi.org/10.1007/s42107-021-00374-z
8. Xu, Z., Ribaudo, N., Li, X.: A genome-scale modeling approach to investigate the antibiotics-
triggered perturbation in the metabolism of pseudomonas aeruginosa[J]. IEEE Life Sci Lett
99, 1 (2017)
9. Mashwani, K., Hamdi, A., Jan, M.: Large-scale global optimization based on hybrid swarm
intelligence algorithm[J]. J Intell Fuzzy Syst 39(1), 1257–1275 (2020)
10. Mortazavi A, Togan V, Moloodpoor M (2019) Solution of structural and mathematical opti-
mization problems using a new hybrid swarm intelligence optimization algorithm[J]. Adv
Eng Softw 127(Jan):106–123
11. Mathew, T., Paul, A., Rojan, A.: Implementation of swarm intelligence algorithms for path
planning[J]. J Phys Conf Ser 1831(1), 012008 (2021)
12. Moon J, Yu C, Gil M (2017) A slave ants based ant colony optimization algorithm for task
scheduling in cloud computing environments[J]. Hum -centric Comput Inf Sci 7(1):28
13. Reddy, G., Phanikumar, S.: Multi objective task scheduling using modified ant colony
optimization in cloud computing[J]. Int J Intell Eng Syst 11(3), 242–250 (2018)
14. Geng, S., Wu, D., Wang, P.: Many-objective cloud task scheduling[J]. IEEE Access 99, 1
(2020)
Research on the Evaluation of Forestry Industry
Technological Innovation Ability Based
on Entropy TOPSIS Method
Shangkun Lu(B)
1 Introduction
The forestry industry is a basic industry for developing the national economy and a key
industry that coordinates ecological, economic, and social benefits. The “five-in-one”
modernization construction pattern clarifies the dominant position of ecological civi-
lization construction and highlights the fundamental support of forestry to social and
economic development. Technological innovation in the forestry industry is an important
guarantee for optimizing the structure of the forestry industry, adjusting the layout of the
forestry industry, improving the efficiency of forest resource utilization, and improving
the quality of forestry industry development. It is also a key factor in realizing the inten-
sive, large-scale, orderly, and sustainable forestry development. Based on the research
results of forestry industry technology innovation at home and abroad, this research will
conduct an empirical analysis of the forestry industry technology innovation capabilities
of 29 provinces in China through literature statistics, objectively evaluate the true level
of my country’s forestry industry technology innovation capabilities, and provide it for
China’s forestry industry Provide pragmatic support for the sustainable development of
the industry.
Changbing Chen proposed that to enhance my country’s industrial technological
innovation capabilities, investment in technological innovation should be increased, and
the efficiency of technological innovation should be improved [1]. Baoming Chen pro-
posed building an evaluation index system of innovation capability based on influencing
factors [2]. Yong Fu explored the connotation, material basis, approach model, basic
management form and industrial attributes of sustainable forestry development, and the
importance of technological innovation in it, thereby in-depth analysis and revealing that
technological innovation is the core driving force of sustainable development forestry
development [3]. Shaopeng Zhang and Hongge Zhu used the spatial Dubin model to
prove that the development of the forestry industry has significant spatial spillover
effects [4]. Zhang Zeng and Wen Fan defined the concept of high-tech industry, and
they constructed an evaluation index system for technological innovation and industrial-
technological innovation capabilities [5]. Yun Chen and Chunfang Tan constructed a
theoretical system of evaluation indicators from two aspects of enterprises’ technologi-
cal innovation capability and innovation efficiency [6]. Yuwei Du and Zhifang Wan used
GEMS model and index system to measure the technological innovation capability of
forestry industry [7]. Zhifang Wan and Xiaolin Ma Use the entropy method to measure
the forestry industry’s technological innovation capability index and propose that tech-
nological innovation is the key to transformation [8]. Baiqing Ye believes that internal
expenditure of R&D expenditure, new product development expenditure, number of
scientific research personnel, full-time equivalent of R&D personnel and expenditure
of digestion and absorption expenditure are important factors affecting technological
innovation capability [9]. Gang Wang and Wei Chen used the Entropy-Topsis model
to empirically measure the competitiveness of the national forestry industry based on
inter-provincial cross-sectional data [10].
To sum up, the development of forestry industry has been related to China’s ecolog-
ical environment construction, social and economic development, the industrial social
responsibility of the benign recycling of resources. It is also a necessary prerequisite
for the protection and sustainable utilization of forest resources. The evaluation of tech-
nological innovation ability of forestry industry is a comprehensive process involving
multi-objective and multi-index, but at present, the index weight mostly adopts subjec-
tive assignment method, which makes the evaluation subjective and objective. In this
paper, the technological innovation ability of forestry industry is taken as the research
point. The entropy weight TOPSIS method is used to empirically measure the innovation
data of forestry industry in 29 provinces of China, objectively analyze the main influ-
encing factors of the technological innovation ability of forestry industry in China, and
put forward countermeasures and suggestions conducive to the sustainable development
of forestry industry.
Research on the Evaluation of Forestry Industry Technological Innovation Ability 683
The smaller the result di∗ , the larger solution di− ,The better the plan.
(5) Calculate the closeness of each index to the optimal index.
The final relative closeness of Scheme Ai to the ideal solution is.
di∗
ci∗ = (1 ≤ i ≤ m).
di− + di∗
3 Empirical Research
The use of the entropy method, because of its objectivity, is conducive to calculation and
evaluation. Therefore, this article uses cross-sectional data to calculate the arithmetic
average of each indicator in the four-year data of 29 provinces in the country from 2015 to
2018 to eliminate the accidental factors of a single year’s data. Calculate the normalized
results of the original data, and get the entropy matrix and weight matrix of each index,
as shown in Table 1 and Table 2.
From this, it can be concluded that the average weight of each level of the forestry
industry’s innovation ability in the country’s provinces from 2015 to 2018. Capability
of the forestry industry, the management innovation ability of the forestry industry has
the largest weight, and the overall entropy is the smallest; the innovation ability of the
forestry industry and the actual utilization of foreign capital in the forestry have the
largest weight; the innovation ability of the forestry market, The timber sales volume
and forest product sales revenue in state-owned forest areas have a greater impact; the
utilization capacity of forestry scientific and technological innovation resources and the
total output value of forest products have the largest impact.
Index D1 D2 D3 D4 D5 D6 D7 D8 D9 D10
Entropy 0.94 0.86 0.90 0.93 0.87 0.91 0.89 0.76 0.93 0.91
Index D11 D12 D13 D14 D15 D16 D17 D18 D19
Entropy 0.91 0.88 0.73 0.86 0.72 0.83 0.43 0.93 0.94
Index D1 D2 D3 D4 D5 D6 D7 D8 D9 D10
Entropy 0.02 0.06 0.04 0.03 0.05 0.04 0.05 0.07 0.02 0.02
Index D11 D12 D13 D14 D15 D16 D17 D18 D19
Entropy 0.03 0.04 0.08 0.05 0.10 0.06 0.19 0.02 0.02
Calculate the above four dimensions of forestry industry innovation capability and the
distances of positive ideal solutions, negative ideal solutions, optimal solution closeness,
and the distribution ratio for decision-making, and rank according to closeness. The final
calculation result is shown in Fig. 1.
It can be seen from Fig. 1 that the static evaluation results of the forestry industry’s
innovation ability in Guangdong Province are the highest. Its resource utilization ability,
market innovation ability, technological innovation support ability and management
innovation ability all rank the highest, indicating that the static state of the forestry
686 S. Lu
1
0.8 Resource ulizaon
0.6
0.4
0.2 Market innovaon ability
0
Inner…
Jiangxi
Beijing
Fujian
Hubei
Hebei
Zhejiang
Shandong
Liaoning
Shaanxi
Tianjin
Yunnan
Chongqing
Gansu
Guangdong
Technical innovaon
support ability
0.6
0.4
0.2
0 Inner…
Guangd…
Heilongji…
Jiangxi
Beijing
Jiangsu
Fujian
Hubei
Shanghai
Guangxi
Guizhou
Shaanxi
Qinghai
Hebei
Sichuan
Zhejiang
Shandong
Liaoning
Jilin
Tianjin
Shanxi
Henan
Hunan
Yunnan
Chongqing
Xinjiang
Gansu
Anhui
Hainan
Results: First, for the forestry industry’s technological innovation resource utilization
capacity index, the optimal solution closeness of most provinces is below 0.3, accounting
for 70% of all samples. The proximity of Guangdong, Sichuan, Zhejiang and Hunan is
above 0.5. It can be found that the output value of the tertiary industry has a high impact
on the resource utilization capacity of the forestry industry. Second, the forestry industry
market innovation capability index, most provinces have an optimal solution closeness
below 0.3, and only Guangxi has a closeness above 0.5 to the optimal solution. In the
forestry industry’s market innovation capability, the most weighted indicators are the
wood sales volume of state-owned forest areas and the total amount of major economic
forest products. Third, it affects the discovery of the forestry industry’s technological
innovation support ability, the sales revenue of forestry products and the amount of
forestry investment. Fourth, analyze the indicators that affect the management innovation
ability of the forestry industry. The actual use of foreign investment in the forestry
industry has a relatively important impact on the management innovation of the forestry
industry.
Research on the Evaluation of Forestry Industry Technological Innovation Ability 687
Panel data from 2006 to 2018 were used for dynamic evaluation. The standardized results
of original data are calculated to obtain the entropy matrix and weight matrix of each
index and the average weight of innovation capability index of the forestry industry in
each province.
The secondary indicators of forestry industry innovation capability, the forestry
industry market innovation capability has the largest weight, and the overall entropy
is the smallest; among the forestry industry’s market innovation capability, the forestry
industry The contribution rate and government financial investment of forestry industry
have relatively large influencing factors; among the indicators of forestry industry inno-
vation capability, the number of state-owned units in the forestry system has obvious
factors; the forestry industry’s technological innovation supportability, the most influen-
tial factor is the amount of forestry investment; forestry industry technology innovation
resources Utilization capacity, the most influential factor is the gross value of tertiary
industry forest output.
By calculating the four dimensions of forestry industry innovation ability and the
distances of positive ideal solutions and negative ideal solutions, the closeness to the
optimal solution and the distribution ratio for decision-making, and ranking accord-
ing to closeness, the final calculation results. First, from 2006 to 2018, the forestry
industry’s innovative resource utilization capacity and the optimal solution approaching
degree increased. In 2012 and 2014, the approaching degree declined, and in 2010, the
approaching degree rose to more than 0.5. Among the forestry industry’s innovative
resource utilization capacity indicators, the main factors leading to the decline are the
growth rate of the total output value of the forestry industry and the total value of the
tertiary industry forestry. Second, in 2013, in addition to timber sales in state-owned
forest areas, the number of forestry enterprises, and educational funding, the proportion
of forestry industry structure, government financial investment in the forestry industry,
the total amount of major economic forest products, and the contribution rate of the
forestry industry were all significantly higher than in previous years.. Third, the degree
of closeness to the optimal solution fluctuates greatly in the forestry industry’s techno-
logical innovation supportability. There was a decline in 2012 and 2014, and a greater
improvement was achieved in 2013. The original data shows that the technical inno-
vation support capacity of forestry products is closely related to product sales revenue
and comes from the market. Fourth, in terms of forestry industry management inno-
vation ability, the degree of closeness to the optimal solution fluctuates greatly, with a
substantial increase in 2013 and a slight decline in 2012.
4 Conclusion
Through the entropy method and TOPSIS method, the forestry industry of 29 provinces
in my country is statically and dynamically evaluated. Find out the advantages and prob-
lems of forestry technology innovation in various provinces. Improve the forestry indus-
try technology innovation evaluation system and assessment indicators. Build a forestry
688 S. Lu
References
1. Chen, C.B.: Measurement and analysis of the degree of industrial technology innovation in
my country. Economic Aspects 6, 76–87 (2019)
2. Chen, B.M.: Research on my country’s industrial technology innovation capability evaluation
index system. Sci Technol Ind 6(11), 22–25 (2006)
3. Fu, Y.: Technological innovation is the core power of sustainable development of forestry
economy. Dev Res 143(S1), 133–135 (2009)
4. Zhang, S.P., Zhu, H.G.: The spatial impact mechanism of China’s forestry industry develop-
ment——a spatial econometric analysis based on panel data of 31 provinces. World For Res
32(2), 61–66 (2019)
5. Zeng, Z.N.: Research on the evaluation index system of technological innovation capability of
high-tech industry. Journal of Xi’an Shiyou University (Social Science Edition) 23(1), 6–10
(2014)
6. Chen, Y., Tan, C.F.: Research on the evaluation index system of technological innovation
capability of small and medium-sized technological SMEs. Science and Technology Progress
and Policy 29(2), 110–112 (2012)
7. Du, Y.W., Wan, Z.F.: Research on the path choice of forestry industry transformation in state-
owned forest regions of heilongjiang province. Issues of Forestry Economics 39(03), 25–33
(2019)
8. Wan, Z.F., Ma, X.L.: Dynamic evaluation of technical innovation ability of wood industry
based on entropy method. Stat Decis 541(01), 74–78 (2020)
9. Ye, B.Q., Wei, W.: Research on the influencing factors of technological innovation capability
of high-tech industries in the three northeast provinces based on PLS. Journal of Harbin
University of Commerce (Social Science Edition) 3, 122–128 (2015)
10. Wang, G., Chen, W.: measurement of forestry industry competitiveness based on entropy-
topsis. Stat Decis 534(18), 57–60 (2019)
Research on the Impact of Temporary Workers’
Psychological Contract Fulfillment on Task
Performance in the Sharing Economy
1 Introduction
The emergence of the sharing economy enables companies to hire labor from online
platforms. According to the “Annual Report on China’s Sharing Economy Development
(2021) [1]”, sharing economy platform enterprises have about 6.23 million staffs, and
platform employees have become a huge new labor group. In the practice of sharing
employment management, there is no substantial labor relationship between employ-
ees and the company. The flexibility of working time and space subvert the traditional
work model under the employment relationship. It is difficult to directly manage and
motivate employees, and they have low loyalty and enthusiasm, leading to inefficiency
and absenteeism. Currently, it is extremely important to form an implicit and unwritten
psychological contract to regulate the relationship between the two parties. The psycho-
logical contract refers to the belief that staff members hold to the exchange agreement
between themselves and the organization. It is a theoretical framework for studying
employee attitudes and behaviors [2]. The psychological contract fulfillment will pro-
mote the employee’s trust in the organization and then affect the staff attitudes and acts,
including organizational identification, satisfaction, and task performance [3]. In social
exchange theory, employees show certain actions according to certain benefit exchange
principles. When staffs feel that the psychological contract is fulfilled well, they tend
to play a higher level of performance in return [4]. In the sharing economy, temporary
workers face unstable employment conditions. Encouraging staff to work and improve
their sense of psychological contract fulfillment is the core of stabilizing employment
relationships and improving performance. It is also one of the most urgent challenges
faced by Chinese companies.
Organizational identity refers to the fact that individuals regard themselves as part of
the organization, which depends on the significance of their relationship to the organiza-
tion [5]. When the psychological contract is destroyed, it will adversely affect the staff’s
sense of belonging [6]. The psychological contract fulfillment affects the organizational
identity and then the service-oriented organizational citizenship behavior [7]. Temporary
workers’ needs for affiliation and social connections positively influence organizational
recognition and commitment, which in turn affects its task performance. With the exten-
sion of working years, staffs have different expectations of the organization. Previous
research has discovered that the relationship between psychological contract fulfillment
and performance changes with increasing working years [8]. In other words, work-
ing years regulate the relationship between psychological contract fulfillment and work
results.
The existing research on shared employment mainly focuses on the characteristics of
the new human resource management model [9]. Few scholars discuss the driving mech-
anism of temporary labor performance from the viewpoint of employee psychological
contracts. Employees on the platform are independent of the organization, and their work
style is different from that of the traditional employees, assumed by the organizational
behavior theory. Because of this point, this article studies the relationship between the
performance of the psychological contract and the task performance of shared tempo-
rary workers from the angle of social exchange theory, divides the psychological contract
into two dimensions, and introduces the mediation variables of organizational identity
and the adjustment of working years. A logical model is established, constructing a
more effective theoretical framework for future research on psychological contract and
task performance and helping corporate managers in the sharing economy make better
strategies conducive to improving employees’ performance. In the face of the new labor
market conditions, this study tends to provide the government with suggestions on how
to guarantee sufficient talent and promote the sustainable development of enterprises.
the moderating effect of working years, and makes assumptions reviewing the above
variables.
employee identity. The perception of the individual and the organization is synchronized
and consistent. When employees perceive their existence in the enterprise, they will have
a perception of group affiliation to the organization.
Psychological contract fulfillment is a vital driving force for staff organizational
identification [21]. Employees and the organization rely on emotional commitment and
mutual respect to maintaining the relationship. The more support and opportunities the
organization provides, the stronger the employee’s sense of belonging is. In transactional
contracts, the most important thing for shared employees is the short-term economic
return to satisfy their purpose of joining the organization and improving their sense of
organizational identity [22]. Therefore, this research believes that temporary workers
perceive the higher level of psychological contract fulfillment in the sharing economy,
the greater the sense of identity with the organization they will have.
H2a: The fulfillment of transactional psychological contracts perceived by temporary
workers in the sharing economy has a significant positive correlation with organizational
identity.
H2b: The fulfillment of the relational psychological contract perceived by temporary
workers in the sharing economy has a significant positive correlation with organizational
identity.
The identification of staffs to the organization depends on their perception that they
are legal members of the organization. A high-level psychological contract can narrow
the distance between the two parties and make the cooperation between the two parties
closer. While participating in organizational affairs, employees will continue to exer-
cise and rebuild their values. With the gradual unification of the values of both sides,
the individual’s sense of identity to the organization will gradually deepen [23]. Staffs
who have a forceful sense of identity with the organization are more likely to show a
supportive attitude towards the organization, have a stronger sense of belonging to the
organization, and tend to take actions that are helpful to the organization [24]. Therefore,
organizational identity is positively correlated with job performance [25]. Organizational
identity plays an important mediating role between psychological contract and staff’s
role performance [26]. The employee’s sense of the psychological contract fulfillment
can generate trust, loyalty, and a sense of identity, motivating employees’ organizational
citizenship behavior [27].
H3: Organizational identification is positively correlated with the task performance
of temporary workers in the sharing economy.
H4a: In the sharing economy, the identification of temporary labor organizations
plays an important mediating role in the relationship between perceived transactional
psychological contract fulfillment and task performance.
H4b: Shared temporary labor organization identity plays a significant mediating role
in the relationship between perceived relational psychological contract fulfillment and
task performance.
Research on the Impact of Temporary Workers’ Psychological Contract 693
Length of service means the number of years staff has worked for a specific organization
[28]. Norris and Niebuhr (1984) believe that length of service regulates the relationship
between job satisfaction and job performance [29]. Wright and Bonett (2002) use meta-
analysis to reveal the moderating effect of working years on organizational commitment
and job performance [30]. Compared with short-term employees, long-term staffs have
a more stable attitude towards the enterprise and have less influence on the performance
of the psychological contract. Acceptance and recognition by organizations are the most
concerning issues for new employees. The improvement of salary and performance is the
most important early signal of new employees’ perception and acceptance. The above
research shows that transactional and relational motivations can increase the organiza-
tional identity of new employees. In shared employment, staffs have a weaker sense
of identity with the organization. With the increase of working years, employees will
adapt to the values of the organization. The relationship between employees’ manner,
perception and action will change with their length of service. As a result, this research
believes that psychological contract fulfillment on the organizational identity of shared
temporary workers with short working years is more significant than employees with
long working years.
H5a: Transactional psychological contract fulfillment has a stronger influence on
the organizational identity of temporary workers in the sharing economy with shorter
service than on employees with longer service.
H5b: Relational psychological contract fulfillment has a stronger influence on the
organizational identity of shared temporary workers with shorter service than employees
with longer service.
Bal et al. (2013) points out that length of service can moderate the relationship
between psychological contract and work outcome [8]. The behavior of employees with
short service is mainly driven by economic exchange, while the action of employees with
long service is mainly driven by emotional factors [30]. New employees’ investment in
work is mainly based on the compensation they receive from the organization. In contrast,
older employees have a weaker perception of the connection between effort and their
compensation [31]. When employees devote themselves to work and are willing to stay in
the organization over time, these employees will be more aware of their responsibilities
than those who do not want to stay, and are more likely to contribute to the organization’s
future. The performance of employees with long working years is related to the relational
psychological contract fulfillment they perceive [32]. Therefore, this research considers
that the relationship between psychological contract fulfillment and task performance is
moderated by working years.
H6a: Transactional psychological contract fulfillment has a stronger impact on the
task performance of temporary workers in the sharing economy with shorter service than
on of employees with longer service.
H6b: Relational psychological contract fulfillment has a stronger impact on the task
performance of temporary workers in the sharing economy with shorter service than on
that of employees with longer service.
694 G. Zhang et al.
3 Research Design
Through a theoretical review, the intermediary moderating effect model of psychological
contract performance and task performance perceived by temporary workers in the shar-
ing economy is shown in Fig. 1. The research is carried out by adopting a questionnaire
survey as the major approach. The survey objects and samples are selected, combined
with the actual situation in China, and the appropriate questionnaire scale is selected.
After modification and improvement, a formal questionnaire is formed.
The reliability test uses Cronbach’s α coefficient, and the results show that the reliability
coefficients of the four scales are 0.881, 0.851, 0.887, 0.816, and the overall questionnaire
reliability is 0.889. All these figures are greater than the high-reliability standard of 0.7,
indicating that the internal consistency of the questionnaire is good.
Since this study is made on domestic and foreign mature scales, the content validity
is good. The construct validity is tested by aggregate validity and discriminative validity
[33]. According to Table 2, all factor loads in the measurement model are more than
0.5, the AVE (average extraction variation) value is more than 0.5, and the combination
reliability is greater than 0.7, indicating good aggregation validity.
696 G. Zhang et al.
This paper compares the average extracted amount of variation with the square of
the correlation coefficient to judge the discriminative validity. Table 3 shows that all
AVE square roots extracted are higher than the correlation between the target variable
and other variables, so the variables have a good discriminative validity.
Path Test of Mediation Effect. In this study, the bootstrap method is used to analyze
the mediation effect, the bootstrap sampling value is set to 2000. As shown in Table 4,
the test findings indicate that the confidence interval corresponding to each path of the
test does not contain 0, p < 0.01. The mediating effects are statistically significant. Com-
pared with the direct effect (β = 0.562, p < 0.01), transactional psychological contract
fulfillment has a relatively weaker influence on task performance through organizational
identification, but the indirect effect is notable (β = 0.071, p < 0.01), which tests the
hypothesis H4a. Relational psychological contract fulfillment has a significantly indi-
rect influence on task performance through organizational identification (β = 0.122, p
< 0.01), which supports hypothesis H4b.
Moderating Effect Test. In order to further explore the moderating effect of length of
service, this study uses Model 59 in SPSS PROCESS 3.3 to test the moderating effect
[34]. This research divides the length of service sample into a high service years group (2–
3 years, >3 years) and a low service years group (≤1 year, 1–2 years). The research results
show that length of service plays a negative moderating role between the performance of
transactional psychological contract and organizational identification (B = -0.169, p <
0.01). The slope in Fig. 3a shows that temporary workers with low working years (simple
slope = 0.611, p < 0.001) have a greater impact on the performance of transactional
psychological contracts than temporary workers with high working years (simple slope
= 0.272, p < 0.001). The organization agrees, therefore, hypothesis 5a is supported.
However, the moderating effect of length of service in relational psychological contract
fulfillment and organizational identification is non-significant (B = -0.024, p = 0.680),
as shown in Fig. 3b, and hypothesis 5b is not valid. Figure 3c shows that the moderating
effect of working years between transactional psychological contract fulfillment and
task performance is non-significant (B = 0.019, p = 0.738), and hypothesis 6a has not
been verified. Length of service has a positive moderating effect on the fulfillment of
relational contract and task performance (B = 0.139, P < 0.01). Relational psychological
contract fulfillment perceived by temporary workers with high working years (simple
slope = 0.592, p < 0.001) has more influence on task performance than that perceived
by temporary workers with low working years (simple slope = 0.312, p < 0.001) does.
As a result, hypothesis 6b is supported.
5 Discussion
The employment relationship in the sharing economy is flexible. There is no clear labor
relationship between employees and employers. Compared with other organizational
Research on the Impact of Temporary Workers’ Psychological Contract 699
behaviors, individuals tend to focus more on their task performance. Most temporary
workers are young people who have no economic savings for the time being, and they
are in urgent need of stable income to meet their own living needs. Therefore, in the short
term, employees’ work motivation largely comes from obtaining labor remuneration and
they are more sensitive to economic exchange. The conclusions of this study find that
both transactional and relational contract fulfillment impact task performance, while
transactional contract has a more greatly direct influence on task performance. This
conclusion demonstrates how different dimensions of psychological contracts have an
impact on task performance. The more employees are satisfied with the organization’s
income and profit distribution, the better their task performance will be.
The research results indicate that the fulfillment of relational psychological contracts
strongly influences task performance through organizational identification. China is a
sensible society. Out of the need for a sense of belonging, employees seek organizational
identity through perceived relational psychological contract fulfillment, thereby improv-
ing task performance. In shared employment, it’s a partnership between the employer
and the employee. The employability of employees is the basis for cooperation between
the two parties, which weakens the attention to attitudes and behaviors, and it is diffi-
cult to establish organizational identity [35]. Therefore, establishing temporary workers’
organizational identity tends to be a relational contract of social exchange.
Most employees with shorter service have just graduated and have no stable income.
The fulfillment of transactional psychological contracts, such as competitive income,
is regarded by these employees as a signal to be accepted by the organization, mak-
ing it easier for them to perceive the close connection with the employer and integrate
the organizational identity and personal identity. Therefore, the fulfillment of a transac-
tional psychological contract has more influence on the organizational identity of short-
term employees than that of long-term employees. At the same time, the fulfillment of
the relational psychological contract has a stronger influence on the task performance
of long-term staffs. Relational psychological contracts include intangible rewards and
exchanges and require a long period of time to establish and develop. The performance
and contribution of older employees are driven by the development of their relation-
ship with the organization. Temporary workers with long service time can better feel
the stable relationship with organizations. Their performance is not only driven by the
economy, but also by emotion.
6 Conclusion
The rapid development of the sharing economy and the emergence of new employment
methods are different from traditional forms of employment. New types of employment
bring higher flexibility, temporary workers’ psychological contracts have higher insta-
bility, and the mutual benefit between enterprises and employees is affected. Therefore,
maintaining and balancing the psychological contract of temporary workers directly
affects their work mentality, workability and performance level, and is also related to the
social identity and long-term development of shared talents. This research systematically
discusses the impact mechanism of temporary workers’ perceived psychological contract
fulfillment on task performance. The results show that under the background of shar-
ing economy, employees’ perceived transactional psychological contract fulfillment and
700 G. Zhang et al.
Acknowledgment. This research was supported by the grants of Shaanxi collaborative innovation
support project No. 2019KRM149 and Shaanxi Provincial Department of Education project No.
20JK0230.
Appendix
The Questionnaire used in this study (Fig. 4).
References
1. National Development and Reform Commission. https://www.ndrc.gov.cn/xxgk/jd/wsdwhfz/
202102/t20210222_1267536.html?code=&state=123 (2021)
2. Robinson, S.L.: Trust and breach of the psychological contract. Adm. Sci. Q. 41(4), 574–599
(1996)
3. Collins, M.D.: The effect of psychological contract fulfillment on manager turnover intentions
and its role as a mediator in a casual, limited-service restaurant environment. Int. J. Hosp.
Manag. 29(4), 736–742 (2010)
4. Wang, Y.S., Chen, H.M.: Research on sharing economic platform based on bilateral market
theory. Modernization of Management 2, 48–50 (2017)
5. Ashforth, B.E., Mael, F.: Social identity theory and the organization. Acad. Manag. Rev.
14(1), 20–39 (1989)
6. Epitropaki, O.: A multi-level investigation of psychological contract breach and organizational
identification through the lens of perceived organizational membership: testing a moderated–
mediated model. J. Organ. Behav. 34(1), 65–86 (2013)
7. Tufan, P., Wendt, H.: Organizational identification as a mediator for the effects of psycholog-
ical contract breaches on organizational citizenship behavior: insights from the perspective
of ethnic minority employees. Eur. Manag. J. 38(1), 179–190 (2020)
8. Bal, P.M., De Cooman, R., Mol, S.T.: Dynamics of psychological contracts with work engage-
ment and turnover intention: the influence of organizational tenure. Eur. J. Work. Organ. Psy.
22(1), 107–122 (2013)
9. Cheng, X.R., Li, P.B., Liang, H.: A study of human resource management patterns in sharing
economy——taking airbnb as an example. Human Resources Development of China 6, 20–25
(2016)
10. Rousseau, D.M., Tuoriwala, S.: Assessing psychological contracts: issues, alternatives and
measures. J. Organ. Behav. 19(51), 679–695 (1998)
11. Soares, M.E., Mosquera, P.: Fostering work engagement: the role of the psychological
contract. J. Bus. Res. 101, 469–476 (2019)
12. Van Hootegem, A., De Witte, H.: Qualitative job insecurity and informal learning: a longitu-
dinal test of occupational self-efficacy and psychological contract breach as mediators. Int. J.
Environ. Res. Public Health 16(10), 1847 (2019)
13. Rousseau, D.M.: Psychological contracts in organizations: understanding written and
unwritten agreements. Sage, Thousand Oaks, CA, USA (1995)
14. Dabos, G., Rousseau, D.M.: Mutuality and reciprocity: psychological contracts in research
teams. J. Appl. Psychol. 89(1), 52–72 (2004)
15. Wayne, S., Shore, L.M., Liden, R.C.: Perceived organizational support and leader-member
exchange: a social exchange perspective. Acad. Manag. J. 40(1), 82–111 (1997)
16. Borman, W.C., Motowidlo, S.J.: Expanding the criterion domain to include elements
of contextual performance. In: Schmitt, N., Borman, W.C. (eds.) Personnel selection in
organizations, pp. 71–98. Wiley, New York (1993)
17. Chiaburu, D.S., Oh, I., Wang, J., Stoverink, A.C.: A bigger piece of the pie: the relative
importance of affiliative and change-oriented citizenship and task performance in predicting
overall job performance. Hum. Resour. Manag. Rev. 27(1), 97–107 (2017)
18. Turnley, W.H., Bolino, M.C., Lester, S.W., Bloodgood, J.M.: The impact of psychological
contract fulfillment on the performance of in-role and organizational citizenship behaviors.
J. Manag. 29(2), 187–206 (2003)
19. Wu, C., Chen, T.: Psychological contract fulfillment in the hotel workplace: empowering
leadership, knowledge exchange, and service performance. Int. J. Hosp. Manag. 48, 27–38
(2015)
702 G. Zhang et al.
20. Long, L.R., Yi, M., Zhang, Y.: The impact of transactional and relational psychological
contract on employees task and contextual performance: the moderating roles of pay for
performance and perceived supervisor support. Forecasting 34(1), 8–14 (2015)
21. Rodwell, J., Ellershaw, J., Flower, R.: Fulfill psychological contract promises to manage
in-demand employees. Pers. Rev. 44(5), 689–701 (2015)
22. Yu, X.D., Liu, R., Chen, H.: Human resource policies in ‘internet +’——a case study based
on Di Di. Human Resources Development of China 6, 6–11 (2016)
23. Rentao, M., Bing, W.: Does not psychological contract breach contribute to organizational
identification? Journal of Capital University of Economics and Business 18(04), 50–57 (2016)
24. Epitropaki, O.: A multi-level investigation of psychological contract breach and organizational
identification through the lens of perceived organizational membership: testing a moderated-
mediated model: psychological contract breach and organizational identification. J. Organ.
Behav. 34(1), 65–86 (2012). https://doi.org/10.1002/job.1793
25. Hong, J., Bin, L.: Relationship between organizational identification and performance of
college teachers. Res. Econ. Manag. 36(12), 75–81 (2015)
26. Liu, W., He, C., Jiang, Y., Ji, R., Zhai, X.: Effect of Gig workers’ psychological contract
fulfillment on their task performance in a sharing economy—a perspective from the mediation
of organizational identification and the moderation of length of service. Int. J. Environ. Res.
Public Health 17(7), 2208 (2020)
27. Christ, O., van Dick, R., Wagner, U., Stellmacher, J.: When teachers go the extra mile: foci of
organizational identification as determinants of different forms of organizational citizenship
behavior among schoolteachers. Br. J. Educ. Psychol. 73(3), 329–341 (2003)
28. Ng, T.W.H., Feldman, D.C.: Does longer job tenure help or hinder job performance? J. Vocat.
Behav. 83(3), 305–314 (2013)
29. Norris, D.R., Niebuhr, R.E.: Organizational tenure as a moderator of the job satisfaction-job
performance relationship. J. Vocat. Behav. 24(2), 169–178 (1984)
30. Wright, T.A., Bonett, D.G.: the moderating effects of employee tenure on the relation between
organizational commitment and job performance: a meta-analysis. J. Appl. Psychol. 87(6),
1183–1190 (2002)
31. Conway, N., Coyle-Shapiro, J.A.: The reciprocal relationship between psychological contract
fulfilment and employee performance and the moderating role of perceived organizational
support and tenure. J. Occup. Organ. Psychol. 85(2), 277–299 (2012)
32. Rousseau, D.M.: New hire perceptions of their own and their employer’s obligations: a study
of psychological contracts. J. Organ. Behav. 11(5), 389–400 (1990)
33. Zhao, F.Q., Huang, H.Y., Chen, Y., Zhang, Q.H.: Effect of work-family balance human
resource practice on job performance: mediation of work-family relationship and moder-
ation of psychological capital. Human Resources Development of China 35(11), 124–140
(2017)
34. Hayes, A.F.: Introduction to mediation, moderation, and conditional process analysis: a
regression-based approach. The Guilford Press, New York (2013)
35. Ling, L., Gu, Y.H.: The relationship between employability and employment stability in share
economy: contradiction and balance. Human Resources Development of China 23, 10–15
(2015)
Research on the Influence of the Investment
Facilitation Level of the Host Country
on China’s OFDI——From the Perspective
of Investment Motivation
Based on the research of Wilson et al. (2003) [1], this article, combining the charac-
teristics of international direct investment and the relevant provisions of the Investment
Research on the Influence of the Investment Facilitation Level 705
Facilitation Action Plan, takes the five aspects including government behavior, infras-
tructure, financial services, market environment and the labor market as First-Level
Indicators. At the same time, in order to measure the host country’s investment facilita-
tion level as systematically and comprehensively as possible, these indicators are further
subdivided into 26 secondary indicators (see Table 1 for details). The relevant data are
from the Global Competitiveness Report (GCR) from 2010 to 2019.
According to Table 2 and Table 3, Asia and Singapore have always ranked first for
many years in terms of investment facilitation. Second, the countries with a high degree
1 The formula is expressed as: Y = X /X
i i max , Yi represents the new value after standardization;
Xi represents the index data obtained after reciprocal treatment; Xmax is the maximum value
that can be obtained in index.
Research on the Influence of the Investment Facilitation Level 707
Country 2010 2011 2012 2013 2014 2015 2016 2017 2018 2019
Oceania 0.76 0.76 0.76 0.75 0.76 0.77 0.77 0.78 0.76 0.77
Africa 0.52 0.53 0.54 0.55 0.56 0.55 0.55 0.55 0.55 0.57
North America 0.63 0.63 0.65 0.67 0.68 0.67 0.67 0.69 0.67 0.68
South America 0.53 0.54 0.55 0.56 0.56 0.55 0.55 0.56 0.54 0.57
Europe 0.68 0.68 0.69 0.67 0.68 0.69 0.70 0.70 0.69 0.71
Asia 0.60 0.62 0.63 0.64 0.64 0.64 0.64 0.65 0.65 0.67
Control variables: trade openness (LnTra), bilateral distance (LnDis). The trade open-
ness of a country is measured by the percentage of country’s imports and exports of goods
and services to the country’s GDP, the data is from the World Bank database; the bilateral
geographic distance is measured by the distance between Beijing and the capital of each
host country, the data coming is from the CEPII database. In addition, logarithms are
taken for all variables to reduce the impact of drastic changes and heteroscedasticity
among variables.
In the above model, β0 represents the constant term; β1 , β2 , γ are the parameters
to be estimated; δt is the time fixed effect; εij is the residual term; i represents the
host country; t represents the year; LnOFDIit is the stock of China’s outward foreign
direct investment; LnIFIit is the comprehensive index of the host country’s investment
facilitation; Controlit is the control variable; Motit represents the specific investment
motive; Motit *LnIFIit is the interaction item of the investment motive and the investment
facilitation comprehensive index of the host country.
the host country are both significantly positive at the level of 1% or 5%, indicating
that China’s OFDI exists obvious Market-Seeking Motivation and Resource-Seeking
Motivation. The coefficients of market scale are greater than the coefficients of natural
resource endowment, to a certain extent, it shows that China’s OFDI is more inclined
to the market scale of the host country between the market scale and natural resource
endowment. The labor productivity level (LnLP) coefficients are all significantly negative
at the level of 1%, indicating that Chinese companies prefer to invest capital in countries
with low productivity and abundant labor populations. The possible explanation for
this is that China’s OFDI flows mainly to low-tech, low-value-added industries. It pays
more attention to the demographic dividend rather than the labor productivity of the host
countries. The strategic asset level (LnTec) coefficient is significantly negative at the level
of 10%, indicating that the impact of the host country’s strategic asset level on China’s
OFDI is still uncertain. It may be because the strategic-seeking motivation of China’s
foreign direct investment is not obvious, or that China’s strategic-seeking motivation
investment for developed countries is smaller than its market-seeking motivation and
resource-seeking motivation investment for other countries.
From the perspective of the interaction, the coefficient of LnIFI*LnGDP in col-
umn (2) is significantly positive, indicating that with the continuous improvement of
the the host country’s investment facilitation level, the market scale has become more
attractive to China’s OFDI. In other words, China’s Market-Seeking OFDI is more
inclined to the host country with the higher level of investment facilitation. The coeffi-
cient of LnIFI*LnRes in column (3) is significantly negative, indicating that the worse
the investment facilitation level in the host country, the stronger the attractiveness of
natural resources to China’s OFDI. In other words, China’s Resource-Seeking Invest-
ment is more inclined to the host country with lower investment facilitation level. The
coefficient of LnIFI*LnLP in column (4) is significantly positive, showing that China’s
Efficiency-Seeking foreign direct investment is more inclined to the countries with a
higher investment facilitation comprehensive index when choosing a location. The coef-
ficient of LnIFI*LnTec in column (5) is significantly positive, indicating that China’s
strategic Asset-Seeking Investment also tends to the countries with higher investment
facilitation level.
From control variables, the trade openness (lnTra) coefficients are significantly pos-
itive at the level of 1%, indicating that the higher the trade openness of the host country,
the more conducive to the inflow of China’s transnational capital. The bilateral distance
(lnDis) coefficients are significantly negative at the level of 1%, indicating that the farther
the bilateral distance is, the less conducive to foreign capital inflow. It is demonstrated
that China’s foreign direct investment location is more inclined to Asia, especially East
and Southeast Asia.
Table 4. Empirical results of the impact of host country investment facilitation on China’s OFDI
to take more measures to improve investment facilitation level. Therefore, this paper
draws on the previous research experience, taking the three lagging periods of investment
facilitation as an Instrumental Variable, and then uses Generalized Method of Moments
(GMM) to test the robustness of the benchmark model. The specific results are shown
in the Table 5. Comparing the results of the benchmark regression in the Table 4, the
coefficients, direction and significance of the main explanatory variables in the GMM
model have not changed significantly, indicating that the conclusion is robust, so we will
not repeat it.
Note: The values in parentheses are z-values, ***, **, and * indicate significant at the level of 1%,
5%, and 10%, respectively
5 Conclusions
This paper constructs a relatively perfect evaluation system of investment facilitation. It
measures the investment facilitation of 101 countries around the world from 2010 to 2019
with the available data, which is found that there are significant differences in the level
of investment facilitation among regions, and there is great room for improvement. On
this basis, the different impact of the host countries’ investment facilitation comprehen-
sive index on China’s OFDI driven by different investment motives is examined. There
following conclusions are: First, China’s OFDI is more inclined to countries with higher
investment facilitation comprehensive index. Second, the host country’s investment facil-
itation comprehensive index has different effects on China’s different Investment-Motive
OFDI. Market-Seeking, Efficiency-Seeking, and Strategic Asset-Seeking OFDI aim at
host countries with higher investment facilitation comprehensive index. Still, Resource-
Seeking OFDI tends to countries with poorer investment facilitation comprehensive
index.
712 C. Niu et al.
Based on the above research conclusions, this paper provides the following measures
and suggestions. First of all, our government should provide some beneficial policy for
enterprises, strive to negotiate bilateral and multilateral investment agreements, and
protect Chinese enterprises’ legitimate rights and interests in the host country. Second,
Chinese enterprises should clarify their position, conduct a comprehensive and detailed
research on the investment environment of the host country, and make location selection
combing with their investment motivation and the characteristics of the host country
to avoid blind investment. Last, “the Belt and Road” initiative and the “Regional Com-
prehensive Economic Partnership” (RCEP) should be fully utilized to actively explore
overseas markets and continuously improve international competitive power.
References
1. Wilson, J.S., Mann, C.L., Otsuki, T.: Trade facilitation and economic development: a new
approach to quantifying the impact. World Bank Econ Rev 17(3), 367–389 (2003)
2. Zhang, Y.: The investment facilitation of “One Belt One Road” and choices of China’s foreign
direct investment-empirical analysis based on cross-panel data and investment gravity model.
J Int Trade 9, 165–176 (2016)
3. Qiao, M.: Will the increase in investment facilitation promote China’s outward foreign direct
investment? —based on panel data analysis of countries along the “Belt and Road.” Inq into
Econ Issues 1, 139–148 (2019)
4. Xietian, Z., Fan, X.: has investment facilitation promoted China’s outward foreign direct
investment? — threshold testing based on host country heterogeneity. Int Bus 6, 59–75 (2019)
5. Liu, Y., Zhao, X.: Investment facilitation in Central and Eastern European Countries and
Its Impact on China’s Foreign Direct Investment. The Journal of Quantitative & Technical
Economics. 1, 83–97 (2021)
6. Dunning, J.H.: Trade, location of economic activity and the multinational enterprise: a search
for an eclectic approach. Theory Transnatl Corp 1, 183–218 (1993)
7. Zhou, C., Liu, X., Guo, Z.: Business environment and China’s ODI—from the perspective of
investment motivation. J Int Trade 10, 143–152 (2017)
8. Wang, Z., Zhou, Q.: how business environment affects Chinese enterprises’ OFDI to countries
along “The Belt and Road.” Collected Essays on Finance and Economics 9, 42–52 (2019)
9. Buckley, P.J., Liu, X.: The determinants of chinese outward foreign direct investment. J Int
Bus Stud 38(4), 499–518 (2007)
Information Technology
and Applications
Adaptive Observer-Based Control for a Class
of Nonlinear Stochastic Systems with Parameter
Uncertainty
Abstract. The adaptive control problem was discussed for stochastic nonlinear
systems with parameter uncertainty in this paper. The actual systems such as
distributed network systems, weather systems or industrial control systems are
constantly changing. There is a correlation between various factors and informa-
tion exchange, material exchange or energy exchange between the system and the
outside world in the system. These exchanges have randomness, more or less. For
these kinds of systems, the model descriptions need the help of stochastic differen-
tial equations. Considering stochastic nonlinear systems with unknown constant
parameters, we introduced the separation theory into the adaptive parameter esti-
mators design in the design process. Using Lyapunov functional and stochastic
analysis technology, we studied the design method of robust adaptive observers
for stochastic systems. So that the gain matrices can be easily obtained in sequence
in the same algorithm. A numerical test was presented to verify the feasibility of
the conclusions obtained.
1 Introduction
For stochastic systems driven by Brown motion, stability analysis and controller design
are topics with high attention. When the system contains unknown parameters, adaptive
control is a very effective method to estimate unknown parameters. It can estimate the
system state and unknown parameters effectively, so it is widely used in fault detection,
fault isolation, and so on [1–3]. However, due to the influence of random factors, the
parameter estimation of a stochastic system with unknown parameters is hard to carry
out. From the analysis of reference [4], it can be seen that the complete separation of
feedback controller gain and state estimation gain is hard to achieve in stochastic systems
due to the effect of parameter fluctuation. Adaptive control for the stochastic nonlinear
systems with unknown parameters and adaptive estimator design problems has become
a very challenging topic when the random factors in the nonlinear system are inevitable.
In [5], the stochastic nonlinear systems without parameters uncertainty were dis-
cussed. On this basis, this paper makes further research. This paper is based on the
filtration Ft{t≥0} with the usual conditions and complete probability space (, F, P).
Denote the transposed matrix of A by AT , moreover,
A1 A2 A1 A2
= .
∗ A3 AT2 A3
2 Preliminaries
here h(t) ∈ Rr represents the output vector, x(t) ∈ Rn represents the state vector, l
represents the unknown constant parameter. A(t) satisfied the formula
A(t) = BF(t)C,
E{dω(t)} = 0.
Assumption 1. The nonlinear part g(x(t), u(t), l) satisfied the following restriction:
g(0, u(t), l) ≡ 0.
Assumption 2. A positive definite matrix R with the condition
E T RY ⊥ = 0
22 = QA + AT Q − XY − Y T X T + δ2 a2 +εC T C,
then the matrices J and W can be found to make stochastic nonlinear systems (4)
asymptotically stable, meanwhile the observer gain matrix satisfies the following
equation
J = R−1 X .
where
ξ T (t) = xT (t) eT (t) elT (t) ,
P = diag{P, Q, D−1 }.
E T Q = WY ,
Thus, we have
Next, utilizing the Lipschitz condition, the following formula can be derived.
LV (ξ(t), t) ≤
xT (t)[P(A + A(t)) + (A + A(t))T P + 1/δ1 PP + δ1 a2 + K T (P + Q)K]x(t)
+ eT (t)[Q((A + A(t)) − JY ) + ((A + A(t)) − JY )T Q + 1/δ2 QQ + δ2 a2 ]e(t).
(8)
T
x(t) 11 0 x(t)
LV (ξ(t), t) ≤ . (9)
e(t) ∗ 22 e(t)
Adaptive Observer-Based Control for a Class of Nonlinear Stochastic Systems 719
with
11 = P(A + A(t)) + (A + A(t))T P + 1/δ1 PP + δ1 a2 + K T (P + Q)K
Z = FT −1 = FS,
Ju = R−1 B.
Proof. Set
P = diag{S, R, D−1 },
V (ξ(t), t) = ξ T (t)Pξ(t),
and u(t) is set to
u(t) = Z x̂(t),
we can obtain
dV (ξ(t), t) = LV (ξ(t), t)dt + 2(xT (t)S + eT (t)R)Kx(t)dω(t). (14)
And we can get the differential operator above is
LV (ξ(t), t) = 2xT (t)S[((A + A(t)) + VZ)x(t) − VZe(t) + g(x(t), u(t), l)]
+ 2eT (t)R[((A + A(t)) − Ju Y )e(t) + g̃ + Eg2 (x̂(t), u(t))el (t)]
− 2elT (t)g2T (x̂(t), u(t))W (y(t) − Y x̂(t)) + xT K T (S + R)Kx(t). (15)
Using the analysis process similar to Theorem 1, the following formula can be obtained
T
x(t) 11 −SVZ x(t)
LV (ξ(t), t) ≤ , (16)
e(t) ∗ 22 e(t)
Where
11 = S((A + A(t)) + VZ) + ((A + A(t)) + VZ)T S
,
+1/δ1 SS + δ1 a2 + K T (S + R)K,
22 = R((A + A(t)) − Ju Y ) + ((A + A(t)) − Ju Y )T R
.
+1/δ2 RR + δ2 a2 .
Set
F = ZT ,
11 −SVZ
one gets that matrix inequality (12) and < 0 are equivalent from Schur
∗ 22
complement Lemma.
By setting
B = RJu ,
then Theorem 2 can be proven from the similar proof process of Theorem 1.
Adaptive Observer-Based Control for a Class of Nonlinear Stochastic Systems 721
5 Numerical Examples
Here we present one numerical example, to verify the effectiveness of the obtained
conclusions.
Example 1. Considering a nonlinear stochastic system (10) as follow with
⎡ ⎤ ⎡ ⎤
−3 1 0 1.2
A = ⎣ 0.3 −4.5 1 ⎦, E = ⎣ 0.6 ⎦,
−0.1 0.3 −3.8 0.5
⎡ ⎤
0.1 −0.1 0.2
K = ⎣ 0.3 0.3 −0.4 ⎦,
0.1 0.1 −0.3
0.8 0.3 0
Y = ,
0.5 1 0.6
g1 (x(t), u(t)) = u(t),
6 Conclusion
This paper considers the estimation method of the state vector of nonlinear stochas-
tic systems with parameter uncertainty. By setting a reasonable Lyapunov–Krasovskii
functional, the corresponding asymptotic stability conclusion of estimation error is
given. The asymptotical stability is guaranteed for the closed-loop system by applying
a state-feedback controller.
References
1. Zhang, K., Jiang, B., Cocquempot, V.: Adaptive observer-based fast fault estimation. Int. J.
Control Autom. Syst. 6, 320–326 (2008)
2. Yan, X.G., Edwards, C.: Fault estimation for single output nonlinear systems using an adaptive
sliding mode estimator. IET Control Theory Appl. 2, 841–850 (2008)
3. Farza, M., M’Saad, M., Maatoug, T., Kamoun, M.: Adaptive observers for nonlinearly
parameterized class of nonlinear systems. Automatica 45, 2292–2299 (2009)
4. Deng, H., Krstić, M.: Output-feedback stabilization of stochastic nonlinear systems driven by
noise of unknown covariance. Syst. Control Lett. 39(3), 173–182 (2000)
5. Miao, X.F., Li, L.L., Yan, X.M.: Adaptive observer-based control for a class of nonlinear
stochastic systems. Int. J. Comput. Math. 92(11), 2251–2260 (2015)
6. Mao, X.R.: Stochastic Differential Equations and Their Applications. Horwood Publishing,
Chichester (1997)
Open Domain Question Answering Based
on Retriever-Reader Architecture
1 Introduction
Question Answering (QA) system aims to produce precise answers to a given user query
directly and concisely. Instead of users searching for keywords and getting a series of
web links, today’s search engines have also added the function of the QA system. Users
can contact quickly without clicking on the relevant links again for most simple queries,
as shown in Fig. 1. OpenQA [1] searches for the final answer from a large text corpus for a
given query. The traditional OpenQA system comprises three modules, namely question
analysis, document retrieval, and answer extraction. Question analysis is subdivided
into question classification and question standardization. This module aims to match
the user’s given query to the question category that has been defined by using natural
language processing tools. According to the processed problems, the document retrieval
module relies on the IR engine to retrieve relevant documents in a small range from a
large-scale text corpus. Answer extraction aims to further.
Extract the final accurate answer from the relevant documents returned in the previous
step. Traditional OpenQA performance is too dependent on the problem analysis phase,
specifically problem classification. Due to the complex nature and polysemy of language,
it is almost impossible to cover all semantic expressions of users, either by classifier or
by handwriting. Therefore, the traditional three-stage architecture is very limited.
As DL and Transfer Learning (TL) have made significant progress, OpenQA is
also undoubtedly driven by this trend. Especially in recent years, DrQA [2] proposed
by Chen et al. simplified the three-stage of traditional OpenQA into “Retriever-Reader”
with only two-stage firstly, as shown in Fig. 2. In the following sections, we will elaborate
on Retriever and Reader.
2 Retriever
The retriever is similar to the IR system, and its function is to return a collection of related
documents, including the correct answer. In general, the current retrievers of OpenQA
contain three categories: sparse retriever, dense retriever, and iterative retriever.
Ranker in [4], the paragraph ranker proposed by Lee et al. [6] focuses on reducing the
recall rate of answers by reducing noise. Kratzwald et al. [7] proposed that RankQA
added an answer rearrangement module. First, this module combines the characteristics
of the retrieval and reading phases. Second, in order to solve the problem of ignoring
other paragraphs due to too much focus on the first relevant paragraph, according to the
conclusion of [5], RankQA aggregates repeated candidate answers together to increase
the amount of information that will provide benefit for subsequent rankings. Finally, the
candidate answers will be reported in this module, and the first answer will be output.
Kratzwald et al. [8] believed that the fixed number of returned documents could not
acquire the optimal solution due to the influence of noise information. They proposed
adaptive document retrieval, which dynamically and adaptively adjusts the number of
returned documents according to the size of information sources and different queries.
Based on BERT [9], Yang et al. [10] proposed BERTserini, a new OpenQA model of
end-to-end training method. The model architecture uses Anserini [11] as a document
retriever, a query expansion-based model that uses BM25 as a rearrangement function.
The experiment found that the paragraph retrieval effect is the best comparing the three
levels of retrieval method. Multi-passage BERT [12] proposed multiple paragraphs to
train the reader in response to the inability to compare the scores of different paragraphs
caused by independent training of related paragraphs. It also uses the sliding window
mechanism and paragraph sorting mechanism to obtain optimal performance. And its
retriever is ElasticSearch based on the BM algorithm.
The above researches are all based on DrQA, and they all use sparse retrievers.
However, sparse retrieval only relies on matching similarity, and it is challenging to
address term mismatch. For this kind of problem, the dense retriever is remarkably
effective.
Dense retriever generally uses dual encoders to encode query and documents, and most
of them are Representation-based at the beginning. For example, OrQA [13], DPR [14],
REALM [15] all use BERT encoders to encode query and documents to generate dense
vector representations. The retrieval score of OrQA [13] is obtained by calculating the
inner product of the dense vector of query and documents and using the Inverse Cloze
Task (ICT) pretraining retriever. Its retrieval accuracy obtained is 19% higher than that
of BM25. Karpukhin et al. [14] proposed DPR. They believe that the ICT calculation
is too large and expensive. DPR realizes the training model only through the question-
answer pair, and the performance of this method has surpassed the traditional sparse
index algorithm.
The above model calculates the document representation in advance and query index
offline, so these representation-based dense retrievers are more efficient. Still, their inde-
pendent representation leads to the loss of due interaction. Retrieve-and-Read [16] stud-
ied the impact of learning answer span from retrieval. Its reader is based on the BiDAF
[17] model, Which utilizes bi-attention to enable effective query interaction with doc-
uments. Nie et al. [18] believed that the purpose of previous research is to find useful
information coverage for reading modules to improve reading ability, rather than to find
726 D. Zheng et al.
accurate information, which deviates from the essence of large-scale reading compre-
hension. They studied the impact of multi-granularity retrieval on the reading module
and realized rich information interaction between query and paragraphs. However, this
method requires a lot of computing power, and the training cost is expensive.
Of course, efficiency and accuracy are not incompatible. DC-BERT [19] mainly
studied the problem of high throughput input caused by the application of PLM in
OpenQA. It has two independent BERT models: an online BERT encodes the question
only once, and an offline BERT pre-encodes all documents and caches their encoding.
After enabling the cache, DC-BERT can immediately read out any document codes.
The decoupling encoding of query and documents, and then fed to the transformer layer
with query and document interaction, Which effectively generates context encoding of
the query-document pair. ColBERT [20] is a high-speed retriever based on BERT. And
proposed a novel sorting model, which uses independent coding to obtain advanced
offline documents. At the same time, a low-cost but powerful post-interaction step is
adopted to achieve a powerful effect of both effectiveness and speed. SPARTA [21]
abandons the inner product of the sequence and uses the fine-grained interaction based
on the token-level level to make retrieval faster and more accurate.
Since a single round of retrieval may miss relevant documents, a lot of research has been
carried out in the past few years based on iterative retrievers, which retrieve documents
related through multiple steps. HotpotQA [22] is a multi-hop inference dataset released
in 2018. Its release has promoted multi-hop OpenQA to a certain extent and provided a
robust multi-hop dataset for subsequent research.
GOLDEN Retriever [23] the query reconstruction task recasting for MRC task,
because they all will be a problem and some context document as input, aims to generate
natural language strings. Das et al. [24] proposed the Multi-step Reasoner model, which
uses unified encoders to encode queries and paragraphs and fixed paragraphs to represent
vectors. In each round of iteration, the query vector is continuously updated so that
different question expression vectors are used to recall different paragraphs. The model
uses reinforcement learning for training. It takes the reader’s output as a reward to
calculate the strategy gradient to update the query vector and can continuously optimize
the retrieval quality. MUPPET [25] uses the MIPS searcher to reconstruct the query
vector from the paragraphs retrieved in the previous iteration. Its iteration is terminated
when it reaches the set maximum number of related documents. Path Retriever [26]
sequentially retrieves associated documents by using RNN. Inference paths are formed
based on the retrieval records, and then these paths are sorted using an MRC model.
The powerful interactivity of the searcher and reader enables more accurate reasoning
and answers to complex questions. The goal of reconstructing and generating the next
query based on the previous search results is to obtain more relevant documents during the
next search. The generated query generally has two forms: Explicit-query and Implicit-
query. [23] generated query is an explicit-query. In other words, it is the representation of
natural language. [24–26] generate implicit-query, which is dense vector representation.
In comparison, both have their own merits. Explicit-query is easy to understand and
Open Domain Question Answering Based on Retriever-Reader Architecture 727
control manually, but it is often limited by the glossary of terms. Implicit-query is not
restricted by terminology but is poor in interpretability.
3 Reader
Reader is the second core module of OpenQA, and it is usually regarded as an MRC task.
Its function is to get the final answer from the document finally returned by the retriever.
What is more challenging is that the given paragraph of ordinary reading comprehension
is usually a single. In contrast, OpenQA needs to find the final accurate answer in multiple
articles or paragraphs, which requires more complex reasoning. There are two categories
of existing readers: extractive reader and generative reader.
4 End-to-End Training
The past few years, a lot of work has begun to study open domain question answering
in an end-to-end training method. The OrQA, as mentioned above, Retrieve-and-Read,
REALM, and RAG all use end-to-end methods to train the model.
Not long ago, Baidu proposed a new RocketQA [36] model, which employs such
technologies as cross-batch negatives, denoised hard negative sampling, and data aug-
mentation. It dramatically improves the effectiveness of the dual retrieval model and
takes an important step for the end-to-end OpenQA. In addition, there is a retriever-only
model, such as Den-SPI [37], which is used for the real-time OpenQA. The dense-sparse
phrase index is proposed, which is set to capture the syntactic, semantic, and lexical infor-
mation to address the problem that some document information is filtered incorrectly
during the process. Discarding reading comprehension, we only use the query to get the
phrase index as the answer, which is very efficient in the end-to-end benchmark test.
Moreover, A new method called retriever-only has surfaced, which are all PLM
based on Seq2Seq, such as GPT-3 mentioned above. Such models have been trained
on a huge corpus, and a large amount of knowledge is stored in parameters. They can
answer questions without any external knowledge. This model can realize the OpenQA
without the retriever, which is a very different change from the past.
Open Domain Question Answering Based on Retriever-Reader Architecture 729
Table 1. Evaluation results of EM on several classic datasets. EM: exact-match. The datasets in
the table are widely used as benchmarks in OpenQA. We abbreviated the name of models and
datasets. S: SQuAD [3], CT: Curated TREC [42], WQ: Web Question [43], NQ: Natural Question
[44], Q-T: Quasar-T [45], H-Q: HotpotQA [20]. We don’t elaborate on these datasets in this paper.
After years of research and development, OpenQA has continuously improved its meth-
ods and related technologies. The ultimate goal of OpenQA is to build a system that
can answer any input question. The achievement of this great goal is a long and arduous
process. The NLP community still needs continuous research and progress. As follows,
we discussed some of the challenges and future development prospects of the existing
OpenQA:
Phrase Index. The phrase index is modular and extensible. The critical difference
between a phrase index and retriever is that the latter index each document by con-
tent, while the former requires an index for each phrase in its context. Although some
work has done related research on the index of the phrase [37–39], constructing phrase
representations in a large corpus of OpenQA is still a mighty challenge. In addition,
existing phrase indexing methods rely heavily on sparse representation to locate related
documents, and their performance is still not as good as the dense index. Seo et al.
[40], published in ACL2021, realized the dense-phrase index in independent learning of
OpenQA firstly. And achieved better results. The future of the dense-phrase index needs
to be further explored.
Retrieval Effectiveness and Efficiency. Generally speaking, sparse retrieval has high
efficiency but low effectiveness, while dense retrieval is the opposite. How to ensure the
effectiveness of retrieval and improve efficiency is a challenge faced by the retriever.
Some studies have made improvements in this area. For example, Seo et al. [41] proposed
a BPR retrieval model. The model combines learning-to-hash techniques on DPR, which
effectively reduces the index space by using binary codes. To get a better and more
effective model, there is still room for exploration and improvement in future work.
References
1. Voorhees, E.M.: The TREC-8 question answering track report. In: Trec, Vol. 99, pp 77–82.
Citeseer (1999)
2. Chen, D.Q., Fisch, A., Weston, J., Bordes, A.: Reading wikipedia to answer open-domain
questions. In: 55th ACL, pp. 1870–1879. ACL, Stroudsburg (2017)
3. Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: SQuAD: 100,000+ questions for machine
comprehension of text. In: Proceedings of the 2016 Conference on EMNLP, pp. 2383–2392.
ACL, Texas (2016)
4. Wang, S., et al.: R3: reinforced ranker-reader for open-domain question answering. In: AAAI-
18. Louisiana (2018)
5. Wang, S., et al.: Evidence aggregation for answer re-ranking in open-domain question
answering. In: 6th ICLR. ICLR, Vancouver (2018)
6. Lee, J., Yun, S., Kim, H., Ko, M., Kang, J.: Ranking paragraphs for improving answer recall
in open-domain question answering. In: Proceedings of the 2018 conference on EMNLP,
pp. 565–569. ACL, Brussels (2018)
7. Kratzwald, B., Eigenmann, A., Feuerriegel, S.: RankQA: neural question answering with
answer re-ranking. In: 57th ACL, pp. 6076–6085. ACL, Florence (2019)
8. Kratzwald, B., Feuerriegel, S.: Adaptive document retrieval for deep question answering. In:
Proceedings of the 2018 conference on EMNLP, pp. 576–581. ACL, Brussels (2018)
9. Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pretraining of deep bidirectional trans-
formers for language understanding. In: Proceedings of the 2019 conference of the NAACL,
pp 4171–4186. ACL, Minnesota (2019)
10. Yang, W., et al.: End-to-end open-domain question answering with BERTserini. In: Proceed-
ings of the 2019 conference of the NAACL, pp. 72–77. ACL, Minnesota (2019)
11. Yang, P., Fang, H., Lin, J.J.: Anserini: enabling the use of lucene for information retrieval
research. In: Proceedings of the 40th international ACM SIGIR conference on research and
development in information retrieval, pp 1253–1256. ACM, Tokyo (2017)
12. Wang, Z., Ng, P., Ma, X., Nallapati, R., Xiang, B.: Multi-passage BERT: a globally normalized
BERT Model for open-domain question answering. In: Proceedings of the 2019 Conference
on EMNLP and 9th IJCNLP), pp 5878–5882. ACL, Hong Kong (2019
13. Lee, K., Chang, M., Toutanova, K.: Latent retrieval for weakly supervised open domain
question answering. In: 57th ACL, pp. 6086–6096. ACL, Florence (2019)
14. Karpukhin, V., et al.: Dense passage retrieval for open-domain question answering. In:
Proceedings of the 2020 conference on EMNLP, pp. 6769–6781. ACL, Online (2020)
15. Guu, K., Lee, K., Tung, Z., Pasupat, P., Chang, M.: Retrieval augmented language model
pre-training. In: Hal, D., III, Aarti, S. (eds.) Proceedings of the 37th international conference
on machine learning, vol. 119, pp. 3929-3938. PMLR, Proceedings of Machine Learning
Research (2020)
16. Nishida, K., Saito, I., Otsuka, A., Asano, H., Tomita, J.: Retrieve-and-read: multi-task learn-
ing of information retrieval and reading comprehension. In: Proceedings of the 27th ACM
international conference on information and knowledge management, pp. 647–656. ACM,
Torino (2018)
17. Seo, M., Kembhavi, A., Farhadi, A., Hajishirzi, H.: Bidirectional attention flow for machine
comprehension. arXiv preprint arXiv:1611.01603 (2016)
18. Nie, Y., Wang, S., Bansal, M.: Revealing the importance of semantic retrieval for machine
reading at scale. In: Proceedings of the 2019 conference on EMNLP and 9th IJCNLP,
pp. 2553–2566. ACL, Hong Kong (2019)
732 D. Zheng et al.
19. Zhang, Y., et al.: DC-BERT: decoupling question and document for efficient contextual
encoding. In: Proceedings of the 43rd international ACM SIGIR conference on research and
development in information retrieval, pp. 1829–1832. Association for Computing Machinery
(2020)
20. Khattab, O., Zaharia, M.: ColBERT: efficient and effective passage search via contextualized
late interaction over BERT. In: Proceedings of the 43rd international acm sigir conference
on research and development in information retrieval, pp. 39–48. Association for Computing
Machinery (2020)
21. Zhao, T., Lu, X., Lee, K.: SPARTA: efficient open-domain question answering via sparse
transformer matching retrieval. In: Proceedings of the 2021 conference of the NAACL,
pp. 565–575. ACL, Online (2021)
22. Yang, Z., et al.: HotpotQA: a dataset for diverse, explainable multi-hop question answering.
In: Proceedings of the 2018 conference on EMNLP, pp. 2369–2380. ACL, Brussels (2018)
23. Qi, P,. Lin, X., Mehr, L., Wang, Z., Manning, C.D.: Answering complex open-domain ques-
tions through iterative query generation. In: Proceedings of the 2019 conference on EMNLP
and 9th IJCNLP, pp. 2590–2602. ACL, Hong Kong (2019)
24. Das, R., Dhuliawala, S., Zaheer, M., McCallum, A.: Multi-step retriever-reader interaction
for scalable open-domain question answering. In: 7th ICLR. ICLR, Louisiana (2019)
25. Feldman, Y., El-Yaniv, R.: Multi-Hop paragraph retrieval for open-domain question answer-
ing. In: 57th ACL, pp. 2296–2309. ACL, Florence (2019)
26. Asai, A., Hashimoto, K., Hajishirzi, H., Socher, R., Xiong, C.: Learning to retrieve reasoning
paths over wikipedia graph for question answering. In: 7th ICLR. ICLR, Louisiana (2019)
27. Lin, Y., Ji, H., Liu, Z., Sun, M.: Denoising distantly supervised open-domain question
answering. In: 56th ACL, pp. 1736–1745. ACL, Melbourne (2018)
28. Min, S., Chen, D., Zettlemoyer, L., Hajishirzi, H.: Knowledge guided text retrieval and reading
for open domain question answering. arXiv preprint arXiv:1911.03868 (2019)
29. Liu, Y., et al.: Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:
1907.11692(2019)
30. Joshi, M., Chen, D., Liu, Y., Weld, D.S., Zettlemoyer, L., Levy, O.: SpanBERT: improving
pretraining by representing and predicting spans. Trans. Assoc Comput. Linguistics 8, 64–77
(2019)
31. Brown, T.B., et al.: Language models are few-shot learners. arXiv preprint arXiv:2005.14165
(2020)
32. Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer.
arXiv preprint arXiv:1910.10683 (2019)
33. Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language
generation, translation, and comprehension. In: 58th ACL, pp. 7871–7880. ACL, Online
(2020)
34. Lewis, P., et al.: Retrieval-augmented generation for knowledge-intensive nlp tasks. arXiv
preprint arXiv:2005.11401 (2020)
35. Izacard, G., Grave, E.: Leveraging passage retrieval with generative models for open domain
question answering. In: 16th EACL, pp. 874–880. ACL, Online (2021)
36. Qu, Y., et al.: RocketQA: an optimized training approach to dense passage retrieval for open-
domain question answering. In: Proceedings of the 2021 conference of the NAACL, pp. 5835–
5847. ACL, Online (2021)
37. Seo, M., et al.: Real-time open-domain question answering with dense-sparse phrase index.
In: 57th ACL, pp. 4430–4441. ACL, Florence (2019)
38. Seo, M., Kwiatkowski, T., Parikh, A.P., Farhadi, A., Hajishirzi, H.: Phrase-indexed question
answering: a new challenge for scalable document comprehension. In: Proceedings of the
2018 conference on EMNLP, pp. 559–564. ACL, Brussels (2018)
Open Domain Question Answering Based on Retriever-Reader Architecture 733
39. Lee, J., Seo, M., Hajishirzi, H., Kang, J.: Contextualized sparse representations for real-time
open-domain question answering. In: 58th ACL, pp. 912–919. ACL, Online (2020)
40. Lee, J., Sung, M., Kang, J., Chen, D.: Learning dense representations of phrases at scale. In:
59th ACL and 11th IJCNLP, pp. 6634–6647. ACL, Online (2021)
41. Yamada, I., Asai, A., Hajishirzi, H.: Efficient passage retrieval with hashing for open-domain
question answering. In: 59th ACL and 11th IJCNLP, pp. 979–986. ACL, Online (2021)
42. Baudiš, P., Šedivý, J.: Modeling of the question answering task in the yodaqa system. In:
International conference of the cross-language evaluation forum for European languages,
pp. 222–228. Springer (2015)
43. Berant, J., Chou, A.K., Frostig, R., Liang, P.: Semantic parsing on freebase from question-
answer pairs. In: Proceedings of the 2013 Conference on EMNLP, pp. 1533–1544. ACL,
Washington (2013)
44. Kwiatkowski, T., et al.: Natural questions: a benchmark for question answering research, vol.
7, pp. 452–466 (2019)
45. Dhingra, B., Mazaitis, K., Cohen, W.W.: Quasar: datasets for question answering by search
and reading. arXiv preprint arXiv:1707.03904 (2017)
Does Network Infrastructure Improve
the Information Efficiency of Regional Capital
Market?—Quasi Natural Experiment Based
on “Broadband China” Strategy
1 Introduction
The expansion of Internet applications makes human activities more and more guided
by it, and the third internet-based technological revolution will become a new driving
force for economic development. China especially benefits from the Internet revolution,
realizing the rapid digital economy and digital finance development. The “Internet plus”
action has promoted the integrated development of the Internet and all industries, show-
ing a good development momentum. The 19th CPC National Congress also proposed
building a cyber power, a digital China, and a smart society. Therefore, in the context
of supply-side structural reform, network infrastructure construction plays a key role
in developing the Internet, which will become a hot academic concern. Therefore, the
quasi-natural experiment implementing the “broadband China” strategy is endowed with
important practical research significance.
So, will be based on the literature analysis and put forward the theoretical assump-
tions, whether to be included in the “broadband China” demonstration city for logo,
build more point of double difference model, at the same time, control the time effect
and industry effects, regional effects, using 1013 a-share listed companies from 2008
to 2019 in 12156 samples of panel data to carry on the empirical research. STATA16
was used for model regression and calculation. Discuss the “broadband China” strategic
impact on stock price synchronicity and company equity concentration, the regulation
of growth and stock market transactions in which the noise effect, finally draw a network
infrastructure to regional capital market information efficiency influence and revelation,
in order to the Internet integration development of regional capital market information
efficiency to supplement the theoretical basis.
Network infrastructure has a long-term impact on the behavior and development of eco-
nomic participants, accelerates the processing and integration of massive information,
and has a significant positive effect on the development of various industries. At the
same time, however, other scholars hold another view. Lin et al. [1] believe that network
infrastructure will not directly improve economic efficiency. Yilmaz et al. [2] find that
network infrastructure construction in other regions will negatively impact local areas,
so the total spill-off effect produced by network infrastructure construction is negative.
Moreover, as for capital market information, on the one hand, network infrastructure
construction accelerates the transmission efficiency of both idiosyncratic information
and market and industry information, which has the opportunity to improve the syn-
chronicity of stock prices. Capital market information disclosure quality, on the other
hand, are under the influence of the surplus radical and smooth [3], the application of
the Internet to improve stock market transaction number and the information diffusion
speed, bring opportunities and challenges, which affect surplus radical and smooth, make
the information quality to drop, so the network infrastructure construction to a certain
extent caused the company to quality information be flooded, And then the synchronicity
of stock prices.
Therefore, the following hypothesis H1 is proposed. There is a significant positive
correlation between the city where the company is registered as the “broadband Chi-
na” demonstration city and the stock price synchronization and network infrastructure
construction harms the information efficiency of the regional capital market.
concentration. The higher the proportion of the largest shareholder, the lower the quality
of capital market information disclosure from earnings management [4]. This means
that ownership concentration plays a partial substitution role in the negative effect of
network infrastructure construction on the quality of information disclosure in the capital
market. Therefore, ownership concentration will weaken the positive effect of network
infrastructure construction on stock price synchronization.
Therefore, the following hypothesis H2 is proposed: The larger the proportion of
the largest shareholder, the weaker the positive correlation between the city where the
company is registered and the “broadband China” demonstration city and the stock
price synchronization, that is, the lower the ownership concentration, the stronger the
negative impact of network infrastructure construction on the information efficiency of
the regional capital market.
Second, the moderating effect of firm growth on the influence of network infrastruc-
ture construction on regional capital market information efficiency. Since the completion
of network infrastructure, the integration of Internet and enterprise goods and services
has played a key role in the growth of enterprises to a large extent. The market-to-book
ratio is an important indicator reflecting the growth of a company. The higher the market-
to-book ratio is, the stronger the enterprise’s incentive to violate the law [5]. Therefore,
the growth of a company increases the possibility of network infrastructure’s influence
on information quality. Therefore, a company’s growth will enhance the positive effect
of network infrastructure construction on stock price coherence.
Therefore, the following hypothesis H3 is proposed: The higher the market-to-book
ratio is, the stronger the positive correlation between the city where the company is
registered and the demonstration city of “Broadband China” and the stock price syn-
chronization is, that is, the greater the growth of the company is, and the stronger the
negative impact of network infrastructure construction on the information efficiency of
regional capital market is.
The moderating effect of stock market transaction noise on the influence of network
infrastructure construction on regional capital market information efficiency. Stock price
synchronicity is influenced by enterprise quality information and noise [6], on the net-
work, information infrastructure construction and the capital market efficiency should be
caused by noise interference is discussed, only when the noise is big, the network infras-
tructure construction and the relationship between stock price synchronicity or signifi-
cantly stronger, to draw a network infrastructure influence the efficiency of information
rather than the noise of the conclusion.
Therefore, the following hypothesis H4 is proposed: The greater the trading noise
in the stock market, the stronger the positive correlation between the city where the
company is registered and the demonstration city of “Broadband China” and the stock
price synchronization, that is, the stronger the negative impact of network infrastructure
construction on the information efficiency of the regional capital market.
Does Network Infrastructure 737
3 Study Design
3.1 Variable Set
Explained Variables. Stock price synchronization is the first choice to measure the
information efficiency of the capital market, and it is also one of the most widely used
indicators. Based on Gul et al. [7] measurement of stock price synchronicity, the market
model is as follows:
Among them, SYNit represents the stock price synchronization of the company i in
t years. The deformation of the model (2) is to expand the index value range to infinity,
and SYN is set as the explained variable. The higher the stock price synchronization is,
the more serious the phenomenon of “rising and falling together” will be, and the lower
the information efficiency of the capital market will be.
Adjustment Variables. According to Chen Donghua and Yao Zhenye [9], the owner-
ship concentration of a company is measured by the proportion of the largest shareholder,
which is set as First, and the growth of the company is measured by the market-to-book
ratio, which is set as MB. The intensity of trading noise is defined by Shen Yongtao and
Gao Yusen [10], and the trading noise in the stock market is measured on the basis of
the negative of the first order autocorrelation coefficient of the regression residual term
of the above market model (1). The greater the variable value, the greater the noise, set
as Noise.
738 G. Yang et al.
Control Variables. Reference dong-hua Chen and Yao Zhenye [9], the selection of
control variable is shown below: The variable of company size is the natural logarithm
of the total assets of the enterprise, set as LNSIZE; The financial leverage variable is
the total liabilities divided by the total assets, set as Lev; The variable of listing years
is the natural logarithm of the listing time of the enterprise, set as LNAGE. Profitability
variable is the return on equity of the enterprise, set as ROE; The variable of market
activity is the natural logarithm of stock market trading volume, set as LNSTV. The
dummy variable of audit quality is whether the Big Four accounting firms audit the
auditor. If the value is 1, if not, it is 0, and set to BIG4. The dummy variable of enterprise
nature is whether it is state-owned holding, if it is 1, if it is not, it is 0, and set as SOE.
the information efficiency of the regional capital information market. Equation (1) is the
result that the synchronicity of stock price is only affected by whether the city where the
company is registered is in the list of “Broadband China” demonstration cities. In order
to ensure the stability of the results, the fixed effects of time, individual, region, and
industry were considered. Equation (2) adds regulating variables and control variables
based on Eq. (1), which is an empirical study of the model (3). Equation (3) considers
the moderating effect of ownership concentration and is an empirical study of the model
(4). Equation (4) considers the moderating effect of company growth and is an empirical
study of the model (5). Equation (5) takes into account the moderating effect of stock
trading noise and is an empirical study of the model (6). At the same time, whether fixed
effects were added, sample size, the goodness of fit and overall significance test results
of the equation were marked at the bottom of the table.
Network infrastructure construction has a negative impact on the information effi-
ciency of regional capital market. In Eq. (1), the influence coefficient of the explanatory
variable on the explained variable is positive 0.065, and the variable passes the T test and
the equation passes the F test at the 1% significance level. It indicates that whether the
city where the company is registered is in the list of “Broadband China” demonstration
cities has a significant impact on the stock price synchronization at the 1% level and a
positive correlation, that is, network infrastructure construction has a negative impact on
the information efficiency of regional capital market. Equation (2) shows that even on
the basis of control variables, the relationship between the two still passes the T-test at
the significance level of 1%, and the coefficient increases to 0.071. After adjustment, R2
increases from 0.364 to 0.396, indicating that the goodness of fit of the equation is higher.
The F value of the significance test of the equation rises from 7.867 to 38.02, indicating
that the significance of the equation is higher. It can be seen that the conclusion obtained
by Eq. (1) is more reliable, so hypothesis H1 is supported by empirical evidence.
The lower the ownership concentration, the stronger the negative impact of network
infrastructure construction on regional capital market information efficiency. In Eq. (3),
the variable “FirstTreatPost” is the product of variable “First” and variable “TreatPost”
after decentralized treatment. After adding the variable “FirstTreatPost” into Eq. (2),
the meaning of “First” and “TreatPost” no longer represents the variable itself. The
purpose of decentralization is to make First significant without unnecessarily affecting
the conclusion. It can be seen that the equation passes the F test at the significance level
of 1%, the FirstTreatPost variable passes the T test at the significance level of 1%, and
the coefficient is negative −0.003. In contrast, the coefficient of TreatPost variable is
significantly positive, indicating that the regulating variable First has a weakening effect
in the influence of the main model. The higher the ownership concentration, The network
infrastructure construction has a weaker negative impact on the information efficiency
of regional capital market. Therefore, hypothesis H2 is empirically supported.
The greater the growth of the company, the stronger the negative impact of network
infrastructure construction on the information efficiency of regional capital market. Same
as the above principle, the variable MBTreatPost in Eq. (4) is the product of variable
MB after decentralized treatment and variable TreatPost. The equation passes the F
test at the significance level of 1%, and the variable MBTreatPost passes the T test
at the significance level of 1%, with a positive coefficient of 0.062. The coefficient of
Does Network Infrastructure 741
TreatPost is the same as that of TreatPost, indicating that the regulating variable MB
has an enhanced effect in the influence of the main model. The greater the growth of
the company, the stronger the negative impact of network infrastructure construction
on the information efficiency of regional capital market. Therefore, hypothesis H3 is
empirically supported.
The greater the noise of stock market, the stronger the negative impact of network
infrastructure construction on the information efficiency of regional capital market.
Equation (5) The variable NoiseTreatPost is the product of variable Noise and vari-
able TreatPost. The equation passes the F test at the significance level of 1%, and the
variable NoiseTreatPost passes the T test at the significance level of 5%, with a posi-
tive coefficient of 0.202. The coefficient of TreatPost is the same as that of TreatPost,
indicating that the moderating variable Noise has an enhanced effect in the influence of
the main model. The greater the Noise of stock market trading, the stronger the negative
impact of network infrastructure construction on the information efficiency of regional
capital market. Therefore, hypothesis H4 is empirically supported.
Table 2. (continued)
Acknowledgment. This research was supported by the 2020 Ideological and Political Work
Research Project of Harbin University of Commerce (2020SZY002)and Humanities and Social
Science Research Planning Foundation of Ministry of Education of China(21YJAZH099).
References
1. Lin, W.T., Shao, B.B.M.: The business value of information technology and inputs substitu-
tion: the productivity paradox revisited. Decis. Support Syst. 42(2), 493–507 (2006)
2. Yilmaz, S., Haynes, K.E., Dinc, M.: Geographic and network neighbors, spillover effects of
telecommunications infrastructure. J. Reg. Sci. 42(2), 339–360 (2002)
3. Bhattacharya, U., Daouk, H., Welker, M.: The world price of earnings opacity. Accounting
Rev. 78(3), 641–678 (2003)
4. Lian, Y.H., Zhang, L.: Ownership concentration, market constraints and earnings management
of commercial banks. J. Financ. Econ. 34(02), 125–138 (2019)
5. Xiao, Q., Shen, H.Y.: Research on analyst concern, growth and corporate violations. Bus.
Res. 10, 116–125 (2017)
6. Zhang, Y.R., Li, X.Y.: Information content measurement in R2 and stock price. J. Manage.
Sci. 13(05), 82–90 (2010)
7. Gul, F.A., Kim, J.-B., Qiu, A.A.: Ownership concentration, foreign shareholding, audit quality,
and stock price synchronicity: evidence from China. J. Finan. Econ. 95(3), 425–442 (2010).
https://doi.org/10.1016/j.jfineco.2009.11.005
8. Xue, C., Meng, Q.X., He, X.J.: network infrastructure construction and enterprise technology
knowledge diffusion, a quasi-natural experiment from “broadband China” strategy. Finan.
Econ. Res. 46(04), 48–62 (2020)
9. Chen, D.G., Yao, Z.Y.: Does government action necessarily increase stock price synchroniza-
tion? – Empirical research based on China’s industrial policy. Econ. Res. J. 53(12), 112–128
(2018)
10. Shen, Y.T., Gao, Y.S.: Is the existence of a-share analysts reasonable? Shanghai Finan. 05,
24–32 (2020)
Optimized Coloring Algorithm Based
on Non-local Neighborhood Search
Abstract. Image coloring is the process of adding color to grayscale images with
the assistance of a computer. This paper proposes an optimal coloring algorithm
based on non-local neighborhood search to better deal with boundary color pen-
etration when colorizing grayscale images. First, the K non-local texture most
similar neighborhoods are searched based on the image texture features. Then,
based on the neighborhood pixels with similar intensity having similar color, the
group of similar neighborhood pixels is further searched. The search results take
into account not only the pixel texture features but also the light intensity distribu-
tion. Finally, the neighborhood pixels are introduced into the optimized coloring
method, and the image is colored. The experimental results show that the way can
effectively reduce boundary color penetration by non-local propagation of color
information. In the case of only a tiny amount of color smearing, the method in
this paper can provide better visual effects.
1 Introduction
Image coloring is a challenging computer vision problem. In recent years, more and
more researchers have shifted their attention to the direction of image coloring. Image
coloring refers to the image processing technology that adds color to gray images or
videos through computers according to the color of the user’s graffiti. This is a term
put forward by Wilson Markel in 1970, used to describe his computer-aided process of
adding color to grayscale movies or TV programs [1]. In recent years, image coloring
is also widely used in science, military, and medical fields [2, 3].
With the further exploration of image coloring by researchers at home and abroad,
many effective algorithms have been put forward, which can achieve the goal of the
accurate coloring of gray images to a certain extent. Among them, the original method
is to segment the picture according to the texture information of the gray image and then
color each region separately [4]. Although this method is simple, the coloring effect
of the image edge parts, such as hair and other complex parts, will have a particular
deviation. Therefore, the segmentation algorithm can only deal with simple images, and
image coloring puts forward higher requirements for the segmentation algorithm. Among
the coloring methods based on color labeling, the representative one is the optimized
coloring algorithm proposed by Levin et al. [5], which assumes a similar relationship
between color and brightness of adjacent pixels in space. Subsequently, Yatziv et al.
[6] proposed a re-coloring method based on weighted distance function derivation and
fusion, and Fattal et al. [7] applied wavelet transform technology to the image editing
method. Wang et al. [8] proposed an algorithm based on neighborhood correlation and
optimization technology. Generally, these methods can’t get ideal results at the boundary,
especially when users doodle a few colors. The algorithms described above are all image
coloring methods based on local neighborhood propagation.
The image coloring method based on global neighborhood propagation can realize
global color propagation. Even if the colored pixels are far away from the user’s location,
they can be better colored. Charpiat et al. [9] proposed a global optimization method
for graph cut for automatic color allocation. Sheng et al. [10] proposed the progressive
coloring optimized by the Gabor filter in the feature space based on the non-local idea.
Musialski et al. [11] proposed a color replacement strategy to keep the distance between
the input image pixels and map the source image sample to the target image. Compared
with the local neighborhood coloring method, global neighborhood propagation reduces
the number of user color graffiti but lacks local or direct color selection. The coloring
method based on local neighborhood propagation or global neighborhood propagation
has certain limitations. When two similar colors are colored into different colors, it will
cause color mixing, resulting in error coloring results. Therefore, this paper proposes an
image coloring method based on a non-local neighborhood search.
Machine learning has been continuously developed in recent years, and it has gradu-
ally penetrated various academic fields. The architecture of the network is also constantly
improving, and network models such as the VGG network, GoogLeNet network, and
ResNet network gradually appear. More and more scholars began to apply machine
learning to image coloring and achieved good results in automatic or semi-automatic
color. The whole process of coloring can be divided into the training stage and the
coloring stage. Model training usually requires many different types of images, which
reduces complex user interaction and reduces the complexity of selecting reference
images. Cheng et al. [12] use many different types of pictures as training sets and CNN
as training models, so extracting image features and color information as training sets is
necessary. Subsequently, Iizuka et al. [13] proposed an end-to-end convolutional neural
network model, which can directly obtain color images from gray images, thus elim-
inating the links of pre-processing and post-processing. But the above network model
needs a lot of pictures for training. The algorithm proposed in this paper can fully play
its advantages when the number of images is minimal.
In this paper, we can better solve the discontinuous or overflow color area problem
based on the optimized coloring algorithm, combined with non-local neighborhood
search. At present, most existing coloring technologies only consider the light intensity
distribution. Although these methods have a good coloring effect on smooth image areas,
they are usually ineffective for vital texture areas unless the user doodles the image areas
in sufficient detail. Therefore, this paper introduces a non-local neighborhood search
method based on texture similarity and spatial coordinate distance. Experimental results
show that the coloring method proposed in this paper is very effective.
746 H. Xin and E. Peng
where s denotes the search time for the number of vectors in the sample data set. b
denotes the time to build the tree, m = mt /md denotes the ratio of the memory used
for the tree (mt ) to the memory used for storing data (md ). The build time weight (wb )
controls the importance of the tree build time relative to the search time.
Philip et al. [15] proved that non-local theory could effectively reduce user input
and maintain accuracy. However, with the increase of the searched pixel neighborhood,
the time of searching k non-local texture most similar neighborhoods increases sig-
nificantly. The KNN method has been widely used because of its simplicity and easy
implementation. Using FLANN, it is easy to search the most similar neighborhoods of k
non-local textures, which is proven effective in [16]. It takes non-local texture similarity
as constraint, and the constructed feature vector is expressed as:
F = {Y (Np ), γ × I , γ × j} (2)
In which Y(Np) is a small luminance area around pixel p in the luminance channel.
Np represents a square neighborhood with a fixed size centered on pixel p. (i, j) is
the coordinate of pixel p, and the parameter γ is used to adjust the search range of a
similar neighborhood. When parameter γ becomes larger, the search area of a similar
neighborhood becomes smaller. When the γ parameter is too large, it does not work
when searching k non-local nearest texture similar neighborhoods. Huang et al. [17]
concluded through experiments that the representative value of parameter γ ranges from
0.5 to 5 and the representative value of parameter K ranges from 5 to 15 in K nearest
neighbor search. Therefore, in this paper, the parameters are fixed at γ = 1 and K = 15,
are used to search K non-local nearest texture-similar neighborhood pixels.
Because the gray image has no color information, only make full use of gray infor-
mation to establish the matching relationship between images. This algorithm can be
applied to images with rich textures. Considering the texture information of the image,
it can better solve the color penetration at the image boundary.
of chroma value. By observation, adjacent pixels with similar intensity have similar
colors. In this algorithm, the neighborhood similarity pixel search method proposed by
Wang et al. [18] is used. The neighborhood pixels are searched by taking a pixel position
and pixel intensity as constraints.
As shown in Fig. 1, purple pixels represent pixels to be estimated. For purple pix-
els, if the distribution of adjacent pixel positions is uniform, such as the green part in
Fig. 1(a) when the chromaticity value of purple pixels is a linear weighting function of
the chromaticity value of green pixels, the final estimated chromaticity value will cause
deviation. If the neighborhood similarity pixel method is used to search, as shown in
Fig. 1(b), the green part is an irregular distribution. It can be seen from Fig. 1(b) that the
chromaticity values of all green pixels are closer to those of purple pixels. Therefore,
based on the neighborhood similarity distribution, the coloring effect of purple pixels is
most relative to its original image.
The parameters σs2 and σr2 adjust the spatial similarity and intensity similarity, respec-
tively. It can be known from formula (2) that when two pixels are close in spatial position
and brightness value, the similarity d i value should be larger. When two pixels are far
away, and the difference in brightness value is larger, the similarity d i value should be
smaller. Order d from high to low, and the first m pixels are the neighborhood d i pixels
of the pixel r that this paper wants to acquire.
Due to the close relationship between the pixels to be colored and the neighborhood
position, the algorithm considers the brightness information of the image. It combines
the location information of the pixels to calculate the neighborhood similarity and realize
748 H. Xin and E. Peng
the non-local neighborhood search. In the case of a small amount of artificial coloring,
it has a good coloring effect.
3 Algorithm Steps
Our algorithm works in YUV color space. YUV [19] is a color-coding method adopted by
European TV systems, in which Y is a luminance channel, which is usually called lumi-
nance or intensity, while U and V are chrominance channels, which are two components
of the color.
In this experiment, firstly, we use FLANN [14] method to search k non-local nearest
neighbor pixels with similar textures. Then, based on the KNN algorithm, pixel location
and pixel intensity are taken as constraints of neighborhood search, and similar neigh-
borhood pixel groups are further searched. Then, the local pixel neighborhood relation
wrs of the pixel to be colored in the luminance channel is calculated, and these weight-
ing coefficients are transferred to the chroma. Finally, the color of pixels is obtained by
solving a quadratic optimization problem.
The brief framework of the whole algorithm is shown in the following Fig. 2.
Where σr2 is the variance of the intensity of the neighboring pixels searched for in pixel
r. wrs is a weighting coefficient whose sum is 1. It can be seen from formula (4) that the
more similar Y(r) is to Y(s), the larger the value of wrs is. When the two intensities are
different, the weighting coefficient is more minor.
Step 3: Calculate Brightness Value. As mentioned in Levin’s method [5], this paper
wants to constrain two adjacent pixels r, s. If their intensities are similar, they should
have similar colors. Therefore, it is desirable to minimize the color weighted average at
the adjacent pixel s at the pixel U(r) and V(r), thereby reducing the difference between
the pixel r and the adjacent pixel s. In this paper, it is calculated by the formula (5) and
formula (6). When the calculated value tends to the minimum, the chroma value of each
pixel is obtained.
2
J (U ) = U (r) − wrs U (s) (5)
r s∈N (r)
2
J (U ) = V (r) − wrs V (s) (6)
r s∈N (r)
insufficient, serious color penetration will occur. Yatziv algorithm, like Levin algorithm,
also has different degrees of color penetration problems. The coloring effect of the new
method proposed in this paper is as shown in Fig. 4(d), which can keep the authenticity
and accuracy of the coloring effect on the premise of using the same user input as the
previous two methods. This method can spread color information globally because the
feature space contains global color’s texture and brightness features. When the marking
range of color lines is insufficient, image color penetration can be better solved.
Tables 1 and 2 lists the PSNR value and SSIM results of the grayscale image after
coloring by different methods. It can be seen from the following table that the output
of the coloring algorithm in this paper is almost superior to the Levin algorithm and
Yatziv algorithm. As shown in a bathroom in Fig. 4, when the user tag area is small and
the boundary is obvious, the Levin and Yatziv algorithms have more boundary color
penetration problems. The proposed method can better solve these problems, and the
corresponding PSNR values are also significantly higher than those of the other two
methods.
Optimized Coloring Algorithm Based on Non-local Neighborhood Search 751
archive
bathroo
m
beach
valley
Table 1. The PSNR results of the new algorithm, Levin algorithm and Yatziv algorithm.
Table 2. The SSIM results of the new algorithm, Levin algorithm and Yatziv algorithm.
5 Conclusion
At present, the optimized coloring algorithm has some problems, such as complicated
user interaction and easy color penetration at the image boundary. To solve these prob-
lems, we propose a new search method that combines non-local neighborhood search
with an optimized coloring algorithm and effectively combines local and non-local color-
ing advantages. When the user mark range is insufficient, the color penetration of image
boundary is effectively solved. The method proposed in this paper can also provide
convenience for other fields, such as animation, film and television, medical treatment,
advertising design, decoration design, etc. Based on the existing algorithm, the follow-
ing improvements can be made. Because the CPU calculates the neighborhood pixels
in the non-local neighborhood search, its execution efficiency needs to be improved.
Therefore, the next step is to use CUDA programming to realize the calculation by GPU
and reduce the running time of this algorithm [20].
Acknowledgment. This work was supported by the grants of the Nature Science Foundation of
Heilongjiang Province (LH2021F035).
References
1. Burns, G.: Colorization [EB/OL]. http://www.museum.tv/archives/etv/
2. Lagodzinski, P., Smolka, B.: Application of the extended distance transformation in digital
image colorization. Multimedia Tools Appl. 69(1), 111–137 (2014). https://doi.org/10.1007/
s11042-012-1246-2
3. Lin, S., Ritchie, D., Fisher, M., Hanrahan, P.: Probabilistic color-by-numbers. ACM Trans.
Graphics 32(4), 1–12 (2013)
4. Jia, Y., Hu, S.: Interactive image dyeing algorithm based on graph segmentation. Chin. J
Compu. 29(3), 508–513 (2016)
5. Levin, A., Lischinski, D., Weiss, Y.: Colorization using optimization. ACM Trans. Graphics
(TOG) 23(3), 689–693 (2004)
6. Yatziv, L., Sapiro, G.: (2006) Fast image video colorization using chrominance blending.
IEEE Trans. Image Process. 15(5), 1120–1129 (2006)
7. Fattal, R.: Edge-avoiding wavelets and their applications. ACM Trans. Graphics (TOG) 28(3),
1–10 (2009)
8. Wang, M., Chen, Z.: A Color Transfer algorithm based on neighborhood correlation and
optimization techniques. In: Fourth International Symposium on Computational Intelligence
and Design, vol. 2, pp. 31–34 (2011)
Optimized Coloring Algorithm Based on Non-local Neighborhood Search 753
9. Charpiat, G., Hofmann, M., Schölkopf, B.: Automatic image colorization via multimodal
predictions. In: Forsyth, D., Torr, P., Zisserman, A. (eds.) ECCV 2008. LNCS, vol. 5304,
pp. 126–139. Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-540-88690-7_10
10. Sheng, B., Sun, H., Chen, S., et al.: Colorization using the rotation-invariant feature space.
IEEE Comput. Graphics Appl. 31(2), 24–35 (2011)
11. Musialski, P., Cui, M., Ye, J., et al.: A framework for interactive image color editing. Visual
Comput. 29(11), 1173–1186 (2013)
12. Cheng, Z.,Yang, Q., Sheng, B.: Deep colorization. In: Proceedings of the IEEE International
Conference on Computer Vision, pp. 415–423 (2015)
13. Lizuka, S., Simo-Serra, E., Ishikawa, H.: Let there be color: joint end-to-end learning of global
andlocal image priors for automatic image colorization with simultaneous classification. ACM
Trans. Graphics (TOG) 35(4), 110 (2016)
14. Muja, M., Lowe, D.G.: Fast approximate nearest neighbors with automatic algorithm
configuration. In: VISAPP, no. 1, pp 331–340 (2009)
15. Lee, P., Wu, Y.: Non-local matting. In: Proceedings of the 2011 IEEE conference on com-
puter vision and pattern recognition, CVPR ’11, pp 2193–2200. IEEE Computer Society,
Washington, DC (2011)
16. Chen, Q., Li, D.: Tang C-K Knn matting. IEEE Trans. Pattern Mach. Intell. 35(9), 2175–2188
(2013)
17. Huang, H., Li, X., Zhao, H., Nie, G., Hu, Z., Xiao, L.: Manifold-preserving image colorization
with non-local estimation. Multimedia Tools Appl. 74(18), 7555–7568 (2015)
18. Wang, H., Gan, Z., Zhang, Y., Zhu, X.: Novel colorization method based on correlation
neighborhood similarity pixels priori[A]. IEEE Beijing Section. In: Proceedings of 2012
IEEE 11th International Conference on Signal Processing (ICSP 2012)[C], pp. 953–956.
IEEE Beijing Section:IEEE BEIJING SECTION (2012)
19. Jack, K.: Video Demystified, 4th edn. Elsevier Science & Technology (2001)
20. Joshi, M., Nkenyereye, L., Joshi, G., et al.: Auto-colorization of historical images using deep
convolutional neural networks. Mathematics 8(12), 2258 (2020)
Research on Digital Twin for Building Structure
Health Monitoring
Su Jincheng1,2(B)
1 Postdoctoral Research Workstation of Northeast Asia Service Outsourcing Research Center,
Harbin University of Commerce, 150028 Harbin, Heilongjiang Province,
People’s Republic of China
2 School of Energy and Civil Engineering, Harbin University of Commerce, 150028 Harbin,
1 Introduction
Buildings are a kind of man-made thing, and man-made things have life. The design
service life of ordinary houses is 50 years. Within the design basis period, only normal
maintenance without overhaul can be used for the intended purpose and complete the
predetermined function. Buildings are vulnerable to damage due to adverse factors such
as multi-field coupling, material degradation, and corrosion. In engineering practice,
structural damage means that the building cannot work in the best state or achieve the
design performance under the target service environment. Generally, structural health
monitoring is used to identify the damage of engineering structures and provide early
warning.
Since the end of the 20th century, civil structures’ large-scale, complex and intel-
ligent development has made structural health monitoring more and more important.
Structural health monitoring extends from single load stress monitoring to structural
damage detection, rapid positioning, and life prediction. Structural health monitoring is
a process of structural damage identification and early warning using the data obtained
2 Digital Twin
The term “digital twin” comes from the US Department of Defense and is used for
aerospace vehicles’ health maintenance and guarantee [1]. In 2011, the US Air Force
756 S. Jincheng
Research Laboratory proposed applying digital twin technology to the life management
of aircraft structure, resulting in the concept of aircraft digital twin, to solve aircraft oper-
ation and maintenance in complex service environments in the future. The application
of the concept of the digital twin has also been further extended to urban management.
Xiong’an New Area proposes to “adhere to the synchronous planning and construction
of digital city and real city”. Based on the theory of urban complex adaptive system and
digital twin technology, the “hidden order” of the urban system can be made explicit
by building a complex system of mutual mapping and collaborative interaction between
physical city and digital city [2].
At present, digital twins have many different definitions and understandings in aca-
demic circles. It is widely accepted that the digital twin uses physical models, sensor
updates, operation history, and other data. It is a simulation process integrating multi-
disciplines, multi-physical fields, and multi-scale and multi-probability. It completes the
mapping in the virtual space and reflects the whole life cycle process of the corresponding
entity equipment.
In other words, the digital twin technology system must support elements such as
virtual space, physical space, and two-way information flow and play a role in its whole
life cycle. Figure 1 shows the technical system of digital twins.
Modeling
Virtual Space
Simulation and Optimization
Operation Management
Technology
System of Physical Space Automation
Digital Twins
Digital Analysis
3 Key Issues
Monitoring is a direct means for humans to scientifically understand the world and the
basis and premise for controlled substances and their development laws. The health
monitoring of engineering structures is to deploy large-scale, multi-species, distributed
sensor networks and acquisition, transmission, management, analysis, and early warning
Research on Digital Twin for Building Structure Health Monitoring 757
systems. It is used to real-time perceive, identify, diagnose and evaluate the damage and
safety state of the structure and its evolution law, reveal the whole life behavior mecha-
nism of the real structure under the coupling action of real load and environment, and the
intelligent function of bionic human self-sensing and self-diagnosis [3]. Figure 2 shows
the theoretical system of structural health management. Figure 3 shows the technical
composition of the structural health monitoring system.
Data Collection
Data Transmission
Data Management
Early Warning
Validate Existing Methods
Evaluation
Developing New Methods
Decision Making
Structural health monitoring includes two major scientific problems: sensing and
data [4]. Since 2005, the application of a structural health monitoring system has pro-
duced more and more monitoring data. These data contain important information such as
structural load and environmental action, behavior mechanism, performance evolution
law, evolution level, etc. A key problem is how to mine and analyze monitoring data to
find structural load and environmental action, structural response, behavior mechanism
and performance evolution law, evaluate and predict structural safety level and change
law, and establish structural health monitoring theory based on monitoring data.
4 Research Paradigm
The digital twin has three meanings for structural health monitoring: deepening under-
standing, strengthening prediction, and in-depth optimization. The main key problems
are multi-physical and multi-scale modeling with high fidelity, prediction and analysis
with high confidence, and high real-time data interaction.
Digital twin emphasizes the use of all data and the diversity of data provided by
sensors and data to drive the acquisition of knowledge to increase cognition. Digital twin
integrates big data, dynamic data-driven application systems, and machine intelligence,
which has changed the research paradigm of structural health management.
758 S. Jincheng
Damage Identification
Basic Information of Structure
Model Modification
Monitoring System Information
Structure Reanalysis Remote Network Monitoring Data Information
Health and Safety Assessment Control Safety Warning Information
Reliability and Life Prediction
Safety Assessment Information
Big data analysis finds hidden patterns, unknown relationships, and other useful
information that can help decision-making from big data and convert the above infor-
mation into executable knowledge. There are three main modes of big data analysis:
descriptive analysis, predictive analysis, and prescriptive analysis. Predictive analysis is
the core of big data analysis by mining patterns or relationships in big historical data,
analyzing trends, production prediction models and scores to evaluate possible events in
the future and predict what may happen in the future. The descriptive analysis summa-
rizes the original data into an understandable format and reveals what happened, which
is the basis of big data analysis. The prescriptive analysis is based on the prediction
model to analyze the possible different results under different decisions and determine
the strategies to be adopted.
The National Natural Science Foundation of the United States formally put forward
the dynamic data-driven application system in 2000, a new research model of cooperative
application and sensing systems. For complex systems such as building structures, the
system state evolves with the system’s operation and has dynamic characteristics. The
dynamic test data is used to modify the analysis model in real-time and adaptively,
eliminate the influence of uncertain factors to the greatest extent, give more accurate
results, and control the implementation of the actual system by participating in the
system decision-making [5].
Driven by big data technology, the digital twin is becoming a reality. A digital twin
can track the system’s state in real-time, carry out online state evaluation, improve the
response to faults, speed up decision-making, and realize system optimization. Unlike
the post-diagnosis mode of diagnosis and maintenance after the system has problems, the
digital twin is transformed into life-cycle management, pre-diagnosis of structural state,
and implementation of preventive maintenance, which greatly increases the availability
and service life of the system and reduces the maintenance cost.
Research on Digital Twin for Building Structure Health Monitoring 759
The uncertainty of structural health monitoring comes from the uncertainty of the
parameters and modeling of the structure itself and the measurement noise of the sensor.
Therefore, the uncertainty of the structure model and the measurement noise can be
considered well by studying the structural uncertainty damage identification method.
At present, structural uncertain damage identification methods can be divided into
two categories: structural damage identification methods based on probability and struc-
tural damage identification methods based on information fusion theory. The former
includes the structural damage identification method based on Bayesian probability the-
ory and the structural damage identification method based on the stochastic finite ele-
ment. The latter includes the structural damage identification method based on Dempster-
Shafer (D-S) evidence theory and the information fusion structural damage identification
method based on Bayesian decision theory.
Unlike finite element modeling, a data-driven (statistical) approach using monitoring
data provides information about the actual operational performance of a specific infras-
tructure [6–9]. The modeling workload related to data-driven technology is relatively
lower than that of finite element modeling. However, even the most complex data-driven
model updated by the artificial neural network, Gaussian process regression, or Bayesian
model requires many actual training data for prediction. Ye et al. described the com-
parison between finite element and data-driven methods based on hybrid or digital twin
methods [10]. These challenges drive research in the field of model updating and system
identification, which attempts to use finite element models and data-driven models (i.e.,
models derived from sensor data) to better predict system behavior [11].However, the
current modeling method seems to be unable to synthesize the uncertainty of measure-
ment data and the prediction from the finite element model specified by inherent error
to generate continuous prediction when the new measurement data is available. This
capability is essential to the realization of digital twins.
5 Conclusions
In this study, the digital twin of building structure health management is studied. The
digital twin for building structure health management has the following meanings:
With the increase of service time of buildings, the demand for real-time and efficient
structural health management is becoming stronger and stronger. The digital twin will
play a great role in this field and form a life cycle management paradigm based on the
digital twin.
760 S. Jincheng
Acknowledgment. This work was supported by University Nursing Program for Young Scholars
with Creative Talents in Heilongjiang Province (UNPYSCT-201720 7) and Heilongjiang Natural
Science Foundation (E2017055).
References
1. Tuegel, E.J., Ingraffea, A.R., Eason, T.G., et al.: Reengineering aircraft structural life
prediction using a digital twin. Int. J. Aerosp. Eng. 2011, 1687–5966 (2011)
2. Zhou, Y., Liy, C.C.: The logic and innovation of building digital twin city in Xiong’an new
area. Urban Dev. Stud. 25(10), 60–67 (2018)
3. Ou, J.: Research and practice of intelligent sensing technologies in civil structural health
monitoring in the mainland of China. In: Nondestructive Evaluation and Health Monitoring
of Aerospace Materials, Composites, and Civil Infrastructure, vol. 6176, pp. 293–304 (2006)
4. Li, H., Bao, Y., Ou, J.: Structural damage identification based on integration of information
fusion and shannon entropy. Mech. Syst. Signal Process. 22(6), 1427–1440 (2008)
5. Frederica, D., Rotea, M.: Dynamic data-driven applications systems. In: Proceedings of the
2006 ACM/IEEE Conference on Supercomputing, p. 2 (2006)
6. Eky, F., et al.: A Self-Sensing Digital Twin of a Railway Bridge Using the Statistical Finite
Element Method. arXiv: Numerical Analysis (2021)
7. Girolami, M., Febrianto, E., Yin, G., et al.: The statistical finite element method (statFEM)
for coherent synthesis of observation data and model predictions. Comput. Methods Appl.
Mech. Eng. 375, 113533:1-113533:32 (2021)
8. Malekzadeh, M., Atia, G., Catbas, F.N.: Performance-based structural health monitoring
through an innovative hybrid data interpretation framework. J. Civ. Struct. Health Monit.
5, 287–305 (2015)
9. Liu, X., et al.: Application of Structural Health Monitoring for Structural Digital Twin.
Offshore Technology Conference Asia (2020)
10. Ye, C., Butler, L., Bartek, C., et al.: A Digital twin of bridges for structural health monitoring.
In: 12th International Workshop on Structural Health Monitoring (2019)
11. Upmanyu, T., Hussain, S., Bharadwaj, S., Saxena, S.: Digital health monitoring and pervasive
healthcare using cloud-connected smart wearable devices. Int. J. u - and e – Serv. Sci. Technol.
10(1), 289–298 (2017)
Research on Modeling and Solution of Reactive
Power Optimization in Time Period
of Power System
Chen Deyu(B)
Abstract. In the reactive power and voltage optimization of the power grid, dis-
crete control equipment such as capacitors and transformer taps act once in a
period to realize the optimization of the integral sum of multi-objective functions
such as line loss, total generator reactive power reserve, and equalization of reac-
tive power reserve of each generator at each time, which can be transformed into
the optimization of the expectation and variance of the objective function at each
time in the period, It is the optimization model of discrete control variables in the
period; As a continuous control variable, the generator terminal voltage can be
adjusted at each time further to realize the optimization of the period objective
function, that is, it is defined as the period continuous control variable optimization
model. The discrete and continuous control variable optimization model is solved
by cross iteration until convergence, and an adaptive weight genetic algorithm is
proposed to solve it. A simulation example verifies the effectiveness of the model
and algorithm.
1 Introduction
In the actual production of the power grid, the service life of static compensation devices,
such as capacitors, will be reduced with frequent switching under working conditions.
At the same time, the static compensation devices are discrete and jump-type control,
and large changes in operating state after control may also affect the reliability of system
operation. Therefore, discrete control devices, such as capacitors, are generally adjusted
once in a period. This value is adapted to the working conditions at each time in this
period as far as possible, and its action frequency is reduced as far as possible. In contrast,
the finer control can be tracked and regulated by continuous control devices such as gen-
erator sets. In production practice, most of the actions of discrete and continuous control
devices are arranged according to operation experience. However, with the complexity
and changeability of power grid structure and operation mode, the control strategies
obtained from operation experience may be inapplicable. A strict mathematical model
⎛ ⎞ ⎤
F3 = − ⎝ QGi,j
m
(x(t), uc (t), ud )⎠dt ⎦
T i,j∈nG
s.t.
⎧ (1)
⎪
⎪ f (x(t), uc (t), ud ) = 0
⎨
h ≤ h(x(t), uc (t), ud ) ≤ h
⎪
⎪ uc ≤ uc (t) ≤ uc
⎩
ud ≤ ud ≤ ud
Where: T Is the time period; t is each time in the time period T , the following
variables related to t refer to the value of t at each time; ud is a discrete control variable,
which is specifically referred to as capacitor switching and transformer tap action, and is
only act once in time period T , ud , ud are the upper and lower limits respectively; uc (t)
is a continuous control variable, such as the terminal voltage setting of the generator,
and which is acted at every moment in the T1 time period, uc , uc are the upper and lower
limits respectively; x(t) is a state variable, such as node voltage amplitude and phase
angle; f(t) is the system power flow equality constraint; h(t) is the inequality constraint
of state variables and function variables, such as node voltage, reactive power output of
generator, h, h are the upper and lower limits of constraints; The Ploss (t) function is the
line loss of each time t in period T , then the optimization objective F1 is the energy
consumption intime period T ; QGim (t) is the reactive power margin of the generator
i at each time, QGim (t) is the sum of reactive power margins of all nG generators at
i∈nG
time t, and the optimization objective F2 is the integral of the sum of generator reactive
power margin in time period T ; QGi,j m (t) is the difference of reactive power margin
between generators i and j at time t, QGi,j m (t) is the sum of reactive power margin
i,j∈nG
differences of any two generators at time t, and the optimization objective F3 is the
integration of the sum of reactive power margin differences of any two generators in
time period T ; MaxF1 indicates the minimum energy consumption in the time period,
and maxF2 indicates the maximum reactive power margin of the generator in the time
period, thus ensuring that the generator can quickly provide more reactive power support
for the system when an emergency occurs, which is beneficial to the stability of the
system; MaxF3 indicates that the reactive power margin of each generator is balanced
as far as possible at each time in the time period, the balanced distribution of reactive
power margin is more conducive to system stability.
2.1 Reactive Power Optimization Model with Discrete Control Variables in Time
Period
Assuming that the continuous control variable uc (t) is known at each time in the model
(1), the discrete control variable ud acts once in the time period T , ensure the optimality
of objective functions F1 , F2 and F3 in the time period T , and the objective function
adopts the integral sum of line loss, generator reactive power margin and generator
reactive power margin equalization function in time period T .
764 C. Deyu
Using the equivalent expression of formula (3) to replace the objective function in
the form of integral sum in optimization model (1), the physical meaning will be more
obvious, that is, the discrete control variables are optimized once in a time period, which
improves the expected value of the objective function at each time in the time period,
that is, the objective function at each time in the time period is optimized as a whole.
At the same time, if the optimization objective only ensures the optimization of
the objective function of the integral sum in time period T , the objective function
may be improved a lot at some times, but the objective function may be much worse
at some times, resulting in the problem of uneven optimization. Therefore, based on
the mathematical expectation expression of the objective function at each time in the
above time period, the mathematical variance of the objective function at each time is
introduced to balance the optimization at each time in a time period:
t1 t2
D(f (t)) = · (f (t1 ) − E(f (t)))2 + · (f (t2 ) − E(f (t)))2
T T (4)
tj
+··· + · (f (tj ) − E(f (t)))2
T
MinD(f(t)) represents the equalization of objective function optimization at each
time in time period.
Assuming that the continuous control variables are given, and the expression forms
of expectation and variance in (3) and (4) are synthesized, the optimization model of
Research on Modeling and Solution of Reactive Power Optimization 765
discrete control variables in time period obtained by decomposition of model (1) can be
expressed as:
⎡ ⎛ ⎞ ⎛ ⎞⎤
max ⎣FE,1 = −E(Ploss (ud )), FE,2 = E ⎝ QGim (ud )⎠, FE,3 = −E ⎝ m (u )⎠⎦
QGi,j d
i∈nG i,j∈nG
⎡ ⎛ ⎞ ⎛ ⎞⎤
min ⎣FD,1 = D(Ploss (ud )), FD,2 = D⎝ QGim (ud )⎠, FD,3 = D⎝ m (u )⎠⎦
QGi,j d
i∈nG i,j∈nG (5)
s.t.
⎧
⎪ ∗
⎨ f (x(t), uc (t), ud ) = 0
h ≤ h(x(t), u∗c (t), ud ) ≤ h
⎪
⎩ ud ≤ ud ≤ ud
Where: u∗c (t) indicates the optimal value of continuous control variable at a given time t,
and the control variable in this model is only discrete control variable ud ; Other variables
have the same meanings as above.
The meaning of model (5) is that the discrete control variable acts once in time period
T , and the expectation E(.) and variance D(.) of the objective function are optimized.
The specific form of the optimization model (5) of discrete control variables in time
period is as follows:
⎡ ⎛ ⎞ ⎛ ⎞⎤
max ⎣FE,1 = −E(Ploss (c, b)), FE,2 = E ⎝ QGim (c, b)⎠, FE,3 = −E ⎝ m (c, b)⎠⎦
QGi,j
i∈nG i,j∈nG
⎡ ⎛ ⎞ ⎛ ⎞⎤
min ⎣FD,1 = D(Ploss (c, b)), FD,2 = D⎝ QGim (c, b)⎠, FD,3 = D⎝ m (c, b)⎠⎦
QGi,j
i∈nG i,j∈nG
s.t. (7)
⎧
⎪
⎪ f θ, V l , ukc (t), c, b = 0
⎪
⎪
⎪
⎪
⎨ Vl ≤ Vl ≤ Vl
⎪ QG ≤ QG ≤ QG
⎪
⎪
⎪
⎪ c≤c≤c
⎪
⎩
b≤b≤b
Where: ukc (t) represents the continuous control variable at each time t of the k-th
iteration, that is, the set value of generator terminal voltage, which is the given value in
the k-th iteration model (7); c is the equivalent susceptance after the capacitor is switched
once in a time period, c, c are the upper and lower limits of the equivalent susceptance of
the capacitor; b is the transformation ratio after the transformer tap is set once in a time
period, b, b are the upper and lower limits of transformation ratio; c and b are discrete
control variables of the model (7); θ is the voltage phase angle of all nodes, V l is the
load node voltage amplitude, V l , V l are the upper and lower limits of the V l , which
are generally 1.05 and 0.95; QG is the reactive power output of generator, QG, QG are
upper and lower limits of reactive power output of generator.
QGim is the reactive power margin of the ith generator, which is:
QG i − QGi
QGim =
QG i − QG i
If the generator reactive power QG is within the upper and lower limits, then the
maximum value of QGim is 1, indicating that the reactive power margin is sufficient; the
minimum value of QGim is 0, indicating that the reactive power margin is scarce and has
reached the upper limit.
The concrete form of the continuous control variable optimization model (6) at each
time t in the time period is as follows, omitting t, including:
⎡ ⎤
max ⎣F1 = −Ploss V g , F2 = QGim (V g ), F3 = − QGi,j
m
(V g )⎦
i∈nG i,j∈nG
s.t.
⎧ (8)
⎪
⎪ f (θ, V l , V g , uk ) = 0
⎪
⎨ V ≤ V ≤dV
l l l
⎪
⎪ V ≤ V g ≤ V g
⎪ g
⎩ QG ≤ QG ≤ QG
Research on Modeling and Solution of Reactive Power Optimization 767
Where: ukd indicates the set value of discrete control variables in the kth iteration,
such as the equivalent susceptance after capacitor switching and the transformer ratio
after transformer tap setting, which are given values; Vg is the generator node voltage
amplitude (including PV node and balance node), which is a continuous control variable
of the model (8). V g and V g are the upper and lower limit constraints of voltage amplitude
of such nodes, generally 1.1 and 0.9; Other variables have the same meanings as above.
The cross iteration of discrete control variable optimization model and continu-
ous control variable optimization model in time period is carried out k times until
maxuck+1 (t) − ukc (t) < εc , maxudk+1 − ukd < εd , the iteration is over, then the opti-
mal solution of the reactive power optimization model in time period is uc (t) and ud of
the k+1th iteration, and t is each time in time period T .
2k−1 < b − a ≤ 2k
The length of the binary string encoded in this way needs at least k bits.
The k-bit binary string (bk−1 bk−2 · · · b0 ) is transformed into the corresponding
integer in the [a, b] interval, which is divided into the following two steps:
1) Convert the binary number represented by this binary string into a decimal number:
k−1
(bk−1 bk−2 · · · b0 )2 = ( bj · 2j )10 = i
j=0
Take the transformer ratio b in the discrete control variable optimization model (7)
as an example to encode the control variable. If the upper and lower limits of b are 1.1
and 0.9 respectively, and the adjustable step size is 0.02, then: (1.1–0.9)/0.02 = 10, 23
10 24 , therefore, for a transformer ratio, the number of binary coding bits is 4. When
there are n transformer ratios controllable, the number of coding bits is 4 × n:
dn,3 dn,2 dn,1 dn,0 |dn−1,3 dn−1,2 dn−1,1 dn−1,0 | · · · · · · |d1,3 d1,2 d1,1 d1,0
In which dn,3 dn,2 dn,1 dn,0 represents a binary code of transformer ratio, and d = 0 or
1, and the binary code is restored to transformer ratio:
3
10 − 0
(d3 d2 d1 d0 )2 = ( dj · 2j )10 = i , and i = 0 + int i · 4
2 −1
j=0
b = 0.9 + i × 0.02
Coding ideas of capacitor susceptance c and generator terminal voltage Vg are the
same.
Optimization models (7) and (8) are all multi-objective optimization, and the key is how
to determine the weight coefficient of each objective in multi-objective optimization.
In this paper, the adaptive weight genetic algorithm proposed in [9] is used to solve
the multi-objective optimization. By using the characteristics and advantages of genetic
algorithm, the weight coefficient is changed continuously with the algebraic evolution,
and the strategy of adaptively determining the weight is given.
Take the model (7) as an example to calculate the fitness function. The maximization
of the three expected objective functions FE,1, FE,2, FE,3 remains unchanged, while the
negative signs are added to the three variance objective functions FD,1, FD,2, FD,3 as
−FD,1, −FD,2, −FD,3 respectively, then the minimization in model (7) is transformed
into maximization.
In genetic evolution, each generation produces a certain number of individuals to
form population P. The maximum and minimum values of all individual optimization
objectives Z FE1 of the current population are defined as:
max
zFE1 = max{FE,1 (V g,j , cj , bj )|V g,j , cj , bj ∈ P}
min
zFE1 = min{FE,1 (V g,j , cj , bj )|V g,j , cj , bj ∈ P}
Where: Vg,j , cj , bj are the terminal voltage setting value, capacitor switching amount
and transformer ratio corresponding to the coding scheme of any individual j in the
population.
max , z min and z max , z min can be obtained.
Similarly, zFE2 FE2 FE3 FE3
Research on Modeling and Solution of Reactive Power Optimization 769
max
zFD1 = max{−FD,1 (V g,j , cj , bj )|V g,j , cj , bj ∈ P}
min
zFD1 = min{−FD,1 (V g,j , cj , bj )|V g,j , cj , bj ∈ P}
max , z min and z max , z min can be obtained.
Similarly, zFD2 FD2 FD3 FD3
In the current generation, the adaptive weight of the expectation objective is expressed
by the following formula:
1
wEk = , k = 1, 2, 3
max
zFEk − zFEk
min
Then, for model (7), the adaptive weight objective function of multi-objective opti-
mization of individual j of each generation in the process of genetic evolution is as
follows:
3
3
z(V g,j , cj , bj ) = wEk (zFEk − zFEk
min
)+ wDk (zFDk − zFDk
min
) (9)
k=1 k=1
In the evolution process, according to some useful information of the current popula-
tion, the target preference is continuously refined, and the weight is adaptively adjusted
to obtain the search pressure towards the forward ideal point, which is different from the
traditional multi-objective optimization with fixed weight [9].
Inequality constraints are given by adaptive penalty function, for example, the
inequality constraints in model (7) can be calculated uniformly gi (V g , c, b) ≤ bi , i =
1, 2, · · · , M . For individuals Vg,j , cj , bj , the adaptive penalty function is:
M
1 bi (V g,j , cj , bj )
p(V g,j , cj , bj ) = 1 − (10)
M bmax
i
i=1
770 C. Deyu
In which:
bi (V g,j , cj , bj ) = max 0, gi (V g,j , cj , bj ) − bi
bmax
i = max ε, bi (V g,j , cj , bj )V g,j , cj , bj ∈ P
bi (V g,j , cj , bj ) is the violation value of the ith constraint of individuals Vg,j , cj ,
bj . bmax
i is the maximum violation value of constraint i of all individuals in the current
population, ε is a small positive number, which is used to avoid division by zero in
penalty function.
Synthesis (9) and (10), for individuals Vg,j , cj , bj , the fitness function with penalty
function is:
4 Simulation Examples
The IEEE30-node system has given the node load data at one time, as well as the active
output and node voltage of each PV generator node, and the node voltage of the balance
node. Based on the data at one time, the daily typical data for simulation analysis is
constructed. Without losing generality, five typical daily load curves are assigned to
each load node. The 24-h daily load curve of each node is obtained. According to the
load curve, the active output of each generator is evenly distributed, and the 24-h active
output curve of the generator is obtained to balance the change of load.
There are 6 generators available for dispatching in IEEE30 nodes. Transformers with
4 adjustable taps; There are two switchable capacitors at nodes 10 and 24, and the related
parameters are shown in reference [10].
Taking the time period consisting of 8, 9, 10 and 11 a.m. as a typical “peak climbing”
working condition, the simulation verification is carried out. In this time period, discrete
control equipment, such as capacitor switching and transformer taps, are only adjusted
once; While continuous control equipment, such as generator terminal voltage setting,
is adjusted at every moment. Therefore, the energy consumption, the reactive power
margin of generators and the balance of the distribution of reactive power margin of
each generator are optimized.
Table 1 gives the parameters of binary coding of control variables solved by genetic
algorithm of optimization models (7) and (8):
Adjustment step Lower limit Upper limit Single coding Total length
length
Generator 0.01 0.90 1.10 5 5 × 6 = 30
Transformer 0.02 0.90 1.10 4 4 × 4 = 16
(continued)
Research on Modeling and Solution of Reactive Power Optimization 771
Table 1. (continued)
Adjustment step Lower limit Upper limit Single coding Total length
length
Capacitor 1 0.01 0.00 0.20 5 5
Capacitor 2 0.01 0.00 0.10 4 4
Continuous 30
variable coding
length
Discrete 25
variable coding
length
The discrete variable optimization model and the continuous variable optimiza-
tion model are solved by interactive iteration respectively. Table 2 gives other relevant
parameters of genetic algorithm, convergence iteration times and other information:
After three times of coordinated optimization of discrete variables and continuous
variables, the optimal solution is obtained by convergence. Table 3 gives the optimal
solution results of control variables:
After optimization, the system energy consumption, reactive power margin and dis-
tribution equilibrium of reactive power margin have been improved in the time period.
Figures 1(a)–(d) show the comparison diagrams before and after the optimization of the
reactive power margin of each generator at each time 8, 9, 10 and 11 of the time period:
QGm
node
(a) Time 8 (b) Time 9
QGm
It can be seen from the above comparison diagram that the expectation and variance
objective function of balanced distribution of reactive power margin are introduced
into the optimization objective, and the reactive power margin of the generator is more
balanced at each moment after optimization.
Figure 2 shows a comparison diagram of the sum of reactive power margins of all
generators before and after optimization at each time. After optimization, the sum of
reactive power margins of generators increases at every time except at time 11. After
optimization at time 11, the reactive power margins decrease, but the decrease is not
large. In order to balance the increase of generator margins at every time in the period,
there may be such optimization results.
Research on Modeling and Solution of Reactive Power Optimization 773
4
beforeoptimization
afteroptimization
3
ΣQGm
2
0
8 9 10 11
time
Fig. 2. Comparison diagram before and after optimization of generator reactive power margin
sum at each time in time period
Figure 3 shows the line loss comparison diagram of the system before and after
optimization at each moment of the time period. After optimization, the line loss at each
time is reduced; And because the expectation and variance description are introduced
into the optimization objective function, the loss reduction effect at each moment is
balanced.
0.1 beforeoptimization
afteroptimization
Ploss
0.05
0
8 9 10 11
time
Fig. 3. Comparison diagram of system line loss before and after optimization at each time in time
period
5 Conclusion
In this paper, the time-period reactive power optimization model is decomposed into
an optimization model considering discrete control variables and an optimization model
considering continuous control variables. The two models are optimized and solved sep-
arately, and cross-iterated until convergence. The modeling problem of discrete control
variable optimization in a period is studied, and the objective optimization functions
such as expectation and variance at each time of period are put forward. The established
model ensures that the optimization objectives at each period are improved overall and
774 C. Deyu
balanced after the discrete control equipment moves once; An adaptive weight genetic
algorithm with discrete control variables and multi-objective optimization objectives is
proposed to solve the optimization model. The weights are given adaptively in genetic
evolution, which makes the optimization solution more reasonable. Simulation examples
verify the rationality of the proposed model and algorithm.
References
1. Hu, Z., Wang, X.: Time-interval based control strategy of reactive power optimization in
distribution networks. Autom. Electr. Power Syst. 26(6), 45–49 (2002)
2. Wang, X., Li, Z., Hu, Z.: Time-interval based comprehensive control strategy for daily
voltage/VAR optimization in distribution systems. Autom. Electr. Power. Syst. 30(7), 5–9
(2006)
3. Fang, X., Guo, Z.: Optimal time-varying reactive power and voltage control in distribution
systems. Autom. Electr. Power Syst. 29(9), 40–44 (2005)
4. Deng, Y., Zhang, B., Tian, T.: A fictitious load algorithm and its applications to distribution
network dynamic optimizations. Proc. CSEE 16(7), 241–244 (1996)
5. Ni, S., Cui, C., Yang, N., et al.: Multi-time-scale online optimization for reactive power
of distribution network based on deep reinforcement learning. Autom. Electr. Power Syst.
45(10), 77–85 (2021)
6. Zhao, J., Zhang, Z., Yao, J., et al.: Heterogeneous decomposition based distributed reactive
power optimization method for global transmission and distribution network. Autom. Electr.
Power Syst. 43(3), 108–114 (2019)
7. Li, Q., Qiao, Y., Zhang, Y.: Continuous reactive power optimization of distribution network
using deep reinforcement learning. Power Syst. Technol. 44(4), 1473–1480 (2020)
8. Zhang, Q., Ding, J., Zhang, D., et al.: Reactive power optimization of high-penetration dis-
tributed generation system based on clusters partition. Autom. Electr. Power Syst. 43(3),
130–137 (2019)
9. Xuan, G., Cheng, R.: Genetic Algorithm and Engineering Optimization. Tsinghua University
Press, Beijing (2004). Translated by Yu, X., Zhou, G.
10. Zhang, B., Chen, S., Yan, Z.: Advanced Power Network Analysis, 2nd edn. Tsinghua
University Publishing House, Beijing (2007)
Research on Task Allocation Method of Mobile
Swarm Intelligence Perception Based on Hybrid
Artificial Fish Swarm Algorithm
Abstract. In this study, the task allocation mechanism in mobile swarm intelli-
gence perception is studied and analyzed. Considering how to allocate tasks under
specified time constraints and optimize for the goal, a hybrid artificial fish swarm
algorithm is proposed by using swarm intelligence algorithm. The inertia index
of particle swarm optimization algorithm is pulled into the typical artificial fish
swarm algorithm, which is proved by simulation experiments. The hybrid arti-
ficial fish swarm algorithm greatly improves the convergence speed, avoids the
shortage that the artificial fish swarm algorithm often stop when it obtains the local
optimization, and this make the algorithm obtains the effect of global optimization.
1 Introduction
The speed of scientific and technological development is getting faster and faster. The
number of mobile smart terminals is constantly increasing. The functions of GPS, sen-
sors, and other equipment have been significantly improved [1], and the sensing meth-
ods of mobile terminals have also been improved. Mobile swarm intelligence takes the
mobile smart terminals carried by most users as the lowest level of sensing information
acquisition unit [2]. It then uses the mobile terminal’s Perception function, which calcu-
lates, stores, and communicates the acquired perception information, realizes real-time
perception and collects surrounding environment information, uploads the information
to the network, and then forwards relevant information to realize the distribution of per-
ception tasks and users perception data Collection. Whether it is the range of perception
and the type of information [3], or the cost and speed of construction, this mobile group
intelligence perception method is far superior to the traditional method of perception.
Under the circumstances, large-scale and complicated perception tasks can be easily
completed [4].
Maximizing the validity and reliability of perception data is the key to mobile per-
ception. This paper proposes a mission assignment method based on mixed artificial fish
swarm group perception. Use the inertia index in the particle swarm algorithm alone
as an index to mediate the speed of convergence, combined with the AFSA. The com-
bination of the two improves the ability of the algorithm to search in the global scope
and prevents the algorithm from finding the local maximum and stopping looking for
good results, and the higher efficiency of the algorithm is verified through simulation
experiments. The main research content and results of this paper are as follows:
2 Related Works
The research of relevant scholars on task allocation is as follows: Chen [5] summa-
rized and proposed a general task allocation framework, summarized and summarized
from three aspects: worker model, task model, and task allocation algorithm, and put for-
ward the development trend of swarm intelligence collaborative task allocation research.
Wang [6] and others proposed a task allocation model combining task requirement fea-
ture extraction algorithm and user label classification method. The category keywords of
perceived tasks are extracted through the task requirement feature extraction algorithm.
Then the classifier is used to predict the user’ s type label, screen users according to the
prediction results, and distribute tasks. Finally, the simulation experiments show that
the proposed model is feasible. Yang [7] et al. Used the fuzzy logic control method to
obtain the density of participants and then proposed to use a global greedy algorithm to
allocate tasks to ensure the maximization of the utility of all tasks. Through simulation
experiments, it was proved that under the conditions of different number of tasks, num-
ber of participants, and maximum workload carried by participants, the global greedy
algorithm is better to other benchmark algorithms in maximizing the utility of all tasks.
Wang [8] et al. Studied the problem of sensor task allocation and proposed the heuristic
greedy method as the baseline solution, combining greedy algorithm and bee algorithm.
Through simulation experiments, it is proved that the hybrid method is better than a
greedy algorithm.
Qiao [9] et al. Proposed a particle swarm majorization method to solve the multi-
objective task allocation problem in a distributed environment and used fuzzy cost selec-
tion to improve the computational efficiency. The algorithm further realizes the Pareto
improvement. Ma [10] et al. Proposed a hybrid bionic self-organizing map neural net-
work algorithm. Due to the principle of AUV individual kinematics, simulation exper-
iments prove that the algorithm has wide application potential. Lu [11] et al. Proposed
PSO-GA-DWPA (discrete wolf swarm algorithm based on particle swarm majoriza-
tion and genetic algorithm) to solve the mission allocation problem of UAV swarm. The
Research on Task Allocation Method of Mobile Swarm Intelligence 777
improved algorithm has superb search quality in high-dimensional space. Deng [12] et al.
Introduced a multi-round allocation algorithm based on EMA prediction, distributed the
workload to the edge node of each time slot according to the capacity and cost prediction
[13] of the edge node, obtained the optimal allocation strategy, and proposed an online
mission allocation algorithm based on Q-Learning. The superiority and effectiveness of
the algorithm are verified by simulation.
To sum up, most scholars only carry out local objective optimization for task alloca-
tion, and lack multi-objective majorization. In this study, the mission allocation question
in mobile swarm intelligence perception is studied, a hybrid AFSA is putted forward. The
inertia index of particle swarm optimization algorithm is introduced into the AFSA to
enhance algorithm convergence speed. Avoid the shortage that the AFSA is easy to get
into local optimization.
The mobile swarm intelligence perception system mainly includes task completers, task
publishers [14], and systems. The task publisher publishes task information in the sys-
tem, including task type, geographic location, completion time, etc. Therefore, when
creating tasks, M = {type, location, time, . . .}, users can participate in tasks by active
participation and passive participation. In the case of active participation, users can
decide whether to perform tasks according to their own abilities, interests, rewards and
other factors; In the case of passive participation, users do not understand the mission,
and the quality of the final mission completion results may be poor. Therefore, it is finally
determined that the set of users participating in the task is. The optimization model of
task allocation in the perception system is shown in Fig. 1:
User capabilities
Task release
Platform
Task publisher
Mobile terminal equipment
The related description of the perception task allocation model built in this paper is as
follows:
(1) In the task assignment in this paper, after the perception task is released, there is
no situation where users cannot be found.
(2) This article aims to accomplish more perception tasks with smaller users. How to
allocate tasks reasonably so that the most tasks can be completed in a short time
and the user’s movement distance is the shortest, so that the cost of perception can
be greatly reduced and the benefits can be maximized.
(3) The optimization goal is mainly to reduce the perception cost and ensure the maxi-
mum completion of the perception task. The corresponding constraints are that the
time for the user to reach the task location and the total time to complete the task
need to be less than or equal to the specified completion time.
Suppose that within the specified t min, the system publishes m perception tasks,
the task set is T = {t1 , t2 , t3 , . . . , tm }, there are n users who are determined to participate
in the perception task, the set is TWk = {tk1 , tk2 , tk3 , . . . , tkn },the distance that the user
needs to move during the task is D(TWk ). In the model constructed in this paper, it
is assumed that each user needs to spend λ min when moving to the perception task
point, the speed is v m min, and each perception task is completed by at most one user.
Therefore, the target value that needs to be optimized is shown in the following formula
(1):
⎧
⎨ min D(TW )
n
k
k=1 (1)
⎩
max|TWk |
1 ≤ |UTm | ≤ η (3)
Among them, the perceived cost of the mobile terminal is related to and directly
proportional to the user’ s expenses, mobile phone power, network expenses, etc. There-
fore, the goal of majorization is to reduce mission completion time and reduce perceived
capitalization costs, so as to improve the revenue of the platform; Then we optimize the
quantity and quality of completed tasks. For the same task completed by multiple users,
we need to optimize it to find the best task result. Therefore, we also need to limit the
number of users completing the task to ensure the load balance of task allocation.
Research on Task Allocation Method of Mobile Swarm Intelligence 779
The AFSA mainly simulates the foraging, clustering and tail chasing behavior of fish
by constructing artificial fish [15], and further find the global optimal value through the
step-by-step optimization process of the fish swarm.
When the fish are foraging, they will swim towards the direction of more food. The
artificial fish Xi will randomly choose the state point within its visible field of vision
visual. After calculating the target value, if the fitness value of the target Yj is greatly
better than the target value Yi , the artificial fish Xi will move towards the point Xj ; If
not, the artificial fish Xi will continue to select new state points Xj within the visible
range and continue to judge. If it does not advance after repeated several times, it will
immediately advance one step to reach a new state point Xj , and its expression is shown
in the following formula (4):
The fish often move forward in groups [16]. The artificial fish Xi searches the position
Xc and number of partner fish within its visible field of vision. If the ratio of fitness value
to the number nf of partner fish is less than the minimum value Yc of, the position of
partner fish is better, and the artificial fish moves to the center of partner fish; If not, the
artificial fish Xi will perform foraging behavior.
When a swarm of fish moves, it makes a choice according to the surrounding envi-
ronment. The behavior of moving in the optimal direction is tail chasing behavior. The
artificial fish Xi searches for the fish Xj with the highest fitness within its visible field
of vision, and its fitness value is Yj . When the optimal position has higher value, the
artificial fish Xi moves one step to the optimal position; If not, the artificial fish Xi will
perform foraging behavior. The objective function of this is as shows (5):
j Xj − Xit
Xit+1 = Xi + ∗ Step ∗ rand {rand ∈ (0, 1)} (5)
Xj − Xjt
780 J. Li et al.
When simulating the fish swarm, AFSA has blind search behavior, takes a long time,
and the overall algorithm efficiency is low. Due to the influence of its field of vision, it
always getting into local optimization. Therefore, the algorithm needs to be improved
to increase efficiency.
ω is the inertia index in the particle swarm, visualnext is the next search range of the
artificial fish swarm, and visual is the visible field of view of the current artificial fish’s
location. The improved expression for the next search is shown in the following formula
(7):
When the artificial fish swarm searches for the next optimal position, the inertia
index increases continuously, the visible area of vision of the artificial fish also increases
gradually, the moving step of the artificial fish will also increase, and the search breadth
of the algorithm will be greatly improved, effectively avoiding the situation of local
optimization; When the algorithm execution to later stage, the inertia index will gradually
decrease to enhance the search accuracy, accurately lock the optimal, the visible area
of view of the artificial fish swarm will be reduced, and the convergence speed of the
algorithm will be greatly accelerated, which makes up for the slow convergence speed
of the AFSA. The improved hybrid AFSA execution flow chart is as shown in Fig. 2:
Research on Task Allocation Method of Mobile Swarm Intelligence 781
Start
Yes
End
5 Simulation Experiment
The experimental hardware environment is: the CPU is AMD A4–7210 APU with AMD
radeon R3Graphics@1.80 GHz, memory 8.00 GB, windows 10 system, Matlab R2016b
simulation platform. The simulated mobile swarm intelligence perception task is ran-
domly distributed in the area of 500 * 500 (unit is km), the maximal number of iterations
L = 200, c1 = c2 = 2, and the scale of artificial fish swarm is 100. Assuming that each
user has the same ability to complete the perception task, other parameter settings of the
simulation experiment are shown in Table 1.
The experiment compares the advantages and disadvantages of hybrid AFSA and typical
AFSA. The results and analysis of the experiment are as follows:
(1) Fitness changes of artificial fish swarm algorithm (AFSA) and mixed AFSA
The maximal quantity of iterations of the algorithm is 200. The fitness change curves
of AFSA and mixed AFSA algorithms are shown in Fig. 3.
The experiment results show that due to the combination of inertia index, the hybrid
AFSA converges faster than the typical AFSA, searches in the global range, avoids local
optimization, and has better stability and convergence.
A reasonable task allocation mechanism can greatly increase the number of tasks
completed simultaneously and reduce the actual time spent by users. The figure of
the number of tasks completed by the algorithm before and after the improvement is
shown in Fig. 4(a). It shows that as the number of iterations continues to increase,
the standard AFSA often gets into the local optimum and stops searching during the
mixed artificial fish swarm. The algorithm can achieve global majorization and search
in the global. Figure 4(b) compares the average time to complete tasks between the
standard AFSA and the hybrid AFSA. The average time consumed by the improved
hybrid artificial fish swarm algorithm is much less than the original time, proving that the
loan balance is greater with the improved algorithm task distribution. The convergence
is not good at the initial stage. Still, after the number of iterations continues to increase,
the convergence becomes better. Therefore, the hybrid artificial fish swarm algorithm
has good advantages.
Research on Task Allocation Method of Mobile Swarm Intelligence 783
AFSA
Mixed AFSA
(a) (b)
Fig. 4. Total number of missions completed and comparison of average time by the two algorithms
The main optimization goal of this paper is to lessen the perceived capitalized cost,
and the reduction of the perceived cost can increase the platform’s revenue. In the
experiment, the standard AFSA and hybrid AFSA are used to complete the same number
of tasks respectively, and the perceptual costs of the two algorithms are compared. The
task ratio of the two algorithms is 100, 200, 300, 400 and 500. As shown from the
experimental results in Fig. 5, with the increasing number of tasks, the perception cost
of the hybrid AFSA is remarkably lower than that of the typical AFSA.
AFSA
Mixed AFSA
6 Conclusion
Firstly, to reduce the mission pass time, reduce the perceived capitalized cost, raise
the revenue of the platform and the quantity of missions passed, this paper limits the
784 J. Li et al.
number of users completing the same task and constructs the task allocation model in
mobile perception; Secondly, based on the task allocation model and combined with
swarm intelligence algorithm, a hybrid AFSA is put forward. The randomly assigned
task allocation scheme is regarded as an “artificial fish swarm”. Combined with the
inertia index in the particle swarm optimization algorithm, the optimal task allocation
scheme in line with the optimization goal is selected; Finally, the simulation experiment
is compared with the original AFSA. The experiment shows that the hybrid AFSA can
effectively improve the convergence speed of the AFSA, avoid the shortage that the
AFSA is simple to get into local optimization and enable the algorithm to search the
optimal task allocation scheme globally.
Acknowledgment. This work is partly supported by the project supported by the National
Social Science Foundation(16BJY125) ,Heilongjiang philosophy and social sciences research
planning project(19JYB026),Key topics in 2020 of the 13th five year plan of Educa-
tional Science in Heilongjiang Province(GJB1320276),Project supported by under-graduate
teaching leading talent training program of Harbin University of Commerce(201907),Key
project of teaching reform and teaching research of Harbin University of Commerce in
2020(HSDJY202005(Z)),Innovation and entrepreneurship project for college students of Harbin
University of Commerce (202010240059),Swarm level scientific research project of Heilongjiang
Oriental University(HDFKY200202),Key entrusted projects of higher education teaching reform
in 2020(SJGZ20200138).
References
1. Wang, L., Zhang, D., Wang, C., Chen, C., Han, X., Abdallah, M.: Sparse mobile crowdsensing:
challenges and opportunities. IEEE Commun. Mag. 54(7), 161–167 (2016)
2. Xiong, Y., Liu, W., Liu, Z.: The key technology of opportunity group intelligence perception
network. ZTE Technol. 21(06), 19–22 (2015)
3. Wang, L., Yu, Z., Guo, B., Xiong, F.: Community task distribution of group intelligence
perception based on mobile social network. J. Zhejiang Univ. (Eng. Sci. Ed.) 52(09), 1709–
1716 (2018)
4. Li, S.: Research on identifying traffic congestion based on group intelligence perception.
Wirel. Internet Technol. 2016(04), 128–139 (2016)
5. Chen, B., Wang, L., Jiang, X., Yao, H.: Summary of research on task allocation of group
intelligence. Comput. Eng. Appl., 1–15 (2021)
6. Wang, X., Liao, Y., Zhao, G., Wang, J., Xie, B.: A task assignment model of group intelligence
perception based on task requirements. Comput. Eng. Sci., 1–10 (2021)
7. Yang, G., Zhang, Y., He, X.: Fuzzy logic control-oriented mobile group intelligence perception
multi-task assignment. Small Microcomput. Syst. 41(10), 2068–2074 (2020)
8. Wang, Z., Huang, D., Wu, H., Deng, Y., Aikebaier, A., Teranishi, Y,: QoS-constrained sensing
task assignment for mobile crowd sensing. In: IEEE Global Commun Conference, pp. 311–
316(2014).
9. Qiao, N., You, L., Sheng, Y., Wang, J., Deng Y.: An efficient algorithm of discrete particle
swarm optimization for multi-objective task assignment. IEICE Trans. Inf. Syst. E99-D(12),
2968–2977 (2016)
10. Ma, X., Chen, Y., Bai, G., et al.: Path planning and task assignment of the multi-AUVs
system based on the hybrid bio-inspired SOM algorithm with neural wave structure. J Braz.
Soc. Mech. Sci. Eng. 43, 28 (2021)
Research on Task Allocation Method of Mobile Swarm Intelligence 785
11. Lu, Y., Ma, Y., Wang, J., et al.: Task assignment of UAV swarm based on wolf pack algorithm.
Appl. Sci. 10(23), 8335 (2020)
12. Deng, X., Li, J., Liu, E., et al.: Task allocation algorithm and optimization model on edge
collaboration. J. Syst. Arch. 110, 101778 (2020)
13. Song, Z., Li, Z., Chen, X.: Mobile group intelligence perception task distribution mechanism
based on compressed sensing. Comput. Appl. 39(01), 15–21 (2019)
14. Wu, G., Chen, Z., Liu, J., et al.: Task assignment for social-oriented crowdsourcing. Front.
Comp. Sci. 15(2), 1–11 (2021)
15. Huang, J., Zeng, J., Bai, Y., et al.: Layout optimization of fiber bragg grating strain sensor
network based on modified artificial fish swarm algorithm. Optical Fiber Technol. 65(1),
102583 (2021)
16. Zhang, L., Fu, M., Li, H., et al.: Improved artificial bee colony algorithm based on damping
motion and artificial fish swarm algorithm. J. Phys: Conf. Ser. 1903(1), 9 (2021)
17. Wu, G., Wan, L.: Research on particle swarm optimization algorithm for robot path planning.
Mech. Sci. Technol., 1–7 (2021)
Research on Task Allocation Model of Takeaway
Platform Based on RWS-ACO Optimization
Algorithm
Abstract. This paper research the task distribution of the takeaway platform,
builds a task distribution model for the takeaway platform, and proposes the
roulette ant colony algorithm (RWS-ACO) that combines the roulette algorithm
with the ant colony algorithm, then conducts simulation experiments to solve the
simulation data. The distribution plan is obtained, which verifies the effectiveness
of the model and algorithm and achieves a win-win situation for multiple parties
in the distribution.
1 Introduction
With the rise of Internet technology and the popularization of mobile smart devices,
online platforms have been deeply integrated with offline businesses. Many industries
seize the opportunity and grow rapidly; among them, the takeaway industry is typical. The
takeaway platform makes full use of “Internet+” and the deep integration of traditional
industries. The Internet to integrate social resource allocation has developed rapidly, and
the number of orders and users has increased rapidly. Task allocation has become a key
point in the study of takeaway platforms.
As consumers develop their consumption habits, the takeaway industry has become
more and more mature. The courier’s capacity is far less than the demand of consumers,
and this has triggered a conflict between consumers and couriers. It is difficult for
consumers to get satisfactory service, and the couriers always violate traffic rules and
cause accidents. The price factor is no longer the primary factor when users make
choices. Now the competition among takeaway platforms is service quality and delivery
speed. Especially during the peak order period, many orders flood into the system.
At this time, the ability of the takeaway platform to process orders is weak, and the
– For the participant selection process and task assignment process, under certain con-
straints, the task allocation model is constructed with the shortest delivery distance as
the optimization goal.
– Proposed roulette ant colony optimization algorithm, solved the model and performed
simulation experiments on simulated data, obtained the task allocation plan, and
verified the model’s effectiveness.
2 Related Work
In terms of task allocation, the research results of related scholars are as follows:
Fu et al. [1] believe that trust in task allocation is directly proportional to the finish
quality and proposed a greedy heuristic algorithm to allocate tasks to workers with low
cost and high trust; Zhang et al. [2] proposed the priority of task assignment depends
on the worker’s ability to complete the task and the possibility of acceptance; LI et al.
[3] proposed the Te matrix formula and used the greedy algorithm to generate the vehi-
cle allocation plan as the initial solution to the problem, combined with the simulated
annealing algorithm, provided a new heuristic algorithm; Alvarez et al. [4] proposed
to adopt BPC algorithm to solve the time window and task assignment problems of
multiple couriers. Experiments to verify that the algorithm has advantages; Wang et al.
[5] proposed to transform the combinatorial optimization problem into a task alloca-
tion function with load capacity and cooperation constraints, solve by the algorithm and
verify effectiveness by experiment; Pan et al. [6] proposed to add an adaptive threshold
algorithm based on the GH algorithm, and experiments to verify that it has lower cost
and higher utility; Chen et al. [7] proposed a dynamic grouping allocation method to
solve the conflicts generated in a dynamic environment, behaved well; Fu et al. [8] pro-
posed a two-stage allocation model to try to distribute tasks fairly and minimize costs,
experiments to verify F-Aware is superior than others; Liang et al. [9] proposed TWD
and design a two-stage competition and cooperation model: competition optimization
and negotiation and cooperation, maximizes personal income and gets the best solu-
tion. Miao et al. [10] proposed an adaptive ant colony algorithm to solve the problem
of local optimal solution, through experimental verification, global optimization can be
achieved; Xia et al. [11] proposed an improved tabu search algorithm to solve a planning
model that takes the number of vehicles and cost as dual goals; Liu et al. [12] used the
greeedy-ot algorithm with optimized threshold to solve the task assignment problem;
Song et al. [13] adopted an online greedy algorithm to solve the task assignment problem
of spatial crowdsourcing, realize distribution with the lowest cost; Li et al. [14] adopted
a bilateral online matching method to match workers with tasks.
To sum up, in terms of task allocation on takeaway platforms, most of them adopt
combinatorial optimization algorithms or swarm intelligence algorithms to solve the
788 L. Jianjun et al.
problem. Therefore, this paper integrates the roulette algorithm in the ant colony
algorithm, and uses the optimized RWS-ACO algorithm to solve the model.
The supply and demand relationship exists in the takeaway platform. There are several
couriers, merchants, and consumers. Request a reasonable arrangement of the delivery
route and time of the courier, and under the given constraints (delivery load, volume,
time, etc.), enable the courier to deliver the goods required by consumers from the
merchant to the consumer on time, and make the objective function (length of the delivery
path) optimal. This paper divides businesses into different business circles by region. It
proposes a task allocation model in response to the requirements of delivery distance,
cost, and overtime in each business circle, to achieve the best match between consumer
orders and couriers. The task allocation model of the takeaway platform is shown in
Fig. 1:
In real life, the task assignment of the takeaway platform is a very complicated problem.
In order to facilitate the construction and solution of the task assignment model, simpli-
fied some problems in the actual delivery process. Now make the following assumptions
about the task assignment problem of the takeaway platform studied:
(1) Assuming that the takeaway platform is divided into different business districts
according to the location of the merchants, the platform selects the appropriate
courier to complete the order delivery task for consumers according to the order
delivery distance and estimated delivery time published by the business circle.
(2) Assuming that the goods required by each consumer can only be delivered by one
courier;
(3) Assuming that the courier must first go to the corresponding business in the business
circle to complete the task of taking meal, and then go to the designated delivery
circle to complete the task of delivering meal;
Research on Task Allocation Model of Takeaway Platform 789
(4) Assuming that each business circle has a fixed delivery person, its starting point
is located in the center of the merchants in the business circle, and approximately
the same distance from each merchant. After completing the delivery task, couriers
will still return to the starting point, and the return distance is not include the total
distance of the delivery.
Let di (n) denote the distance from the i-th customer to the n-th business circle, where
n = 1,2, …, N, N represents the total number of business circles. Let the set Di (n) =
{d i (n)}, n = 1,2, …, N means the collection of distances from customer i to each business
circle. d i(m) = minDi (n), then the i-th customer will be delivered by the m-th district
distributor.
Based on the above analysis, construct a corresponding order distribution model for
multi-commercial takeaway platforms. The relevant parameters are shown in Table 1
below:
Parameter Means
ni Business circle
P Number of consumers
c The distance between the courier’s starting point and each merchant
Li The number of couriers in the i-th business circle
W ij The load of the j-th courier in the i-th business circle
T ij The maximum time of the j-th courier in the i-th business circle
qk The weight of the k-th consumer’s goods
tk Delivery time of the k-th consumer
d ij Distance from consumer i-th to consumer j-th
nij The number of consumers delivered by the j-th courier in i-th business circle
Rij The delivery route of the j-th courier in the i-th business circle
r ijk The k-th consumer delivered by the j-th courier in i-th business circle
This paper takes the shortest delivery path as the objective function, and establishes the
following mathematical model:
N Lj nij
minZ = [ drij(k−1) rijk + drij0 rijnij δ(nij ) + cP] (1)
i=1 j=1 k=1
nij
s.t. qrijk ≤ Qij (2)
k=1
790 L. Jianjun et al.
nij
trijk ≤ Tij (3)
k=1
N Li
nij = P (4)
i=1 j=1
m Q
τij = τijk , τijk = Lk , If ant k passes client i, j in iteration
(10)
k=1
0, others
Step1: Initialize, t = 0;NC = 0;τ ij (t) = C;τ ij (t) = 0; put m ants on n business circles;
Step2 Set the taboo table index s = 1; add starting point to the respective taboo table;
Step3 Determine whether the taboo table is full: If full, calculate the travel length
that all ants have traveled, update the optimal path; else, set s = s + 1, the m ants select
the next business circle according to pijk Obtained by selecting the customer through the
roulette algorithm, adding the business circle to the taboo table, then re-inserting it into
the judgment of the full taboo table. Until it reaches the full state;
Step4 Calculate τijk , update the pheromone; t = t + n; NC = NC+1;
Step5 Determine whether the termination conditions are met, if met, output; else,
clear taboo tables and return to (2) until the termination conditions are met (Fig. 2).
5 Simulation Experiment
5.1 Experimental Environment and Data Description
Verify the effectiveness of the constructed task allocation model and the optimized ant
colony algorithm through simulation experiments. Experiment with MATLAB R2016a
as the simulation platform. The heuristic factor Eta is set as the reciprocal of the distance,
the initial pheromone matrix is the identity matrix, and the parameter settings are shown
in Table 2:
Table 3. (continued)
In the above figure, the asterisk locations are the business circle locations, and the
circled place is where the order is requested to be delivered. Each closed path is the order
794 L. Jianjun et al.
delivery path of the corresponding courier. Without overtime and overload, to achieve
the goal and make the delivery distance the shortest, it required four delivery personnel
to complete the tasks in the two business circles. The specific order task allocation plans
are as follows:
No. 1 courier: responsible for the delivery tasks of business circle 7, the orders that
need to be delivered are 5, 8, 13 and19;
No. 2 courier: responsible for the delivery task of business circle 10, the orders that
need to be delivered are 1, 2, 4, and 20;
No. 3 courier: responsible for the delivery tasks of business circle 10, the orders that
need to be delivered are 9, 15, and 18;
No. 4 courier: responsible for the dispatch task of business district 10. The orders
that need to be dispatched are 3, 6, 11, 12, 14, 16, and 17.
The adaptability comparison chart of the ACO algorithm and the RWS-ACO
algorithm is shown in Fig. 4 below:
Simulation experiments verify that the model and algorithm can reasonably allocate
the orders posted on the platform to the corresponding couriers and successfully deliver
the goods to consumers within the specified time under limited constraints. The destina-
tions to be delivered can be reasonably allocated optimally and save algorithm running
time.
6 Conclusion
This article first builds a task allocation model that takes the shortest delivery distance as
the optimization goal and takes the delivery load, volume, and time as constraints. Sec-
ondly, based on the ant colony intelligence algorithm, the roulette algorithm is combined
with the ant colony algorithm, and the RWS-ACO algorithm is used to solve the model.
Finally, the simulation experiment verifies that the algorithm can effectively achieve task
allocation, ensure that randomness is not lost, ensure that all consumers are likely to be
Research on Task Allocation Model of Takeaway Platform 795
selected, and avoid falling into local optimality. There are still some shortcomings in this
paper. In a follow-up study, the actual distance between the merchant and the courier
should be further considered.
Acknowledgment. This work is partly supported by the project supported by the National Social
Science Foundation(16BJY125), Heilongjiang philosophy and social sciences research planning
project(19JYB026), Key topics in 2020 of the 13th five year plan of Educa-tional Science in
Heilongjiang Province(GJB1320276), Project supported by under-graduate teaching leading talent
training program of Harbin University of Commerce (201907), Key project of teaching reform and
teaching research of Harbin University of Commerce in 2020(HSDJY202005(Z)), Innovation and
entrepreneurship project for college students of Harbin University of Commerce (202010240059),
School level scientific research project of Heilongjiang Oriental University (HDFKY200202), Key
entrusted projects of higher education teaching reform in 2020(SJGZ20200138).
References
1. Fu, D., Liu, Y.: Trust-aware task allocation in collaborative crowdsourcing model. Comput.
J. 64(6), 929–940 (2021)
2. Zhang, X., Su, J.: An approach to task recommendation in crowdsourcing based on 2-tuple
fuzzy linguistic method. Kybernetes 47(8), 1623–1641 (2018)
3. Li, B., Yang, X., Xuan, H.: A hybrid simulated annealing heuristic for multistage het-
erogeneours fleet scheduling with fleet sizing decisions. J. Adv. Transp. 2019, 1–19
(2019)
4. Alvarez, A., Munari, P.: An exact hybrid method for the vehicle routing problem with time
windows and multiple deliverymen. Comput. Oper. Res. 83, 1–12 (2017)
5. Wang, J., Wang, J., Han, Q.: Multivehicle task assignment based on collaborative neurody-
namic optimization with discrete hopfield networks. IEEE Trans. Neural Networks Learn.
Syst. 1–13 (2021)
6. Pan, Q., Pan, T., Dong, H., Gao, Z.: Research on task assignment to minimize travel cost for
spatio-temporal crowdsourcing. Wireless Com Network 2021, 59 (2021)
7. Chen, X., Zhang, P., Du, G., Li, F.: A distributed method for dynamic multi-robot task
allocation problems with critical time constraints. Robot. Auton. Syst. 118, 31–46 (2019)
8. Fu, D., Liu, Y., Yan, Z.: Fairness of task allocation in crowdsourcing workflows. Math. Probl.
Eng. 2021, 1–11 (2021)
9. Liang, D., Cao, W., Xu, Z., Wang, M.: A novel approach of two-stage three-way co-opetition
decision for crowdsourcing task allocation scheme. Inf. Sci. 559, 191–211 (2021)
10. Miao, C., Chen, G., Yan, C., Wu, Y.: Path planning optimization of indoor mobile robot based
on adaptive ant colony algorithm. Comput. Ind. Eng. 156, 107230 (2021)
11. Xia, Y., Fu, Z.: Improved tabu search algorithm for the open vehicle routing problem with
soft time windows and satisfaction rate. Cluster Comput. 22(1), 1–9 (2018)
12. Liu, J., Xu, K.: Budget-aware online task assignment in spatial crowdsourcing. World Wide
Web 23(1), 289–311 (2020)
13. Song, T., Xu, K., Li, J., Li, Y., Tong, Y.: Multi-skill aware task assignment in real-time spatial
crowdsourcing. GeoInformatica 24(1), 153–173 (2020)
14. Li, Y., Fang, J., Zeng, Y., Maag, B., Tong, Y., Zhang, L.: Two-sided online bipartite matching
in spatial data: experiments and analysis. GeoInformatica 24(1), 175–198 (2020)
Utilizing Vias and Machine Learning
for Design and Optimization of X-band
Antenna with Isoflux Pattern
for Nanosatellites
1 Introduction
The satellites antennas have an important role in Earth observation missions,
where a high transmission rate is necessary to acquire sensed images in different
spectral bands for an enormous number of remote sensing applications. The wide
coverage pattern or Isoflux pattern is one of the requirements for the antenna
system in satellites. The Isoflux antenna on the satellites provides a uniform
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2022
A. E. Hassanien et al. (Eds.): BIIT 2021, LNDECT 107, pp. 796–804, 2022.
https://doi.org/10.1007/978-3-030-92632-8_75
Utilizing Vias and Machine Learning for Design and Optimization 797
The basic structure of designed antenna consists of two choke rings fed by a
radiator in order to get the desired isoflux pattern is shown in Fig. 1. The antenna
798 M. A. Maged et al.
The patch and the circuit are printed on two stacked RO4003c substrates
which creates a embedded ground plane between the patch and the feeding net-
work’ board. The patch antenna is connected to the coupler through four via
holes crossing the embedded ground plane. The resulting microstrip assembly is
placed on a metal cylinder (4 mm high and 33 mm in diameter), which is sur-
rounded by twelve parasitic crossed dipoles placed on a 45 mm-diameter circle,
printed on both sides of a second RT5880 1.524 mm-thick substrate. Figure 2
illustrates the structure of the X-Band Isoflux antenna from front and back view.
Fig. 2. The antenna front and back view of shows the two calendars of vias and feeding
network.
(1)
Where x is the vector of d design variables, X = [l, u] represents a feasible
d
region of the design variable x, f denotes the response results of the antenna of
interest and F (x) is the antenna optimization objective function and gi (x) are
constraints. The vector x∗ is the optimal parameters of the design. Normally, f is
obtained through computationally expensive electromagnetic simulation. Since,
our main goal is synthesis of the Isoflux pattern, applying the machine learning
concepts by learning the desired gain pattern at X-band target frequency for
the designed antenna. In this case, the objective function measures the distance
between a specified pattern and the desired, it is commonly called error function
in supervised learning [8–10]. The objective function (F ) corresponding to the
design is given by:
N
x∗ ∈ arg min F (x) = (Pd (fi (x) − Pa (fi (x))2
x
i
(2)
g(x) = v(x) − vmin ≥ 0,
x ∈ [l, u]d
Where (Pd ) is the desired pattern, (Pa ) is the actual pattern values are
expressed in dB and v(x) is the antenna volume. Where xu , xl are upper and
lower limits of design variables, respectively.
800 M. A. Maged et al.
4 Simulation Results
The designed X-Band Isoflux vias choke rings antenna has been successfully sim-
ulated and designed using advance EM simulation and Python programming lan-
guage. The CST microwave studio [11], GA optimization and Python libraries sim-
ulation tools are employed to efficiently evaluate the candidate antenna designs
to determine their gain.
Fig. 3. GA-based machine learning algorithm framework for synthesis Isoflux pattern
the objective functions, and the visualization of the performance, are preset
in the main program as follows: maximum number of iterations = 20, number
of chromosomes = 5, Pcross = 0.5, Pmu = 0.1 and range bound of two radii
[l, u]2 = [20, 22], [26, 30].
Table 1. The dimensions of optimized vias rings radii for synthesis Isoflux pattern
Parameter Value
r1 20.4 mm
r2 28.6 mm
Using those dimensions, simulated return loss and gain are presented in Fig. 4
and Fig. 5, respectively. For the operating frequency, return loss is −20 dB and
gain > 3 dB at desired beam-width also shows excellent cross polarization mea-
surements. Figure 6 shows the searching for best obtained learned Isoflux pat-
terns that nearly matched desired pattern through the learning and optimization
process.
802 M. A. Maged et al.
Fig. 5. The best optimized radii gain pattern of designed antenna at φ = 45◦
Fig. 6. The Isoflux gain patterns of designed antenna through learning of best radii
values at φ = 45◦ .
Also, Fig. 7 shows the learning curve of the GA for optimizing the radii of
rings of vais. The fitness value of the best chromosome stabilizes and fixed for all
the generations and converged to a solution is obtained by the five independent
runs. As can be seen in Fig. 8, the axial ratio (AR) of the antenna is below 3 dB
within the whole 65◦ range in plane cuts of φ equal to 45◦ .
Utilizing Vias and Machine Learning for Design and Optimization 803
Fig. 7. The learning curve of the genetic algorithm for fives runs.
5 Conclusion
This paper presents a new design for X-Band Isoflux choke ring antenna uti-
lizing vias for nanosatellite applications. This design introduces the utilization
of supervised learning and optimization for the radius of vias-based choke rings
using GA which results not only in gain improvement but also learning the
Isoflux pattern.
The optimized X-band antenna has wider impedance bandwidth of > 200
MHz, higher gain > 3 dB in entire coverage range, return loss of less than
−20 dB at resonant frequency, axial ratio less than 3 dB and VSWR less than 2
in desired frequency range of X-band. The final X-band antenna design will be
used in the nanosatellites of Egyptian Space Agency (EgSA).
804 M. A. Maged et al.
References
1. Colantonio, D., Rosito, C.: A spaceborne telemetry loaded bifilar helical antenna for
LEO satellites. In: Proceedings of the International Microwave and Optoelectronics
Conference, Belem, Brazil, pp. 741–745 (2009). https://doi.org/10.1109/IMOC.
2009.5427481
2. Nawaz, W., Ali, A.: Improvement of gain in dual fed X band isoflux choke horn
antenna for use in LEO satellite mission. In: Fourth International Conference on
Aerospace Science and Engineering (ICASE), pp. 1–4. IEEE (2015)
3. Fouany J., et al.: New concept of telemetry X-band circularly polarized antenna
payload for CubeSat. IEEE Antennas Wirel. Propag. Lett. 17, 2987–2991 (2017).
https://doi.org/10.1109/LAWP.2017.2756565
4. Maldonadoa, A., Panduroa, M., del Rio Bociob, C., Mendez, A.: Design of concen-
tric ring antenna array for a reconfigurable isoflux pattern. J. Electromagn. Waves
Appl. 27(12), 1483–1495 (2013). https://doi.org/10.1080/09205071.2013.816877
5. El Misilmani, H., Naous, T., Al Khatib, S.: A review on the design and optimization
of antennas using machine learning algorithms and techniques. Int. J. RF Microw.
Comput. Aided Eng., 1–28 (2020). https://doi.org/10.1002/mmce.22356
6. Haupt R.: Genetic Algorithms in Electromagnetics, 1st edn. Wiley, Hoboken (2013)
7. Koziel, S., Ogurtsov, S.: Simulation-driven design in microwave engineering: meth-
ods. In: Koziel, S., Yang, X.S. (eds.) Computational Optimization, Methods and
Algorithms. SCI, vol. 356, pp. 153–178. Springer, Heidelberg (2011). https://doi.
org/10.1007/978-3-642-20859-1 8
8. Hagglund, A.: Using genetic algorithms to develop a conformal VHF antenna.
Master’s thesis, Stockholm-Sweden (2014)
9. Ledesma, S., Ruiz-Pinales, J., Cerda-Villafan̈a, G., Garcia-Hernandez, M.: A
hybrid method to design wire antennas: design and optimization of antennas using
artificial intelligence. IEEE Antennas Propag. Mag. 57(5), 23–31 (2015). https://
doi.org/10.1109/MAP.2015.2453912
10. Smith, J., Baginski, M.: Thin-wire antenna design using a novel branching scheme
and genetic algorithm optimization. IEEE Antennas Propag. Mag. 67(5), 2934–
2941 (2019). https://doi.org/10.1109/TAP.2019.2902960
11. CST Team: CST Studio Suite: Computer Simulation Technology. https://www.
cst.com/products/cstmws
Research on Automatic Generation System
of the Secondary Safety Measures of Smart
Substation Based on Archival Data
Sun Liye(B)
Electric Power Science Research Institute of State Grid Heilongjiang Provincial Electric Power
Co., Ltd., Haerbin 150030, Heilongjiang, China
Abstract. This article summarizes and analyzes the secondary safety measures
and regulations archival data of smart substation work ticket, establishes electri-
cal archives database of typical safety measures, and converts the relay protection
language of safety measures in archives database into semantics identifiable by
computer, which matches implementation and operation objects modeled in main-
tenance work of substation analyzing SCD file archives to realize the automatic
generation of operating maintenance and repair safety measures for smart sub-
station secondary system. With the aid of archival data of safety measures, it has
effectively improved efficiency and accuracy of safety measure compilation, pro-
vided the significant decision-making basis for safe operations, and improved the
technical level of safe production.
1 Introduction
and realize the automatic generation of typical safety measures. It demonstrates various
implementation objects and operation objects via a graphical interface to form the step-
by-step demonstration of safety measures [2]. Work tickets’ archival data will improve
the accuracy and efficiency of safety measures compilation and provide a significant
basis for safe operations.
Table 1. The relationships between the object of implementation and the object of the operation
The final complete safety measures are the orderly combination of operation objects.
Table 1 indicates the relationship between the implementation object and operation object
related to the equipment A to be maintained.
Hence, the typical safety measures in the archive database could be formalized,
forming into semantic rules understandable by the computer. Then when the maintenance
people are writing specific safety measures, it can combine the voltage class and wiring
mode of the substation, type of the equipment to be maintained, the primary operational
condition of the equipment, maintenance type, and other elements to match appropriate
the appropriate semantic rules. That is to model the general implementation object and
Research on Automatic Generation System 807
operation object in semantic rules into specific implementation object and operation
object examples in the substation to automatically generate safety measures tickets that
are supposed to be carried out during maintenance operations.
The functional pressure plate, soft pressure plate, and maintenance pressure plate
in the specific examples of implementation object and operation object all have been
illustrated in the SCD file in line with standard methods, which can be easily computer
modeled and used to match general semantic rules. However, although optical fire, the
secondary connection, and other operation objects are not described in the SCD file and
could not achieve the model matching, optical fiber and the secondary connection on site
are all physical, visible, and convenient for on-site verification, which could be marked
during matching. And in this way, they could be taken as items to be confirmed in the order
of automatically generated safety measures to avoid a large amount of supplementary
modeling work and provide guidance to on-site compilation and implementation of
safety measures tickets.
The formalization of the typical secondary safety measures consists of 3 parts: defini-
tion of substation design scheme, the definition of maintenance mode, and formalization
of implementation object and operation object of safety measures. The design scheme
of substation mainly concerns the following dimensional elements: voltage class, main
wiring mode, protection configuration, sampling form (electronic voltage and current
transformer or general voltage and current transformer), whether adopting the consoli-
dation unit, and whether adopting the smart terminal. Maintenance mode considers the
power off the scope of primary equipment (power failure at the whole station, voltage
class power failure, interval power failure, no power failure), maintenance type (initial
inspection, routine test, defects elimination, and upgrading).
In the operation procedure, in the substation for each design scheme, every type of
equipment (busbar protection, circuit protection, main transformer protection, consoli-
dation unit, smart terminal, spare power automatic switching, etc.) has its detailed rules
on safety measures under the determined maintenance mode where combination with
the operation procedure must identify the implementation objects and operation objects
in safety measures, model and object them one by one.
The formalization of implementation object can be expressed as:
Primary Blackout
Range Equipment
Maintenance Type .
to be Inspected
{ Start
Interval Type Start Device
Type to Target Interval Type (1)
Target Device Type of Implementation Object Semantics
The main process of substation archive object modeling is to analyze the SCD file archive
and model various data objects. The specific process is to: determine the substation design
scheme type, that is, to determine substation voltage class, design scheme, protection
configuration, sampling form (Electric voltage and current transformer or general volt-
age and current transformer), whether adopting the consolidation unit, whether adopting
the smart terminal [6]. The current IED naming practice on SCD by individual domes-
tic integrators extracts the models that define various types and intervals of secondary
equipment type in the SCD file. For example, PM stands for busbar protection, and PL
stands for circuit protection.
When analyzing SCD files, there is a communication relationship between any two
devices, including GOOSE transceiver and SV transceiver, which are defined as imple-
mentation objects between devices. Then the similar virtual signals are then combined
and formed into one group of virtual signals, defining one group of virtual signals
as an implementation object with a grouping method based on matching keywords in
semantics, refer to Table 3.
3.3 The Semantic Rules and Semantic Matching with Case Objects in Substation
Firstly, for semantic rules and semantic matching with case objects in substation, it
needs to identify the equipment to be inspected, maintenance mode, which is based on
the typical safety measures semantic rules used for information matching in maintenance
and repair scenarios. Then apply the typical secondary maintenance’s safety measures
semantic rules to match each implementation object and each operation object modeled
in a substation. The specific process is as follows:
Research on Automatic Generation System 809
Semantics of Initial end to terminal Virtual signal type Key words for fuzzy
implementation object end matching
Current sampling Circuit consolidation SV Current
unit to circuit
protection
Current sampling Main transformer SV (High voltage *
consolidation unit to Current) or (Medium
main transformer voltage * Current) or
protection (Low voltage *
Current) or (Common
winding * Current) or
(Drivepipe * Current)
Start failure Circuit protection to GOOSE (Circuit branch *
busbar protection Failure) or (Main
transformer * Failure)
Start failure Main transformer GOOSE (Failure of united
protection to busbar jumping)or (Failure of
protection turn-in) or (Failure of
start)
Start failure Busbar protection to GOOSE (Circuit branch *
busbar protection Failure) or (Buscouple
* Failure)
…… …… …… ……
soft pressure plate] exit [PL612A 220 kV × × Circuit 57211 Circuit protection A
set] connecting [ML612A 220 kV × × Circuit 57211 Circuit consolidation unit A
set]”.
1. In the SCD file, the modeled operation object in the semantic rule base can match
with the data object in the SCD file, such as the trip outlet soft pressure plate.
The matching result directly replaces the general operation object with the specific
operation object case.
2. In SCD file, the modeled, but due to reasons such as nonstandard description, the
general operation object in semantic rules failed to match with operation case object,
this kind of circumstance should be marked as “to be confirmed” so that for staff
on-site to confirm whether the operation object to be matched exists. If the operation
object exists, it is an option for safety measures. If not, unplug the optical fiber for
disconnection.
3. Because of SCD file is unmodeled, semantic rules of some general operation objects
in semantic rules could not carry out model matching. The automatically generated
safety measures tickets will be listed directly in which the general operation objects
are defined in the typical safety measures semantics and identified as “to be con-
firmed” for staff on-site to confirm the operation case object, such as optical fiber
the secondary connection.
Node technology, built on the framework of NW.js, and provided in the form of PC client
software. The overall structure is shown in Fig. 1:
The overall structure comprises the data, business analysis, service, human-computer
interaction, and management layers.
Data Layer. The data layer is divided into smart substation archive databases, including
public database, typical scheme database, SCD information database, semantic rule base,
maintenance rules, sequential verification rules, etc.
Business Analysis Layer. The business analysis layer mainly analyzes the association
between SCD files, substation data models, and typical schemes, and the generation of
safety measures.
Service Layer. The service layer mainly provides graphical services and sequential
calibration services.
of busbar protection (A4 maintenance) shall be carried out under the condition of power
failure of the primary equipment at all busbar intervals, and the single device outage
maintenance mode shall be adopted. As shown in Fig. 2, the corresponding typical
tickets can be screened by selecting the devices to be inspected and the type of work
(Figs. 3, 4 and 5).
Fig. 2. Busbar protection initial safety measures under the condition of power failure of the
primary equipment
In the case of 220 kV circuit protection, the completeness of the safety measures
of the initial and secondary inspection can reach 100% under the condition of primary
equipment power failure. In the case of 220 kV circuit protection, the completeness of
the safety measures of the initial and secondary inspection can reach 100% under the
condition of primary equipment power failure. The generated secondary safety measures
cover all operations such as soft pressing plate, functional pressing plate, maintenance
pressing plate, and pulling out optical fiber, and ensure that all operations are free from
omission. It can find the virtual circuit, soft pressing plate, maintenance pressing plate,
optical fiber, etc., contained in the typical ticket but not found in the SCD file. It can
identify virtual circuits in SCD documents but not be involved in typical tickets. The
recognition rate shall be more than 95%, which automatically prompts the operation and
maintenance personnel to verify safe and stable work.
814 S. Liye
6 Conclusion
It proposes the semantic rules of object extraction to establish the complete safety mea-
sure database. The file data define the division of responsibilities, items, implementation
and recovery sequence, format, and content of safety measures for on-site work of sec-
ondary circuit. The implementation object and operation object model of the SCD file
automatically match the implementation object and operation object in the typical ticket.
The current operation safety measures are automatically generated based on the historical
archival data. The efficiency and accuracy of preparing safety measures are significantly
20% higher than the existing technology and reference.
References
1. Luo, Z.K.: Study on secondary safety measures of smart substation. Electr. Eng. 15, 144–146
(2019)
2. Sheng, H.H., Wang, D.L., Ma, W., Luo, S.J., et al.: Exploration of intelligent operation manage-
ment system of relay protection based on big data. Power Syst. Prot. Control. 47(22), 168–175
(2019)
3. Geng, S.B., Zhao, C.L., Gao, X., et al.: Research and application of the strategy for repair
security measures based on formalized description mechanism. Power Syst. Prot. Control.
46(22), 178–186 (2018)
4. Xu, J.Y., Song, F.H., Lu, Z., et al.: Design and realization of the secondary maintenance safety
measures visualization and one-touch operation system in smart substation. Power Syst. Prot.
Control. 45(16), 136–144 (2018)
816 S. Liye
5. Hou, K.Y., Shao, G.H., Wang, H.M., et al.: Research on practical power system stability analysis
algorithm based on modified SVM. Prot. Control Mod. Power Syst. 3(3), 119–125 (2018)
6. Bu, Q.S., Gao, L., Yan, Z.W., et al.: Anti-maloperation strategy and realization of soft-
pressing board of relay protection in intelligent substation. Electr. Power Autom. Equipment
36(12),156–160+168 (2016)
Decision-Making Framework of Supply Chain
Service Innovation Based on Big Data
Haibo Zhu1,2(B)
1 Harbin University of Commerce, Harbin 150028, China
2 Heilongjiang Provincial Laboratory of Electronic Commerce and Information Processing,
Individual and Whole Service Innovation in Supply Chain. The relationship between
the supply chain and enterprises is bidirectional. The supply chain is composed of enter-
prise nodes, and the supply chain affects enterprise performance and operation. There-
fore, service innovation can not be realized by a single enterprise or a single channel;
only by the joint action of all enterprises can it be successfully implemented, and we
should make full use of the knowledge reserve, profitability, and resources of all parties,
and fully grasp its internal mechanism to integrate and improve the supply chain.
Service Innovation in the Context of Big Data. The development of data and infor-
mation technology has triggered the global information revolution and become the core
competitive highland of the next global production revolution. The international compet-
itiveness of countries and regions with high levels of information technology is relatively
strong. In the current era, service innovation should be a new service model that can
increase comprehensive competitiveness. It also requires paying attention to customer
needs, using information technology and super network to explore the value of knowl-
edge service, and increasing enterprise profits. Chen et al. [3] proposed that big data
can be used for service innovation, such as analyzing customer behavior and preference,
predicting customer consumption. Wang et al. believe that improving innovation ability
is to mine valuable information about customers hidden in big data by constructing an
analysis framework. From this, we can see that in the past, most of the researches focused
on the innovation of a single enterprise, but little on the customer experience, and rarely
integrated and innovated the whole supply chain [4].
Service Innovation Ability. Service innovation needs the joint action of individual and
team, and its ability is also composed of individual innovation ability and team inno-
vation ability. Individual mainly refers to the induction education, continuous training
in work, accumulated work experience, and the common interaction behavior among
individuals, teams, and individuals and teams. Flexible use of individual innovation and
individual thinking and skills can form team innovation ability and improve service
innovation. Compared with intangible service innovation, tangible service innovation
is more explicit and easier to code. In the process of transformation to team innova-
tion ability, the problems encountered are relatively small. According to Tobias et al.
[5], we should examine the factors conducive to improving the service innovation ability
from six aspects: perceived demand and technology, cultivating service thinking, flexibly
mobilizing service, cooperation, expanding service scope, and learning ability.
Service innovation is mainly divided into three stages: concept, development, and
introduction. The first stage of service value realization is the concept stage. The main
task is to build a conceptual framework and initially complete all aspects of this process,
such as conceptual creativity, preliminary development, analysis, and inspection, etc. The
second stage is the development stage, which mainly includes feasibility demonstration,
operation link design, training, etc. The third stage is the introduction stage, which is
the evaluation and estimation of service quality, the use of customers, and suggestions
for improvement. This is also the final stage. In the three stages, service innovation
involves integrating and cooperating innovation ability in six aspects: resource input
in production, production, management, marketing and staff, and customers in person-
nel. At the same time, based on the characteristics of the development and evolution
Decision-Making Framework of Supply Chain Service Innovation 819
process of the service industry, the new path and new strategy including four dimen-
sions are formulated, namely service-related page interaction, informatization, lean, and
inventory.
Operation Strategy. In recent years, with the development of the Internet, online pay-
ment and shopping play a more important role in customers’ consumption and greatly
impact offline stores. By exploring the impact, this paper analyzes the mechanism of
customers’ online and offline choices to improve the online and offline sales strategy
and think about how to change the means of payment or delivery to attract customers
and improve customer satisfaction.
Price. How to price is the core problem of multi-channel supply chain and the main
factor affecting customer consumption demand and the research focus of scholars.
Channel Supply Chain Service Level. The service level of the channel supply chain
is usually measured by sales staff performance, customer service, and innovation. In
measuring the service level, the demand for the same product with the same price and
different channels is analyzed, and the relationship between service decision-making and
price is studied. It can be found that among many factors, the degree of effort of sales staff
is the most important. But in reality, it is difficult to evaluate the performance of sales
staff, and it is also difficult to monitor the sales efforts of retailers. Giri (2019) found that
retailers and manufacturers compete for market share by advertising in specific regions,
so advertising investment is an alternative factor [8]. Through the comparison of the
retailers advertising in the same area, it is found that when the price of multi-channel
is the same, the sales volume is less and the sales unit price is higher; Compared with
no advertising, advertising can alleviate the conflict of interest of each channel in the
supply chain and improve the profits of enterprises.
In summary, driven by e-shopping, enterprises will put more energy into improving
service level and innovation to obtain higher profits and occupy more market share. In
the past, in the traditional single-channel environment, offline interaction could provide
a good shopping experience and other value-added services to improve the attractive-
ness of customers. However, in today’s dual-channel model, price is no longer the only
competitive factor, and service level is also included in consideration of retailers. By
analyzing the equilibrium pricing strategies in different markets, it is found that service
is not considered as the influencing factor of decision-making. And the way of factory
direct sales can promote enterprises to improve service and carry out service innovation.
Further analysis based on the customer utility theory shows that when the service level
of retailers is compared with that of manufacturers, the former can increase the profits
820 H. Zhu
of manufacturers, all links in the whole channel and the whole supply chain, which is
an effective means to alleviate the contradictions among enterprises in the channel, and
channel competition is very beneficial to improve their service level. However, in terms
of the current research situation, few scholars are involved in the service innovation and
level of the Omni-channel supply chain.
Fig. 1. Decision framework of supply chain service innovation based on big data
factors, not limited to the past trend, such as marketing, after-sales, policy, etc., which
are the missing dimensions of mainstream forecasting methods. However, we can make
improvements by mining information from big data to improve the accuracy of mar-
ket demand forecasting, improve retailers’ competitiveness, help formulate marketing
strategies to adapt to market development and changes, improve after-sales service and
shopping experience, and so on make customers satisfied.
In this paper, data mining combines data mining and a Bayesian network, first through
data mining to obtain useful data and then the Bayesian algorithm to extract refined data.
The specific process is as follows:
Firstly, collect data, that is, obtain data related to service demand, including customer
purchase records, early consumer search information on related products, and post-use
evaluation of products, which can reflect consumer preferences. That is to say, in the
observed supply chain, the appropriate samples are selected according to the previous
experience and experts’ suggestions, and the influencing factors are identified to become
Bayesian network nodes, and each node is assigned a value.
Secondly, the sample data is preprocessed to facilitate the next step of data analysis. The
original data should be processed by clustering analysis, hierarchical analysis, and other
statistical methods to make its assignment discontinuous to model the algorithm easily.
Thirdly, we design and construct a Bayesian network, namely the belief network, a
closed-loop graph composed of all links and directed edges. Each node is composed of
random variables, and the relationship between variables is represented by the connection
direction of the parent node to the child node. The conditional probability is used to
express the connection strength, and the prior probability without parent node is used
to express the relationship between information. The uncertainty of related variables is
determined by P((X |π (X ))), where π (X ) represents the parent node X . Therefore, under
the relevant assumptions of Bayesian network, it can be expressed as joint probability
distribution, as follows:
P(X1 , X2 , . . . , Xn ) = P(Xi |π (Xi ))i = 1, 2, . . . , n (1)
822 H. Zhu
environment, and the distribution range is wider and wider, and the scale of data col-
lected is also becoming larger and larger. The cost of obtaining this information is very
high, but most of the internal value is ignored. Therefore, managers need to understand
the Omni-channel supply chain’s information to reduce the cost of obtaining informa-
tion and improve work efficiency. Therefore, in the primary stage of data experiment
management, making storage requirements and scientific computing conditions of many
data sets.
Clean Up and Integrate Data. In the process of general data analysis, the scale of data
to be processed is often too large, and the data source is very complex. There are serious
inconsistencies in database information, data loss, and noise instability. Therefore, before
analyzing the data, we must preprocess the data collected in advance. We can choose
data cleaning, integration, conversion, restoration, and other technologies to solve data
instability or inconsistency. Using data mining-related technologies, we can find useful
information for decision-makers, such as helping enterprises find single or composite
skills in product and service requirements. We can also use visualization techniques to
get more information we want, such as integrating resources and capabilities of other
companies and the customer needs to be faced by enterprises. These pieces of information
are very important for the development of the analysis model in the next stage.
Data Analysis. The data preprocessing in the second stage is more suitable for analysis.
In this paper, the network optimization reasoning model combined with other enterprise
capability sets shows product and service innovation and expansion. Such as “E” rep-
resents the problem to be solved, “TR” represents the ability set needed to solve the
problem, “SK” represents the ability set to get a solution, and “I” represents the media-
tion skills. Then the mediation skills “I” can help managers obtain the real needs from
the ability set. This method is based on the inference network graph SK and uses integer
programming to find the optimal solution from the beginning node to the end node.
Moreover, E, TR, SK, I, and related skills and data can be obtained from the data
processing in the second stage. The obtained information can be input into the reasoning
diagram to carry out logical reasoning processes. A model to solve a specific problem can
be constructed under the background of big data and Omni-channel supply chain service
innovation behavior evolution theory. Then, the expansion process of the reasoning
network is visualized by MATLAB and other programming software to help decision-
makers get the optimal solution [12].
Interpret Data and Make Decisions. First, collect data and preprocess it, then use a
reasoning network for data analysis, build a set of capabilities that can be continuously
improved, and obtain the best solution that Omni-channel managers want by seeking
the optimal solution. Then the best decision is obtained through analysis. The network
relationship obtained in this process is sorted and applied to the whole Omni-channel
supply chain service process. In a word, the above process can fully refine the demand of
classified products and services, build a specific analysis process, improve and coordi-
nate the Omni-channel supply chain, and help manufacturers and retailers continuously
improve and innovate in products and services.
824 H. Zhu
References
1. Feng, Z., Guo, X., Zeng, D.: Some frontier topics of business management research under
the background of big data. J. Bus. Rev. 16(01), 7–12 (2013)
2. Robert, A., Stephen, M.: The services science and innovation series. Eur. Manag. J. 66(02),
57–67 (2008)
3. Chen, G.: Research on business model innovation mechanism of supply chain enterprises. J.
Sci. Res. Manage. 26(12), 12–16 (2018)
4. Wang, J.: Research on channel optimization in supply chain of retail enterprises. J. Bus. Econ.
Res. 20(4), 26–32 (2018)
Decision-Making Framework of Supply Chain Service Innovation 825
5. Tobias, S., Cheri, S.: Data science, predictive analytics, and big data in supply chain
management: current state and future potential. J. Bus. Logist. 33(01), 44–53 (2015)
6. Darrell, R.: The future of shopping. Harv. Bus. Rev. 89(12), 64–75 (2011)
7. Zhang, Z., Xue, C.: Research on supply chain optimization of retail enterprises under channel
mode. J. Ind. Econ. 21(4), 32–36 (2016)
8. Giri, B., Bardhan, S.: Sub-supply chain coordination in a three-layer chain under demand
uncertainty and random yield in production. Int. J. Prod. Econ. 56(09), 16–33 (2019)
9. Yang, B., Park, B., Chang, B.: Optimal pricing and advertising decisions in a dynamic multi-
channel supply chain. Int. J. Control Autom. 11(10), 119–134 (2018)
10. David, S., Nikolaos, T.: Operations management in multi-channel retailing. Oper. Res. 36(05),
66–78 (2016)
11. Pattnayak, P., Patra, S., Pradhan, J.: Optimizing the network flow in cloud supply chain
management. Int. J. u - and e-Serv. Sci. Technol. 10(5), 43–54 (2017)
12. Niranjan, T., Parthiban, P., Sundaram, K.: Designing an omnichannel closed loop green supply
chain network adapting preferences of rational customers. J. Oper. Res. Soc. 38(11), 89–99
(2019)
Determination of Headways Distribution
Between Vehicles on City Streets
Abstract. This paper discusses a work that contains a study to solve the optimal
saturation of cities’ street network elements by vehicles. The work is based on
experimental-theoretical studies of patterns for headway determination between
vehicles on the city arterial streets. The main approaches to determining the traffic
capacity of roads and carriageways of city streets are considered and analyzed. The
lanes capacity of city arterial streets having uninterrupted traffic is determined,
taking into account the regularities of traffic flows. Regularities of vehicles dis-
tribution on lanes of the arterial streets are proved, and parameters of headways
distribution between the vehicles moving in a dense traffic flow are found. Based
on the obtained data, it has been found that the minimum headway for each lane
is different. This is influenced by the speed and composition of the traffic flow
and the patterns of vehicles distribution between the lanes. The actual minimum
headway between vehicles moving on the city arterial streets is found.
1 Introduction
A modern city is a concentration of residential buildings, industrial enterprises, admin-
istrative, cultural, medical institutions, etc., and many people in a relatively small area.
The comfortable and efficient level of life in such type complicated urban system is based
on the efficiency level interaction of all its components through the proper organization
of transport system as the main element uniting other components into a single whole
structure.
Nowadays, in multi-purpose cities with a well-developed division of social labor,
the residence place is less and less tied to the workplace. Need in the trips is growing
due to the growth of mobility requirement through the number of trips (traffic volume)
and traffic distance increase.
Urban traffic is a result of the need to interact between elements of the city’s planning
structure. Such interaction is the most effective when the modes of transport providing
traffic flows are most in line with the size of the city. Urban traffic in today’s major and
largest cities has grown into an important transport and urban planning problem. The
correct solution depends on people’s level of living conditions and the further territorial
and economic development of the cities themselves.
Daily trips of thousands of people create high-density traffic and pedestrian flows
on the city street network, which leads to a significant increase of time taken for trips
and forms so-called “transport fatigue” from uncomfortable travel conditions.
Analyzing the traffic process on the carriageway lanes of the city arterial streets, we
see that the uneven distribution of interspace between vehicles in the lanes is observed
in the interrupting traffic flow. The distribution of interspaces is significantly affected
by public transport stops the distance between signalized intersections, the number of
lanes, signalized intersections, where it is allowed to change the traffic direction, etc.
Accordingly, this explains the essence of the carriageway efficiency. It should be noted
that in these conditions, the main parameters of the carriageway and the factors that
affect the increase of its capacity are almost completely the same for all traffic lanes
on the certain carriageway section, in particular the geometric dimensions of each lane
(carriageway width, transverse and longitudinal slope, the radius of the vertical and
horizontal curve, etc.), the same are the parameters and properties of the pavement and
the road structure itself, as well as the environmental conditions.
Based on the principles of headways distribution between vehicles moving on the
road network, it is possible to determine the range of certain theoretical models applica-
tion to describe the patterns of the traffic flow formation on the city’s arterial streets
depending on the lanes number. Surveying the traffic patterns on multi-lane streets
and roads enables to improve the theoretical basis of their design and develop specific
measures to improve and enhance traffic conditions on them.
Many scientific works of national and foreign scientists are devoted to the study of
determining the distribution of time headways between vehicles moving in a dense flow,
among which there are works [1–6]. But, analyzing these scientific works, it should be
noted that the study of the time headways between vehicles on the arterial streets of
Ukrainian cities with 4–8 lanes was not paid attention.
The purpose of the work is to improve the approaches and principles of determining
the capacity of the lane of the arterial streets and roads for uninterrupted traffic, taking
into account their traffic flow patterns.
3 Results
The urban streets and roads capacity is the most important task among the all urban
transport problems of the city. It is well known that the city street network capacity (SNC)
is the main indicator characterizing the effectiveness conditions of the city transport
system operation.
The contradictions that arise in the traffic flows organization on the city streets,
create the corresponding difficulties in their operation. The vehicle-to-vehicle interaction
and the need to ensure urban traffic safety require the introduction of restrictive and
828 A. Bieliatynskyi et al.
prohibitive measures, which, in turn, affect the speed of traffic flows and the street
network capacity.
Each section of the city street has its planning features that allow the vehicle to move
at a speed determined by traffic rules. It is known that the lower the speed, the greater the
time spent, and it is the increase of travel time that almost always bothers the citizens.
Speed is one of the most important traffic indicators because it characterizes its target
function. Any reduction of vehicles speed compared to the permitted one, and even more
so the traffic delays, lead to economic losses. Therefore, the urban transport operation
efficiency contains organizational measures directed on traffic delay minimization.
When researching the principles of traffic delays and places of traffic congestion
formation [7–9], the main attention should be paid to improving the traffic conditions
on the city arterial streets, reducing delays and thus increasing the traffic speed. One of
the main tasks to ensure continuous, comfortable, and safe vehicles on the city street
network is to minimize vehicle delays under heavy traffic flows.
It is known that the theoretical maximum traffic capacity of the city road lane, taking
into account the maximum dynamic length of the platoon moving at a speed of 40 km/h
is 1450 vehicles/hour. But today, this theoretical value does not correspond to the actual
number. As a result of the observations performed, we have found that the maximum
traffic intensity in one lane of arterial streets of Kyiv city is 1869 units/hour, which
is 29% higher than theoretical indicators. This is because the theoretical calculation is
made taking into account all the factors of influence, which have inflated indicators that
ensure the safety of vehicles and pedestrians. According to works [10–12], the maximum
traffic capacity on the city road lane at the optimal speed values at the uninterrupted traffic
conditions and maximum traffic density is about 2500 vehicles/hour.
Such value can be achieved at a speed of 100 km/h, as is the case on highways in the
United States, England, and other countries [13]. It should be noted that the maximum
traffic capacity in real conditions on urban arterial streets is almost impossible to achieve.
When organizing the city streets traffic, it is clear that it is necessary to consider the
current road regulations, which clearly define the maximum speed value on the city
streets and roads. Therefore, the real conditions create certain restrictions to achieve the
maximum result.
But it should be noted that taking into account the real situation in the lane of the
street or road. An individual approach to traffic management on them, will provide better
operation of this particular carriageway, which, in general, will help to improve the traffic
conditions within the whole city street network and will increase its traffic capacity.
In our opinion, today, the proposal to use the indicator of specific traffic intensity is
very interesting. In [13], it is noted that according to the theory of saturated traffic flows,
it is advisable to determine the traffic capacity not in the cross-section of the lane but
space, so it is necessary to determine the specific intensity U (vehicle/h·km), which is
distributed in lane space and is the ratio of intensity to the length of the lane L:
U = N /L (1)
The specific traffic intensity allows to compare city highways and roads depending on
the distribution of traffic flow intensity on them and thus to characterize their efficiency
in time and space.
Determination of Headways Distribution Between Vehicles 829
This indicator can allow the rational distribution [12, 14, 15] of traffic flows on the
city street network, thus ensuring its efficient operation.
Specific traffic intensity determines the headway between vehicles according to
which they pass the lane’s cross-section. At the headway l.44 s – 2500 vehicles/h [9], at
the headway of 1 s, there is an intensity of 3600 vehicles/h. Thus, at the optimal level of 50
vehicles per kilometer at the headway 2 s, the traffic capacity reaches 1800 vehicles/hour.
Thus, the minimum headway between vehicles moving in the lane is a decisive factor
in determining its maximum traffic capacity.
For our further study, which determines the headways between vehicles moving in
different lanes on the city street network, we took sections of the arterial street of Kyiv
city having uninterrupted traffic. Such sections have been chosen following the following
requirements: absence of road intersections of any kind; absence of any road elements
causing to change the lane, the direction of traffic or having direct influence on traffic
flow.
At this stage of the study, the main task was to determine the headways between
vehicles moving in the traffic flow. The fourth lane of the city arterial street has been
chosen during the field observations because it is the most saturated lane on the four-lane
highways. The data had been fixing in the condition of heavy traffic, where the situation
was close to traffic congestion. Vehicles were moving in a dense uninterrupted flow. This
condition is characterized by the maximum density and almost uniform distribution of
vehicles on available lanes. The average speed of traffic was 18.7 km/h.
The method of mathematical statistics was used to process the obtained data. To
calculate the mean value and standard deviation, the data obtained from the observation
represented in the Table 1.
Table 1. Calculation of headways between vehicles on the arterial street for the fourth lane.
Table 1. (continued)
To determine the average headway between vehicles in a dense traffic flow, we take
the accuracy = 1 s.
Based on the experimental indicators calculations, it can be stated that the average
headway between vehicles under dense flow conditions ranges from 2.81 to 3.21 s.
According to the obtained average headway, we determine the traffic intensity
of vehicles, which is approximately 1200 vehicles/hour or within the range (1120–
1280 vehicles/hour). It should be noted that this is the value of the design traffic intensity
of the lane (in the given units) that is used for designing arterial streets of the city with
uninterrupted traffic (ASCUT) in accordance with current regulations. Then, taking into
account the coefficients of use of traffic lanes (in the conditions of absence of public
transport stops on the road stage or if they are arranged outside the carriageway), that is
1.9 for two lanes, 2.7 for three lanes, and 3.5 for four lanes, the traffic capacity of city
arterial streets can be determined as 4200 vehicles/hour.
When designing the arterial streets, we first consider its traffic capacity at the intensity
that is characteristic for the road condition close to the congestion state, where the
vehicle’s speed is minimal. The last, according to the theory of traffic flows, does not
provide maximum traffic intensity, which is exactly that parameter that characterizes the
effective operation of the arterial street network section.
To ensure the efficient operation of the arterial street, it is necessary to set the optimal
headway between vehicles, which will enable the maximum number of vehicles to pass
through for a certain time, ensuring at the same time the safe and convenient conditions
of the traffic. Hence, the task is to determine the maximum number of vehicles each lane
can pass in a partially uninterrupted flow where traffic is regulated by vehicle-vehicle
interactions and interactions between vehicles and the roadway.
The main task of the study was to establish the average and minimum possible
headways between vehicles moving in the lane and establish the average traffic speed on
this section of the arterial street. Therefore, the relevant observations were performed on
four lanes of the arterial street. These observations involved determining the headway
between vehicles moving one after another. The observation was performed for 20 min
for each lane. The procedure of the experiment was to record the number of vehicles
that pass through the cross-section of the lane for a 10 s period. The obtained data were
grouped and listed in Table 2. The shape of the distribution curve depends on the intensity
and varies depending on the lane number (Fig. 1).
Determination of Headways Distribution Between Vehicles 831
Fig. 1. Distribution of headways in time on four lanes of the city arterial streets with uninterrupted
traffic
Table 2. The distribution of headways on the arterial streets of the city for uninterrupted traffic.
0 >10,0 32 0 0 0 0 0 0 0 0 0 0 0
1 5,0–10,0 42 42 26 26 26 8 6 6 1 8 8 1
(continued)
832 A. Bieliatynskyi et al.
Table 2. (continued)
2 3,33–5,0 26 52 33 30 60 18 20 40 9 10 20 4
3 2,5–3,33 16 48 30 26 78 23 34 82 18 16 48 9
4 2,0–2,5 2 8 5 24 96 29 18 92 21 28 100 19
5 1,67–2,0 2 10 6 12 60 18 26 130 29 20 112 21
6 1,43–1,67 0 0 0 2 12 4 16 96 22 24 144 27
7 1,25–1,43 0 0 0 0 0 0 0 0 0 10 70 13
8 1–1,25 0 0 0 0 0 0 0 0 0 4 32 6
Total 120 160 100 120 332 100 120 446 100 120 534 100
The difference between the average, actual and minimum headway for each lane of
the carriageway is: for the first lane 5 s, the second one −1.9 s, the third one – 1.2 s, and
the fourth one −1 s. As far as the optimal headway is concerned, we see that the first
lane is underloaded 1.8 times, the second 1.6 times, the third and fourth 1.5 times.
Fig. 2. Minimum and average headway between vehicles on each separate traffic lane
The difference between the average and optimal headways is explained by the fact
that when the distance between vehicles moving in the traffic flow is short and a speed
is over 60 km/h, it is very difficult for a driver to change the lane. In [9] it is noted
that the headways between vehicles that enable you to change the lanes in traffic flow
should be 2.5 s, and at headways of 1.6 s, such a maneuver becomes dangerous. As you
can see from Table 2, the percentage of vehicles moving in the corresponding lane with
headways less than 2.5 s, are: for the first lane 11%, the second lane 51%, for the third
Determination of Headways Distribution Between Vehicles 833
and fourth lanes 72%, and 86%, respectively. Therefore, at this intensity, the change of
the current lane to the third and fourth lanes is almost impossible, which, in turn, does
not allow to achieve the optimal headway for each lane.
4 Discussion
The traffic capacity of the city arterial street lanes for public passenger transport uninter-
rupted traffic is: 850 vehicles/hour for the first lane, 1550 vehicles/hour for the second
lane, 2000 vehicles/hour for the third lane, 2400 vehicles/hour for the fourth lane, and
the total intensity is 6800 vehicles/hour. According to the data represented in [9], the
traffic capacity for the four-lane street is 7235 vehicles/hour. The following data were
obtained for the two-lane streets: 850 vehicles/h for the first lane the 2000 vehicles/h for
the second lane. For the three-lane streets, such type data represent 850 vehicles/h for
the first lane, 1600 vehicles/h for the second lane, and 2000 vehicles/h for the third lane.
Based on the observations performed, the proportions of vehicles distribution on the
road lanes depending on lanes number and the headways between vehicles on each lane
where they move (Table 3).
Table 3. The proportions of vehicles distribution on the road lanes depending on lanes number
and the headways between vehicles.
It should be noted that the vehicles speed has a significant effect on the headway
between vehicles moving in separate lanes on the roadway.The current task of the study
was to determine the speed of traffic in the lanes and to determine compliance between
the actual speed distribution and “normal distribution law”. The value of the average
actual speed, which was obtained as a result of experimental studies on the arterial
streets of Kyiv, is shown in Fig. 3.
According to the survey performed and the results obtained, the following values
were obtained: for the first lane the characteristic speed is 43.2 km/h, for the second –
54.2 km/h, for the third – 65.1 km/h, and for the fourth – 78.9 km/h. The obtained data
indicate that the vehicles on the third and fourth lanes exceed permanently a speed limit
approved by traffic rules.
One of the reasons for the speed decrease and its uneven distribution between the
lanes is the change of traffic lanes by vehicles, as well as the presence of freight and
public passenger transport vehicles on the third and even fourth lanes.
The results of data processing correspond to the well-known principles, where the
speed in the first lane is always less than in the others. For example, the average speed
834 A. Bieliatynskyi et al.
on the first lane of a four-lane carriageway is less than on a three-lane road of 43.2 and
45.2 km/h, respectively. This is significantly affected by the width of the first lane, the
safe gap, vehicles parked along the sidewalk, the composition of the traffic flow, and
in particular, by the public passenger transport. It should be noted that the analysis of
the relationship between intensity and velocity is very well described and performed in
[3], where the corresponding regression equations are defined. Given that the intensity
for lanes, which we obtained as a result of observations, almost coincides with the data
obtained in work [9], so for further calculations and generalization of the study results,
we will use the appropriate regression equations proposed in [9].
Based on the observations discussed above, it can be argued that ensuring the effective
use of different sections of the city arterial streets is to organize the vehicle traffic on
them with the optimal headway, which is individual for each lane.
5 Conclusion
Comparing the obtained data, it can be noted that all lanes are, in fact, not fully saturated,
but, at the same time, it must be said that achieving maximum traffic capacity on the
arterial streets of Ukrainian cities is currently almost impossible. Therefore, as already
mentioned, the traffic capacity is affected by many different external and internal factors
with a probabilistic effect. In the observation, we have fixed the maximum traffic inten-
sity on lanes on city arterial streets. The obtained results show that transport network
efficiency increase due to the rational distribution of vehicles in the lanes, and the traffic
capacity increase by reducing the headway between vehicles is possible only at specific
sections of the arterial street network.
References
1. Semchenko, N.: The relationship of speed with the intensity and density of traffic flows on
multi-lane highways. Road Transport. 19, 38–40 (2006)
2. Krasnikov, A.: Regularities of distribution of time headways between vehicles on multi-lane
roads. Works of MADI 95, 74–83 (1975)
Determination of Headways Distribution Between Vehicles 835
3. Petrov, E.: Modelling of high-intensity traffic flow. Omsk Sci. Bull. 21, 137–138 (2002)
4. Zaporozhtseva, E.: Traffic capacity of highways. Socio-economic problems of development
and functioning of transport systems of cities. 19, 225–229 (2010)
5. Jinhwan, J., Changsoo, P., Byunghwa, K., Namkuk, C.: Modeling of time headway distribution
on suburban arterial: case study from South Korea. Procedia Soc. Behav. Sci. 16, 240–247
(2011)
6. Rupali, R., Pritam, S.: Headway distribution models of two-lane roads under mixed traffic
conditions: a case study from India. Eur. Transp. Res. Rev. 10(1), 1–12 (2018)
7. Stepanchuk, O., Bieliatynskyi, A., Pylypenko, O., Stepanchuk, S.: Surveying of traffic
congestions on arterial roads of Kyiv city. Procedia Eng. 187, 14–21 (2017)
8. Nguyen-Phuoc, D.Q., Young, W., Currie, G., De Gruyter, C.: Traffic congestion relief
associated with public transport: state-of-the-art. Public Transp. 12(2), 455–481 (2020)
9. Zaporozhtseva, O.V.: Improving the Principles of Determining the Capacity Of Multi-Lane
Highways. Dissertation of the Candidate of Technical Sciences: 05.22 .11/O.B. Zaporozht-
seva. – Kharkiv National Automobile and Road University of the Ministry of Education and
Science of Ukraine, Kharkiv, 145 p. (2016)
10. Stepanchuk, O., Bieliatynskyi, A., Pylypenko, O., Stepanchuk, S.: Peculiarities of city street-
road network modelling. Procedia Eng. 134, 276–283 (2016)
11. Guk, V.I.: Elements of the theory of traffic flows and design of streets and roads: Tutorial/V.I.
Hooke. – K.: UMK VO, 255 (1991)
12. Stepanchuk, O., Bieliatynskyi, A., Pylypenko, O.: Modelling the bottlenecks interconnection
on the city street network. In: Popovic, Z., Manakov, A., Breskich, V. (eds.) TransSiberia
2019. AISC, vol. 1116, pp. 889–898. Springer, Cham (2020). https://doi.org/10.1007/978-3-
030-37919-3_88
13. Guk, V.I., Shkodovsky, Y.M.: Transport Flows: The Theory of Their Application In Urban
Planning: Monograph. – X.: Golden Pages, 232 (2009)
14. Stepanchuk, O., Bieliatynskyi, A., Pylypenko, O.: Regularities of City Passenger Traffic Based
on Existing Inter-district Links. In: Murgul, V., Pukhkal, V. (eds.) International Scientific Con-
ference Energy Management of Municipal Facilities and Sustainable Energy Technologies
EMMFT 2019. EMMFT 2019. Advances in Intelligent Systems and Computing, vol. 1258.
Springer, Cham (2019)
15. Hossain, M.T., Hasan, M.K.: Assessment of traffic congestion by traffic flow analysis in Pabna
Town. Am. J. Traffic Transp. 4(3), 75–81 (2019)
Does Digital Finance Promote the Upgrading
of Industrial Structure
Harbin Vocational College of Science and Technology, Harbin 150300, Heilongjiang, China
quy@hrbcu.edu.cn
Abstract. The growth of digital finance improves SMEs’ financing conditions and
business circumstances and has a crucial role in upgrading the industrial structure
of China. This paper takes China’s 30 provinces from 2011 to 2019 as samples and
studies the effect of digital finance on upgrading China’s industrial structure from
both theory and demonstrative two aspects. The results reflect that digital finance
stimulus the optimization and upgrading of China’s industrial structure as a whole.
Furthermore, digital finance in the eastern region has a significant positive impact
on the industrial structure. However, digital finance plays an inhibitory role in
upgrading the central and western regions in the industrial structure.
The upgrading of industrial structure is a major measure to change the extensive mode
of economic growth and symbol of development modern social and economic [1]. Since
the reform and opening up, China has rapidly developed from a backward agricultural
country to manufacturing. The three industrial structures have been constantly adjusted,
and industrial development has made great achievements. However, there is still a big
gap between industrial structure and high-quality, sustainable growth of the economy.
Financial development is a key way to promote the upgrading of industrial structures.
Still, there is a serious problem of financial exclusion in the development process of
traditional finance, which can not play effectively to the ability of financial services to
the substantial economic growth. With the development of the Internet, digital finance
emerges as the times require. Digital finance debases the problem of information asym-
metry and the threshold of financial services through big data so that SMEs can also
access financial services. Digital finance has significantly improved the coverage and
use depth of financial services, optimized the resource allocation pattern, and promoted
upgrading industrial structures. Besides, digital finance can improve the credit system
and make the capital flow to high-quality industrial projects to boost the upgrading of
industrial structure [2]. Therefore, this study explores the influence of digital finance on
upgrading industrial structures based on theory.
Independent Variable. This paper selects digital finance (DF) as the independent vari-
able. Referring to the research of Guo et al. (2020) [11], inclusive digital finance are
selected as the index to measure digital finance. To further explore the effect of digital
finance on upgrading industrial structure, this paper also chooses three sub-dimensions
of inclusive digital finance (coverage breadth, usage depth, and digitization level) as
independent variables.
838 Y. Qu et al.
Control Variable. This paper selects the economic development level, fixed assets
investment level, government intervention level, and the R&D investment level as the
control variables to increase the robustness of the empirical model. Economic devel-
opment level is measured by regional per capita GDP. For the stability of the data, the
logarithm of regional per capita GDP is taken and recorded as lnPGDP. The growth rate
of fixed assets is used as the proxy index for the level of fixed assets (FA). The level of
government intervention is benchmarked by the natural logarithm of government gen-
eral budgetary defray (lnGov). R&D level is assessed by the natural logarithm of R&D
investment (lnR&D).
This work selects panel data of 30 areas (including provinces, autonomous regions,
and municipalities) between 2011 and 2019, excluding Tibet, Taiwan, Hong Kong, and
Macao. The data of inclusive digital finance originates China digital finance research cen-
ter of Peking University, and the rest of the data comes from China Statistical Yearbook.
To eliminate the effect of price, the index data related to price factors are deflated. For
individual missing values, the interpolation method is used for information processing.
The descriptive statistics of each variable involved in this study are recorded in Table 1.
According to the above analysis, this paper constructs model (1) to explore the
relationship between digital finance and industrial structure upgrading.
3 Empirical Research
Firstly, this paper uses the Hausman test to judge which model is used in benchmark
regression. Results show that the adjoint probability value of the Hausman test is 0.0002
in the regression of digital inclusive finance and breadth coverage to industrial structure
upgrading. Therefore, the fixed-effect model is selected for regression. In the regression
of the usage depth of digital finance and the digitization level to upgrade industrial struc-
ture, the adjoint probability value of the Hausman test is 0.1913 and 0.0887, respectively.
So for the regression of these two sub-dimensions, a random effect model is used.
Table 2 reports the regression results of digital finance and its three sub-dimensions
on industrial structure. Column (1) (3) (5) (7) are the regression results of digital finance
and its sub-dimensions on the upgrading of industrial structure. Columns (2) (4) (6) (8)
are the regression results after the introduction of control variables.
From the result of column (1), it can be seen that digital finance has a promoting
effect on upgrading of industrial structure and has passed the significance test under 1%
level. And every 1% increase in digital finance will lead to 0.032 units in upgrading
industrial structure. After introducing control variables, the promotion effect is slightly
weakened but still has a significant promotion role. This means that digital finance can
improve the aggregation quality among industries, promote the overall coordination and
resource allocation of industrial structures, and promote upgrading industrial structures.
As for control variables, regional economic development level can promote the indus-
trial structure upgrading. Economic growth can promote capital accumulation, weaken
the resource constraints in structural adjustment, provide a broader market for indus-
trial structure transformation, and accelerate the restructuring of new industries. The
level of fixed assets investment, government intervention, and R&D investment levels
all have inhibitory effects on upgrading the industrial structure. And the level of fixed
asset investment has no significant effect on the upgrading of industrial structure. The
government intervention level and the level of R&D investment are significant at 10%
and 5%, respectively. Under the condition of the market economy, there is large informa-
tion asymmetry. The government cannot comprehensively grasp and accurately analyze
all the information in the market, leading to the distortion of industrial policy in the
implementation and implementation, which has no help to the upgrading of industrial
structure. In terms of scientific research and development, China may pursue more eco-
nomic growth, and technological innovation is more inclined to projects with a high
output value. These projects are not concentrated in the tertiary industry. And the trans-
formation rate of scientific results in China has always been low, which may lead to
the investment in scientific research and development inhibiting the industrial structure
upgrading.
840 Y. Qu et al.
From columns (3)–(8), it can be seen that the coverage breadth of digital finance
has no obvious impact on the upgrading of industrial structure. The usage depth of
digital finance significantly promotes the upgrading of industrial structure and passed
the significance test at the level of 1%. However, the digitization level of digital finance
significantly inhibited the upgrading of industrial structure at level of 1%. It shows that
digital finance plays an active role in upgrading industrial structure in the usage depth.
Table 2. The influence of digital finance and sub dimensions on the industrial structure upgrading
Notes: *, **, *** indicate significance at the 10%, 5% and 1% level. For the estimated coefficients,
the standard errors are in parentheses.
parts, the development of digital finance is more reflected in promoting the growth of
the secondary industry. After the second industry gets the dividend, it may turn the third
industry into the second industry. This enlightens that reasonable policies should be
inclined to the central and western areas to avoid expanding development differences
between the eastern and western regions.
Table 3. Heterogeneity analysis of the impact of digital finance on the upgrading of industrial
structure
To ensure the reliability of empirical results, this work uses various methods to check
out the robustness of the regression model.
First, the core explanatory variables are delayed for a period and then regressed. As
shown in Table 4 (1) and (2), digital finance with lag period still produces a significant
active role in promoting the upgrading of industrial structure.
Secondly, due to the municipalities directly under the central government have
more superiorities in economic growth, political policy, science and culture, and trans-
portation location than other provinces, which may lead to inconsistent effects of digital
842 Y. Qu et al.
finance on industrial upgrading. Therefore, this paper excludes the data of four munici-
palities directly under the central government of China and returns again. The results are
shown in Table 4 (3) and (4), and the digital finance has a strong promoting correlation
to the upgrading of industrial structure.
In conclusion, the empirical result of this paper has certain robustness.
Acknowledgments. We thank the funding sponsored by the Key Project of Educational Science
Planning in Heilongjiang Province (ZJB1421217) and General Project of Teaching Reform of
Higher Vocational Education in Heilongjiang (SJGZY2019172).
References
1. Kuznets, S.S.: Modern Economic Growth : Rate, Structure, And Spread. Yale University
Press, New Haven (1966)
2. Guo, W.L., Chen, J.Y.: Research on the effect of digital inclusive finance development and
industrial structure upgrading in China. Mall Modern. 11, 146–148 (2020)
3. Goldsmith, R.W.: Financial Structure and Development. Yale University Press, New Haven
(1969)
4. Wurgler, J.: Financial markets and the allocation of capital. J. Financ. Econ. 58(1), 187–214
(2000)
5. Da Rin, M., Hellmann, T.: Banks as catalysts for industrialization. J. Financ. Intermed. 11(4),
366–397 (2002)
6. Aghion, P., Howitt, P., Mayer-Foulkes, D.: The effect of financial development on conver-
gence: theory and evidence. Q. J. Econ. 120(1), 173–222 (2005)
7. Chava, S., Oettl, A., Subramanian, A., Subramanian, K.V.: Banking deregulation and
innovation. J. Financ. Econ. 109(3), 759–774 (2013)
8. Sasidharan, S., Jijo Lukose, P.J., Komera, S.: Financing constraints and investments in R&D:
evidence from indian manufacturing firms. Q. Rev. Econ. Finance 55, 28–39 (2015)
9. Fan, X.C., Xia, Y., Hao, Y.M.: Research on the impact of inclusive finance development on
industrial structure upgrading in China. Zhejiang Finance 5, 23–30 (2017)
10. Cai, H.Y., Xu, Y.Z.: Does trade openness affect the upgrading of China’s industrial structure.
Res. Quant. Econ. Tech. Econ. 34(10), 3–22 (2017)
11. Guo, F., Wang, J.Y., Wang, F., Kong, T., Zhang, X., Cheng, Z.Y.: Measuring the development
of digital inclusive finance in China: index compilation and spatial characteristics. Econ. Q.
19(4), 1401–1418 (2020)
Electricity Consumption Prediction Based
on Time Series Data Features Integrate
with Long Short-Term Memory Model
Jiaqiu Wang1(B) , Hao Mou2 , Hai Lin1 , Yining Jin3 , and Ruijie Wang1
1 Beijing China-Power Information Technology Co., Ltd., Haidian District Beijing, China
wangjiaqiu@alu.hit.edu.cn
2 State Grid Sichuan Electric Power Company, Gaoxin District, Chengdu, China
3 Harbin University of Commerce, Songbei District, Harbin, China
1 Introduction
It is of great significance to improve the accuracy of electricity consumption prediction.
First of all, electricity sales, as one of the most concerning indicators of power grid enter-
prises, can be accurately predicted to help power supply enterprises effectively adjust
the power supply plan, optimize the power supply structure, and improve the security
of power system operation. Secondly, as a barometer of economic development, the
electricity consumption also reflects economic development in a region. The govern-
ment, the power supply company, can adjust the region’s industrial structure according
to electricity consumption data [1] to promote economic development.
In essence, establishing an electricity consumption prediction model is a method to
forecast future electricity consumption based on the historical data of electricity con-
sumption. Therefore, the problem of electricity consumption prediction belongs to the
time series prediction problem. Many studies have taken weather, temperature, holidays,
and other factors affecting the establishment of the electricity consumption prediction
model. But as a typical time series problem, few works take the trend, seasonality, and
periodicity of time series prediction problem as the main factors to establish the electric-
ity consumption prediction model, which makes the established model lack persuasive.
At the same time, the existing electricity consumption prediction methods, such as the
traditional empirical method, regression analysis method, gray prediction method and
so on, do not carry out unified and logical treatment for different types of electricity
consumption prediction, and the stability of the model is poor. Although the artificial
neural network prediction method and SVM prediction method can effectively use his-
torical data, their adaptive ability and self-learning ability are relatively insufficient, and
the prediction effect is relatively unsatisfactory.
Therefore, given the above problems, a Long Short-Term Memory Model based on
time series feature fusion is proposed. The model takes into account the influence of three
data features of time series on time prediction results and effectively integrates time-
series features into the Long-short term memory model with strong self-learning ability.
The actual historical data of electricity consumption is used to verify the effectiveness,
and the proposed method is compared with other methods. The experimental results
show that the proposed power consumption prediction model based on the fusion of
time series features has higher accuracy, which indicates that the fusion of time series
features in the power consumption prediction model is applicable and obtains better
prediction results.
The organizational structure of this article is as follows. Section 2 introduces the
related work. Section 3 gives relevant concepts and formal expressions of the model.
Section 4 introduces the structure of the model proposed in this paper. Section 5 intro-
duces the use of data sets, experimental environment, and evaluation indexes of exper-
imental results, and compares and analyzes the experimental results obtained by the
proposed method. Section 6 summarizes this paper and looks into the future work.
2 Related Work
In essence, The prediction of the future electricity consumption based on historical elec-
tricity consumption data belongs to a time series prediction problem. At present, To
divide the electricity consumption prediction model based on the time series predic-
tion method into the electricity consumption prediction model based on the traditional
algorithm, machine learning, and neural network.
Many scholars have practiced the prediction model of electricity consumption based on
traditional methods and achieved good results. Fan Jiao [2] and Wang Tao [3] et al. took a
holiday, temperature, and other factors into consideration in the time series of electricity
consumption. At the same time, they used traditional electricity consumption prediction
models such as a combination model of GM (1,1) model, ARIMA model, Winters
model, and moving average model to forecast monthly electricity consumption. The
results showed that improving the prediction effect of the combination model. Guefano
846 J. Wang et al.
Serge [4] and Laxmi Rajkumari [5] et al. used a new GM (1, 1) – Var (1) hybrid model
based on VAR and gray models and Holt-Winter smoothness (no seasonal) technique to
predict the future electricity consumption in a region. The rationality and applicability
of the model are verified. However, the traditional method of electricity consumption
prediction model of the curve is limited, which cannot fit the changing trend beyond the
curve rule.
For the large-scale industrial users who need to use high voltage grades in the power
grid, the daily electricity consumption time series of the past period is used to predict
the daily electricity consumption of the future period. It should be noted that the high
voltage level of large-scale industrial users is different from the low voltage level of
residential users. For some time in the future, the predicted electricity consumption
is not a prediction of the total amount of electricity, but a forecast of daily electricity
consumption. Therefore, the prediction problem can be regarded as a multi-steps time
series prediction.
Definition 1. Time series of daily electricity consumption. It Refers to the ordered array
of electricity used by users every day according to the time sequence of its occurrence.
Given xt represent a variable varying with time t(t ≥ 1), which means that on the day t,
electricity consumption is xt . In a period of time (t1 , t2 , ..., tN ) t1 < t2 < · · · < tN , N >
2, the discrete sequence (xt1 , xt2 , ..., xtN ) composed by the measurement of electricity
constitutes a time series of electricity consumption.
Time series characteristics contain the trend, seasonality, and periodicity of daily
electricity consumption time series.
Definition 2. Trend Characteristic: the daily electricity consumption shows a relatively
slow and long-term trend of continuous rise, decline, or stays with the same nature over
time, but the range of change may not be equal.
Definition 3. Periodicity Characteristic: When the time series of daily electricity con-
sumption has the rise and fall of irregular frequency, that is to say, the fluctuation of
daily electricity consumption is irregular, it means that the series is periodic.
Definition 4. Seasonality Characteristic: The frequency of electricity consumption fluc-
tuation in the time series of daily electricity consumption is constant and related to a
period of fixed length, indicating seasonality in the series.
Forecasting problem of power consumption of high voltage users. For high-voltage
electricity users, given their historical time series of daily electricity consumption,
the values of future daily electricity consumption series with multiple time steps
are predicted. Formally, given the time series (xt1 , xt2 , ..., xtN ) of users’ historical
daily electricity consumption in the past time interval (t1 , t2 , ..., tN ), the time series
(xtN +1 , xtN +2 , ..., xtN +H ), H > 2 of daily electricity consumption in the future time
interval is predicted. The formula is expressed as follows.
unit. The gate lets information selectively pass through mainly through a function and a
dot product. LSTM is often used to capture the length of data Gating units make it have
strong generalization ability and well performance in time series data. LSTM is often
used to capture the long-term and short-term time dependence of data, and the gating
unit makes it have a strong generalization ability and has a good performance in time
series data.
As RNN adopts the backpropagation algorithm in training when the sequence with
long interval. To find the gradient by chain rule, the phenomenon of gradient disap-
pearance will appear. The gradient disappearance will become more serious with the
longer period of time, making it difficult for RNN to train effectively for the sequence
with long-range. The memory unit structure of LSTM controls the memory of historical
information by three gate structures: forgetting gate, input gate, and output gate. The
calculation process of time t is as follows.
ht = ot tanh(ct ) (6)
In the formula, i, o and f represent input gate, output gate and forgetting gate respec-
tively. c stands for memory unit. h represents hidden layer output, and its subscript
represents time t. W represents the connection weight, and its subscript represents the
weight associated items. b represents the offset item. According to the values of i, o and
f in the range of [0, 1], the proportion of historical information passing through the gate
structure is controlled. So tanh, sigmoid is the activation function.
This paper proposes a time-series features fusion long-short time memory network
model to learn the fused multi-sequence features information and to capture the change
law of sequence features in time series and the relationship between multiple sequence
feature signals at the same moment. The LSTM model is shown in Fig. 1. Power con-
sumption data are contained in time series according to three data characteristics of
Trend, Seasonality and Periodicity for vectorial processing with Word2vec [16]. Then,
the three serial feature signals are serialized into multi-channel input data according
to time, and three serialized LSTM blocks are used to extract deep features from the
multi-channel input. The time step of the LSTM block is set as 1.
The third LSTM block only returns the output of the last moment, reducing the
number of depth characteristics and the number of parameters. Through three groups of
LSTM blocks, the model uses a 4-layers fully connected network to classify the depth
characteristics.
Electricity Consumption Prediction Based on Time Series Data 849
Tendency
LSTM LSTM LSTM
Periodicity FC FC FC FC Sigmoid
Block Block Block
Seasonality
Fig. 1. The proposed time-series feature fusion long-short time memory network.
5 Experiment
In this chapter, we verify the proposed prediction method. To prove the effectiveness of
our algorithm, we introduce it from four aspects: a data set and experimental environment,
experimental results, and evaluation criteria. At last, we analyze the experimental results.
5.2 Evaluation
To verify the validity of the method proposed in this paper for Electricity consumption
forecast, we will compare the experimental results by using the error indexes commonly
used in the world, Root means square error (RMSE) and mean absolute percentage error
(MAPE). At the same time, the Mean Squared Log Error (MSLE) is also an important
index to verify the power consumption prediction results. Expressing the formula as
follows:
850 J. Wang et al.
yi represents the actual value, yi represents the predicted value.
We compared Time Series Features Fusion Long Short-Term Memory Model with basic
LSTM model, Auto-Regressive Integrated Moving Average (ARIMA) on the evaluation
metrics. Showed the experimental results in Fig. 2, Fig. 3, and Fig. 4, and Compared
with other methods, Our proposed algorithm has fewer mistakes in predicting electricity
consumption.
Electricity Consumption Prediction Based on Time Series Data 851
Fig. 2. The comparison between predicted results and the actual results.
Fig. 3. The comparison between the proposed method and other methods on RSLE and RMSE.
Figure 2 shows the comparison between the predicted and actual electricity con-
sumption from August to December 2020. The papers’ predicted rend of electricity
consumption is consistent with the actual trend of electricity consumption from Fig. 2.
The experimental results in Fig. 3 and Fig. 4 show the electricity consumption prediction
model established in this paper compare with the basic long-short term memory network
model and the autoregressive moving average model (ARIMA) under the three evalua-
tion indexes. From the experimental results in the figure, we can see that the electricity
consumption prediction model established in this paper can achieve better results in the
mean absolute percentage error, root means square error and means square log error.
852 J. Wang et al.
Fig. 4. The comparison between the proposed method and other methods on MAPE.
Acknowledgment. The authors thank the editor and anonymous reviewer for their helpful
comments and suggestions that have led to their improved version of this paper.
References
1. Wu, Q., Lu, S.F., Yang, S.H., et al.: Research on household electricity management under
corresponding conditions on demand side. Electric Power Eng. Technol. 35(5), 28–31 (2016)
2. Fan, J., Feng, H., Niu, D.X., Wang, X.Y., Liu, F.Y.: Monthly electricity consumption prediction
based on wavelet analysis and GM-ARIMA model. J. North China Electric Power Univ. Nat.
Sci. Ed. 42(4), 101–105 (2015)
3. Wang, T., Zhang, W., Lu, M.M., Wang, T.M.: Research on prediction method of power
consumption of regional power grid based on time series. Electr. Technol. 11, 9–12 (2010)
4. Guefano, S., Tamba, J.G., Azong, T.E.M., Monkam, L.: Forecast of electricity consumption
in the Cameroonian residential sector by Grey and vector autoregressive models. Energy 214
(2021)
Electricity Consumption Prediction Based on Time Series Data 853
5. Laxmi, R.: Relation between electricity consumption and economic growth in Karnataka,
India: An aggregate and sector-wise analysis. Electricity J. 33(5), (2020)
6. Cao, M., Ju, J., Bai, Z.Y., Liu, J.T.: Prediction method of electricity consumption based on
ARMA-SVM combination model. Energy Environ. 1, 49–51 (2021)
7. Banik, R., Das, P., Ray, S., Biswas, A.: Prediction of electrical energy consumption based
on machine learning technique. Electr. Eng. 103(2), 909–920 (2020). https://doi.org/10.1007/
s00202-020-01126-z
8. Tay, K.G., Sim, S.E., Tiong, W.K., Huong, A.: Forecasting electricity consumption using the
second-order fuzzy time series. IOP Conf. Ser. Mater. Sci. Eng. 932(1), 12056 (2020)
9. Yasir, H.A., Gurudeo, A.T.: The relationship between electricity consumption, peak load and
GDP in Saudi Arabia: a VAR analysis. Math. Comput. Simul. 175 (2020)
10. Ramos, D., Faria, P., Vale, Z., Mourinho, J., Correia, R.: Industrial facility electricity con-
sumption forecast using artificial neural networks and incremental learning. Energies 13(18),
4774 (2020). https://doi.org/10.3390/en13184774
11. Ali, S.H., Zhang, J.R., Azeem, A., Asif, M.: Impact of electricity consumption on economic
growth: an application of vector error correction model and artificial neural networks. J. Dev.
Areas 54(4), (2020)
12. Jayanth, B.A., Harish Ram, D.S., Binoy, B.N.: A deep learning approach to electric energy
consumption modeling. J. Intell. Fuzzy Syst. 36(5), 4049–4055 (2019)
13. Hakan, Y., Umut, U., Oktay, T.: A long-short term memory application on the turkish intraday
electricity price forecasting. Press Academia Procedia (1) (2018)
14. Manowska, A.: Using the LSTM network to forecast the demand for electricity in Poland.
Appl. Sci. 10(23), 8455 (2020)
15. Ya’u, G.A., Lawal, A.M., Jimmy, N.J., Nannin, R.E., Chidi, E.O., Jafaru, B.: LSTM network
for predicting medium to long term electricity usage in residential buildings (Rikkos Jos-City,
Nigeria). Comput. Sci. Eng. 9(2), (2019)
16. Rong, X.: Word2vec Parameter Learning Explained. Comput. Sci. (2014)
Intelligent Search Method in Power Grid Based
on the Combination of Elasticsearch
and Knowledge Graph
Jiaqiu Wang1(B) , Xinhua Yang1 , Yining Jin2 , Xueyong Hu1 , and Le Sun1
1 Beijing China-Power Information Technology Co., Ltd., Haidian District, Beijing, China
wangjiaqiu@alu.hit.edu.cn
2 Harbin University of Commerce, Songbei District, Harbin, China
1 Introduction
With the advent of big data, the power system with large and complex assets, knowledge-
intensive, and consumption characteristics has introduced smart devices and dispatchable
networks in the era of big data, which together constitute a powerful smart grid. In
the smart grid operation, generating a large amount of electric power business data is
inevitable, so it isn’t easy to search the data by traditional manual means. Therefore, a
fast and accurate method is needed to search target data from massive data to improve
the production efficiency of the power industry.
At present, the dispatching work of the power grid is mainly decided and operated
by dispatchers. The dispatcher needs to analyze the state and parameter changes of the
power network and take measures based on experience. This artificial scheduling mode is
subjective [1]. With the rapid development of power systems, the structure and operation
mode of the power grid become more and more complex, and the difficulty of power grid
disposal is increasing [2]. Although the traditional text search method based on keyword
search can get paragraph search results, the search results lack fragmentation. They even
appear to be irrelevant answers, which reduces the work efficiency [3]. Electric power
enterprises as manufacturing enterprises, the low efficiency of the work schedule directly
affects the benefits of enterprises. At present, most of the existing methods focus on the
paragraph search of the technical standard documents of electric power enterprises, and
few can give an accurate answer to the question.
Power business document data includes power dispatching systems, enterprise tech-
nical standards, etc. To accurately find the dispatch knowledge, equipment knowledge,
equipment troubleshooting problems, and equipment maintenance history information,
this paper proposes an intelligent search method based on the combination of elastic
search and knowledge graph to help the dispatcher make decisions and improve the
operation efficiency of the power grid. The Elasticsearch method is used to select the
contents of paragraphs in advance to reduce the search scope. Secondly, in a small range
of paragraph content, the triples in the knowledge graph are used to match the problems
in fine granularity to realize the intelligent and accurate search of the technical stan-
dard documents of electric power enterprises. The experimental results of the proposed
method are compared with those of other methods in terms of evaluation index by using
the real-world data set of power enterprises’ technical standard documents. Experimen-
tal results show that the accuracy and recall of the proposed method are better than other
comparison methods, which indicates that the combination of knowledge graph search
in the Elasticsearch engine can achieve a more accurate search.
The organizational structure of this article is as follows. Section 2 introduces the
related work, and Section 3 describes the proposed method in detail. In Section 4, exper-
imental setup, data set and experimental environment, evaluation indexes, and exper-
imental results obtained by the proposed method. Section 5 introduces the functional
logic structure and function of the prototype knowledge retrieval system of the power
grid dispatching system. Section 6 summarizes this paper and looks into the future work.
2 Related Work
This chapter introduces the relevant work based on Elasticsearch and Knowledge Graph
search methods and explains the differences between the proposed method and the related
work.
power enterpri, Yang [5] et al. used Elasticsearch search engine, the development tech-
nology of front, rear end separation, and the design pattern of microservice architecture
to establish a specific data application aiming at the difficulty of locating the target in
massive data, and the results showed that this method could effectively improve the work
efficiency. In campus networks, Xue [6] and Qin [7] et al. took full-text searches as the
research problem. They proposed combining Elasticsearch full-text search engine with
Scrapy framework and optimizing Elasticsearch Chinese index, respectively. Experi-
ments show that applying the Elasticsearch engine in the campus network to search the
full text can get better results. The Elasticsearch search engine is also widely recognized
in other areas. Aleksei, V. et al. [8] proposed a system based on the Elasticsearch engine
and MapReduce model to solve the problem of user authentication. Prashant, K. et al.
[9] applied the Elasticsearch engine to the search application of the e-book portal.
In the field of the smart grid search, search results are required to be more accurate.
Compared with the related work, the proposed method does not rely on Elasticsearch
alone but only uses the ES method to prefilter the relevant paragraph content. This
approach is expected to improve the accuracy of search results.
combines the Elasticsearch search method with the knowledge graph search method,
which can quickly and accurately locate the target content in the standard documents of
electric power enterprises. It can not only browse large-scale text data comprehensively
and quickly in the full text of standard documents of electric power enterprises, but also
make up for the low search precision of single target words by traditional methods, and
improve the precision and efficiency of accurate search of standard documents. The steps
for this method are described below.
Step 1: Building a Knowledge Graph. In the unit of paragraph content, the enterprise
technical standard document is split. Each paragraph’s content is indexed and parsed
into a knowledge graph representation. According to the corresponding content of the
technical standard documents of electric power enterprises, the word segmentation algo-
rithm of POS (Part-of-speech) [17] and part of speech tagging algorithm are firstly used
for word segmentation and tagging processing of the text content of each paragraph.
Secondly, a regular extraction method is used to extract numbers and units. Finally,
the dependency syntax analysis method is used to identify the dependency relations of
context relations, analyze and extract the data of different power grid business fields,
and build the data knowledge map of power enterprises with the extracted entities and
relations.
The knowledge graph elements corresponding to the content of the construction
paragraphs are composed of triples, as shown in Table 1. Each row is represented as a
triple, each triple consists of an entity, a relationship, and an entity or attribute value.
The triple ID of 2010000 is a triad (scope, contains, Excludes certain small and special
transformers).
Step 2: Paragraph Segmentation. In this process, a power domain word divider con-
verts each text into a series of words. The word segmentation consists of three com-
ponents, namely Character Filter, Tokenizer, and Token Filter. The Character Filter
component removes the HTML tags for each triple, The Tokenizer component divides
words according to the rules of the word splitter, The Token Filter modifies the shard
words into the lowercase format, deletes stop words and adds synonyms. After being
processed by three components, the paragraph content is broken down into words.
For example, the entity in the triple is “excluding certain small and special transform-
ers.“. If the entity contains HTML markup, the Character Filter component removes it.
The result of the Tokenizer component’s word segmentation is: “Excludes certain small
and special transformers.“ The result after processing by the Token Filter component is
“excluding small special transformers”.
Step 3: Build the Elasticsearch Inverted Index File. Inverted indexes are relationships
that lookup document IDs by words [18]. You can search the corresponding paragraph
content in the document of enterprise technical standard by searching the keywords in
the electric power field. The inverted index measures four statistical indexes of enter-
prise technical standard documents, including Doc id, word frequency TF, Position, and
Offset. For example, the sentence after paragraph segmentation is “ Excludes small spe-
cial transformers”. The inverted index statistical index of the word “small” is shown in
Table 2.
858 J. Wang et al.
Step 4: Search for Paragraph Content Based on Elasticsearch. Enter the question
into the search service, and word the question by paragraph segmentation. The result of
the word segmentation is returned to the retrieval service to find the matching enterprise
technical standard document ID in the inverted index file. According to the similarity
measurement method of term frequency-inverse document frequency (TD/IDF) [19], the
relevance score of each word in the problem is calculated. Then the relevance score of the
whole problem is obtained after adding them together. Finally, all enterprise technical
standard documents are sorted according to their relevance from high to low.
Step 5. Accurate Search Based on the Knowledge Graph. To search the target answer
more accurately, within the range of the search results of paragraph content, an accurate
search was carried out by combining the data knowledge graph of electric power enter-
prises. After information extraction, entity link and knowledge fusion, the problem is
represented as a problem triad. The problem triad is accurately matched with the power
enterprise data knowledge graph to achieve the accurate search of the target answer.
Intelligent Search Method in Power Grid 859
For example, the question “What should the thickness of the oil passage of the
radiator be less than?” The thickness of the oil passage of the radiator should not be less
than 9 mm. The distance between plates of the self-cooling type should not be less than
45 mm”, The data knowledge graph of power enterprises is expressed as (oil channel
thickness, less than, 9 mm) and (distance between slices, less than, 45 mm). According
to the CYPHER retrieval statement of the knowledge graph, the correct answer was
9 mm.
4 Experiment
In this chapter, we verify the effectiveness of our proposed method through experiments.
The experimental data set, experimental setting, experimental environment and exper-
imental evaluation are introduced. Finally, the experimental and analytical results are
given.
In the experiment, the proposed method is verified by electric power enterprises’ real-
world technical standard document dataset. Electric Power Enterprise Technical Stan-
dard Document (EPETSD) data set comprises 200 documents of national standards,
Enterprise standards, and rules and regulations of the Electric Power Industry. Each
standard technical document of electric power enterprise contains several paragraphs,
and the sample data set after splitting each standard technical document of electric power
enterprise is shown in Table 3. Table 3 contains several paragraphs of a document, and
the corresponding split inverted index and knowledge graph triples.
Experimental Setup. All data sets are taken as validation data sets, and the proposed
method is run on the same data set with other comparison methods. The search results of
all the methods were compared with the real content, and the values of all the methods
in the evaluation indexes were calculated. Other comparison methods include traditional
Elasticsearch based on Keyword search and knowledge graph based on semantic search
(KGSS) [20].
To ensure the stability of the search results, and the contingency of results can be
avoided. We perform 10-fold cross-validation [21] of all the methods.
Table 3. The contents of several paragraphs of a document and the corresponding split inverted
index and knowledge graph triples.
Precision and recall were used to evaluate the effectiveness of the proposed method and
compared with the evaluation indexes of other methods.
Precision is the ratio of the number of real answers to the retrieved questions to the
number of all retrieved paragraphs.
1 Ti
n
Precision = (1)
n Ri
i=1
Ti is the number of real answers to the retrieved questions, Ri is the number of all
retrieved paragraphs, and n is the number of all problems.
The recall is the ratio of the number of real answers retrieved to the number of
relevant paragraphs.
1 Ti
n
Recall = (2)
n AN
i=1
We compare our proposed method SEK with ESKS, KGSS on the evaluation metrics. The
experimental results are shown in Fig. 1, which indicates that the proposed method SEK
Intelligent Search Method in Power Grid 861
has better experimental results in terms of precision and recall. This is mainly because of
the combination of Elasticsearch search and knowledge graph search method, which can
select the contents of the paragraphs in advance and the finer-grained search scope. In the
small range of paragraph content, triples in the knowledge graph can match the problem
in fine granularity, which improves the precision. Therefore, the method proposed in this
paper effectively searches problems in the data set of technical standard documents of
electric power enterprises.
This chapter introduces the functional logic structure and functional module description
of the prototype system for knowledge search of power grid dispatching system.
The functional logic structure of the system can be divided into three layers: text
parse layer, data storage layer, and application layer from bottom to top. In the text
parse layer, paragraph segmentation is carried out for the dispatching documents and
technical documents of the power grid enterprise. Then the Elasticsearch inverted index
and knowledge graph are constructed, and these data output to the data storage layer.
In the data storage layer, the Neo4j graph and MySQL relational databases are used to
store knowledge graphs and Elasticsearch inverted index. When the question is entered,
the system will query the data in the database and output the relevant paragraph content ID
and knowledge graph triples to the stored content. Data from the stored content is output
to the application layer for display. The application layer includes information retrieval,
accurate question and answer of knowledge graph and professional recommendation
functions. The functional logic structure of the system is shown in Fig. 2.
The information retrieval function of the application layer is used to return documents
related to the user’s search problem. The exact question and answer function of the
Knowledge Graph can return the answer of the question more accurately within the
scope of relevant documents. The answer is in a specific string and number, rather
862 J. Wang et al.
Store content
Fig. 3. The main interface of the prototype system for knowledge search of power grid dispatching
system.
than the most relevant paragraphs. The professional recommendation function is used
to recommend documents that are associated with the search question. It is important
to note that the recommendation document is not the same as the paragraph content
returned by the information retrieval but is related to the paragraph content in some
other way.
In Fig. 3, the problem is that the thickness of the radiator walkway should not be less
than what. The information retrieval returned three paragraphs, and the exact question
and answer of the knowledge graph returned 9 mm. The knowledge graph showed the
Intelligent Search Method in Power Grid 863
References
1. Guo, R., Yang, Q., Liu, S.H., Li, W., Yuan, X., Huang, X.H.: Research and application on
construction of knowledge graph of power system fault handling. Power System Technol.
45(6), 2092–2100 (2021)
2. Qiao, J., Wang, X.Y., Min, R., et al.: Framework and key technologies of knowledge-graph-
based fault handling system in power grid. Proc. CSEE 40(18), 5837–5848 (2020)
3. Gao, H.X., Lu, M., Liu, J.N., et al.: Review on knowledge graph and its application in power
systems. Guangdong Electric Power 33(9), 66–76 (2020)
4. Quan, L.X., Ma, P.: Talk about a kind of intelligent search engine technology for big data
industry. Comput. Technol. Autom. 39(2), 170–176 (2020)
5. Yang, Q.: Research and practice of innovative data application based on elasticsearch in
Tianwan nuclear power plant. Power Big Data 24(2), 41–46 (2021)
6. Xue, T., Zhang, T.C., Zhuang, X.F., He, X.: Research and implementation of campus network
search engine based on scrapy framework and elasticsearch. Northeastern University, Editorial
Department of Control and Decision, 6 (2020)
7. Qin, J.C., Shen, H.L.: Research and implementation of elasticsearch based full-text search
platform on campus. Mod. Comput. Prof. Ed. 34, 96–100 (2018)
8. Aleksei, V., Aleksei, S., Shamil, M.I.: Big data processing for full-text search and visualization
with elasticsearch. Int. J. Adv. Comput. Sci. Appl. (IJACSA) 8(5) (2017)
9. Prashant, K., Aishwarya, S., Komal, D., Manaswini, M., Meghna, M.: Search engine for
Ebook portal. Int. J. Sci. Technol. Res. 4(8) (2014)
10. Liu, Q., Li, Y., Duan, H., Liu, Y., Qin, Z.G.: A review of knowledge graph construction. J.
Comput. Res. Dev. 53(3), 582–600 (2016)
11. Xu, H., Hong, Q., Yao, X.M., Li, X.L., Lu, S.Y.: Semantic search algorithm based on
knowledge graph in smart grid. Lab Res. Explor. 40(4), 71–74+86 (2021)
12. Wang, Q., Wei, J., Yan, R.Z., Qian, X.D., Yang, B.: Application of knowledge graph in smart
grid. Electron. Compon. Inf. Technol. 4(1), 135–137+147 (2020)
13. Liu, J., et al.: Application and research of knowledge graph in electric power field. Electric
Power Inf. Commun. Technol. 18(1), 60–66 (2020)
864 J. Wang et al.
14. Wang, M., Chen, W., Wang, S., Jiang, Y., Yao, L., Qi, G.: Efficient search over incomplete
knowledge graphs in binarized embedding space. Future Gener. Comput. Syst. 123, 24–34
(2021). https://doi.org/10.1016/j.future.2021.04.006
15. Xin, H., Duan, J., Dang, D.: Natural language question answering over knowledge graph: the
marriage of SPARQL query and keyword search. Knowl. Inf. Syst. 63(4), 819–844 (2021).
https://doi.org/10.1007/s10115-020-01534-4
16. Maulik, K., et al.: Text snippets to corroborate medical relations: an unsupervised approach
using a knowledge graph and embeddings. In: AMIA Joint Summits on Translational Science
proceedings 2020, vol. 5 (2020).
17. Mou, Q.L.: Research and Application of Deep Transfer Learning Algorithm for Text
Classification Problem. Northwest Normal University (2020)
18. Qu, Z.J., Fan, M.M., Zhou, R.L., Wang, H.L., Zhu, D.: Inverse index query technology of non-
primary row key for mass distribution network dispatching monitoring information. Power
Syst. Prot. Control 46(23), 162–168 (2018)
19. Fu, T.Y.: Application of Semantic Relevance in Patent Text. Jiangsu University of Science
and Technology (2020)
20. Xu, H., Hong, Q., Yao, X.M., Li, X.L., Lu, S.Y.: Semantic search algorithm based on
knowledge graph in smart grid. Lab. Res. Explor. 40(4), 71–74+86 (2021)
21. Liang, Z.C., Li, Z.W., Lai, K., Lin, Z.C., Li, T.G., Zhang, J.X.: Evaluation of general-
ization capability of predictive models using 10-fold cross validation and its R software
implementation. Chin. J. Hosp. Stat. 27(4), 289–292 (2020)
Longitudinal Guidance Control of Landing
Signal Officer Based on Variable Universe
Fuzzy Logic
Ming Zhao1,2 , Yang Liu1 , Hui Li1,2 , Yun Cao1(B) , Jian Xu1,2 , and Guilin Yao1,2
1 Harbin University of Commerce, Harbin 150028, China
hrbcu_lh@163.com
2 Heilongjiang Provincial Key Laboratory of Electronic Commerce and Information Processing,
1 Introduction
Landing Signal Officer (LSO) guidance and evaluation technology play an important
role in ensuring carrier-based aircraft’s landing safety [1–3]. Due to the complexity of
the lending environment, the artificial landing of aircraft needs LSO guidance and super-
vision. The success rate of landing could be increased significantly, and the incidence of
accidents could be reduced with control of LSO [4, 5]. As a research object, LSO guid-
ance technology, wave-off decision technology, and evaluation technology are analyzed
and researched widely at present [6–8].
In order to analyze the influencing factors of carrier-based aircraft landing safety
comprehensively, after discussing the landing guidance characteristics of LSO, the guid-
ance prediction models of the longitudinal loop would be established, respectively.
Based on variable universe fuzzy logic, a coupling fuzzy guidance model is imple-
mented according to the priority principle, and an LSO landing guidance control system
is realized.
Very High
High
A Little High
OK
A Little Low
Low
Very Low
Desired touchdown
Carrier
The ideal landing longitudinal sink rate is constant. However, there is pitch, yaw and roll
motion during the landing process; the pilot sometimes changes the attitude by altering
the sink rate or throttle.
When the sink rate remains too low, the aircraft’s position could gradually higher than
the ideal glideslope, and the wave-off maneuver would be employed to ensure landing
safety. If the sink rate continues to be higher, the aircraft may be below the glideslope,
and there is a risk of falling into the sea or the ramp, which should be avoided.
Longitudinal Guidance Control of Landing Signal Officer 867
Predicion
information Coupling fuzzy Longitudinal fuzzy
guidance system guidance system
The initial input universe at k = 0 moment of controller is Ui0 = [−ηCIi , CIi ], and
the initial fuzzy sets are A0ij (j = 1, 2, ..., m) in Ui0 , and the initial output universe is
V 0 = [−ξ YI , YI ].
At k moment, the center of Akij is xkij , −ηCIik = xki1 < xki2 < ... < xkimi < CIik , and
the center of Bjk is ykj , and −ξ YIk = yk1 < yk2 < ... < ykq < YIk , in the meantime, the
relationship between Akiri (ri = 1, 2, ..., ni ) and Akij (j = 1, 2, ..., m) is shown as (3), the
input and output universes change to
Through the effect of scaling factor, the universe shrinks with the decrease of vari-
ables. The controller that changes universe on-line is called variable universe fuzzy
controller, and the change process is shown in Fig. 3 [13–17].
(1) When the longitudinal input deviation is larger, the output signal should satisfy the
purpose of reducing the deviation as soon as possible. Especially there is a great
upward longitudinal deviation of carrier-based aircraft. The LSO should send the
“NB” instruction of downward direction to the pilot.
(2) When the longitudinal input deviation is smaller, based on considering the elimi-
nation of deviations, the system’s stability should be maintained as far as possible
to avoid excessive overshoot and oscillation.
1.5
Glideslope
1
0.5
Deviation
-0.5
-1
Sink rate
-1.5
0 1 2 3 4 5 6 7 8 9 10
t/s
(a) Input signals of longitudinal system
2
1.5
1 Longitudinal
0.5 instruction
Amplitude
0
-0.5
-1
-1.5
-2
0 1 2 3 4 5 6 7 8 9 10
t/s
is [−3.5, 3.5], and the initial range of drift rate deviation is [−1, 1]. The sinusoidal signal
with a period of 6.3 s of centering deviation and the signal with a period of 3.2 s of drift
rate deviation is input into the lateral controller. The output curve is shown in Fig. 5.
4
Centering
3
2
1
Deviation
0
-1
Drift rate
-2
-3
-4
0 1 2 3 4 5 6 7 8 9 10
t/s
(a) Input signals of lateral system
4
3
2
Lateral
Amplitude
1 Instruction
0
-1
-2
-3
-4
0 1 2 3 4 5 6 7 8 9 10
t/s
(b) Output signals of lateral system
Fig. 5. Input/output curves of lateral system
6 Conclusions
In the landing process of carrier-based aircraft, based on the characteristics of LSO
deviation instructions, the glideslope deviation, and sink rate deviation of the longitudinal
loop are determined as the influencing factors of carrier-based aircraft landing safety.
The deviations above are input into the landing guidance model of LSO and an intelligent
guidance prediction system fuzzy intelligent guidance system based on variable universe
fuzzy logic. Simulation results show that the guidance prediction model established by
fuzzy logic conforms to the operating characteristics of the true LSO, and the output
872 M. Zhao et al.
results accord with the correction of deviation effect under the real environment. In
the meantime, establishing the system model also provides an effective solution for the
objects with uncertainty, nonlinearity, and task environment complexity, especially for
the object is an individual.
References
1. Rudowsky, T., Cook, S., Hynes, M.: Review of The Carrier Approach Criteria for Carrier-
Based Aircraft. Techinical Report NAWCADPAX/TR-2002/71, 52–58, 124–165, USA (2002)
2. Li, H., Jiang, H.T., Su, X.: Modeling landing signal officer instruction associated with
operation guide system. Int. J. Control Autom. 8(2), 373–382 (2015)
3. Wang, J., Wu, W., Jia, L.: Study and simulation analysis on wave-off capability of carrier-based
airplane. Aircr. Des. 30(4), 13–18 (2010)
4. Duan, Z., Wang, W., Geng, J., He, D.: Precision trajectory manual control technologies for
carrier-based aircraft approaching and landing. Acta Aeronaut. Astronaut. Sin. 40(4), 22328
(2019)
5. Zhu, Q., Yang, Z.: Dynamic recurrent fuzzy neural network-based adaptive sliding control
for longitudinal automatic carrier landing system. J. Intell. Fuzzy Syst. 37(1), 53–62 (2019)
6. Zuo, Z., Wang, L., Liu, H., Wang, Y.: Similarity for simulating automatic carrier landing
process of full-scale aircraft with scaled-model. Acta Aeronaut. Astronaut. Sin. 40(12), (2019)
7. Zhou, J., Jiang, J., Yu, C., Xiao, D.: Carrier aircraft dynamic inversion landing control based
on improved neural network. J. Harbin Eng. Univ. 39(10), 1649–1654 (2018)
8. Hess, R.: Simplified approach for modeling pilot pursuit control behaviour in multi-loop flight
control task. Inst. Mech. Eng. 220(2), 85–102 (2006)
9. Wang, L., Zhu, Q., Zhang, Z., Dong, R.: Modeling pilot behaviors based on discrete-time
series during carrier-based aircraft landing. J. Aircr. 53(6), 1922–1931 (2016)
10. Li, H.: Modeling landing signal officer instruction associated with operation guide system.
Int. J. Control Autom. 8(2), 373–382 (2016)
11. Shi, M., Cui, H., Qu, X.: Modeling landing signal officer for carrier approach. J. Beijing Univ.
Aeronaut. Astronaut. 32(2), 135–138 (2016)
12. Li, H.: Integrated evaluation technology of landing signal officer for carrier-based aircraft.
Int. J. Multimedia Ubiquit. Eng. 11(1), 169–178 (2016)
13. Shi, P., Xu, Z., Wang, S.: Variable universe adaptive fuzzy PID control of active supension.
Mech. Sci. Technol. Aerospace Eng. 38(5), 713–720 (2019)
14. Du, E., Wang, S., Chang, L.: Variable universe fuzzy controller design of missile mark
trajectory with feed-forward compensation. J. Acad. Armored Force Eng. 31(2), 84–89 (2017)
15. Yang, Z., Wang, H.: Maximum power point tracking for photovoltaic power system based on
asymmetric fuzzy control. Mech. Autom. 41(2), 153–156 (2012)
16. Liu, J., Zhang, Y.: Variable universe fuzzy PID control method in piezoelectric ceramic
precision displacement system. Autom. Instrum. 32(2), 45–49 (2017)
17. Li, D., Shi, Z., Li, Y.: Sufficient and necessary nonditions for Boolean fuzzy systems as
universal approximators. Inf. Sci. 178(2), 14–24 (2008)
Author Index
A G
Abdulghany, Nayera M., 796 Gao, Mengyu, 643
Alieksieiev, Volodymyr, 594 Geng, Wenli, 643
Gerguis, Ramy A., 796
Guo, Min, 152, 181, 214
B
Bai, Yu, 169 H
Bi, Huolong, 653 Han, Ping, 3, 61
Bieliatynskyi, Andrii, 826 Han, Tongtong, 152, 181, 214
Han, Wang, 70
Han, Xuena, 344, 441
C Han, Yong, 141
Cao, Yun, 482, 865 He, Kaiwen, 527
Heiden, Bernhard, 594
Chen, Ming, 537
Hong, Peng, 247, 321
Hu, Liangjin, 461
Hu, Wen, 537
D
Hu, Xueyong, 276, 854
Daizhi, Jin, 70
Huadong, Sun, 366, 392
Deyu, Chen, 761
Huang, Chongli, 15, 288
Dong, Xiaohong, 547, 567 Huang, Yao-yao, 661
Du, Yong, 247, 321
J
Jiang, Fan, 15, 288
E Jiang, Shu-bo, 88, 661
El-Telbany, Mohammed, 796 Jiang, Zaiyu, 276
Jianjun, Li, 786
Jin, Hao, 482
F Jin, Yining, 276, 844, 854
Fan, Jingtao, 257, 501 Jincheng, Su, 754
Fan, Ying, 267 Jing, Fan, 160
Fan, Zhipeng, 537 Jinpeng, Jiao, 98
Fang, Yang, 786 Junji, Liu, 51
© The Editor(s) (if applicable) and The Author(s), under exclusive license
to Springer Nature Switzerland AG 2022
A. E. Hassanien et al. (Eds.): BIIT 2021, LNDECT 107, pp. 873–875, 2022.
https://doi.org/10.1007/978-3-030-92632-8
874 Author Index
K R
Kang, Chengwen, 703 Ren, Wanli, 78
Kewen, Liu, 51 Rong, Xin, 557
Kong, Leilei, 141
S
L Shen, Yang, 190
Lei, Shuying, 703 Song, Rong, 401
Li, Ang, 511 Song, Yi, 653
Li, Dengping, 734 Stepanchuk, Oleksandr, 826
Li, Hui, 234, 309, 441, 482, 865 Su, Fangyuan, 670, 775
Li, Jianjun, 557, 670, 775 Su, Xiaodong, 332, 379
Li, Jiao, 61 Sun, Huadong, 415
Li, Juan, 613 Sun, Jianming, 344
Li, Lin, 25 Sun, Le, 854
Li, Shizhou, 332, 356, 379 Sun, Ming, 461
Li, Yanlai, 106 Sun, Xu, 141
Li, Zeming, 234, 441 Sun, Yutong, 141
Liang, Hongyu, 332, 379
Lin, Hai, 844 T
Liu, Junjun, 670, 775 Tan, Xue, 234, 309, 441
Liu, Kewen, 106 Tang, Meng, 257, 501
Liu, Qiuming, 577 Tang, Zheqing, 247
Liu, Wei, 537 Tian, Linlin, 689
Liu, Xiangbin, 577 Tian, Shaoqing, 15, 288
Liu, Xing, 836 Tonino-Heiden, Bianca, 594
Liu, Yang, 482, 865
Liye, Sun, 805 W
Lu, Shangkun, 681 Wang, Chaoyu, 379
Wang, Hongxin, 309
Wang, Jiaqiu, 276, 461, 844, 854
M
Wang, Jiaying, 511
Ma, Jian, 415
Wang, Peng, 276
Ma, Ling, 301
Wang, Ruijie, 844
Ma, Zhiwei, 429
Wang, Shiyu, 247
Maged, Maha A., 796
Wang, Tao, 131
Miao, Xiufeng, 715
Wang, Wenjing, 152, 181, 214
Mou, Hao, 844
Wang, Xiaoyu, 35
Wang, Yamin, 471
N Wang, Yanchun, 461
Newagy, Fatma, 796 Wang, Yang, 321
Ning, Hui, 141 Wang, Yu, 78, 247, 321, 450
Ning, Shi Yong, 206 Wei, SiFan, 450
Ning, Shiyong, 152, 181, 214 Wei, Sifan, 78
Niu, Chaoqun, 703 Wei, Tianshi, 247
Wen, Yan, 527, 734
P Wu, Jinpeng, 344
Pan, Wei, 321 Wu, Shirui, 119, 332, 356
Pang, Shi-kun, 605
Peng, Ezhen, 744 X
Pengfei, Zhao, 366, 392 Xiao, Yuanhao, 567
Pylypenko, Oleksandr, 826 Xiaodi, Xu, 786
Xie, Hai, 836
Q Xie, Jie, 689
Qiu, Zeguo, 511 Xin, HaiTao, 491
Qu, Yi, 653, 836 Xin, Haitao, 744
Author Index 875