Professional Documents
Culture Documents
The Generalization Negation of Probability Distribution and Its Application in Target Recognition Based On Sensor Fusion
The Generalization Negation of Probability Distribution and Its Application in Target Recognition Based On Sensor Fusion
Abstract
Target recognition in uncertain environments is a hot issue. Fusion rules are used to combine the sensor reports from dif-
ferent sources. In this situation, obtaining more information to make correct decision is an essential issue. Probability distri-
bution is one of the most used methods to represent uncertainty information. In addition, the negation of probability
distribution provides a new view to represent the uncertainty information. In this article, the existing negation of probability
distribution is extended with Tsallis entropy. The main reason is that different systems have different parameter q. Some
numerical examples are used to demonstrate the efficiency of the proposed method. Besides, the article also discusses the
application of negation in target recognition based on sensor fusion to further demonstrate the importance of negation.
Keywords
Probability distribution, negation, Tsallis entropy, Gini entropy, sensor fusion
Creative Commons CC BY: This article is distributed under the terms of the Creative Commons Attribution 4.0 License
(http://www.creativecommons.org/licenses/by/4.0/) which permits any use, reproduction and distribution of the work without
further permission provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/
open-access-at-sage).
2 International Journal of Distributed Sensor Networks
X X
when q ! 1, which corresponds to the Boltzmann- H1 (P) =
H1 (P) (1 Pi )ln(1 Pi ) + Pi lnPi
Gibbs entropy, namely !
XN
X H1 (P) + l
D0 = H1 (P) Pi 1
H(P) = H1 (P) = Pi lnPi ð4Þ i=1
i=1 H1 (P)
∂H1 (P) 1 1 Pi
= lnPi + 3 ln +l
One can rewrite Tsallis entropy as follows29 ∂Pi n1 n1
ø H(P)
H(P)
k XW
Hq = Pi (1 Pq1
i ) ð5Þ
q1 1 From the above calculation, it can be seen that
Tsallis entropy can increase after negation.
where q is a real parameter sometimes called entropic-
index which generalizes the usual exponential and plays
the important role in description of thermodynamic. k Property 2. There is maximum Tsallis entropy with uni-
is the Boltzmann constant (or some other convenient form distribution after multiple negations.
value, in areas outside physics, such as information the- The general formula of the negation method is as
ory, cybernetics and others).28,29 Besides, when q = 2, follows
the Tsallis entropy degenerates the Gini entropy.
1 n 3 x1 1
xi + 1 = + ð7Þ
n n 3 (n 1)i1
The proposed method
where xi + 1 represents the ith negation. After multiple
From the above, it can be seen that Yager’s method negations, the probability distribution is as follows
would be the special case. In order to expand the appli-
cation of negation, the article proposed a method that 1
lim xi + 1 =
uses Tsallis entropy to measure the uncertainty of i!‘ n
NPD. Using Tsallis entropy to measure the uncertainty
From the above, it can be seen that Tsallis entropy
of NPD can expand the application of negation
reaches the maximum after multiple negations.
Besides, in nature, any operation can cause energy
1 X 1 Pi 1 Pi q1
Hq = 1( ) ð6Þ consumption. The change of entropy means the con-
q1 n1 n1 sumption of energy, and consumed energy cannot be
In the following part, the property of negation-based reused. Similarly, the variation of information entropy
proposed method can be discussed. is accompanied by the consumption of information. Of
various entropies, Tsallis entropy is non-extensive
entropy and applied in artificial and social complex sys-
Property 1. Tsallis entropy can increase after negation. tems. Hence, using Tsallis entropy to measure uncer-
Specific proof is as follows tainty of NPD can expand the applications of negation.
1 X q
Hq (P) = 3 (1 (Pi ))
q1 Examples and discussion
X q
= 1 3 (1
Hq (P) i ))
(P Example
q1
Example 1. Given the event space X = fx1 , x2 , x3 g,
when q 6¼ 1 P(x1 ) = 0:2, P(x2 ) = 0:5, P(x3 ) = 0:3, the NPD can be
1 X q X 1 Pi q obtained as follows
Hq (P) =
Hq (P) 3( Pi )
q1 n1 1 ) = 0:4, P(x
P(x 2 ) = 0:25, P(x
3 ) = 0:35
!
XN
Hq (P) + l
D0 = Hq (P) Pi 1 Comparing P(xi ) with its negation P(x i ), it can be
i=1 seen that the event xi which attaches to lower probabil-
q
∂D0 q q 1 ity acquires higher probability after negation process.
= q1
p + ð1 Pi Þq1 + l i)
∂Pi q1 q1 n1 Attractively, what happens if negation is taken to P(x
Hq (P) ø 0
Hq (P) ? Second negation is made and the results are obtained
as follows
when q = 1
P(x ) = 0:375, P(x
) = 0:3, P(x ) = 0:325
1 2 3
4 International Journal of Distributed Sensor Networks
) 6¼ P(x )
P(x i i
Further discussion Figure 1. The trend of Tsallis entropy after each iteration of
negation process.
Again considering the Example 1, the Table 1 is used
to show the change of probability after each iteration
of negation process. It can be clearly known that prob-
ability is reallocated after negation. Gradually, the Table 1. The change of probability after each iteration of
probability distribution becomes more and more close negation process.
to the uniform distribution. What is the cause of this
phenomenon? Frequency of negation P(x1 ) P(x2 ) P(x3 )
The concept of entropy is derived from physics.47,48
0 0.2000 0.5000 0.3000
With the increasing application of entropy,49,50 infor-
1 0.4000 0.2500 0.3500
mation entropy has become an indispensable part of 2 0.3000 0.3750 0.3250
modern scientific development.51,52 Shannon53 first 3 0.3500 0.3125 0.3375
proposed the concept of information entropy to 4 0.3250 0.3437 0.3313
describe the uncertainty and has applied in a lots of 5 0.3375 0.3281 0.3344
6 0.3310 0.3360 0.3330
fields.54,55 However, as in typical physical problems, 7 0.3345 0.3320 0.3335
there are some examples where the Boltzmann- 8 0.3330 0.3340 0.3330
Shannon entropy is not suitable.56,57 In 1988, Tsallis 9 0.3333 0.3332 0.3335
proposed a non-extensive entropy called Tsallis 10 0.3333 0.3333 0.3334
entropy. Subsequently, non-extensive statistical
mechanics which is Generalization of Boltzmann-
Gibbs Statistics emerged based on Tsallis entropy.
gradually in the process of negation and is almost
More importantly, the Boltzmann-Gibbs statistics is
invariable after the fifth negation, regardless of the
recovered as the limitation when q ! 1.58 The value of
value of q. It is clearly shown that the uncertainty gra-
q is hidden in the microscopic dynamics of the system.
dually increases in any system after each iteration of
It can be obtained by some experiments. A relevant
negation process. When there are multiple iterations of
improvement on the inference accuracy by adopting
negation process, the probability distribution would be
non-extensive entropies is proposed by Tsallis,57 where
uniform distribution and Tsallis entropy has maximum
q equals 2.5 obtains a better result. Besides, in
entropy, which is consistent with Property 1 and
article,59q = 2:5 is used to obtain mutual information
Property 2. Hence, using Tsallis entropy to measure
during DREAM4 data. As a consequence, using Tsallis
uncertainty of NPD is reasonable.
entropy to measure the uncertainty extended the
method of Yager.27 Therefore, it is of great significance
to measure uncertainty of negation with Tsallis Application of negation based on sensor
entropy.
fusion
Let’s calculate the uncertainty using Tsallis entropy
with different q to observe how Tsallis entropy change Target recognition is paid great attention in military
after negation, as shown in Figure 1. There are changes applications.60 There are some methods to make deci-
of Tsallis entropy with five different q from 12 to 3. It sions.61–63 Fusion rules can help us make better deci-
can be easily seen that the Tsallis entropy increases sions.64,65 However, using existing information to make
Gao and Deng 5
more accurate decision is also an open issue. The nega- From the above, it can be seen that the Tsallis
tion provides a new view to obtain more accurate jud- entropy after fusion becomes less, showing that the
gement with the collected information. The specific fusion can decrease the uncertainty of information.
example is as follows. However, fusion rules don’t work in extreme prob-
There are three sensors to recognize the target which ability distribution. Using negation can provide another
maybe A, B and C. The results from the sensors are as view to analyse this phenomenon.
follows There is a special example to better explain the appli-
cation of negation. There are two sensors to recognize
p1 (A) = 0:7, p1 (B) = 0:2, p1 (C) = 0:1 the targets which maybe A, B and C. The results from
p2 (A) = 0:6, p2 (B) = 0:3, p2 (C) = 0:1 the sensors are as follows
p3 (A) = 0:5, p3 (B) = 0:3, p3 (C) = 0:2
p1 (A) = 0:8, p1 (B) = 0, p1 (C) = 0:2
64 p2 (A) = 0, p2 (B) = 0:8, p3 (C) = 0:2
Dempster rule is widely used in sensor data fusion.
The combination results are shown as follows p2 (A) = 0, p2 (B) = 0:8, p3 (C) = 0:2
K = 0:77, p(A) = 0:91, p(B) = 0:078, p(C) = 0:087 Using fusion rules, one can get the results as follows
Using the method of negation, new information is p(A) = 0, p(B) = 0, p(C) = 1
obtained as follows
From the result, it can be seen that the target must
p1 (A) = 0:15,
p1 (B) = 0:4,
p1 (C) = 0:45 be C as p(C) = 1. Obviously, the result is not very reli-
2 (A) = 0:2, p
p 2 (B) = 0:35, p
2 (C) = 0:45 able. Besides, it can be seen that when evidence is highly
conflicting, Dempster rule doesn’t work. However,
3 (A) = 0:25, p
p 3 (B) = 0:35, p
3 (C) = 0:4 negation can provide another information, which is
useful to make decision
Similarly, fusion rule is used as follows
KN = 0:8625,
p(A) = 0:055; ,
p(B) = 0:356,
p(C) = 0:375 p1 (A) = 0:1, p1 (B) = 0:5, p1 (C) = 0:4
p2 (A) = 0:5, p2 (B) = 0:1, p2 (C) = 0:4
It is well known that conflicting coefficient K plays The fusion results are as follows
an essential role in information fusion.8 As a result,
analysing the difference of K between the original prob- p(A) = 0:19, p(B) = 0:19, p(C) = 0:62
ability and its corresponding negation is necessary. It is
easily found that the conflict after negation becomes Next, considering the changes of entropy between
bigger. The reason is that p(A) = 1 p(A) represents the original and negation, assume that q = 2 as there
the probability the target is not A, namely p(A) is the are two sensors
probability that target is B or C. Besides, the method
of negation contains the uncertain class (AorB). Hence, T1 = 0:4800, T2 = 0:4800, T = 0
the conflict becomes bigger due to the existing uncer- T 1 = 0:8100, T 2 = 0:8100, T = 0:7480
tain class after negation.
Obviously, negation gives us some new information.
More importantly, by comparing the results between
From the above result, the probability that target is not
the original and negation, the decision has more sup-
C is 0.62. It can be seen that the result of negation is
port to target A. p(A) = 0:055 reflects the probability
more reasonable than the original result. More impor-
that the target is not A. That is to say, it can increase
tantly, negation provides another view to analyse prob-
the probability of target A from the other side. Hence,
lem and can better handle conflicting evidence. Hence,
it can get two results from the two sides based on data,
negation is useful to make decision.
which is better to make a more reasonable decision in
In addition, analysing the Tsallis entropy between
target recognition.
original and negation, it can be found that Tsallis
Next, considering the changes of entropy between
entropy becomes bigger. Besides, according to the sec-
the original and negation, assume that q = 3 as there
ond law of thermodynamics, the entropy of an isolated
are three sensors
system never decreases. The negation can increase
Tsallis entropy. Moreover, entropy is irreversible.
T1 = 0:324, T2 = 0:378, T3 = 0:42, T = 0:123
From this view, the negation can make systems sponta-
T 1 = 0:421, T 2 = 0:429, T 3 = 0:439, T = 0:375 neously evolve towards thermodynamic equilibrium,
6 International Journal of Distributed Sensor Networks
the state with maximum entropy. That is to say, nega- article: The work is partially supported by National Natural
tion tends to make the system uniformly distributed, Science Foundation of China (Grant Nos 61573290,
and in fact it does. Hence, the negation not only pro- 61503237).
vides a new way to understand problems, but also pro-
vides better performance of decision support system. ORCID iD
Besides, because the Tsallis entropy is highly applied in Yong Deng https://orcid.org/0000-0001-9286-2123
many applications, negation provides a new view to
obtain information. Hence, using Tsallis entropy to
measure the uncertainty can enlarge the application of References
negation. 1. Liu C, Grenier D, Jousselme AL, et al. Reducing algo-
rithm complexity for computing an aggregate uncertainty
measure. IEEE T Syst Man Cyb: Part A Syst Human
Conclusion 2007; 37(5): 669–679.
2. Han Y and Deng Y. An evidential fractal AHP target
Probability distribution is efficient to represent knowl- recognition method. Defence Sci J 2018; 68(4): 367–373.
edge. However, everything in nature and society has its 3. Liu ZG, Pan Q and Dezert J. A new belief-based K-near-
negation, which shows that negation is very essential. est neighbor classification method. Pattern Recogn 2013;
Similarly, probability distribution also has its negation. 46(3): 834–844.
The article extends the proposed negation by using 4. Mohamed AB, Gunsekaran M, Doaa ES, et al. Integrat-
Tsallis entropy. Besides, numerical example is used to ing the whale algorithm with Tabu search for quadratic
calculate the uncertainty by different q of Tsallis assignment problem: a new approach for locating hospi-
entropy after negation. It can be found that there is tal departments. Appl Soft Comput 2018; 13: 530–546.
maximum Tsallis entropy after taking many iterations 5. Chen L and Deng X. A modified method for evaluating
sustainable transport solutions based on AHP
of negation no matter what the value of q is, meanwhile,
and Dempster–Shafer evidence theory. Appl Sci 2018;
the probability distribution becomes uniform distribu- 8(4): 563.
tion. Hence, it is reasonable to choose Tsallis entropy in 6. Deng X and Deng Y. D-AHP method with different
NPD after q is determined. From the above, it can be credibility of information. Soft Comput 2019; 23(2):
known that the negation can consume some informa- 683–691.
tion and increase the Tsallis entropy. Finally, the article 7. Seiti H and Hafezalkotob A. Developing pessimistic-
also discusses the application of negation in target rec- optimistic risk-based methods for multi-sensor fusion: an
ognition based on sensor fusion, which shows that interval-valued evidence theory approach. Appl Soft
negation can obtain a more reasonable decision when Comput 2018; 72: 609–623.
the conflict is high. In sum, negation not only provides 8. Wang Y, Zhang K and Deng Y. Base belief function: an
a new view to obtain information from another side efficient method of conflict management. J Ambient Intel
Human Comput 2018; 2018: 1–11.
based on known information, but also can include
9. Axenie C, Richter C and Conradt J. A self-synthesis
imprecise class to make better decision. Combining the approach to perceptual learning for multisensory fusion
known information and the information of negation in robotics. Sensors 2016; 16: 101751.
can increase the accuracy of decision-making. 10. Han Y and Deng Y. An enhanced fuzzy evidential
This article is a preliminary study to obtain the DEMATEL method with its application to identify criti-
NPD based on uncertainty measurements. This work is cal success factors. Soft Comput 2018; 22(15): 5073–5090.
done mainly to design a more efficient negation process 11. Su X, Mahadevan S, Han W, et al. Combining dependent
and uncertainty measurement, and expand the applica- bodies of evidence. Appl Intell 2016; 44(3): 634–644.
tion of negation. Besides, it is essential to determine the 12. Han Y and Deng Y. A novel matrix game with payoffs
uncertainty related to the negation and study its of maxitive belief structure. Int J Intell Syst 2019; 34(4):
properties. 690–706.
13. Abdel-Basset M, Mohamed M and Smarandache F. An
extension of neutrosophic AHP-SWOT analysis for stra-
Declaration of conflicting interests tegic planning and decision-making. Symmetry-Basel
The author(s) declared no potential conflicts of interest with 2018; 10(4): 116.
respect to the research, authorship, and/or publication of this 14. Lachaize M, Le Hgarat-Mascle S, Aldea E, et al. Eviden-
article. tial framework for error correcting output code classifica-
tion. Eng Appl Artif Intel 2018; 73: 10–21.
15. Behrouz M and Alimohammadi S. Uncertainty analysis
Funding of flood control measures including epistemic and alea-
The author(s) disclosed receipt of the following financial sup- tory uncertainties: probability theory and evidence the-
port for the research, authorship, and/or publication of this ory. J Hydrol Eng 2018; 23(8): 04018033.
Gao and Deng 7
16. Wu B, Yan X, Wang Y, et al. An evidential reasoning- 36. Atli BG, Miche Y, Kalliola A, et al. Anomaly-based
based CREAM to human reliability analysis in maritime intrusion detection using extreme learning machine and
accident process. Risk Anal 2017; 37(10): 1936–1957. aggregation of network traffic statistics in probability
17. Chen L and Deng Y. A new failure mode and effects space. Cognitive Comput 2018; 10(5): 848–863.
analysis model using Dempster-Shafer evidence theory 37. Li Y and Deng Y. Generalized ordered propositions
and grey relational projection method. Eng Appl Artif fusion based on belief entropy. Int J Comput Commun
Intell 2018; 76: 13–20. Control 2018; 13(5): 792–807.
18. Chatterjee K, Pamucar D and Zavadskas EK. Evaluating 38. Cetinkaya MC, Ustunel S, Ozbek H, et al. Convincing
the performance of suppliers based on using the R’AMA- evidence for the Halperin-Lubensky-Ma effect at the N-
TEL-MAIRCA method for green supply chain imple- SmA transition in alkyloxycyanobiphenyl binary mix-
mentation in electronics industry. J Clean Prod 2018; 84: tures via a high-resolution birefringence study. Eur Phys
101129. J E 2018; 41(10): 129.
19. Gao X and Deng Y. The negation of basic probability 39. Dong Y, Zhang J, Li Z, et al. Combination of evidential
assignment. IEEE Access 2019; PP(99): 1932. sensor reports with distance function and belief entropy
20. Janghorbani A and Moradi MH. Fuzzy evidential net- in fault diagnosis. Int J Comput Commun Control 2019;
work and its application as medical prognosis and diag- 14(3): 293–307.
nosis models. J Biomed Inform 2017; 72: 96–107. 40. Ye J and Cui W. Exponential entropy for simplified neu-
21. Seiti H, Hafezalkotob A, Najafi S, et al. A risk-based trosophic sets and its application in decision making.
fuzzy evidential framework for FMEA analysis under Entropy 2018; 20(5): 357.
uncertainty: an interval valued DS approach. J Intell 41. Mardani A, Nilashi M, Zavadskas EK, et al. Decision
Fuzzy Syst 2018; 35(2): 1419–1143. making methods based on fuzzy aggregation operators:
22. Liu Z, Pan Q, Dezert J, et al. Combination of classifiers three decades review from 1986 to 2017. Int J Inform
with optimal weight based on evidential reasoning. IEEE Technol Decis Making 2018; 17(2): 391–466.
T Fuzzy Syst 2018; 26(3): 1217–1230. 42. Jugade SC and Victorino AC. Grid based estimation of
23. Sun R and Deng Y. A new method to identify incomplete decision uncertainty of autonomous driving systems using
frame of discernment in evidence theory. IEEE Access belief function theory. IFAC-PapersOnLine 2018; 51(9):
2019; 7: 15547–15555. 261–266.
24. Deng X, Hu Y, Deng Y, et al. Environmental impact 43. Dutta P. Modeling of variability and uncertainty in
assessment based on D numbers. Expert Syst Appl 2014; human health risk assessment. MethodsX 2017; 4: 76–85.
41(2): 635–643. 44. Pan L and Deng Y. A new belief entropy to measure uncer-
25. Rosenberg E. Maximal entropy coverings and the infor- tainty of basic probability assignments based on belief func-
mation dimension of a complex network. Phys Lett A tion and plausibility function. Entropy 2018; 20(11): 842.
2017; 381(6): 574–580. 45. Dahooie JH, Zavadskas EK, Abolhasani M, et al. A
26. Srivastava A and Maheshwari S. Some new properties of novel approach for evaluation of projects using an
negation of a probability distribution. Int J Intel Syst interval–valued fuzzy additive ratio assessment (ARAS)
2017; 33(5): 1133–1145. method: a case study of oil and gas well drilling projects.
27. Yager RR. On the maximum entropy negation of a prob- Symmetry 2018; 10(2): 45.
ability distribution. IEEE T Fuzzy Syst 2015; 23(5): 46. Yin L, Deng X and Deng Y. The negation of a basic
1899–1902. probability assignment. IEEE T Fuzzy Syst 2019; 27(1):
28. Tsallis C. Nonadditive entropy: the concept and its use. 135–143.
Eur Phys J A 2009; 40(3): 257–266. 47. Clausius R. The mechanical theory of heat: with its appli-
29. Tsallis C. Possible generalization of Boltzmann-Gibbs cations to the steam-engine and to the physical properties
statistics. J Stat Phys 1998; 52(1–2): 479–487. of bodies. Wakefield: John Van Voorst, 1867.
30. Tsallis C. Introduction to nonextensive statistical 48. Preda V, Dedu S and Gheorghe C. New classes of Lorenz
mechanics: approaching a complex world. New York: curves by maximizing Tsallis entropy under mean and
Springer, 2009. Gini equality and inequality constraints. Physica A 2015;
31. Kang B, Deng Y, Hewage K, et al. Generating Z-number 436: 925–932.
based on OWA weights using maximum entropy. Int J 49. Yin L and Deng Y. Toward uncertainty of weighted net-
Intell Syst 2018; 33(8): 1745–1755. works: an entropy-based model. Physica A 2018; 508:
32. Pramanik S, Dalapati S, Alam S, et al. NC-cross entropy 176–186.
based MADM strategy in neutrosophic cubic set environ- 50. Deng W and Deng Y. Entropic methodology for entan-
ment. Mathematicas 2018; 6(5): 67. glement measures. Physica A 2018; 512: 693–697.
33. Xiao F. An improved method for combining conflicting 51. Deng X and Jiang W. Dependence assessment in human
evidences based on the similarity measure and belief func- reliability analysis using an evidential network approach
tion entropy. Int J Fuzzy Syst 2018; 20(4): 1256–1266. extended by belief rules and uncertainty measures. Annal
34. Jaynes ET. Gibbs vs Boltzmann entropies. Am J Phys Nuclear Energ 2018; 117: 183–193.
1965; 33(5): 391–398. 52. Xiao F. Multi-sensor data fusion based on the belief
35. Yager RR. Interval valued entropies for Dempster–Sha- divergence measure of evidences and the belief entropy.
fer structures. Knowl Based Syst 2018; 161: 390–397. Inform Fusion 2019; 46: 23–32.
8 International Journal of Distributed Sensor Networks
53. Shannon CE. A mathematical theory of communication. 60. Han Y and Deng Y. A hybrid intelligent model for
Bell Syst Tech J 1948; 27(3): 379–423. assessment of critical success factors in high risk emer-
54. Khammar AH and Jahanshahi SMA. On weighted cumu- gency system. J Ambient Intel Human Comput 2018; 9(6):
lative residual Tsallis entropy and its dynamic version. 1933–1953.
Physica A 2018; 491: 678–692. 61. Huynh VN, Nguyen TT and Le CA. Adaptively entropy-
55. Khammar A and Jahanshahi S. Quantile based Tsallis based weighting classifiers in combination using Demp-
entropy in residual lifetime. Physica A 2018; 492: 994–1006. ster–Shafer theory for word sense disambiguation. Com-
56. Wilk G and Wodarczyk Z. Example of a possible interpreta- put Speech Lang 2010; 24(3): 461–473.
tion of Tsallis entropy. Physica A 2008; 387(19): 4809–4813. 62. Khan M, Son LH, Ali M, et al. Systematic review of deci-
57. Lopes FM, Oliveira EAD and Cesar RM Jr. Inference of sion making algorithms in extended neutrosophic sets.
gene regulatory networks from time series by Tsallis Symmetry-Basel 2018; 10(8): 314.
entropy. BMC Syst Biol 2011; 5(1): 61–61. 63. Dutta P. An uncertainty measure and fusion rule for con-
58. Brito LRJD, Silva S and Tsallis C. Role of dimensionality flict evidences of big data via Dempster–Shafer theory.
in complex networks. Sci Report 2016; 6: 27992. Int J Image Data Fusion 2018; 9(2): 152–169.
59. Koike CY and Higa CHA. Inference of gene regulatory 64. Zhang H and Deng Y. Engine fault diagnosis based on
networks using coefficient of determination, Tsallis sensor data fusion considering information quality and
entropy and biological prior knowledge. In: Proceedings evidence theory. Adv Mech Eng 2018; 10(10): 809184.
of the IEEE international conference on bioinformatics and 65. Li M, Zhang Q and Deng Y. Evidential identification of
bioengineering, Taichung, 31 October–2 November 2016, influential nodes in network of networks. Chaos Soliton
pp. 64–70. New York: IEEE. Fract 2018; 117: 283–296.