Download as pdf or txt
Download as pdf or txt
You are on page 1of 11

Online Social Networks and Media 14 (2019) 10 0 050

Contents lists available at ScienceDirect

Online Social Networks and Media


journal homepage: www.elsevier.com/locate/osnem

Rumors and their controlling mechanisms in online social networks: A


survey
Mohammad Ahsan∗, Madhu Kumari
Computer Science and Engineering Department, National Institute of Technology, Hamirpur, Himachal Pradesh, India

a r t i c l e i n f o a b s t r a c t

Article history: Due to the rapid exchange of information and large user base of social networking sites, these are focused
Received 8 April 2019 for gathering the latest information or news from people over the world. Anyone having an internet con-
Revised 3 September 2019
nected device can share thoughts or update on real-time events. Social media helps reporters as well as
Accepted 25 September 2019
common men in sharing useful information, but at the same time, it also leads to deliberate or accidental
spread of rumors, i.e. pieces of information having uncertain truth at the time of posting. During social
Keywords: crisis, people access these platforms to get relevant information. In the rush of being early responders
Rumor spread to a critical event users post the information even without checking its veracity and that further used
Rumor veracity by other users to fill-in their informational gap. So flagging out the unverified information can be useful
Rumor detection
in maintaining a distance from spreading the information that may end up being false. The openness of
Rumor control
online social networking platforms (i.e. Twitter or Facebook), presence of machine learning and NLP (Nat-
Twitter
ural Language Processing) based techniques give us a chance to inspect the conduct of people in posting
rumorous information. In this work, we summarize and present the efforts and achievements so far to
combat the spread of rumorous information. These efforts composed of analyzing the content of rumors,
properties of users who share rumors and network structure that favor the spread of such information.
© 2019 Elsevier B.V. All rights reserved.

Contents

1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1. Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2. Multidisciplinary analysis of rumor diffusion. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2.1. Psychology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2.2. Sociology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2.3. Epidemiology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2.3.1. SI model. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2.3.2. SIS model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.3.3. SIR model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
3. Features categorization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
3.1. User-related features: emotional pressure, intellectual pressure, personal involvement, physical proximity, emotional proximity,
follow ratio (log10 ## ff ol ol l owers
l owing
), age, verified (whether the user is verified by Twitter or not) and tweet frequency . . . . . . . . . . . 5
3.2. Content-related features: importance, informational ambiguity, information with no clear source, believability, information
overload, information dearth and veracity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
3.3. Network related features: communities, degree distribution and hubs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
3.4. Characteristics of extreme events: static or dynamic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
4. Rumor veracity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
4.1. Experts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
4.2. Websites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6


Corresponding author.
E-mail address: ahsan@nith.ac.in (M. Ahsan).

https://doi.org/10.1016/j.osnem.2019.10 0 050
2468-6964/© 2019 Elsevier B.V. All rights reserved.
2 M. Ahsan and M. Kumari / Online Social Networks and Media 14 (2019) 100050

4.3. Certainty and evidentiality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6


4.4. Communities, source ambiguity, and corroboration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
4.5. Cognitive psychology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
4.6. Machine learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
4.6.1. Message based features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
4.6.2. User based features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
4.6.3. Topic based features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
4.6.4. Propagation based features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
4.7. Web services. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
5. Rumor detection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
5.1. Identifying relevant posts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
5.2. Regular expressions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
5.3. Speech acts. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
5.4. Tweets timeline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
5.5. Semi-automatic approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
6. Rumor control. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
6.1. Anti-rumor models. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
6.1.1. Delayed start model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
6.1.2. Beacon model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
6.1.3. Neighborhood model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
6.2. Rebuttals. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
6.3. Credible sources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
6.4. Influential users . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
6.5. Emergency officials and news media . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
7. Conclusion. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
Declaration of Competing Interest . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Acknowledgment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Further Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

1. Introduction have to follow their protocols before reporting the news and that
make them slow. On one hand, they have to catch the breaking
Social media has a profound effect on the way we communicate news and on the other hand, they have to wait for veracity check.
nowadays. In the past, people had to use verbal and non-verbal This situation slow down the reporting process of mainstream me-
media to share information with other individuals, e.g. town criers dia and people move to social media for obtaining relevant in-
and drum. These mediums of communication were not sufficient to formation. The openness and flexibility of social media applica-
convey information beyond few miles. So, to overcome these limi- tions like Twitter enable its users to be a first responder of an
tations, people switched to print media (i.e. newspaper and books) extreme event and cope up post-event situation by utilizing the
and mainstream media (i.e. televisions and radio). But there exists local knowledge [2]. This real-time reporting on Twitter is also ac-
a crisp boundary between journalists and ordinary citizens which knowledged by mainstream media and in many cases, social media
restricts people from sharing their experience and ideas. The ad- is cited as their source of information. For example, in 2010, after
vent of social media broke that boundary and enables individuals a few moments of an incident, pictures and videos of earthquake
to act as an audience and reporters as well. It broadens the di- stricken Haiti were shared by Twitter users and mainstream me-
versity of information that people can receive from others. Some- dia cite that information as their news source [3]. So, the speed
times it is hard to quantify how much our lives has been impacted and ability of providing local information are the advantages of so-
by social media until we check the scenario of a few years back. cial media. Despite these benefits, social media are questioned for
At the beginning of this decade, envisioning the communication its information quality. From the examination of Google’s search
paradigm that offered to us by social media applications like Face- results, it was found that Tweeter and blogs’ real-time informa-
book and Twitter was quite unimaginable. Now, with the advent of tion mainly constitute of “fabricated content, lies, misinterpreta-
social media it is possible that information posted by a single user tion, and unverified events” [4].
receive the attention of millions of users within few seconds. The lack of accountability and veracity checking techniques at
The applications of social media i.e. Twitter or Facebook, not social media platforms make them a perfect place to perform ill
only assist the people to keep in touch with their friends and fam- practices i.e. spreading propaganda, spam, misinformation, and ru-
ily members, but also to get informed with breaking news and on- mors. The Oxford dictionary defines rumor as “a currently circulat-
going events. Due to huge user-base and trust among social media ing story or report of uncertain or doubtful truth”. After an earth-
users, these platforms have become a main source of information quake strike in Haiti, rumors circulated that “JetBlue and American
for many users. If a social media user has mobile device with inter- Airlines are offering free air travel to Haiti for doctors and nurses”
net connectivity, then he/she can post any information from any- and “UPS will ship any package under 50 lbs to Haiti for free”. A
where at any time. This low cost of information exchange at online spokeswoman from JetBlue refute this information by saying that
platforms allows information to spread quickly and more widely they are flying only Haiti Consulate scrutinized relief workers, not
than ever before [1]. just anyone who is willing to fly [5]. A co-founder of Twitter, Biz
During unexpected extreme events (i.e. natural disasters and Stone, fairly recognized this problem and said that credibility is
terrorist attacks), people get tense and to remove their tension a key for social media [3]. Social media information rarely con-
they turn to mainstream media for getting a trustworthy and rel- tains a warning label and knowing the correctness is not easy for
evant information. But mainstream media, i.e. televisions or radio, people until they refer other sources like social circles, televisions,
M. Ahsan and M. Kumari / Online Social Networks and Media 14 (2019) 100050 3

microblogs or newspapers. So, before taking the ownership of (Rumor∼ =Ambiguity × Importance). The multiplicative nature of this
shared information and forwarding it to others there is need of function represents that for the growth of a rumor, it must be
checking the credibility of information and its source. shrouded with some level of ambiguity and recipients must find it
important. If the content of the rumor is not important then there
1.1. Motivation is no incentive in sharing it with others and without ambiguity the
information is a fact and do not need any subjective interpretation.
The rumors hold the ability of doing a significant amount of In 1973, Anthony [16] argued that measuring the importance of
harm, i.e. financial loss, reputation loss and in extreme cases even an information is a difficult task and she used a psychometric tool
danger of lives. This potential has prompted a long history of en- “Taylor Manifest Anxiety Scale” [17] to comprehend importance fac-
deavors to comprehend rumor process and devise some combat- tor under the concept of anxiety. The reasoning behind this change
ting methods. The primary technique utilized by government and is – if a person feels anxious about a piece of information then it
different authorities is communicating messages to the whole pop- signals the importance of information for him. In this formulation,
ulace to debunk rumors [1]. But this technique is difficult to adopt rumor is conceived as a method of releasing emotional strain by
in a situation where individuals believe that government has a per- explaining ambiguous information.
sonal stake in debunking such rumors, i.e. maintaining the lawful-
ness. For example, in 2003 a rumor was surfaced in Nigeria that 2.2. Sociology
Polio vaccine is contaminated to kill fertility. These rumors were
spread broadly due to several factors, i.e. few years back the death Allport et al. [13] and Anthony [16] studied the affective aspect
of 11 children due to a clinical trial and support of influential peo- of anxiety whereas Bordia et al. [14] examined its cognitive angle
ple like politicians and religious leaders to these rumors. Nigerian and recognize rumor as a collective process of reducing informa-
authorities spent around two years to combat these rumors. It was tional uncertainty. In the absence of adequate information from le-
the case of single governance, which took that much of the time. gitimate sources, people continually attempt to extract meaningful
In case of social media, national borders are crossed and a con- information about uncertain situations. From a sociology perspec-
founded situation exists where some users trust the authority and tive, rumor is viewed as a “collective process that arises when ade-
others doubt it. So, to combat the spread of rumors on social me- quate information is unavailable, from formal or legitimate sources,
dia there is a need of more robust approach. to interpret a problematic situation or event” [18]. It means rumors
develop when an unexpected or unfamiliar situation arises and an
2. Multidisciplinary analysis of rumor diffusion interpretation is required to know how to respond to the situation.
This approach strikingly differs from psychological theory of
Rumor diffusion has been extensively studied in the field of Allport and Postman [13] which considers rumor as a false in-
psychology, sociology and epidemics from last 70 years. In this sec- formation. However, both domains agree with the finding that
tion, we present key findings of these areas to understand the ru- extreme events such as natural disaster [13,18] and terrorism
mor diffusion process from a multidisciplinary viewpoint. [19] provide the optimal condition for rumor spread by creating
an anxious environment and generating ambiguous information.
2.1. Psychology Rumor researchers of both domains provide insights for reduc-
ing anxiety and controlling rumor spread in affected communities.
Most of the online social networks became a part of our life They analyzed that timely provision of credible information from
in the last fifteen years. However, in the field of psychology, ru- trustworthy sources play a key role in subsiding anxiety and com-
mors’ propagation and their counter measures have been exten- batting rumor spread.
sively studied from seventy years [6]. Researchers in psychology
mainly carried out their work in offline manner and very few 2.3. Epidemiology
of their studies are there to examine the online content [1]. The
main focus of prior research which examine the online content was The process of rumor spread and epidemic spread is often com-
to understand the dynamics of information diffusion and not to pared by the researchers. In epidemiology, robust framework has
check the veracity of information [7–9]. Social psychologists have been developed for modeling the spread of pathogens. Each indi-
found some rumor causing factors like information ambiguity and vidual is assumed to belong to one of these states: susceptible (S),
scarcity [10,11]. These characteristics are commonly presented by infectious (I), and recovered (R). Susceptible individuals are healthy
extreme events like earthquake and terrorist attacks. When people people, infectious are contagious who can infect other people, and
experience these events, they get anxious to understand the uncer- recovered class represents those who have been infected prior but
tainty of the situation. Initially they turn to reliable news sources returned to a healthy state, and hence not infectious. Similarly, in
such as mainstream media for getting relevant information and if rumor diffusion, susceptible nodes are unaware of the rumor and
these sources do not timely provide the required information then known as ‘ignorants’, infected nodes spread rumors to the neigh-
people use social media platforms to release their anxiety. Anxiety boring nodes and referred as ‘spreaders’, and recovered nodes have
hinders cognitive ability of people and make them to believe an known the rumor but do not spread it further, these nodes are
information that would otherwise not to be believed under nor- called as ‘stiflers’. There are few well-studied models for epidemic
mal conditions. The more anxious a person is, more likely he/she spread i.e. SI, SIS and SIR model. In these models, the number of
spreads the rumors [12]. The ambiguity of information leads peo- susceptible, infected and recovered individuals, at time t, is de-
ple to share it further because “our minds protest against chaos” noted by S(t), I(t) and R(t) respectively.
[13]. Therefore, more ambiguous an information, more frequent the
communication among people for extracting a clear meaning out 2.3.1. SI model
of it. Researchers have also explored some other psychological fac- In this model, an individual can be in one of two states: sus-
tors, i.e. information accuracy and its importance that affect rumor ceptible (healthy) state or infected (sick) state. It is assumed that
spread [14,15]. susceptible individuals become infected at the rate β after coming
After exploring rumors’ characteristics that were pervasive dur- in contact with infected individuals. In this model, if an individ-
ing the Second World War , Allport and Postman [13] mod- ual becomes infected, then it cannot recover to a healthy state and
eled rumor spreading as a function of ambiguity and importance remains infected forever. Assuming each individual has k contacts
4 M. Ahsan and M. Kumari / Online Social Networks and Media 14 (2019) 100050

and equal probability of being in contact with an infected individ- phenomenon and provides a more realistic explanation of informa-
ual. The chance of an infected individual to contact the susceptible tion diffusion in online social networks. There is a common struc-
nodes is k S(Nt ) . Since I(t) individuals are transmitting the disease ture in online social networks known as community which shows
at the rate β , the total spread in time frame dt becomes: information trapping behavior and often prohibits large scale infor-
mation diffusion [21,22]. These communities are formed on a par-
S (t )I (t )
dI (t ) = βk dt ticular theme or interest and that is why members of a community
N
prefer to share information with each other rather than with non-
dI (t ) S (t )I (t ) community members. Due to this behavior information get trapped
= βk in a local region and do not spread globally. Independent spread-
dt N
ers break this barrier by carrying information from one network
di to another network. This model explains why some rumors spread
= βksi, globally even in the presence of communities.
dt
Xia et al. [23] have considered official sources of information
where s and i represents a fraction of susceptible and infected in- that release authoritative information during the spread of ru-
dividuals i.e. s = S(Nt ) and i = I (Nt ) . mors. The classical SIR model is extended to a two-stage model
di – SIAR model. At first stage, 0 ≤ t < τ each node can be in
= βk i (1 − i ) one of three states: susceptible (S), infected (I) and stiflers (R).
dt
At t = 0, a random node is chosen from susceptible nodes as in-
di di fected (spreader) and all other nodes remain susceptible (ignorant).
+ = βkdt Ignorant nodes become spreaders at a spreading rate λ(0 ≤ λ <
i (1 − i )
1) after coming in contact with spreaders. In second stage, t = τ (τ
On integrating both sides, we get > 0) a random node is chosen from susceptible nodes as authori-
tative which spread authoritative information. Whenever an igno-
ln i − ln (1 − i ) + C = βkt
rant node comes in contact with authoritative node, it becomes
At t = 0, we get: authoritative at a diffusion rate β (0 < β ≤ 1). If a spreader con-
tacts other spreader, authoritative or stifler nodes, then the con-
io ioeβkt 1
C= , i= and τ= , tacting spreader becomes a stifler at the rate α . It happens because
( 1 − io ) 1 − io + ioeβkt βk spreading node finds other sharing or refuting the shared infor-
where τ is the characteristic time that required in reaching a 1/e mation which represents that news lost its value. Spreaders also
fraction of susceptible individuals. Hence, increasing either the turn to stifler state, even without contacting other nodes at a for-
likelihood of disease transmission (β ) or degree of individuals (k) getting rate δ (0 < δ ≤ 1). This study examined the critical thresh-
enhances pathogen spreading speed. old (λc ) of SIR and SIAR models and found an increment in λc of
Here, at the beginning, pathogen spreads exponentially, but SIAR model. It means it is tough to spread rumors if authoritative
with the passage of time infected individuals encounter a less sources are there to inform the ignorant nodes about rumors. If the
number of susceptible nodes and spread slows down for large t. rumor spreading rate of a network is less than critical threshold, λ
At the end of the epidemic, everyone becomes infected i.e. < λc , then rumor cannot spread in that network. They observed
i(t → ∞) = 1 and s(t → ∞) = 0. that sooner the authoritative information is available, lesser is the
negative impact of rumors. So, authority should release the infor-
2.3.2. SIS model mation as soon as possible.
An individual can only be in one of two states: susceptible and Borge et al. [24] proposed two modifications to classical rumor
infected; but infected individuals can recover to a susceptible state models to capture activity pattern and apathy of individuals. Hu-
at a fixed rate μ. So, total spread of pathogen in time frame dt be- man beings have a heterogeneous (i.e. not always active) activity
comes: pattern and their activities are not randomly distributed over time
[25–27]. It has been observed that activity distributions are cor-
di
= βki(1 − i ) − μi, related with degree distributions [28]. Rumors die out soon if they
dt
are posted by low degree nodes and shared persistently if triggered
where μi is the rate at which individuals recover from the disease. by high degree nodes. The effective way of implementing this cor-
relation is to assign each node an activity probability ai = ki /kmax ,
2.3.3. SIR model where ki is the degree of node i and kmax is the maximum node-
For many pathogens, individuals develop immunity after they degree in the network. The first modification assigns these activ-
recover from infection. Hence, instead of returning to susceptible ity probabilities and second considers the possibility of apathetic
state, they get removed from the population. These recovered indi- ignorant node which shows no interest in the rumor and directly
viduals do not count any longer from the perspective of pathogen goes to stifler state. This is a common behavior in online social net-
as they cannot be infected, nor they can infect others. At the end, works where received messages are rarely spread further. In this
everyone recovers from infected states i.e. i(∞) = 0. model, when an ignorant node comes in contact with spreaders,
These epidemic models have been extensively used by rumor the ignorant node becomes spreader with a probability λp and sti-
researchers. Qian et al. [20] present SIR-like rumor model which fler with a probability λ(1 − p), where p is probability that the ig-
captures the scenario of independent spreaders. These spreaders are norant node is interested in diffusing the rumor.
ignorant nodes who learn a rumor without interacting with neigh- From psychology, we learn that informational ambiguity and
bors within the network. In epidemics, susceptible individuals re- scarcity are the roots of a rumor generation on the online so-
ceive pathogen from infected neighbors, but in social media, igno- cial networks. For the growth of a rumor, it need to be masked
rant nodes can receive rumor without interacting with neighboring with some level of ambiguity and recipients must find it impor-
nodes, for example: a Twitter user can learn about a rumor from tant. Sociologists analyzed rumor as a collective process of releas-
Facebook instead of his/her followees within Twitter. It means ru- ing anxiety when adequate information is unavailable from legiti-
mor diffusion on one social network can be affected by other net- mate sources during an unclear situation. Finally, due to a resem-
works. The introduction of independent spreaders captures this blance between epidemic spread and rumor spread, the concepts
M. Ahsan and M. Kumari / Online Social Networks and Media 14 (2019) 100050 5

Table 1
Multidisciplinary factors that affect rumor spread on online social networks.

Area of research Factors Findings



Psychology Ambiguity, scarcity [10,11] Information ambiguity and its scarcity lead to rumors spread as our minds
protest against chaos.

Anxiety [12] The more ambiguous a person is, the more likely he/she spreads the
rumors.

Ambiguity, importance [13] Rumor spread is a multiplicative function of information ambiguity and its
importance. If any of these becomes zero, there will be no rumor.
(Rumor ∼= Ambiguity × Importance)

Ambiguity, anxiety [16] Anxiety signals the importance of information. If a person feels anxious
about a piece of information then it is important for him.
(Rumor ∼= Ambiguity × anxiety)

Sociology Absence of adequate information from legitimate sources Rumor is a collective process that arises when adequate information is
[18] unavailable from legitimate sources for interpreting a problematic
situation.

Natural disaster [18] and terrorism [19] Natural disaster and terrorism provide optimal conditions for rumor
spread by creating an anxious environment and generating ambiguous
information.

Epidemiology Independent spreaders [20] These spreaders learn rumors without interacting with users of their
network. They overcome the information trapping capabilities of online
communities by carrying information from one network to another.

Authoritative information [23] In the presence of authoritative nodes (SIAR model), critical threshold (λc )
of a network get increased. It means rumor spreading is tough if
authoritative sources are these to inform ignorant nodes about rumors.

Activity pattern and apathy [24] Activity distributions are correlated with degree distributions.

Rumors posted by high degree nodes shared persistently and by low
degree nodes die out soon.

When an ignorant node comes in contact with spreaders, it becomes
spreader with probability λp and stifler with probability λ(1 − p).

of epidemic modeling are used to understand spreading and for- sources physically close to the event. Apart from physical proximity
getting mechanism of rumors on social media. Table 1 provides a – a physical distance between the people and disaster event, emo-
brief summary of key findings. tional proximity – a sentimental association with disaster affected
people or disaster location, also played a significant role in rumor
3. Features categorization mongering. The follow ratio of a person represents his/her reputa-
tion and determine whether a person is likely to support rumors
With the popularity of Twitter as an important source of infor- or deny them. In the context of news reporting, users with high
mation, it is also highly abused for the spread of rumors. During an follow ratio are news organizations which are assumed to provide
emergency or serious event, people get anxious to understand the well-grounded information with certainty and evidences. But the
uncertainty of the situation and start spreading information with- pressure of journalism makes them to shift their priorities from
out checking its factual accuracy. There exist multiple features that pre-publication verification to post-publication verification. It cre-
contribute to the birth and growth of rumors on Twitter i.e. anxi- ates the possibility of rumor posting by these reliable sources of
ety of users, ambiguity of information, personal involvement in the information. So, it is concluded that high follow ratio tends the
event, physical and emotional proximity to the event. These fea- users to support rumors and low follow ratio makes them to deny
tures are broadly categorized as: rumors [32]. The age of a user, verified status and tweet frequency
do not show a significant difference in deciding whether a person
3.1. User-related features: emotional pressure, intellectual pressure,
is rumor supporting or not.
personal involvement, physical proximity, emotional proximity,
follow ratio (log10 ## ff ol
ol l owers
l owing
), age, verified (whether the user is
verified by Twitter or not) and tweet frequency 3.2. Content-related features: importance, informational ambiguity,
information with no clear source, believability, information
Emotional pressure represents anxiety and intellectual pressure overload, information dearth and veracity
points to ambiguity. During extreme events (i.e. earthquake or ter-
rorist attack), official news sources do not provide the required in- Allport and Postman [13] define rumor as a multiplicative func-
formation in a timely manner and people get anxious to compre- tion of informational ambiguity and its importance. A rumor can-
hend the situation. To alleviate the tension, people share the un- not be there if any factor becomes zero. Oh et al. [12] found that
official information with their friends and interpret that to get a information with no clear source is more important rumor caus-
coherent understanding of the unclear situation. There are a num- ing factor than anxiety. It is contrary to the findings of traditional
ber of features which affect the information sharing behavior of research on rumors which considers anxiety as most important ru-
social media users. The more anxious a person is, more likely mor causing factor. The absence of informational source produces
he/she spreads rumors [12,29,30]. The ambiguity of the informa- anxiety among users and motivate them to further share the in-
tion source represents the veracity of information. The more am- formation in order to find the actual truth. The believability of a
biguous an information source is, the more frequent is the com- rumor does not affect its subsequent spread during natural disas-
munication for information seeking and sharing [19]. The feeling ters [33,34]. But, when a rumor is important to users and its be-
of personal involvement in a disaster event is positively related to lievability is high then it spreads more frequently [35]. So, without
the spread of rumors [12]. Huang et al. [31] interviewed the par- importance the believability of a rumor does not affect its spread.
ticipants to examine the effect of physical and emotional proxim- Information overload hampers the emergency responders in effi-
ity on the spread of rumors during 2013 Boston Marathon Bomb- ciently providing the credible information to the right people in a
ings. Several participants conveyed that they put more trust on the timely manner [36]. Information dearth is a lack of local informa-
6 M. Ahsan and M. Kumari / Online Social Networks and Media 14 (2019) 100050

tion required by people to form localized decisions [12]. The ve- 4.2. Websites
racity of rumors also affects their spread; false rumors spread for
a longer time than true rumors. The veracity resolution of false The alternate approach is to take the help of websites like
rumors takes more time than true rumors because proving the Snopes.com and FactCheck.org. These websites document online
counter evidences is a more challenging task than proving some- rumors and stories of uncertain truth. They provide ground-truth
thing true. Before resolution people tend to share rumors, but after annotations regarding rumors veracity.
resolution their interest is decreased. So, false rumors get spread
for a longer time than true rumors. 4.3. Certainty and evidentiality

3.3. Network related features: communities, degree distribution and Zubiaga et al. 2016 [32] annotated rumors by examining three
hubs dimensions of rumors: (i) support and response type, (ii) certainty,
and (iii) evidentiality. Support reflects whether the message given
The structure of connections among social media users matters in a source tweet is supporting, denying, or presenting an un-
extensively in the spread of rumors [37]. There is a common struc- clear view about the tweet’s content. Response type is a support
ture in online social networks known as community which hinders that response tweets provide to the source tweets. This factor is
the information spread. Due to dense intra-community connectiv- extremely useful in determining rumors’ veracity. It has the fol-
ity and sparse inter-community connectivity, a user most likely lowing annotations: agreed – when a response tweet supports its
shares the rumors with members of his/her own community. This source tweet, disagreed – when a response tweet denies its source
reinforcement leads the rumors get trapped in one community tweet, appeal for more information – when a response tweet ap-
and decrease their reach. Communities also helpful in spreading peals for additional information about its source tweet, and com-
rumors as they impose a social influence on the people to trust ment – when the author of a response tweet is not interested in
the received information without checking its factual accuracy or assessing the veracity and just makes a comment. Certainty is the
source credibility [38]. Degree distribution is the probability that a degree of confidence that authors expressed regarding their tweets.
randomly selected node from a network has degree k and it affects These are the annotations of certainty: certain – when an author is
the rumor spreading dynamics. Online social networks are scale- confident about the content of his/her tweet, somewhat certain –
free and follow power-law degree distribution where some nodes when an author is partially confident, uncertain – when an author
have a very high degree, known as hubs. These hub nodes rapidly is totally unsure.
spread the rumors by taking them to a large number of ignorant Evidentiality reflects the type of evidence provided by the au-
nodes (people unaware of the rumor), but at the same time, it can thors regarding the statements made in their tweets. It has these
hinder rumors spread [37]. After knowing that every friend already annotations: attachment of a URL, attachment of a picture, em-
knows about the rumor, people lose interest in the rumor and do ployment of reasoning, first-hand experience, no evidence, quota-
not spread that further. So, hubs play both roles: supporting as tion of a person or organization, and quotation of an unverifiable
well as hindering rumor spread. source. There is not much of a distinction in certainty, but evi-
dentiality reflects a significant change in different veracity values
of rumors (true, false, or unverified). All three dimensions are an-
3.4. Characteristics of extreme events: static or dynamic
notated through a crowdsource platform named as Crowd Flower
[39]; in order to get done in a timely manner. But the usage of
The characteristics of extreme events help in understanding the
crowdsource platforms is not suitable for the task of assessing the
dynamics of rumor spread [19]. Natural disasters have static post-
rumors veracity [40]. This approach favors to give ‘true’ annota-
event impact and authorities find it feasible to provide the rele-
tions to the false rumors.
vant information to the affected community in a timely manner.
The timely availability of certain information from credible sources
4.4. Communities, source ambiguity, and corroboration
subside the anxiety of people and leads to rumor control. But, dur-
ing terrorist attacks, terrorists are dynamic and they keep moving
Communities affect the information spreading behavior by im-
from one location to another. In this situation, it becomes infeasi-
posing a social trust on the users. The members of same com-
ble for the authority to process the available information and re-
munity share strong connections with each other and tend to be-
turn back with some related information. So, these events lead to
lieve an information shared by other members, without checking
the growth of rumors.
its factual accuracy and source ambiguity. But, in Twitter, commu-
nity members do not share such strong connections and not eas-
4. Rumor veracity ily accept a dubious information. Here, source ambiguity is use-
ful in evaluating the veracity of circulating information. Huang
Most of online social networks suffer from the spread of ru- et al. [31] interviewed people after Boston Marathon Bombings and
mors. These rumors seem credible to social media users, but ver- found that corroboration is a method of determining the credibility
ifying their veracity is a challenging job. The existing approaches of circulating information.
for solving this task are:
4.5. Cognitive psychology
4.1. Experts
The spread of rumors or misinformation depends on the assess-
It is usually done with the help of experts like journalists who ment of truth-value of this information by the receivers. Cognitive
analyze the claims made in rumors and information released by psychology provides the following factors to determine the authen-
authorities. If the claims are found in accordance with authorita- ticity of information: consistency of message, coherency of mes-
tive information then rumors are annotated as ‘true’, otherwise, sage, credibility of source and general acceptability. An information
‘false’. Sometimes, these experts do not find sufficient information is easily accepted when it is consistent and compatible with the
to decide the veracity of rumors and annotate the rumors as ‘un- pre-existing beliefs of the people. If the facts or statements con-
verified’. This approach is expensive and sometimes it is not easy tained in a message are coherent and have no contradiction, then
to include third-party (i.e. journalists) in our research projects. there are more chances of the message acceptance. The perceived
M. Ahsan and M. Kumari / Online Social Networks and Media 14 (2019) 100050 7

credibility of the information sender and repetitive exposure to the Table 2


Regular expressions to identify verification and
information increase the acceptability of information. These factors
correction patterns.
are highly utilized by the researchers for measuring tweets’ credi-
bility [21,22,41,42]. Pattern Regular expressions

Verification is(that|this|it)true
4.6. Machine learning wh[a]∗ t[?!][?1]∗
(real?|really?|unconfirmed)
Correction (rumor|debunk)
Castillo et al. [21] estimate tweets credibility by utilizing su-
(that|this|it) is not true)
pervised learning techniques. They used four types of features for
training their supervised models: workers. Castillo et al. [21] used MTurk to annotate the credibility
of collected tweets and then utilized the annotated data to train
4.6.1. Message based features machine learning models for the prediction of tweets credibility.
The presence of question marks or exclamation marks in a mes- Truthy: This web service was developed by the researchers of
sage, presence of hashtags, message is a retweet or not, length of Indiana University, USA. It tracks political memes to detect the
the message, and the number of words with positive or negative spread of misleading information during elections. It visually rep-
sentiment. resents how misleading memes spread during election hours and
enables the users to sort them on the basis of the number of
4.6.2. User based features tweets, number of retweets, number of users, and the size of the
These are the properties of users who have posted the mes- largest connected component [22].
sages i.e. number of followers, number of followees, number of
tweets posted by a user, and account age (time since a user cre- 5. Rumor detection
ated his/her account).
There are very serious consequences of the spread of false ru-
4.6.3. Topic based features mors on peoples’ lives and their reputation. During terrorist attacks
These features are derived from the previous feature sets or natural calamities access to credible information is very much
(message-based and user-based), for example, the fraction of required for the well-being of people. But, on social media, infor-
tweets containing hashtags, the fraction of tweets containing URLs, mation never come with a label of rumor or non-rumor and even a
and the fraction of tweets having words of positive and negative rumorous information seems credible. Rumor detection helps peo-
sentiment. ple by identifying whether an information is spreading a rumor or
not. It is a binary classification where a set of social media posts
4.6.4. Propagation based features are passed to the rumor detection system and each post is assigned
These features are related to a propagation tree generated from a class out of these two classes: rumor (R) and non-rumor (NR).
the retweets of messages. These features are included in this cat- There are various approaches which are followed by the rumor re-
egory: depth of the retweet-tree and number of initial tweets of a searchers for detecting rumors on social networks, as represented
topic. in Fig. 1.
Initially, data samples were assigned their labels (or classes) by
using Mechanical Turk [43]. Then the labelled data were used to 5.1. Identifying relevant posts
train a number of supervised models like Bayes networks, decision
trees, and Support Vector Machine (SVM). The best performance Early research on rumor detection was only limited to identify
achieved through these models was given by decision tree (with relevant posts of pre-identified rumors [44]. The relevancy of posts
J48 algorithm), an accuracy of 89%. can be measured by training a machine learning model on the pre-
The fraction of tweets containing URLs is the root of the deci- viously identified rumors. For example, by training a classifier on a
sion tree and hence most important feature for accessing tweets’ set of rumors i.e. ‘Obama is a Muslim’, we can identify that ‘Obama
credibility. Tweets with no URL tend to be related with a non- is not Muslim’ is a relevant post and ‘Obama was talking to a group
credible information. The sentiment of tweets is also a relevant of Muslims’ is an irrelevant post. But this approach only works for
feature in the task of credibility prediction. A credible informa- identifying long-standing rumors, where new unseen rumors are
tion is very much related to the fraction of tweets with negative relevant to the posts that were used for training the classifier. It
sentiments, rather than positive sentiments, which tend to express fails when new unseen rumors have a different context as there is
non-credibility. Tweets with more number of retweets are consid- no training about these rumors. So, with every context-switching
ered more credible. Users who post non-credible information tend of rumors, this approach needs the researchers to identify new ru-
to have a small number of followers and only posted a very few mors and re-train the classifiers. To overcome this issue and iden-
tweets so far. The number of followers, followees and tweet activ- tify rumors during fast-paced contexts i.e. breaking news, classi-
ity (number of tweets posted by a person) represent the reputation fiers need to be trained on some generalizable patterns.
score of the users. Credible users tend to have a high reputation
score and the credibility of tweets increase as they pass through 5.2. Regular expressions
highly reputed users.
There are certain needs that led the social media users to share
4.7. Web services information with each other i.e. the need of verifying a contro-
versial information. These needs not only assist in rumor spread,
Amazon Mechanical Turk (MTurk): It is a web service that pro- but also provide a clue of rumor detection. Rumors provoke skep-
vides human intelligence. Even after so much advancement in tic users to post tweets for enquiring their veracity. Zhao et al.
computing technology, there are many jobs that can be done more [45] have utilized these findings and manually curated a list of five
effectively by human beings than computers, like identifying ob- regular expressions to identify tweets that contain verification and
jects in a photo. Companies or developers who need Human Intel- correction patterns (Table 2).
ligence Tasks (HITs) to be done can use MTurk API for accessing Tweets with these patterns are clustered on the basis of a sim-
the services of thousands of global, high quality, and on-demand ilarity score given by Jaccard coefficient and a single statement
8 M. Ahsan and M. Kumari / Online Social Networks and Media 14 (2019) 100050

Fig. 1. Methods to detect long-standing and fast-paced rumors on Twitter.

is extracted from each cluster. These statements are used to pull assertions and gives 91% accuracy. After detecting tweets contain-
more tweets that discuss these statements. Then they employ two ing assertions, the next step is to pass them to a hierarchical clus-
human annotators to manually annotate tweets as R or NR on the tering module. This module clusters tweets by using tf − idf cosine
basis of collected tweets, statements extracted from clusters, and similarity and produce a set of clusters, each containing tweets of
related information from Web search. In this work, the speed of similar assertions. In tf − idf, tf counts the number of times a term
rumor detection using regular expressions is compared with trend- occurs in a tweet and idf measures the number of tweets that con-
ing topics and popular memes. Trending topics and popular memes tains a term and represent how much informative it is – general
can reveal some rumors, but they both produce many false posi- or informative.
tives and have a low precision. They analyzed that the use of en-
quiry tweets (tweets with verification patterns) not only helpful in
5.4. Tweets timeline
detecting more rumors but it also contributes to earlier detection.
Detection of rumors by using enquiries and correction pattern is
Zubiaga et al. [51] identify rumors by enlisting the help of ex-
found to be 4.3 h earlier than correction only, 3.6 h earlier than
perienced annotators (i.e. journalists). In the process of rumor de-
trending topics, and 2.8 h earlier than popular memes. There is a
tection, a single tweet is not always enough to determine whether
considerable room for improving this approach of rumor detection
it is a rumor or not. So, they developed a tool which displays
i.e. automatic filtering of correction and enquiry posts rather than
the timeline of event specific tweets. They have used #ferguson
using some manually curated list of regular expressions, automatic
keyword with Twitter’s streaming API to collect tweets associated
detection of rumors, and finding features that are potentially re-
with Ferguson unrest. There are a large number of tweets associ-
lated to the veracity of the rumors.
ated with this event and manual annotation of all is not feasible.
So, tweets not having a specific number of retweets are discarded
5.3. Speech acts
as rumor tends to spark a decent number of retweets. The next is-
sue is the limited context of a rumor. To enhance the context of
Vosoughi [46] have developed a rumor detection model by us-
a rumor they also collect replies associated with the tweets be-
ing the pragmatics of tweets. The semantic (i.e. opinion words,
cause replies help in determining the nature (R or NR) of a post.
emoticons, n-grams, etc.) and syntactic (i.e. #, @, punctuations,
They examined that rumors percentage decreases with a decrease
part-of-speech etc.) features of tweets have been extensively an-
in the number of retweets and rumors sparked more replies than
alyzed by the researchers for detecting rumors, but sometime lit-
non-rumors due to controversial nature. It also observed that ru-
eral meaning is not sufficient and context is required to understand
mors take longer time for annotation than non-rumors as annota-
the exact meaning of an information. Pragmatics consider how the
tors have to explain the reason of considering tweets as rumors.
context of an information contributes to its meaning. In this work,
speech acts (assertion, expression, question, recommendation, re-
quest, and miscellaneous) are used as pragmatics. The recognition 5.5. Semi-automatic approach
of these speech acts is a multi-class classification problem and four
classifiers have been trained on a manually labelled dataset with On social media, although it is possible to manually annotate
3312 binary features – 1647 semantic and 1666 syntactic. Seman- the post as R/NR but it encounters certain challenges in the form
tic features used in this task are: opinion words [47], vulgar words of label quality and time. Social media data come in non-structural,
[48], emoticons [49], English speech act verbs, and n-grams; and noisy and multi-modal (i.e. text, pictures, and video) form. The an-
syntactic features are: punctuations, # (hashtag), @ (user mention), notation of new unseen rumors at these platforms needs a deep
RT (retweet), part of speech, and abbreviations [50]. The purpose of knowledge of the domain. But it is not always possible to get the
identifying speech acts is to detect assertions which form the basis experts of a domain in a timely manner and low quality labels are
of rumors. By training and testing these classifiers: naïve Bayes, de- assigned to the posts under investigation. Due to these low qual-
cision tree, logistic regression and SVM, they observed that logis- ity labels, rumor detection systems fail to provide an accurate de-
tic regression outperforms the other three classifiers in detecting tection of rumors. Twitter overcome this issue by implementing a
M. Ahsan and M. Kumari / Online Social Networks and Media 14 (2019) 100050 9

semi-automatic strategy which uses an algorithm [52] to automat- The other rumor controlling approaches are: rumor rebuttals,
ically assign a credibility rating to each post and lets the users to reliable information from credible sources, exposing people to in-
give their feedback on these ratings. formation that refutes rumors, dropping messages from highly in-
fluential users, news media, and frequent update from emergency
6. Rumor control officials.

The spread of rumors causes a significant amount of harm in


6.2. Rebuttals
the form of financial loss, reputation loss and lives of people. This
potential has led to a long thread of research for providing ru-
It has been advocated by a number of studies that rumor re-
mor combatting mechanisms. The primary mechanism that gov-
buttals and refutation eliminate reliance on rumors to a degree
ernments have utilized to combat rumors is centralized where an
[13,14,56]. But according to psychological studies, these are ineffec-
authority broadcasts anti-rumor messages to the whole popula-
tive in decreasing belief on rumors. People build a mental model
tion [1]. But it is ineffective in situations where people mistrust
on the received information when alternative information is ab-
the motives of authority. Therefore, a strategy where anti-rumor
sent; any correction presented after model generation has no sig-
messages are not propagated by a centralized authority is the only
nificant effect on the people and they may prefer to keep their
feasible solution for combatting rumors. Tripathy et al. [53] have
models [57]. This persistent reliance is termed as ‘continued influ-
formulated three anti-rumor models: delayed start model, beacon
ence effect’. So, to enhance the effectiveness of rumor control, a
model and neighborhood model.
repetitive rebuttal or correction is required at the time of initial
exposure to rumorous information [58].
6.1. Anti-rumor models

These models utilize the trust (links) of social media users to 6.3. Credible sources
propagate anti-rumor messages rather than the power of authori-
ties. Oh et al. [19] have analyzed the effect of information released
from credible sources. They suggest that a credible information
6.1.1. Delayed start model contribute to subside anxiety of Twitter communities during ex-
This model is known as the decentralized method of rumor treme events. The reduction of anxiety, improves information qual-
combatting. It models the situation where a single authority de- ity and suppress groundless rumors.
tects the rumor spread after n time units and decides to start
anti-rumor process. First, a random infected node Vi is selected 6.4. Influential users
by the authority to start anti-rumor process and then Vi spreads
anti-rumor messages to its neighboring nodes Ni . Each neighbor- Sathanur et al. [59] study rumor control by dropping messages
ing node w|w ∈ Ni accepts anti-rumor with probability p (p is 0.05 from influential users. These users generally have a large num-
in [54]). ber of followers and mainly utilized by the marketing companies
The time between the start of rumor and start of an anti-rumor for rapidly distributing their information to the targeted audience.
process is a limitation of this model. In this time lag, rumors During elections, these are also misused for spreading propagandas
spread freely in the network. To overcome this problem, beacon and rumors. But in this study, they are targeted in an opposite way
model is developed. – to combat the rumor spread rather than spreading rumors. It has
been analyzed that dropping messages from these influential users
6.1.2. Beacon model
is helpful in controlling a rapid spread of rumorous information
It is a semi-centralized method and more proactive than the
from Twitter-like social media platforms.
delayed start model. Here, authority selects a set of vigilant nodes,
beacons, and embeds them in the network to continually lookup
6.5. Emergency officials and news media
for the rumors. The selection of beacons may be random or topic
specific (i.e. a beacon for technology may not be a beacon for
Jones et al. [60] suggest that emergency officials should spread
sport). On receiving a rumor, beacon immediately starts spread-
required information to the affected people in real time. It pre-
ing anti-rumor to its neighbors. So, in the beacon model, starting
vents people from getting anxious and spreading rumors. When-
time of anti-rumor process is not fixed and depends upon the time
ever possible, new information should be added in each update as
when a beacon receives the rumor.
it makes them feel that the information is not auto-generated and
In the delayed start model, the starting time of anti-rumor pro-
coming from official sources. News media are also advised to share
cess is fixed but in Beacon model, the location of the beacons is
a responsibility of spreading accurate information.
fixed in the network and start of anti-rumor process depends on
the time when beacons encounter some rumor.
7. Conclusion
6.1.3. Neighborhood model
It is fully decentralized where each node can detect rumor with With the rise in popularity of online social networks, the pen-
some probability and spread anti-rumor messages to the neighbor- etration of these platforms is also increased for spreading rumors.
ing nodes. In this model, beacons are not selected by the authority, Rumors can cause a significant harm to people in the form of fi-
but self-created during the spread of rumors. A beacon model with nancial and reputation loss. This survey provides a summary of ru-
b nodes as beacons and n total nodes is comparable with a neigh- mor causing factors and methods that are utilized to check the ve-
borhood model which have nodes refuting rumors with probability racity of shared information, to detect rumors and to control the
b
n. spread of rumors. From psychology, sociology and epidemiology
The spread of anti-rumors do not dominate rumors spread un- we came to know that anxiety, absence of adequate information
til a wide distribution of rumors [55]. This is the main reason of from legitimate users and activity pattern affect rumor diffusion
rumors spread even when non-rumors are present on social me- among people. This survey discusses how the veracity of the ru-
dia platforms because people more likely receive rumors than non- mors is checked with the help of human experts, machine learning
rumors. techniques and web services. It also explains the rumor detection
10 M. Ahsan and M. Kumari / Online Social Networks and Media 14 (2019) 100050

methods that are highly utilized for detecting rumors at online so- [23] L. Xia, G. Jiang, Y. Song, B. Song, Modeling and analyzing the interaction be-
cial networking platforms, especially Twitter. Finally, we listed ru- tween network rumors and authoritative information, Entropy 17 (1) (2015)
471–482, doi:10.3390/e17010471.
mor combatting methods in order to protect the people from the [24] J. Borge-Holthoefer, S. Meloni, B. Gonçalves, Y. Moreno, Emergence of influen-
chaotic situation created by the spread of rumorous information. tial spreaders in modified rumor models, J Stat Phys 151 (1–2) (2013) 383–393,
doi:10.1007/s10955- 012- 0595- 6.
[25] A.L. Barabasi, The origin of bursts and heavy tails in human dynamics, Nature
Declaration of Competing Interest 435 (7039) (2005) 207–211, doi:10.1038/nature03459.
[26] B. Gonçalves, J.J. Ramasco, Human dynamics revealed through web analytics,
Phys. Rev. E 78 (2) (2008) 026123, doi:10.1103/PhysRevE.78.026123.
None.
[27] A. Vazquez, B. Racz, A. Lukacs, A.L. Barabasi, Impact of non-Poissonian activity
patterns on spreading processes, Phys. Rev. Lett. 98 (15) (2007) 158702, doi:10.
Acknowledgment 1103/PhysRevLett.98.158702.
[28] N. Perra, B. Gonçalves, R. Pastor-Satorras, A. Vespignani, Activity driven mod-
eling of time varying networks, Sci Rep 25 (2) (2012) 469, doi:10.1038/
This work was supported by the Ministry of Electronics and srep00469.
Information Technology (MEITY), Government of India. We would [29] R.M. Tripathy, A. Bagchi, S. Mehta, Towards combating rumors in social net-
works: models and metrics, Intell. Data Anal. 17 (1) (2013) 149–175, doi:10.
like to thank the editor and the anonymous reviewers for their in- 3233/IDA-120571.
sightful comments. [30] R.L. Rosnow, Inside rumor: a personal journey, Am. Psychol. 46 (5) (1991) 484–
496 http://dx.doi.org/10.1037/0 0 03-066X.46.5.484.
[31] Y.L. Huang, K. Starbird, M. Orand, S.A. Stanek, H.T. Pedersen, Connected
References through crisis: emotional proximity and the spread of misinformation on-
line, in: Proceedings of the 18th ACM Conference on Computer Supported Co-
[1] P. Ozturk, H. Li, Y. Sakamoto, Combating rumor spread on social media: the operative Work & Social Computing, Vancouver, Canada, 2015, pp. 969–980,
effectiveness of refutation and warning, in: IEEE 48th Hawaii International doi:10.1145/2675133.2675202.
Conference on System Sciences (HICSS), Kauai, HI, USA, 2015, pp. 2406–2414, [32] A. Zubiaga, M. Liakata, R. Procter, G.W.S. Hoi, P. Tolmie, Analysing how peo-
doi:10.1109/HICSS.2015.288. ple orient to and spread rumours in social media by looking at conversational
[2] J. Li, H.R. Rao, Twitter as a rapid response news service: an exploration in the threads, PLoS ONE 11 (3) (2016) e0150989, doi:10.1371/journal.pone.0150989.
context of the 2008 China earthquake, Electron. J. Inf. Syst. Dev. Ctries. 42 (1) [33] J. Prasad, The psychology of rumour: a study relating to the great Indian earth-
(2017) 1–22, doi:10.1002/j.1681-4835.2010.tb0 030 0.x. quake of 1934, Br. J. Psychol. 26 (1) (1935) 1–15, doi:10.1111/j.2044-8295.1935.
[3] R. O’Connor, Facebook and Twitter are reshaping journalism as we know it, tb00770.x.
https://www.alternet.org/story/121211/, 2009. (Accessed 28 Jan 2019). [34] D. Sinha, Behaviour in a catastrophic situation: a psychological study of reports
[4] E. Mustafaraj, P.T. Metaxas, From obscurity to prominence in minutes: political and rumours, Br. J. Psychol. 43 (3) (1952) 200–209, doi:10.1111/j.2044-8295.
speech and real-time search, in: Proceedings of the WebSci 10: Extending the 1952.tb00343.x.
Frontiers of Society On-Line, Raleigh, NC: US, 2010. [35] M.E. Jaeger, S. Anthony, R.L. Rosnow, Who hears what from whom and with
[5] Online fact checking site. https://www.snopes.com. (Accessed 15 Mar 2019). what effect: a study of rumor, Soc.Cul. South Asia 6 (3) (1980) 473–478, doi:10.
[6] H.B. Dunn, C.A. Allen, Rumors, Urban legends and internet hoaxes, in: Proceed- 1177/014616728063024.
ings of the Annual Meeting of the Association of Collegiate Marketing Educa- [36] N. Bharosa, J. Lee, M. Janssen, Challenges and obstacles in sharing and
tors, 2005, p. 85. coordinating information during multi-agency disaster response: proposi-
[7] E. Bakshy, I. Rosenn, C. Marlow, L. Adamic, The role of social networks in in- tions from field exercises, Inf. Sys. Front. 12 (1) (2010) 49–65, doi:10.1007/
formation diffusion, in: Proceedings of the 21st ACM International Conference s10796- 009- 9174- z.
on World Wide Web, Lyon, France, 2012, pp. 519–528, doi:10.1145/2187836. [37] X. Qiu, L. Zhao, J. Wang, X. Wang, Q. Wang, Effects of time-dependent diffusion
2187907. behaviors on the rumor spreading in social networks, Phys. Lett.A 380 (24)
[8] D.M. Romero, B. Meeder, J. Kleinberg, Differences in the mechanics of informa- (2016) 2054–2063, doi:10.1016/j.physleta.2016.04.025.
tion diffusion across topics: idioms, political hashtags, and complex contagion [38] R.K. Garrett, Troubling consequences of online political rumoring, Hum. Com-
on Twitter, in: Proceedings of the ACM 20th International Conference on World mun. Res. 37 (2) (2011) 255–274, doi:10.1111/j.1468-2958.2010.01401.x.
Wide Web, Hyderabad, India, 2011, pp. 695–704, doi:10.1145/1963405.1963503. [39] Figure Eight – a high quality data annotation platform. https://www.
[9] S. Goel, D.J. Watts, D.G. Goldstein, The structure of online diffusion networks, figure-eight.com. (Accessed 26 Mar 2019).
in: Proceedings of the 13th ACM Conference on Electronic Commerce, Valencia, [40] A. Zubiaga, H. Ji, Tweet, but verify: epistemic study of information verifi-
Spain, 2012, pp. 623–638, doi:10.1145/2229012.2229058. cation on Twitter, Soc. Netw. Anal. Mining 4 (1) (2014) 163, doi:10.1007/
[10] G.W. Allport, L. Postman, An analysis of rumor, Publ. Opin. Q. 10 (4) (1946) s13278- 014- 0163- y.
501–517, doi:10.1093/poq/10.4.501. [41] J. Ratkiewicz, M. Conover, M. Meiss, B. Gonçalves, S. Patil, A. Flammini,
[11] N. DiFonzo, P. Bordia, Rumor psychology: social and organizational approaches, F. Menczer, Detecting and tracking the spread of astroturf memes in microblog
American Psychological Association, 2007, doi:10.1037/11503-000. streams, in: Proceedings of the 20th International Conference Companion on
[12] O. Oh, M. Agrawal, H.R. Rao, Community intelligence and social media ser- World Wide Web, 2010, pp. 249–252, doi:10.1145/1963192.1963301.
vices: a rumor theoretic analysis of Tweets during social crises, MIS Q. 37 (2) [42] A. Gupta, P. Kumaraguru, Credibility ranking of Tweets during high impact
(2013) 407–426. events, in: Proceedings of the ACM 1st Workshop on Privacy and Security in
[13] G.W. Allport, L.J. Postman, The psychology of rumor, J. Clin. Psychol. 3 Online Social Media, Lyon, France, 2012, p. 2, doi:10.1145/2185354.2185356.
(4) (1947) 402 New York, NY: Holt, Rinehart, & Winston, doi:10.1002/ [43] Amazon Mechanical Turk – a crowdsourcing marketplace where businesses
1097-4679(194710)3:4402::AID-JCLP22700304213.0.CO;2-T. and individuals outsource their jobs and processes to a distributed workforce
[14] P. Bordia, N. DiFonzo, Problem solving in social interactions on the internet: in a virtual manner. https://www.mturk.com. (Accessed 26 Mar 2019).
rumor as social cognition, Soc. Psychol. Q. 67 (1) (2004) 33–49, doi:10.1177/ [44] V. Qazvinian, E. Rosengren, D.R. Radev, Q. Mei, Rumor has it: identifying
019027250406700105. misinformation in microblogs, in: Proceedings of the Conference on Empiri-
[15] R.L. Rosnow, E.K. Foster, Rumor and gossip research, Psychol. Sci. Agenda 19 cal Methods in Natural Language Processing, Edinburg, United Kingdom, 2011,
(4) (2005) 1–2. pp. 1589–1599.
[16] S. Anthony, Anxiety and rumor, J. Soc. Psychol. 89 (1) (1973) 91–98, doi:10. [45] Z. Zhao, P. Resnick, Q. Mei, Enquiring minds: early detection of rumors in so-
1080/00224545.1973.9922572. cial media from enquiry posts, in: Proceedings of the 24th International Con-
[17] Taylor manifest anxiety scale – a test of anxiety as a personality trait. https: ference on World Wide Web, Florence, Italy, 2015, pp. 1395–1405, doi:10.1145/
//psychology-tools.com/taylor-manifest-anxiety-scale. (Accessed 22 Mar 2019). 2736277.2741637.
[18] J.M. Dahlhamer, J.M. Nigg, An Empirical Investigation of Rumoring: Anticipat- [46] S. Vosoughi, D. Roy, Tweet acts: a speech act classifier for Twitter, in: Proceed-
ing Disaster Under Conditions of Uncertainty, Disaster Research Centre, Univer- ings of the 10 th AAAI Conference on Web and Social Media ( ICWSM), Cologne,
sity of Delaware Library, 181 South College Avenue, Newark, USA, 1994. Germany, 2016, pp. 711–715.
[19] O. Oh, K.H. Kwon, H.R. Rao, An exploration of social media in extreme events: [47] Harvard inquirer – a computer assisted for content analysis of textual data.
rumor theory and Twitter during the Haiti earthquake 2010, in: ICIS 2010 Pro- http://www.wjh.harvard.edu/∼inquirer/. (Accessed 26 Mar 2019).
ceedings – Thirty First International Conference on Information Systems, 231, [48] Swear word list and curse filter – an API to filter swear words & adult lan-
Saint Louis, MO, US, 2010, pp. 7332–7336. guage. http://www.noswearing.com/dictionary. (Accessed 26 Mar 2019).
[20] Z. Qian, S. Tang, X. Zhang, Z. Zheng, The independent spreaders involved SIR [49] Emoticons – text based expressions. https://pc.net/emoticons/. (Accessed 26
rumor model in complex networks, Phys. A 429 (2015) 95–102, doi:10.1016/j. Mar 2019).
physa.2015.02.022. [50] NetLingo – list of internet acronyms & text message jargon. https://www.
[21] C. Castillo, M. Mendoza, B. Poblete, Information credibility on Twitter, in: Pro- netlingo.com/category/acronyms.php. (Accessed 27 Mar 2019).
ceedings of ACM 20th International Conference on World Wide Web, Hyder- [51] A. Zubiaga, M. Liakata, R. Procter, K. Bontcheva, P. Tolmie, Towards detecting
abad, India, 2011, pp. 675–684. rumours in social media, AAAI Workshop on AI for Cities, 2015.
[22] J. Ratkiewicz, M. Conover, M. Meiss, B. Gonçalves, S. Patil, A. Flammini, [52] A. Gupta, P. Kumaraguru, C. Castillo, P. Meier, TweetCred: real-Time credibility
F. Menczer, Truthy: mapping the spread of astroturf in microblog streams, in: assessment of content on Twitter, in: International Conference on Social Infor-
Proceedings of ACM 20th International Conference Companion on World Wide matics, Cham, Springer, 2014, pp. 228–243, doi:10.1007/978- 3- 319- 13734- 6_
Web, Hyderabad, India, 2011, pp. 249–252, doi:10.1145/1963192.1963301. 16.
M. Ahsan and M. Kumari / Online Social Networks and Media 14 (2019) 100050 11

[53] R.M. Tripathy, A. Bagchi, S. Mehta, Towards combating rumors in social net- [58] S. Lewandowsky, U.K. Ecker, C.M. Seifert, N. Schwarz, J. Cook, Misinformation
works: models and metrics, Intell. Data Anal. 17 (1) (2013) 149–175, doi:10. and its correction: continued influence and successful debiasing, Psycholo. Sci.
3233/IDA-120571. Pub. Interest 13 (3) (2012) 106–131, doi:10.1177/1529100612451018.
[54] R.M. Tripathy, A. Bagchi, S. Mehta, A study of rumor control strategies on social [59] A.V. Sathanur, M. Sui, V. Jandhyala, Assessing strategies for controlling viral
networks, in: Proceedings of the 19th ACM International Conference on Infor- rumor propagation on social media - A Simulation approach, in: IEEE Inter-
mation and Knowledge Management, Toronto, Canada, 2010, pp. 1817–1820, national Symposium on Technologies for Homeland Security (HST), Waltham,
doi:10.1145/1871437.1871737. USA, 2015, pp. 1–6, doi:10.1109/THS.2015.7225278.
[55] A. Friggeri, L.A. Adamic, D. Eckles, J. Cheng, Rumor cascades, The 8th Interna- [60] N.M. Jones, R.R. Thompson, C.D. Schetter, R.C. Silver, Distress and rumor expo-
tional AAAI Conference on Weblogs and Social Media (ICWSM), 2014. sure on social media during a campus lockdown, Proc. Nat. Acad. Sci. (PNAS)
[56] C.M. Seifert, The continued influence of misinformation in memory: what 114 (44) (2017) 11663–11668, doi:10.1073/pnas.1708518114.
makes a correction effective? Psychol. Lear. Moti. 41 (2002) (2002) 265–294,
doi:10.1016/S0 079-7421(02)80 0 09-3. Further Reading
[57] H.M. Johnson, C.M. Seifert, Sources of the continued influence effect: when
misinformation in memory affects later inferences, J. Exp. Psychol. 20 (6) M. Ahsan, M. Kumari, T.P. Sharma, Detection of context-varying rumors on Twitter
(1994) 1420–1436 http://dx.doi.org/10.1037/0278-7393.20.6.1420. through deep learning, Int. J. Adv. Sci. Technol. 128 (2019) 45–58, doi:10.33832/
ijast.2019.128.05.

You might also like