Professional Documents
Culture Documents
User Preference-Aware Fake News Detection: Yingtong Dou, Kai Shu, Congying Xia, Philip S. Yu, Lichao Sun
User Preference-Aware Fake News Detection: Yingtong Dou, Kai Shu, Congying Xia, Philip S. Yu, Lichao Sun
User Preference-Aware Fake News Detection: Yingtong Dou, Kai Shu, Congying Xia, Philip S. Yu, Lichao Sun
{ydou5,cxia8,psyu}@uic.edu,kshu@iit.edu,james.lichao.sun@gmail.com
2051
Short Research Paper II SIGIR ’21, July 11–15, 2021, Virtual Event, Canada
}
v1
!"#$%&'$()*#+,-#)%./01%234)56789:;<=><?@ABC"7>DCE01FGGG=H9'I-JKLMN6=OPQ+R%@S.$5L?BT/BS)=:UV7'EIW4N5TOX?Y)EX7%MZQ)/RZ4B@E%>5&.D[/\R]IT5D7&W^R2__-H_/2/I&?`33a68#`$Y[aB`D>3?VK2]VM;UYL_>7Ca'M^^\aT)^RIM[_:Q5L#MCC>W<^:5?`D4`4H7SJ*M8Hb#cTKN?%M&?&N..dab[LVN"X"9;SHJ9E=$O.BYT7*d.6A]<26>Oe>"bbZ-OZ`E=@-b"*-d8'[>9<f27eX?Q;e9cI">$MOa%LIR-f:;NK>MS<M>'`c`&@b>)b@d%7.=-?8)]S6&A]E-XS8aQ=8@ee$"UM4^O/4)NS6A:IKf'X&AP2WTf_4U&Ob??aT[7X4KWLJ8\[T_+?6S_D]?.+IDEDC^@G49@GJ=5GK[9'G=8M=Z='%/d9%VPU\%P3%&HT+ZE$YQ<GVD@EAG&aP]G00!3"#$%&'$F
Loss Function
Social Context Extraction v2 !"#$%&'$()*#+,-#)%./01$2345"6"78924$53:*;<;=>?@A201B>>>8CD'E-F5GHIJ8<KL+M%NO.$9GPQ:/QO)8;R4S'7ETUI9:<VPW)7VS%HXL)/MXUQN7%Y9&.Z3/6M[E:9ZS&T?M=\\-C\/=/E&P]^^_J`#]$W3_Q]ZY^P45=[4HaRWG\YSb_'H??6_:)?MEH3\;L9G#HbbYTc?;9P]ZU]UCSOF*H`C@#d:5IP%H&P&I..A_@3G4#OJ?Z+b_\^/Q#:;<KH"5`2#^6Jd-?:T??']*_:*c/QE_LGQ_`9>.?^@HS&JN&W*Tc/U?G?I:)7RUN]be">M?"DP"GF\=ZaEaGTa@')*+QV&9@$R-'^6U]3bM$_9M@::#?)"ad@&a@8R#^//MbYJ'[3SAb5AL$aTaX*=b.&:aMSHO9.[bF\;#DYf6&+\Ufe%WdP8LcT>OQK>M.Z>>5RIQIb<7Xac7%EeC73.6`_DO*ef4WN<f%:7^8??K[S00!^"#$%&'$B
User Engagement L
!"#$%&'$()*#+,-#)%./012345678729:;6<=;.>?@@A<B>.C01DCCCEF:'G-HI34JA&K3+8LM?NO'MPEKQEHRST'7.3-"U/OVC2=C/"SW-#=7XBY*9*I6=ZKW5[:6@&3Q=-AN=)$6HC/;IZQPU8\F4G#%Z.:=5*-L+]G.^/LP_R:P)`aEF%$L?[a\Y-YZab.JRc[F/S-b"9ZZ:V'9dbG>@YY]%R::9'4_>d6T*T8;WEb5aK]VE]Y-`#Ke'A7%-T:#<.aK2G9]);?'H%8'OX@BS@2`?;\[S\"cR<_;L6GW6]T+JRb]'P"2OYT]\9\BU'6[)"K"8M.8B=IY+RRc]74>SfIRO_QBP4334%')EZR*f+)T%Y`:9e&dbL2)TSfEJOU%YX9;ESdTX+JdRA[?3/+EYAK3>R\b"]*c9&6PSZQZ[2_5YWQ%RP+>BM%OH&f-Pb*EA/*L6</Ed#GC$>#CVEb:'VH:*B\^2K%LGNKOAKYeF5*]e<6fFC$F%b)C00!`"#$%&'$D
… EU
!"#$%&'$()*#+,-#)%./0123435)/6789:&;<9./=>?@A)?B)01CDDD7?E'F-G73H6I&JKL$E?GMGN+.O-#OP?K-;2Q)'RO&S$)TL=G>UL6-V>LTKORWPNXOS&FG<X/%8X/-UY4ZL5PS/Z2%[[Z\6Q]\-G\YL^")-_9$?&`YHY5?7/B29%aN$P#Pb$]@>>5G[-_V@D@=_L4/?H#;/88>$73a-RBM9/>=\#`/33aJ78J%O`4&SQLGa$T?]%-Z=Q3^&#L9O.7+/RT>D8"#SXKEc^*K_"3*OIB#d\6YIW__"T_JU+VP@L=#4[A7bd=\GaO4.'7-@[$[:+4eQ>S=Tf;S*cP?b)a&+$I/$KU&3Vad.)Y:X?&%#=A<b_I'H]UGeA>QH8>^J)+X`UaV;[:f#S>V'$"-J`"8*4V&OB>I53TZ+UV?E5I#3c%*EV@%9_>W/]\a/`bfRbbVYJIAcOD6STB/*NB2a'XaRXa&-OU/MF6PXb2\M]ZRRQ?<!X"#$%&'$C
News Piece vn
!"#$%&'$()*#+,-#)%./01234*56#789$7#:;#$<=>7?@A5B=01CDDDEFA'G-HE5IJ=&3:AKL@MN+OPQ8RS&=$T"UQVT%NW='QJX7:EY.Z:S7+A-L[\]@^)@%K?R+]4R_Q`R[I_V6:::Q$.->G>@=P]UW`a`b\:;.&Y-24\$J)FYANBE@a^V9#533bV[3X`ZP@^#G']<JcSS#I$2K=YD_+XJY<AB;IKHdTQ>JZ&'.[)@T^&TEE);e]-?NH8%%55>;+*>O6/G#:H/Wf"E*\9N65Q$8>M8HE;^)bXbc/75\7:"WPK_2Y[dd)#XRe^Db'>H2KS$<4PS<6RUO\_Z8)925W'Y4<B=)bE.=[L-O-D].dHN:NF\b8=IQN)<TAK'9^fL82?^TX\MF?NZ]/VfJ>%2XOOXN828=H#XW:Y8cGE-4\'H5O)HFFL'A"bL$]))RI>GJ^dG[JbQZ=UF93US^D*/*"b/GB.GK6ab6H'_B<&)B*R6Je\@DI>3R.G0!4"#$%&'$C
1
Embedding
vn !"#$%&'$()*#+,-#)%./0123456789-':;<=>?5@47A9BC<;)01DEEEFGH'I-J3KL6MF9N>+O%PB.$QKRST/SB)FU:VW'XICY6QT9ZR[)XZW%L7>)/O7YSPX%\Q&.]=/^O@ITQ]W&C_O`aa-Ga/`/I&R<bbAM2#<$[=AS<]\bRV3`@VLc:[Ka\WdA'L__^AT)_OIL=aU>QK#Ldd\C4_UQR<]Y<YGWBJ*L2Ge#;T36R%L&R&6..5Ae=KVY_>LKYH^aBXE:'=Fd)aW`^9]88;O'aLSLf)/MaR-@CfMP.7$CY>N'B<;UELdP5GLBaPLSABS-QQ43\U.eRF'?J+T'&-aG/fPSANZC[_=E$45:8HZ/[^%MGH&-e6&Z#L:LBTI\_II:RHZ].PFUY5RV[;99X<IBR'P)P-94`5IF)Qa&@4fJ#)O++Hb)J#4<@J?\NEIK4E_EB'FN2&EEY?ERE>VXE?][]T%]Q8]R^Z./%;>ebEQ<4IVHX.VMW00!b"#$%&'$D
Representation
News Textual
y !"#$%&'$()*#+,-#)%./0123456789:;#:<.=>42?7@)6'A:B01CAAA=.D'B-E=FGHIAJIK'?LMD.$NO;3P&;?Q8Q2%'?/HR=STUMV$"W*@XXX<K#YMZ[QMNOX\HLQ)HLTYBQAB?D]OI^K)]T<6DIMI7/_7[GT?$[:F]\_*VX$`VX]HQN&.+V<a\*'2:'+*+A\b;BM6$a/KAG\5Z;QNA&X=Hcdc[`+=bN))N6^DQ;KZQ%BU<FE#.7X-E[7`'+$ae^W7&B"5=Nc+LL#)K'?6#QG;cS5X_`b)R[25E/S]U$F;OF8#S3%6Z@e<5f\NXIdFGbLGc5Uc8&)\:Ff8@db.R2J37OAb)<2GF8WY_cdJd7e@#Hc\[BIT6='E-OAbGZS&c<`LGAE[M3cU)QS&&%^$*M.QQc]RVT/fKd4J\#VV7K6DF9#$Df$Vef6S/UEc/=aH@QZfK2MZ:c=3Wc_]WT]FXDb&KeX4;:`P^T:4/AL[]=&2M3S)0!d"#$%&'$C
Figure 1: The proposed UPFD framework for user preference-aware fake news detection. Given the news piece and its engaged users on social
media, we extract the exogenous context as a news propagation graph and encode the endogenous information based on user historical posts
and news texts. The endogenous and exogenous information are fused using a GNN encoder. The final news embedding, composed of user
engagement embedding and news textual embedding, is fed into the neural classifier to predict the news’ credibility.
• We study a novel problem of user preference-aware fake news the FakeNewsNet dataset [25] which contains news content and its
detection on social media; social engagement information on Twitter. Then we use the Twitter
• We propose a principled way to exploit both endogenous prefer- Developer API [4] to crawl historical tweets of all accounts that
ence and exogenous context jointly to detect fake news; and retweeted the news in FakeNewsNet.
• We conduct extensive experiments on real-world datasets to To obtain rich historical information for user preference model-
demonstrate the effectiveness of UPFD for detecting fake news. ing, we crawl the recent two hundred tweets for each account, so
as to near 20 million tweets being crawled in total. For inaccessible
2 OUR APPROACH users whose accounts are suspended or deleted, we use randomly
In this section, we present the details of the proposed framework sampled tweets from accessible users engaging the same news as
for fake news detection named UPFD (User Preference-aware Fake its corresponding historical posts. Because deleting the inaccessible
News Detection). As shown in Figure 1, our framework has three users will break the intact news propagation cascading and results
major components. First, given a news piece, we crawl the historical in a less effective exogenous context encoder. We also remove the
posts of the users engaged in the news to learn user endogenous special characters, e.g., “@” characters and URLs, before applying
preference. We implicitly extract the preferences of engaged users text representation learning methods.
by encoding historical posts using text representation learning tech- To encode the news textual information and user preferences, we
niques (e.g., word2vec [16], BERT [5]). The news textual data is employ two types of text representation learning approaches based
encoded using the same approach. Second, to leverage user exoge- on language pretraining. Instead of training on the local corpus,
nous context, we build the news propagation graph according to its the word embeddings pretrained on a large corpus are supposed
engagement information on social media platforms (e.g., retweets to encode more semantic similarities between different words and
on Twitter). Third, we devise a hierarchical information fusion pro- sentences. For pretrained word2vec vectors, we choose the 680k
cess to fuse the user endogenous preference and exogenous context. 300-dimensional vectors pretrained by spaCy [10]. We also employ
Specifically, we obtain the user engagement embedding using GNN pretrained BERT embeddings to encode the historical tweets and
as the graph encoder, where the news and user embeddings en- news content as a sequence [5] using bert-as-a-service [32].
coded by the text encoder are used as their corresponding node Next, we elaborate the details of applying the above text rep-
features in the news propagation graph. The final news embeddings resentation learning models. spaCy includes pretrained vectors
are composed by the concatenation of user engagement embedding for 680k words, and we average the vectors of existing words in
and news textual embedding. combined recent 200 tweets to get user preference representation.
Next, we will introduce how we encode endogenous preference, The news textual embedding is obtained similarly. For the BERT
extract the exogenous context, and fuse both information. model, we use the cased BERT-Large model to encode the news and
user information. The news content is encoded using BERT with
maximum input sequence length (i.e., 512 tokens). Due to BERT’s
2.1 Endogenous Preference Encoding
input sequence length limitation, we could not use BERT to encode
It is non-trivial to explicitly model the endogenous preference of a 200 tweets as one sequence, so we resort to encode each tweet sep-
user only using his/her social network information. Similar to [2, 13, arately and average them afterward to obtain a user’s preference
20] which model the users’ personality, sentiment and stance using representation. Generally, the tweet text is way shorter than the
their historical posts, we leverage the historical posts of a user to news text, we empirically set the max input sequence length of
encode his/her preference implicitly. However, none of the previous BERT as 16 tokens to accelerate the tweets encoding time.
fake news datasets contain such information. In this paper, we select
2052
Short Research Paper II SIGIR ’21, July 11–15, 2021, Virtual Event, Canada
2053
Short Research Paper II SIGIR ’21, July 11–15, 2021, Virtual Event, Canada
Politifact Gossipcop
Table 2: The fake news detection performance of baselines 0.86
0.96
ACC ACC
and our model. Stars denote statistically significant under 0.84
F1
0.94
F1
Value
Value
0.80
0.90
0.78
0.88
0.76
POL GOS 0.86
Model 0.74
2054
Short Research Paper II SIGIR ’21, July 11–15, 2021, Virtual Event, Canada
REFERENCES zenodo.1239675
[1] Anton Abilov, Yiqing Hua, Hana Matatov, Ofra Amir, and Mor Naaman. 2021. [20] Jing Qian, Mai ElSherief, Elizabeth M Belding, and William Yang Wang. 2018.
VoterFraud2020: a Multi-modal Dataset of Election Fraud Claims on Twitter. Leveraging intra-user and inter-user representation learning for automated hate
arXiv preprint arXiv:2101.08210 (2021). speech detection. arXiv preprint arXiv:1804.03124 (2018).
[2] Nadeem Ahmad and Jawaid Siddique. 2017. Personality assessment using Twitter [21] Francisco Rangel, Anastasia Giachanou, Bilal Ghanem, and Paolo Rosso. 2020.
tweets. Procedia computer science 112 (2017), 1964–1973. Overview of the 8th Author Profiling Task at PAN 2020: Profiling Fake News
[3] Shantanu Chandra, Pushkar Mishra, Helen Yannakoudakis, and Ekaterina Spreaders on Twitter. In CLEF.
Shutova. 2020. Graph-based Modeling of Online Communities for Fake News [22] Lee Ross and Andrew Ward. 1995. Naive realism in everyday life: Implications
Detection. arXiv preprint arXiv:2008.06274 (2020). for social conflict and misunderstanding. (1995).
[4] Twitter Developer. 2021. Twitter API. https://developer.twitter.com/. [23] Natali Ruchansky, Sungyong Seo, and Yan Liu. 2017. Csi: A hybrid deep model
[5] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: for fake news detection. In CIKM.
Pre-training of deep bidirectional transformers for language understanding. arXiv [24] Kai Shu, H Russell Bernard, and Huan Liu. 2019. Studying fake news via net-
preprint arXiv:1810.04805 (2018). work analysis: detection and mitigation. In Emerging Research Challenges and
[6] Matthias Fey and Jan E. Lenssen. 2019. Fast Graph Representation Learning with Opportunities in Computational Social Network Analysis and Mining. Springer,
PyTorch Geometric. In ICLR Workshop on Representation Learning on Graphs and 43–65.
Manifolds. [25] Kai Shu, Deepak Mahudeswaran, Suhang Wang, Dongwon Lee, and Huan Liu.
[7] William L Hamilton, Rex Ying, and Jure Leskovec. 2017. Inductive representation 2020. Fakenewsnet: A data repository with news content, social context, and
learning on large graphs. In NeurIPS. spatiotemporal information for studying fake news on social media. Big Data 8,
[8] Yi Han, Shanika Karunasekera, and Christopher Leckie. 2020. Graph Neural 3 (2020), 171–188.
Networks with Continual Learning for Fake News Detection from Social Media. [26] Kai Shu, Deepak Mahudeswaran, Suhang Wang, and Huan Liu. 2020. Hierarchical
arXiv preprint arXiv:2007.03316 (2020). propagation networks for fake news detection: Investigation and exploitation. In
[9] Naeemul Hassan, Gensheng Zhang, Fatma Arslan, Josue Caraballo, Damian Proceedings of the International AAAI Conference on Web and Social Media.
Jimenez, Siddhant Gawsane, Shohedul Hasan, Minumol Joseph, Aaditya Kulkarni, [27] Kai Shu, Amy Sliva, Suhang Wang, Jiliang Tang, and Huan Liu. 2017. Fake
Anil Kumar Nayak, et al. 2017. ClaimBuster: the first-ever end-to-end fact- News Detection on Social Media: A Data Mining Perspective. KDD exploration
checking system. VLDB (2017). newsletter (2017).
[28] Lichao Sun, Kazuma Hashimoto, Wenpeng Yin, Akari Asai, Jia Li, Philip Yu, and
[10] Matthew Honnibal and Ines Montani. 2017. spacy 2: Natural language understand-
Caiming Xiong. 2020. Adv-BERT: BERT is not robust on misspellings! Generating
ing with bloom embeddings, convolutional neural networks and incremental
nature adversarial samples on BERT. arXiv preprint arXiv:2003.04985 (2020).
parsing. To appear 7, 1 (2017).
[29] Lichao Sun, Congying Xia, Wenpeng Yin, Tingting Liang, Philip S Yu, and Lifang
[11] Rohit Kumar Kaliyar, Anurag Goswami, and Pratik Narang. [n.d.]. FakeBERT:
He. 2020. Mixup-Transfomer: Dynamic Data Augmentation for NLP Tasks. arXiv
Fake news detection in social media with a BERT-based deep learning approach.
preprint arXiv:2010.02394 (2020).
Multimedia Tools and Applications ([n. d.]), 1–24.
[30] William Yang Wang. 2017. " Liar, Liar Pants on Fire": A New Benchmark Dataset
[12] Hamid Karimi, Proteek Roy, Sari Saba-Sadiya, and Jiliang Tang. 2018. Multi-
for Fake News Detection. arXiv preprint arXiv:1705.00648 (2017).
Source Multi-Class Fake News Detection. In COLING.
[31] Yaqing Wang, Fenglong Ma, Zhiwei Jin, Ye Yuan, Guangxu Xun, Kishlay Jha, Lu
[13] Anupam Khattri, Aditya Joshi, Pushpak Bhattacharyya, and Mark Carman. 2015.
Su, and Jing Gao. 2018. EANN: Event Adversarial Neural Networks for Multi-
Your sentiment precedes you: Using an authorâĂŹs historical tweets to pre-
Modal Fake News Detection. In KDD.
dict sarcasm. In Proceedings of the 6th workshop on computational approaches to
[32] Han Xiao. 2018. bert-as-service. https://github.com/hanxiao/bert-as-service.
subjectivity, sentiment and social media analysis. 25–30.
[33] Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. 2018. How powerful
[14] Thomas N Kipf and Max Welling. 2017. Semi-supervised classification with graph
are graph neural networks? arXiv preprint arXiv:1810.00826 (2018).
convolutional networks. In ICLR.
[34] Zhitao Ying, Jiaxuan You, Christopher Morris, Xiang Ren, Will Hamilton, and Jure
[15] Yi-Ju Lu and Cheng-Te Li. 2020. GCAN: Graph-aware Co-Attention Networks
Leskovec. 2018. Hierarchical graph representation learning with differentiable
for Explainable Fake News Detection on Social Media. In Proceedings of the 58th
pooling. In Advances in neural information processing systems.
Annual Meeting of the Association for Computational Linguistics.
[35] Ye Zhang and Byron Wallace. 2015. A sensitivity analysis of (and practitioners’
[16] Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient
guide to) convolutional neural networks for sentence classification. arXiv preprint
estimation of word representations in vector space. arXiv preprint arXiv:1301.3781
arXiv:1510.03820 (2015).
(2013).
[36] Xinyi Zhou, Jindi Wu, and Reza Zafarani. 2020. SAFE: Similarity-Aware Multi-
[17] Federico Monti, Fabrizio Frasca, Davide Eynard, Damon Mannion, and Michael M
modal Fake News Detection. In Pacific-Asia Conference on Knowledge Discovery
Bronstein. 2019. Fake news detection on social media using geometric deep
and Data Mining. Springer, 354–367.
learning. ICLR Workshop (2019).
[37] Xinyi Zhou and Reza Zafarani. 2020. A survey of fake news: Fundamental
[18] Raymond S Nickerson. 1998. Confirmation bias: A ubiquitous phenomenon in
theories, detection methods, and opportunities. ACM Computing Surveys (CSUR)
many guises. Review of general psychology 2, 2 (1998), 175.
53, 5 (2020), 1–40.
[19] Martin Potthast, Johannes Kiesel, Kevin Reinartz, Janek Bevendorff, and Benno
Stein. 2018. BuzzFeed-Webis Fake News Corpus 2016. https://doi.org/10.5281/
2055