ArticlePDF Available

Verifying information with multimedia content on twitter: A comparative study of automated approaches

Springer Nature
Multimedia Tools and Applications
Authors:

Abstract and Figures

An increasing amount of posts on social media are used for dissem- inating news information and are accompanied by multimedia content. Such content may often be misleading or be digitally manipulated. More often than not, such pieces of content reach the front pages of major news outlets, having a detrimental eect on their credibility. To avoid such eects, there is profound need for automated methods that can help debunk and verify online content in very short time. To this end, we present a comparative study of three such methods that are catered for Twitter, a major social media platform used for news sharing. Those include: a) a method that uses textual patterns to extract
Content may be subject to copyright.
Noname manuscript No.
(will be inserted by the editor)
Verifying Information with Multimedia Content on
Twitter
A Comparative Study of Automated Approaches
Christina Boididou ·Stuart E.
Middleton ·Zhiwei Jin ·Symeon
Papadopoulos ·Duc-Tien Dang-Nguyen ·
Giulia Boato ·Yiannis Kompatsiaris
Received: date / Accepted: date
Abstract An increasing amount of posts on social media are used for dissem-
inating news information and are accompanied by multimedia content. Such
content may often be misleading or be digitally manipulated. More often than
not, such pieces of content reach the front pages of major news outlets, having
a detrimental effect on their credibility. To avoid such effects, there is profound
need for automated methods that can help debunk and verify online content
in very short time. To this end, we present a comparative study of three such
methods that are catered for Twitter, a major social media platform used for
news sharing. Those include: a) a method that uses textual patterns to extract
C. Boididou
Information Technologies Institute, CERTH
E-mail: boididou@iti.gr
S. E. Middleton
University of Southampton IT Innovation Centre
E-mail: sem@it-innovation.soton.ac.uk
Z. Jin
University of Chinese Academy of Sciences
E-mail: jinzhiwei@ict.ac.cn
S. Papadopoulos
Information Technologies Institute, CERTH
E-mail: papadop@iti.gr
D. Dang-Nguyen
University of Trento; Dublin City University
E-mail: dangnguyen@disi.unitn.it; duc-tien.dang-nguyen@dcu.ie
G. Boato
University of Trento
E-mail: giulia.boato@unitn.it
Y. Kompatsiaris
Information Technologies Institute, CERTH
E-mail: ikom@iti.gr
2 Christina Boididou et al.
claims about whether a tweet is fake or real and attribution statements about
the source of the content; b) a method that exploits the information that same-
topic tweets should be also similar in terms of credibility; and c) a method
that uses a semi-supervised learning scheme that leverages the decisions of two
independent credibility classifiers. We perform a comprehensive comparative
evaluation of these approaches on datasets released by the Verifying Multime-
dia Use (VMU) task organized in the context of the 2015 and 2016 MediaEval
benchmark. In addition to comparatively evaluating the three presented meth-
ods, we devise and evaluate a combined method based on their outputs, which
outperforms all three of them. We discuss these findings and provide insights
to guide future generations of verification tools for media professionals.
Keywords Fake Detection ·Verification ·Credibility ·Veracity ·Trust ·
Social Media ·Twitter ·Multimedia
1 Introduction
Recent years have seen a tremendous increase in the use of social media plat-
forms such as Twitter and Facebook as a means of sharing news content and
multimedia, and as a source and sensor of trends and events [1]. The simplicity
of the sharing process has led to large volumes of news content propagating
over social networks and reaching huge numbers of readers in very short time.
Especially multimedia posts (images, videos) can very quickly reach huge au-
diences and become viral due to the fact that they are easily consumed.
Given the speed of the news spreading process and the competition of news
outlets and individual news sources to publish first, the verification of informa-
tion and content is often carried out in a superficial manner or even completely
neglected. This leads to the appearance and spread of large amounts of fake
media content. In particular, when a news event breaks (e.g., a natural dis-
aster), and new information and media coverage is of primary importance,
news professionals turn to social media to source potentially interesting and
informative content. It is exactly this setting, when the risk of fake content be-
coming widely disseminated is the highest. By fake, we refer to any publication
or post with multimedia content that does not represent accurately the event
that it refers to. It may be reposted content falsely associated with a current
event, digitally manipulated content, computer-generated imagery presented
as real imagery or speculations regarding the association of persons with a
current event. In a similar way, by real we refer to posts with content that
rightly represent the event they claim to. There are also posts that, despite
sharing fake multimedia content, explicitly report that the content is fake (e.g.
to warn readers) or they refer to it with a sense of humour; those are excluded
from our study.
An example of reposted content is a widely shared photo of two young
children hugging each other (Fig. 1 (a)). The image was claimed to depict
a brother who is protecting his little sister during the earthquake in Nepal
(April 2015), but it was later reported that it was taken a decade ago in
Verifying Information with Multimedia Content on Twitter 3
Fig. 1: (a) Siblings’ photo from Vietnam reposted as being from Nepal earth-
quake. (b) Fake media content spread during the Malaysian airlines breaking
news story.
a province of Vietnam by a professional photographer. In some cases, the
consequences of fake content reaching a very large part of the population can
be quite severe. For example, fake images became popular on social media
after the Malaysia Airlines passenger flight disappeared on 8th March (Fig.
1 (b) illustrates an example). During the investigation of the plane trace,
false alarms that the plane was detected came up. Taking into account how
sensitive the case was, the circulation of this content deeply affected the people
directly involved in it, such as the families of the passengers, causing emotional
distress. An extended analysis on rumour propagation during the London riots
[32] concluded that rumours typically start with someone tweeting about an
incident, which then gets re-tweeted and reposted in a number of variations.
An interactive representation1of the riots’ rumours across time shows the
velocity with which fake information propagated and the fact that hours, even
days are needed to debunk such false claims.
There are several challenges that journalists face in the process of assess-
ing the veracity of user-generated content. Notably, “traditional” digital media
verification techniques employed by journalists [34], e.g. looking into the Exif
metadata2of content or getting in touch with the person that published it,
are often not possible or very slow due to the characteristics of social me-
dia platforms. For instance, Twitter and Facebook remove the Exif metadata
from posted images, and Twitter accounts in most cases provide no contact
information (e.g., email, telephone number). Furthermore, conventional image
forensics approaches are hardly applicable due to the image resizing and re-
compression operations that are automatically applied by these social media
platforms to all uploaded content [38].
1http://www.theguardian.com/uk/interactive/2011/dec/07/london-riots- twitter
2Exif metadata contain information about the date, time and location an image was
taken, the model of the device, and copyright information, which can be very useful when
assessing the credibility of multimedia content [34].
4 Christina Boididou et al.
The above challenges highlight the need for novel tools that can help news
professionals assess the credibility of online content. To this end, we present
and compare three automated approaches to solve this problem on Twitter.
This study is based on the Verifying Multimedia Use (VMU) task [4] that was
organized as part of the 2015 and 2016 editions of the MediaEval initiative3,
with the goal to benchmark methods on the problem of automatically pre-
dicting whether a tweet that shares multimedia content is misleading (fake)
or trustworthy (real). The presented approaches [7,21,26] are the ones that
competed in this task and include a) a method that uses attribution in tan-
dem with fake and genuine claim extraction, b) a method that verifies tweets
by exploring inter-tweet information, and c) one that uses a semi-supervised
learning scheme. This study conducts and presents a comprehensive compar-
ison between them in a more extended experimental setting, and draws ac-
tionable insights regarding the strengths and weaknesses of each method. In
particular, the main contributions of this article include the following:
We describe a new benchmark dataset based on a revised and extended
version of the MediaEval dataset. Our revised dataset has duplicate and
near-duplicate tweets removed and its cross-event balance improved to re-
duce content bias towards more popular events. This dataset is publicly
available for other researchers to use for benchmarking.
We report a new experimental analysis using this new benchmark dataset
following a leave-one-event-out cross-validation scheme. Each of our fake
classification approaches is evaluated and its ability to predict the veracity
of content analysed and contrasted.
We present results derived from an ensemble of the three fake classification
approaches. We analyze the advantages of the ensemble-based approach
and show that it is more effective in classifying fake content than each
individual method on its own. This result provides a benchmark for other
researchers to compare against in the future.
2 Background
Classifying online content with respect to its credibility and veracity is a highly
complex problem that has been studied in multiple settings and using a variety
of approaches. Our work focuses on the problem of single post verification,
i.e. classifying an individual content item as being fake or real. This is in
contrast to the related problem of rumour detection [41], which considers that a
piece of false information is spreading across social networks. Although rumour
detection is a highly relevant research problem and several methodological
aspects are common with those arising in our problem setting, the following
discussion is mostly focusing on single post verification approaches.
Several of the previous studies in the area focused on the statistical analysis
on social media with the goal of extracting features that can be used as robust
3http://multimediaeval.org/
Verifying Information with Multimedia Content on Twitter 5
verification indicators for classifying social media content (sec. 2.1). Another
field of study concerns methods for assessing the credibility of the source (or
user account), where a post of interest originates (sec. 2.2). A different re-
search area concerns the development of image forensics approaches that can
potentially provide valuable complementary signals to the verification process
(sec. 2.3). We also present in sec. 2.4 a few systems that attempt to solve
the problem by leveraging methods and results from the works of sec. 2.1-2.3.
Finally, we provide a description of the Verifying Multimedia Use (VMU) task
(sec. 2.5) that constitutes the basis for our experimental study.
2.1 Verification cues for social media content
Several previous studies focus on the automatic extraction of credibility cues
from the content of social media posts, either by using Natural Language
Processing from the posts’ text or by extracting other features. For instance,
Castillo et al. [11] presented a supervised learning method to assess content
credibility in Twitter. They extract discussion topics that are categorized as
news or chat by human annotators and they build a model to automatically de-
termine which topics are newsworthy by assigning a credibility label to them.
Martinez-Romo et al. [25] retrieve tweets associated with trending topics and
use blacklists to detect spam URLs in them. Then, they introduce language
models based on probability distributions over pieces of text and by adding
content features, they apply several models to evaluate their approach, which
achieves high accuracy in classifying tweets with spam content. O’Donovan et
al. [29] performed an analysis of the utility of various features when predict-
ing content credibility. First, they collected Twitter data derived from very
different contexts and they defined a set of features including content-based
features, user profile features, and others that focus on the dynamics of infor-
mation flow. Then, by checking the distribution of each feature category across
Twitter topics, they concluded that their usefulness can greatly vary with con-
text, both in terms of the occurrence of a particular feature, and the manner
in which it is used. The work in [15], which is very similar in terms of objective
to the VMU task that we study in this paper, tries to distinguish between fake
and real images shared on Twitter by using decision tree-based classification
models on tweet text and Twitter account features. Using Hurricane Sandy as
the evaluation dataset, they report a 97% detection accuracy.
2.2 Source and user credibility on social networks
Several approaches focus on the study of user behaviour on social networks as
well as on the detection of spam accounts. Stringhini et al. [36] investigated
techniques for automated identification of spam accounts on Twitter by detect-
ing anomalous behaviour. They used six features (friend-follower ratio, URL
ratio in messages, similarity of messages sent by a user, friend choice, mes-
sages sent, friend number) in a classification model and managed to identify
6 Christina Boididou et al.
about 15,000 spam accounts on Twitter. Canini et al. [9] proposed a method,
that, given a particular topic, identifies relevant users, based on a combina-
tion of their expertise and trust. The authors employed an LDA topic model,
using keywords extracted from the user profile, to estimate the association
between a user account and a topic and then rank them based on their credi-
bility. Additionally, Starbird et al. [35] examined an automated mechanism for
identifying Twitter accounts providing eyewitness testimony from the ground.
They used profile features, such as number of statuses and of followers, and
features that describe how the Twitter community interacts with the user
during the event. Finally, they applied an SVM classifier with asymmetric soft
margins and they managed to achieve promising results on the task. Two of
the methods proposed in this work also use features derived from social media
accounts (including newly proposed features) with the aim of finding common
characteristics of users that tend to share misleading content.
2.3 Image forensics
Image forensics has been long used for assessing the authenticity of images by
detecting whether a digital image has been manipulated. Image manipulation
is typically classified as splicing (transferring an object from an image and
injecting it into another), copy-move (copying an object from the same image
to a different position) or retouching (enhancing contrast, sharpening edges or
applying color filters). These manipulations normally leave digital traces that
forensics methods try to detect. The method in [14] exploits inconsistencies
in the Color Filter Array (CFA) interpolation patterns, allowing for accurate
splice localization. Since the most common format of digital images is JPEG,
numerous methods try to exploit traces left by the JPEG compression pro-
cess. In [24] and [31], different methods are proposed to determine whether an
image was previously JPEG compressed. In [3], original and forged regions are
discriminated in double compressed images for both aligned (A-DJPG) and
non-aligned JPG (NA-DJPG). Double and triple compressions are detected
and discriminated in [30], by exploiting the analysis of the Benford-Fourier
coefficients. A method for tampered regions detection was proposed in [23] on
the block artifact grids (BAG), which are caused by the block-based processing
during JPEG compression and are usually mismatched in copy-move or splic-
ing manipulations. Other methods aim to detect non-native JPEG images by
analyzing quantization tables, thumbnails and information embedded in Exif
metadata [22]. For copy-move detection, many methods have been proposed
in the recent years, mainly based on the matching of keypoints [2], or regions
[18]. For detecting image retouching, most current methods exploit illumina-
tion or shadow inconsistencies [28], or geometric relations disagreement [12]
within an image. Despite the proliferation of image forensics methods, a recent
experimental study [39] has concluded that many of them are ineffective on
real cases of manipulated images sourced from the Web and social media due
to the fact that such images typically undergo multiple resaving operations
Verifying Information with Multimedia Content on Twitter 7
that destroy a considerable part of the forensic traces. Nonetheless, one of
the compared methods in our study makes use of forensics features (extracted
from the image accompanying the tweet) as additional verification signals.
2.4 Systems for assessing content credibility
In the context of assessing content credibility, Ratkiewicz et al. developed the
Truthy system [33] for real-time tracking of political memes on Twitter and
for detecting misinformation, focusing on political astroturf. Truthy collects
tweets, detects memes and introduces a web interface that lets users anno-
tate the memes they consider truthful. Another system for evaluating Twitter
content is TweetCred [16], a system that computes for each tweet a credibil-
ity score. It takes the form of a Web application that can be installed as a
Chrome extension. The system encourages users to give feedback by declaring
whether they agree or not with the produced score. We include TweetCred in
our comparative experimental study as a state-of-the-art method.
2.5 Verifying Multimedia Use (VMU) task
To assess the effectiveness of automated tweet verification methods, we rely on
resources produced by the VMU task [4], which was introduced in 2015 as part
of the MediaEval benchmarking initiative. The definition of the task is the fol-
lowing: “Given a tweet and the accompanying multimedia item (image or video)
from an event of potential interest for the international news audience, return
a binary decision representing verification of whether the multimedia item re-
flects the reality of the event in the way purported by the tweet.” In practice,
participants received a list of tweets that include images or video and were
required to automatically predict, for each tweet, whether it is trustworthy
or deceptive (real or fake respectively). An unknown label is also accepted
in case that there is no available prediction for a tweet. In addition to fully
automated approaches, the task also considered human-assisted approaches
provided that they are practical (i.e., fast enough) in real-world settings, such
as manually identifying the veracity of a multimedia item by searching on
trustworthy online websites or resources. The following considerations should
be made in addition to the above definition:
A tweet is considered fake when it shares multimedia content that does not
faithfully represent the event it refers to. The variety of untrustworthy and
misused content appearing in the context of past events led us to devise a
small typology of misleading use of multimedia content (see Fig. 2).
A tweet is considered to be real when it shares multimedia that accurately
represents the event it refers to.
A tweet that shares content that does not represent accurately the event
it refers to but reports the false information or refers to it with a sense of
humour is neither considered fake nor real (and hence not included in
the datasets released by the task).
8 Christina Boididou et al.
Fig. 2: Different types of misleading multimedia use. From left to right: a) re-
posting an old photo showing soldiers guarding the Tomb of the Unknown Sol-
dier claiming it was captured during the Hurricane Sandy in 2012, b) reposting
digital artwork as a photo from the solar eclipse in March 2015, c) specula-
tion of depicted people as being suspects of the Boston Marathon bombings
in 2013, d) spliced sharks on a photo captured during the Hurricane Sandy.
For each tweet, the task has also released three types of feature:
tweet-based (TB-base): Extracted from the tweet itself, e.g. the number of
terms, the number of mentions and hashtags, etc. [8].
user-based (UB-base): Based on the Twitter profile, e.g. the number of
friends and followers, the account age, whether the user is verified, etc. [8].
forensics (FOR): Forensic features extracted from the visual content of the
tweet image, and specifically the probability map of the aligned double
JPEG compression, the potential primary quantization steps for the first
six DCT coefficients of the non-aligned JPEG compression, and the PRNU
(Photo-Response Non-Uniformity) [13].
3 Description of verification approaches
3.1 Using attribution, fake and genuine claim extraction (UoS-ITI)
This approach is motivated by an established journalistic process for verifying
social media content [34]. The central hypothesis is that the “wisdom of the
crowds” is not really wisdom when it comes to verifying suspicious content.
Instead it is better to rank evidence from Twitter according to the most trusted
and credible sources in a way similar to the one practiced by journalists.
A trust and credibility model was created based on an NLP pipeline involv-
ing tokenization, Part-Of-Speech (POS) tagging, Named Entity Recognition
(NER) and Relation Extraction (RE). The novelty of this approach lies within
the choice of regex patterns, which are modelled on how journalists verify fake
and genuine claims by looking at the source attribution for each claim, and
the semi-automated workflow allowing trusted lists of entities to be utilized.
A novel conflict resolution approach was created based on ranking claims in
order of trustworthiness. To extract fake and genuine claims, a set of regex
patterns were created (see Fig. 3) matching both terms and POS tags. Claims
of an image being fake or genuine occur infrequently, and by themselves are
Verifying Information with Multimedia Content on Twitter 9
not sufficient. If an image is claimed to be real without any supporting at-
tribution we assume it is fake, since from our own analysis strong debunking
posts almost always contain attribution. We combine all fake and real claims
with trusted source attribution (e.g. via BBC News) to discover strong evi-
dence. To extract attribution, a combination of Named Entity (NE) matching,
based on noun and proper noun sequences, and regex patterns for source cita-
tion was used. Other researchers have published linguistic patterns that were
used to detect rumours [8,10,40], but the combination of fake/genuine claims
and source attribution used by the UoS-ITI approach is novel in that it uses
insights from well-established journalistic processes for social media content.
First, an NLP pipeline is executed (see Fig. 4) that takes in each tweet
from the test dataset and tokenizes it using a Punkt sentence tokenizer and
a Treebank word tokenizer. To help the POS tagger, no stemming is applied
and text case is preserved. Each tokenized sentence is passed to a Treebank
POS tagger, which supports more European multi-lingual tagsets than other
taggers (e.g., Stanford POS tagger), an important consideration for future
Named Entity Patterns
Examples
@ <ANY>
(NOUN | PROP_NOUN | NAMESPACE)
(NOUN | PROP_NOUN | NAMESPACE) (NOUN | PROP_NOUN
| NAMESPACE)
@bbcnews, BBC News, CNN.com, CNN
Attribution Patterns
<NE> <SYMBOL> <URl>
<NE> *{0,1} <lMAGE> *(0,2) <URl>
<NE> *{0,1} <FROM> *{0,2} <URl>
<RT> <SYMBOL>{0,1} <NE>
<FROM> *{0,2} <NE>
What a great picture! @bbcnews: http://bit.|y/1234
@bbcnews image - http://bit.|y/1234
@bbcnews releases photo http://bit.|y/1234
RT: @bbcnews "|ove|y picture of
eyewitness report via @bbcnews
Faked Patterns
<lMAGE> *{0,1} ^<POSnot> <FAKED>
<POSis> <POSa>{0,1}^<POSnot> <FAKED> *{0,1} <lMAGE>
<POSis>{0,1} <POSnot> <POSa>{0,1} <REAL>
image is fake! ...
is a fake image
… Is not a real …
Genuine Patterns
<lMAGE> <POSis> *{0,1} ^<POSnot> <REAL>
<POSis> <POSa>{0,1} ^<POSnot> <REAL> *{0,1} <lMAGE>
<POSis>{0,1} <POSnot> <POSa>{0,1} <FAKE>
image is totally genuine
is a real image
Is not a fake
Key
(mm) = n to m matches allowed
<NE> = named entity
<SYMBOL> = symbols (e.g. : = -)
<POSnot> = POS adverbs RB (e.g. not, never)
<POSis> = POS verbs VBZ, VBD (e.g. is, was)
<POSa> = POS determiner DT (e.g. a, the)
Fig. 3: Verification Linguistic Patterns in UoS-ITI. These patterns are encoded
as regex patterns matching on both phrases in content and their associated
POS tags (e.g. NN = noun, NNP = proper noun).
10 Christina Boididou et al.
work as breaking news can happen anywhere in the world, not just English
speaking locations. Namespaces and URI’s are extracted prior to tokenization
and re-inserted after POS tagging as explicit NEs so they can be matched
using regex expressions later.
The employed NER strategy is based on a regex POS expression that
matches unigrams and bigrams with nouns, proper nouns, namespaces and
Twitter usernames. This is a high recall-low precision approach to NER as we
want to capture all relevant NEs at this stage. Next, a lookup table is used
to filter candidate NEs into sets of trusted, untrusted and unknown entities;
this allows the removal of known false positive NE values and the application
of blaklist and whitelist values. Finally, candidate NEs are used to create the
POS and NE-labelled sentence, which is passed to a set of regex expressions
encoding typical relationship phrases for fake, real and attribution claims.
The approach is semi-automated in that it exploits a list of a priori known
trusted and untrusted sources. All news providers have long lists of trusted
sources for different regions around the world so this information is readily
available. For this task, a list of candidate NEs was created by first running
the NER regex patterns on the test dataset. Then, each NE was manually
checked via Google search (e.g. looking at Twitter profile pages) and NEs were
removed that were considered as irrelevant for inclusion in a list of trusted
or untrusted sources by a journalist. Instead, NEs were kept that included
news organizations, respected journalists and well cited bloggers and experts.
Creating these lists took under two hours (570 NEs checked, 60 accepted).
The employed RE approach uses a set of regex expressions that match seri-
alized POS and NE tagged sentence trees. These regex expressions were man-
ually created after a detailed analysis of the linguistic patterns from Twitter,
YouTube and Instagram around a number of previously crawled event types
(e.g., hurricanes, tornados, earthquakes, blackouts, conflicts, etc.). For finding
attributed NEs, all the common ways were studied, in which social media users
Social Media Text
(e.g. Tweet) Clean & Tokenize
(sent and word) POS Tagging Named Entity
Recognition
Compute
Sentence Tree
Variants
Relational
Extraction
(regex patterns)
(Un)Trusted
entity Matching
Image Content
Cross-check
Trustworthiness
Assessment
Decision
(fake, real)
Multi-lingual
Grammer
Named Entity
Regex
Multi-lingual
Grammer
- Attribution
Regex
- Faked Regex
- Genuine Regex
Information Extraction
Trust and Credibility Analysis
Stoplist Entities
Trusted Entities
Untrusted Entities
TreeTagger
Entity Recognition
Fig. 4: NLP pipeline for regex-based NER and RE in UoS-ITI.
Verifying Information with Multimedia Content on Twitter 11
attribute sources. This is typically either as a reference to a NE followed by
a link or image reference, or a statement claiming that a link or image refer-
ence is created and/or verified by an NE. For fake and genuine claims, regex
patterns are created from the most commonly phrased claims about images or
links being fake or real. Examples of the regex patterns can be seen in Fig. 3
and make heavy use of the POS tags to avoid being overly term specific.
Once all the attributed sources, fake and genuine claims have been iden-
tified, a trustworthiness decision for each image is made. Just like human
journalists, claims are ranked by trustworthiness based on whether the claim
comes directly from a trusted author (top rank), is attributed to a trusted
source (second rank) or is from an unknown source (third rank). Claims di-
rectly by, or attributed to, untrusted sources are ignored. The final decision
for each image is taken based on only the most trustworthy claims. A con-
servative claim conflict resolution approach is used, where a fake claim by a
source supersedes a real claim by an equally trusted source.
3.2 Using a two-level classification model (MCG-ICT)
Existing approaches often formulate the tweet verification problem as a binary
classification task [8]. Features from tweet text and users are extracted to train
a classifier at the message (tweet) level. One problem of this training strategy
is that tweets are trained and tested individually. However, tweets in the real
world have strong relations among each other, especially, tweets of the same
topic will likely have the same credibility: real or fake.
Rather than classifying each tweet individually, the MCG-ICT approach
verifies tweets by leveraging inter-tweet information, such as whether they
contain the same multimedia content. It was empirically observed that even
such simple implications among tweets would be useful to boost the original
message-level predictions. In fact, in recent work [19, 20], links among tweets
were built by clustering tweets into sub-events or topics. Thus, credibility eval-
uation can be performed at different scales to provide more robust predictions.
3.2.1 Two-level classification model
As illustrated in Fig. 5, the MCG-ICT approach comprises two levels of clas-
sification: a) The message-level, which learns a credibility model per message
(tweet). Features extracted from the text content, user information and other
components of a tweet are used for training a classifier. b) The topic-level, i.e.
a specific subject of discussion or sub-event under the broader unfolding event.
By assuming tweets under a same topic likely have similar credibility values,
tweets are clustered into different topics. Compared with raw tweets, topics
eliminate variations of tweets by aggregating message-level credibility classifi-
cations. The topic-level feature is computed as the average of the tweet-level
feature vectors around the topic. The following processing steps take place for
topic-level classification:
12 Christina Boididou et al.
Topic clustering : In [19], a clustering algorithm is used to cluster tweets
into sub-events. But this algorithm performs poorly in forming topics in
the target dataset as it is difficult to decide the optimal number of clusters.
However, in the studied verification setting, each tweet contains an image
or video, and each image/video can be contained in more than one tweets.
This intrinsic one-to-many relation is used to form topics: each image/video
corresponds to a topic and tweets containing it are assigned to this topic.
Topic labeling: Each topic is labelled using the majority of the labels of its
tweets. These labels are used for training the topic-level classifier. In fact,
with the proposed topic formation method, almost all tweets in a topic
have the same label, resulting in topics labelled with very high confidence.
Topic-level feature aggregation: Message-level features (section 3.2.2) of all
tweets in a topic are aggregated by averaging them to derive a single topic-
level feature vector. By taking the average of all tweets, the impact of noisy
or outlier tweets is suppressed.
Fusing topic-level result: After topic-level classification, a probability value
is computed for each topic representing the likelihood of it being fake.
Then, for each tweet in the topic, this value is added as a feature to its
original feature vector. Finally, a message-level classifier is trained with
this extended feature in order to produce the final results.
In terms of classification model, several options from the state of the art were
tested, and the selection was based on the performance on the development
set using cross validation. J48 Decision Trees were selected for the topic-level
classification, and Random Forests for the message-level classification.
3.2.2 Feature extraction
At the message level, we use as base features the ones shared by the task,
TB-base and UB-base (Sec. 2.5). Some additional features were also tested
Tweets
Feature Set
Content Features User Features Other Features
Message Level
Classifier
Final Results
Topics
Topic Labeling
Topic Level
Features
Topic Level
Classifier
Pre-results
Clustering
Feature Extraction
Feature Aggregation
Fig. 5: Overview of the MCG-ICT two-level classification model. Topic-level
classifications are fused with the message-level ones to produce the final result.
Verifying Information with Multimedia Content on Twitter 13
but not included in the final approach configuration: word term features and
several image features.
The commonly used term frequency (tf) and tf-idf features were tested. Ex-
periments on the training (development) set indicated that such features could
lead to overfitting, since they led to very high performance on general cross-
validation and very low performance on leave-one-event-out cross-validation.
Since few words co-occur across different events, one may assume that other
keyword-based features (e.g., LDA) would also contribute little to this task.
Several image-based features (e.g. image popularity, resolution) were also
tested. Such image features could replace the topic level features to train clas-
sifier at topic level, because a topic is generated for each image as mentioned
earlier. Experiments on the development set showed that these features led
to slightly worse performance for the topic-level classification, compared to
content-based ones, and to much worse performance when combined with
message-level features. Moreover, image features cannot be applied directly
on videos included in the test set. Hence, those were not further considered.
3.3 Using an agreement-based retraining scheme (CERTH-UNITN)
This approach combines different sets of tweet-based (TB), user-based (UB)
and forensics (FOR) features in a semi-supervised learning scheme. A more
detailed exposition of this method is presented in [5]. The approach builds on
supervised classification models and an agreement-retraining method that uses
part of its own predictions as new training samples with the goal of adapting
to tweets posted in the context of new events.
Fig. 6 depicts an overview of the method. It relies on two individual clas-
sification models, one based on the combination of TB and FOR features and
a second based on UB features. Bagging is used to ensure higher reliability in
the training process of the classifiers (CL11 ... CL1nand CL21 ... C L2n), and
an agreement-based retraining strategy (fusion) is employed with the goal of
improving the accuracy of the overall framework. All classifiers are based on
Random Forests of 100 trees.
3.3.1 Feature extraction
The approach uses the TB-ext and UB-ext features, which are an extended
version of the TB-base and UB-base released by the MediaEval task. For the
FOR features, we also include additional ones.
TB-ext: These are binary features extracted from the tweet text, e.g. the
presence of a word, symbol or external link. Language-specific binary features
are also used corresponding to the presence of specific terms; for languages, in
which such terms are not available, the values of these features are set to null
(missing). Language detection is performed with a publicly available library4,
4https://code.google.com/p/language-detection/
14 Christina Boididou et al.
Tweets
TB CL11 CL12 CL1n
CL21 CL22 CL2n
+
+
Prediction 1
Fusion
Prediction 2
UB
Feature Extraction Bagging in model building Agreement-based retraining technique
Fig. 6: Overview of the CERTH-UNITN method.
and a feature is added for the number of slang words in a text, using slang
lists in English5and Spanish6. For the number of nouns, the Stanford parser7
is used to assign POS tags to each word (only in English). For the readability of
text, the Flesch Reading Ease method is used8, which computes the complexity
of a piece of text as a score in [0,100] (0: hard-to-read, 100: easy-to-read).
UB-ext: User-specific features are extracted such as number of media content,
account age and others that refer to information that the profile shares. In
addition, these include whether the user declares a location and whether this
can be matched to a city name from the Geonames dataset9.
For both TB and UB features, trust-oriented features are computed for the
links shared, through the tweet itself (TB) or the user profile (UB). These include
the WOT metric10, a score indicating how trustworthy a website is according
to reputation ratings by Web users, the in-degree and harmonic centrality,
which are rankings based on the links of the web forming a graph11, and web
metrics provided by the Alexa API12.
FOR: For each image, additional forensics features are extracted from the pro-
vided BAG feature based on the maps obtained from AJPG and NAJPG. First, a
binary map is created by thresholding the AJPG map (we use 0.6 as thresh-
old), then the largest region is selected as object and the rest of the map is
considered as the background. For both regions, seven descriptive statistics
(max, min, mean, median, most frequent value, st. deviation, and variance)
are computed from the BAG values and concatenated to a 14-d vector. Figure 7
illustrates the feature extraction process. We apply the same process on the
NAJPG map to obtain a second feature vector.
5http://onlineslangdictionary.com/word-list/0- a/
6http://www.languagerealm.com/spanish/spanishslang.php
7http://nlp.stanford.edu/software/lex-parser.shtml
8http://simple.wikipedia.org/wiki/Flesch_Reading_Ease
9http://download.geonames.org/export/dump/cities1000.zip
10 https://www.mywot.com/
11 http://wwwranking.webdatacommons.org/more.html
12 http://data.alexa.com/data?cli=10&dat=snbamz&url=google.gr
Verifying Information with Multimedia Content on Twitter 15
Input Image (real) AJPG Object/Background Mask
BAG
1234567
0
0.5
1Descriptive Statistics
Object
Background
Input Image (fake) AJPG Object/Background Mask
BAG
1234567
0
0.5
1Descriptive Statistics
Object
Background
Fig. 7: Illustration of forensics feature extraction process.
3.3.2 Data pre-processing and bagging
To handle the issue of missing values on the features, Linear Regression (LR) is
used for interpolating missing values. This is applied only to numeric features
as the method is not applicable for boolean values. Only feature values from
the training set are used in this process. Data normalization is performed to
scale values to the range [-1, 1]. Furthermore, bagging is used to improve the
accuracy of the method. Bagging creates mdifferent subsets of the training set,
including equal number of samples for each class (some samples may appear
in multiple subsets), leading to the creation of minstances of CL1and C L2
classifiers (m= 9), as shown in Fig. 6. The final prediction for each of the test
samples is calculated using the majority vote of the mpredictions.
3.3.3 Agreement-based retraining
Agreement-based retraining is used to improve the prediction accuracy for un-
seen events. This is motivated by a similar approach implemented in [37] on
the problem of polarity classification. To this end, two classifiers are built
CL1,C L2, each on different types of feature, and their outputs are combined
16 Christina Boididou et al.
as follows: The two outputs for each sample of the test set are compared,
and depending on their agreement, the test set is divided in two subsets, the
agreed and disagreed sets. Assuming that the agreed predictions are correct
with high likelihood, they are used as training samples to build a new classifier
for classifying the disagreed set of instances. To this end, in the subsequent
step, the agreed samples are added to the best performing of the two initial
models, CL1,CL2(comparing them on the basis of their performance using
cross-validation on the training set). The goal of this method is to retrain the
initial model and adapt it to the specific characteristics of the new event. In
that way, the model can predict more accurately the values of the samples for
which CL1,C L2did not agree in the first step.
4 Experiments and Evaluation
4.1 Datasets
The conducted experiments were based on the benchmark dataset released
by the VMU task in 2015 (Sec. 2.5). We refer to the original version of the
dataset as dataset#1. This has been collected over a number of years using a
crowd-sourcing approach. Images are found by volunteers (including the au-
thors of the article), and paid micro-workers (via Amazon Mechanical Turk).
Each suggested content item was provided with associated news reports or de-
bunking articles by journalists, offering evidence regarding its veracity. These
were then used to provide ground truth labels (real/fake) for each tweet.
The dataset consists of tweets relating to 17 events listed in Table 1, com-
prising in total 197 cases of real and 191 cases of misused images, associated
with 6,225 real and 9,404 fake tweets posted by 5,895 and 9,025 unique Twitter
users respectively. Note that several of the events, e.g., Columbian Chemicals,
Passport Hoax and Rock Elephant, were actually hoaxes, hence all multimedia
content associated with them was fake. For several real events (e.g., MA flight
370) no real images (and hence no real tweets) are included in the dataset,
since none came up as a result of the conducted data collection process.
For further testing, we additionally created dataset#2, which is a subset of
dataset#1 by first performing near-duplicate tweet removal: we empirically set
a minimum threshold of similarity and computed the Levenshtein Distance13
for each pair of texts. A small amount of near-duplicate texts exceeding the
threshold were manually removed. Note that in dataset#1 the number of
unique fake and real multimedia items, which the tweets are associated with,
is highly unbalanced. As the aim of dataset#2 is to create a balanced dataset,
we randomly selected a subset of fake and real multimedia items as well as the
tweets associated with them.
Finally, to assess the stability of results, we also compared the methods
on the dataset that was used in the VMU task sequel in 2016 [6]. This is a
superset of dataset#1, using the latter as development set, while it contains
13 http://rosettacode.org/wiki/Levenshtein_distance#Java
Verifying Information with Multimedia Content on Twitter 17
an additional 998 real and 1,230 fake tweets in the test set, organized around
64 cases of real and 66 cases of misused multimedia items. The tweet IDs and
image URLs for all of the above datasets are publicly available14.
Table 1: Upper part: MediaEval’15 events and derivative datasets (#1,#2):
For each event, we report the numbers of unique real (if available) and fake
images (IR,IFrespectively), unique tweets that shared those images (TR,TF)
and unique Twitter accounts that posted those tweets (UR,UF). Bottom
part: MediaEval’16 events and corresponding statistics.
ID Event dataset#1 dataset#2
IRTRURIFTFUFIRTRURIFTFUF
E1 Hurricane Sandy 150 4,664 4,446 53 5,558 5,432 60 838 825 16 376 369
E2 Boston Marathon bombing 29 344 310 35 189 187 18 131 120 13 56 56
E3 Sochi Olympics - - - 14 274 252 - - - 9 76 74
E4 MA flight 370 - - - 23 310 302 - - - 13 88 87
E5 Bring Back Our Girls - - - 7 131 126 - - - 6 35 33
E6 Columbian Chemicals - - - 15 185 87 - - - 6 124 64
E7 Passport hoax - - - 2 44 44 - - - 1 5 5
E8 Rock Elephant - - - 1 13 13 - - - 1 4 4
E9 Underwater bedroom - - - 3 113 112 - - - 2 4 4
E10 Livr mobile app - - - 4 9 9 - - - 3 6 6
E11 Pig fish - - - 1 14 14 - - - 1 4 4
E12 Solar Eclipse 5 140 133 6 137 135 2 55 54 3 53 53
E13 Girl with Samurai boots - - - 3 218 212 - - - 2 16 16
E14 Nepal Earthquake 11 1004 934 20 356 343 6 113 107 8 178 176
E15 Garissa Attack 2 73 72 2 6 6 2 40 39 2 4 4
E16 Syrian boy - - - 1 1786 1692 - - - 1 197 195
E17 Varoufakis and zdf - - - 1 61 59 - - - 1 29 28
Total 197 6,225 5,895 191 9,404 9,025 88 1,177 1,145 88 1,255 1,178
Event IFTFIRTREvent IFTFIRTR
Gandhi Dancing 1 29 - - Woman 14 children 2 11 - -
Half of Everything 9 39 - - American Soldier Quran 1 17 - -
Hubble Telescope 1 18 - - Airstrikes 1 24 - -
Immigrants fear 5 33 3 18 Attacks in Paris 3 44 22 536
ISIS children 2 3 - - Ankara Explosions - - 3 19
John Guevara 1 33 - - Bush b ook 1 27 - -
Mc Donalds Fee 1 6 - - Black Lion 1 7 - -
Nazi Submarine 2 11 - - Boko Haram 1 31 - -
North Korea 2 10 - - Bowie David 2 24 4 48
Not Afraid 2 32 3 35 Brussels Car Metro 3 41 - -
Pakistan Explosion 1 53 - - Brussels Explosions 3 69 1 9
Pope Francis 1 29 - - Burst in KFC 1 25 - -
Protest 1 30 10 34 Convoy Explosion Turkey - - 3 13
Refugees 4 35 13 33 Donald Trump Attacker 1 25 - -
Rio Moon 1 33 - - Eagle Kid 1 334 - -
Snowboard Girl 2 14 - - Five Headed Snake 5 6 - -
Soldier Stealing 1 1 - - Fuji Lenticular Clouds 1 123 1 53
Syrian Children 1 12 1 200 Total 66 1,230 64 998
Ukrainian Nazi 1 1 - -
4.2 Measuring accuracy by leave-one-event-out cross-validation
The conducted experiments aimed at evaluating the accuracy of each method
on a variety of unseen events. The features of fake tweets may vary across
different events, so the generalization ability of automated methods is consid-
ered an important aspect of its verification performance. To this end, we used
14 https://github.com/MKLab-ITI/image- verification-corpus/
18 Christina Boididou et al.
each time one of the events Ei,i={1,2, ..., 17}for testing, and the remaining
ones for training. For example, for evaluating the performance on event E1,
we used the tweets of E2,E3, ..., E17 for training, and the tweets of E1for
testing. This is in contrast to the MediaEval task, where events E1-11 were
used for training, and E12-17 for testing. To evaluate the approaches, we used
the established measures of precision (p), recall, and F1 -score (F1). Assuming
that the positive class is the case that a tweet instance is fake, and negative
that a tweet instance is real, we define these metrics as:
p=tp
tp +f p , F 1 = 2·tp
2·tp +fp +f n (1)
where tp refers to true positives (correctly detected fake), fp to false positives
(real misclassified as fake), and fn to false negatives (fake as real).
Table 2 presents the precision and F1-scores that the approaches achieved
in the context of the MediaEval tasks and on dataset#1 and dataset#2. We
also compare our results with those presented by the TweetCred method [16].
We have re-implemented a variation of the method described in the paper: we
identified the common tweet- and user-based features that the CERTH-UNITN
and TweetCred methods use, and we built classification models for each of the
datasets. We evaluated common state-of-the-art classifiers on each dataset:
SVM, RandomForest and AdaBoost. For the first two datasets (MediaEval
’15 and dataset#1), SVM achieved the highest performance, while for the
rest RandomForest worked best. As can be seen, TweedCred ranks second on
two of the datasets (dataset#2 and MediaEval’16), third on MediaEval’15
and first on dataset#1.
Table 2: Precision and F1-scores achieved on a) MediaEval VMU 2015
task, with E1-11 used for training and E12-17 for testing, b) dataset#1, c)
dataset#2, and d) MediaEval VMU 2016 task. For b and c, the leave-one-
event-out cross-validation method was used for measuring performance.
Method MediaEval ’15 dataset#1 dataset#2 MediaEval ’16
p F1 p F1 p F1 p F1
UoS-ITI 1.000 0.830 0.938 0.224 0.917 0.244 0.520 0.468
MCG-ICT 0.964 0.942 0.804 0.756 0.816 0.750 0.563 0.504
CERTH-UNITN 0.861 0.911 0.755 0.693 0.690 0.635 0.980 0.911
TweedCred [16] 0.680 0.800 0.810 0.820 0.810 0.650 0.580 0.720
Fig. 8 illustrates the F1-scores of the tested approaches for each event of
dataset#1 and dataset#2. The mean performance of the approaches is also
illustrated in red. Events are ordered based on the F1-score achieved by the
highest-scoring method (MCG-ITI). On average, it is clear that on dataset#1
the two-level classification method (MCG-ITI) outperforms the other two.
However, given that it strongly relies on the fact that tweets are correlated
when they share the same multimedia content, it seems that it performs better
on events that comprise only one or few unique multimedia cases (e.g. Passport
Verifying Information with Multimedia Content on Twitter 19
Fig. 8: F1-scores for dataset#1 and dataset#2 per event and approach.
hoax, Syrian boy). When a single event includes numerous multimedia cases
(e.g., Nepal earthquake), its performance decreases. The UoS-ITI approach
seems to accurately predict the posts associated with video content (Syrian
boy, Varoufakis and zdf). Similar results are obtained on dataset#2 (Fig. 8).
In addition to F1-scores, we also report the corresponding precision scores
in Fig. 9 for each event on dataset#1 and dataset#2. In both cases, the
UoS-ITI approach outperforms the other two, providing a very high precision
(>0.9) for the small set of tweets it was able to classify. This is an important
observation, since it allows us to use the results of UoS-ITI as a type of pre-
classifier in an ensemble-based approach (Sec. 4.4).
Another striking finding concerns the stability of performance of the three
methods when testing them on a different dataset (Mediaeval ’16). The CERTH-
UNITN approach is a clear winner in this case, since it is the only approach
that manages to retain its performance at comparable levels, while the other
two approaches perform considerably worse. This provides evidence in support
of the generalization ability of the agreement-based retraining method.
4.3 Measuring verification performance per multimedia item
A further experiment explores the verification accuracy of methods on each
unique multimedia item. Given that dataset#1 contains 388 unique multime-
dia items, we divided tweets in groups according to the multimedia item they
are associated with. Then, we calculate the performance of each approach on
each of those unique multimedia cases. Fig. 10 illustrates the achieved F1-
scores. In the horizontal axis, we present the unique multimedia items. The
20 Christina Boididou et al.
Fig. 9: Precision scores for dataset#1 and dataset#2 per event and approach.
figure reveals that the MCG-ITI and CERTH-UNITN approaches perform
similarly in the majority of cases, while the UoS-ITI achieves quite low per-
formance compared to them, due to the fact that it avoids producing a result
in cases where there is not sufficient information to make this decision.
Another key question of this analysis is the level of overlap between the
method results. For this reason, we conduct pairwise comparisons between
the previously generated F1-score distributions. After calculating the number
of multimedia items for which the methods have an F1-score equal to zero
and equal to one, we report the percentage of items, for which the methods’
predictions agree. In addition, we calculate the Pearson correlation between
these distributions. Table 3 presents the results of this analysis. These make
clear that the three approaches are highly correlated in terms of the cases
they fail to predict (F1-score = 0) and less correlated in the cases where
they succeed (F1-score = 1). The pairwise Pearson correlations demonstrate
that MCG-ICT and CERTH-UNITN approaches are much more similar in
their predictions compared to UoS-ITI. Overall, these experiments reveal that
there is potential for improving upon the results of the individual methods by
fusing their results, which we investigate in the next section.
4.4 An ensemble verification approach
We investigate three ensemble methods for fusing the results of the three
approaches:
– ENS-MAJ: For each tweet, we aggregate individual predictions by ma-
jority vote. However, in the UoS-ITI approach, several of the predictions
Verifying Information with Multimedia Content on Twitter 21
Table 3: Pairwise comparison of F1-score distributions. Reporting percentages
(%) where the F1-score of both methods is equal to 0 and 1, and the Pearson
correlation between them.
F1=0 F1=1 pearson corr
UoS-ITI vs MCG-ICT 53.3 0.7 0.295
UoS-ITI vs CERTH-UNITN 59.5 6.7 0.332
MCG-ICT vs CERTH-UNITN 51.5 9.5 0.707
Fig. 10: F1-score across unique images for each approach. To make the visu-
alization of results cleaner, images in the xaxis are sorted based on the mean
F-score across the three approaches.
are marked as unknown, which is a problem in case the decisions of the
other two approaches disagree. To overcome this tie, we assign the tweet
as fake, assuming that it is preferable to falsely consider a case to be fake
than falsely consider it as real.
– ENS-HPF: This takes advantage of the high precision (low fp rate) of
the UoS-ITI method. If the UoS-ITI prediction for a tweet is other than
unknown, we adopt it as correct; otherwise, we check the other methods’
predictions and in case of disagreement we consider the item to be fake.
– ENS-ORA: This is a hypothetical (oracle) ensemble method, which se-
lects the correct prediction if at least one of the methods’ predictions is
correct. This provides an upper-bound of the performance that could be
theoretically possible if we could optimally combine the three approaches.
Fig. 11 presents the resulting F1-scores and precision per event of the ensemble
methods. On average, these achieve higher performance than the individual
approaches. This result stems from the complementarity of the approaches’
outputs, which was illustrated in sec. 4.3, and the effectiveness of the pro-
posed schemes in combining their outputs. ENS-ORA delineates the maxi-
mum possible performance that is achievable by combining the three methods.
Out of the two practical fusion schemes, ENS-HPF produces more accurate
predictions compared to ENS-MAJ as the former uses the highly accurate
UoS-ITI approach as the preferred approach for performing the classification
and falls back to the other two approaches in case UoS-ITI produces no result.
22 Christina Boididou et al.
Fig. 11: F1-scores and Precision of the ENS-MAJ,ENS-HPF and ENS-ORA en-
semble methods for dataset#1, mean score of each method on the dataset of
MediaEval ’16 and individual mean scores of the approaches.
4.5 Relevance of experimental results for real-world use cases
These promising results from our ensemble verification approach should also
be seen in the context of typical use cases for automated verification of images
and videos trending on social media.
An example use case involving semi-automated verification is in support of
journalists who are trying to verify social media content for use in news stories.
Breaking news in particular has competing objectives to publish content first
(i.e. as quickly as possible) and get the verification right (i.e. take enough
Verifying Information with Multimedia Content on Twitter 23
time to ensure that the content is not fake). Publishing eyewitness content
before rivals will gain a journalist much kudos. Publishing a fake image, then
later being forced to retract the story, can seriously damage a journalist’s
reputation. Let us consider a use case where a journalist wants to receive
a live feed (e.g. every 5 or 10 minutes) of the top 500 trending images on
Twitter, classified and filtered using our ensemble of fake classifiers. Our best
reported result (F1=0.79, p=0.82, r=0.81) means that on average for 500
trending images, only 90 would be classified in error. During news events such
as the Hurricane Sandy 2012 fake social media images on Twitter [27] [17]
outnumbered real images by two to one. In the context of our use case this
means that of the 500 images considered, 333 would on average be fake, and
of those 333, 273 would be classified as fake and filtered. This represents a
significant reduction in the images the journalist needs to consider in real-time,
something which is very useful when working under breaking news deadlines
where a story must be verified and published within minutes.
Another example user case, this time involving fully automated verification,
is where an automated news summarization platform wants to aggregate news
feeds from sources such as popular social media bloggers in real-time. Readers
of such news summarization platforms typically accept a higher number of
false stories than they would from a journalist-based news service. In this case
the volume of trending images that need checking would be much larger, with
tens of thousands of images being checked as candidates for aggregation into
news alert summary snippets. The fully automated nature of our approach
makes this a viable proposition. Even for platforms like Storyful, where news
snippets are passed to a human checkdesk for a final verification step, our
approach could have significant utility as a pre-filter.
5 Conclusions
In this article, we presented a comparative study for automated approaches
for verifying online information with multimedia content. By presenting three
different in nature methods, we showed that there are several ways to deal
with the challenging problem of verification. To measure verification accuracy,
we evaluated these methods by using leave-one-out cross-validation and by
reporting their scores per multimedia case. In the MediaEval’15 dataset and
dataset#2, the MCG-ICT method achieved the highest F1-scores, particularly
for events with few cases of unique multimedia items. The UoS-ITI achieved
the highest precision scores with a very low false positive rate for the events
it could classify. The CERTH-UNITN method led to consistently high results
on average, and in particular on events with many tweets. Importantly, it
managed to retain its performance on the MediaEval’16 dataset, clearly out-
performing the other two and the TweetCred method, and demonstrating the
potential of the agreement-based retraining scheme for making the approach
applicable to new datasets.
24 Christina Boididou et al.
By combining approaches into an ensemble method we were able to further
increase the F1-score of the best performing method by approximately 7%
(ENS-HPF) with a theoretical upper bound of improvement of approximately
20% (ENS-ORA), which is a very encouraging finding, and provides a state
of the art benchmark for other researchers in this field. To improve results even
further we feel that more sophisticated use of visual and contextual features are
needed. This is a very challenging area and will need a combination of image
forensics, computer vision and cross-referencing of contextual information (e.g.
weather, maps, etc.) about the event of interest.
Acknowledgements This work has been supported by the REVEAL and InVID projects,
partially funded by the European Commission (FP7-610928 and H2020-687786 respectively).
References
1. Aiello, L.M., Petkos, G., Mart´ın, C.J., Corney, D., Papadopoulos, S., Skraba, R.,
oker, A., Kompatsiaris, I., Jaimes, A.: Sensing trending topics in twitter. IEEE
Trans. Multimedia 15(6), 1268–1282 (2013). DOI 10.1109/TMM.2013.2265080. URL
http://dx.doi.org/10.1109/TMM.2013.2265080
2. Ardizzone, E., Bruno, A., Mazzola, G.: Copy move forgery detection by matching tri-
angles of keypoints. Information Forensics and Security, IEEE Transactions on 10(10),
2084–2094 (2015)
3. Bianchi, T., Piva, A.: Image forgery localization via block-grained analysis of jpeg arti-
facts. IEEE Transactions on Information Forensics and Security 7(3), 1003–1017 (2012)
4. Boididou, C., Andreadou, K., Papadopoulos, S., Dang-Nguyen, D.T., Boato, G., Riegler,
M., Kompatsiaris, Y.: Verifying multimedia use at mediaeval 2015. In: MediaEval 2015
Workshop, Sept. 14-15, 2015, Wurzen, Germany (2015)
5. Boididou, C., Papadopoulos, S., Apostolidis, L., Kompatsiaris, Y.: Learning to detect
misleading content on twitter. In: Proceedings of the 2017 ACM on International Con-
ference on Multimedia Retrieval, pp. 278–286. ACM (2017)
6. Boididou, C., Papadopoulos, S., Dang-Nguyen, D., Boato, G., Riegler, M., Middleton,
S.E., Petlund, A., Kompatsiaris, Y.: Verifying multimedia use at mediaeval 2016. In:
Working Notes Proceedings of the MediaEval 2016 Workshop, Hilversum, The Nether-
lands, October 20-21, 2016. (2016). URL http://ceur-ws.org/Vol- 1739/MediaEval_
2016_paper_3.pdf
7. Boididou, C., Papadopoulos, S., Dang-Nguyen, D.T., Boato, G., Kompatsiaris, Y.: The
certh-unitn participation@ verifying multimedia use 2015 (2015)
8. Boididou, C., Papadopoulos, S., Kompatsiaris, Y., Schifferes, S., Newman, N.: Chal-
lenges of computational verification in social multimedia. In: Proceedings of the Com-
panion Publication of the 23rd International Conference on World Wide Web Compan-
ion, pp. 743–748 (2014). DOI 10.1145/2567948.2579323. URL http://dx.doi.org/10.
1145/2567948.2579323
9. Canini, K.R., Suh, B., Pirolli, P.L.: Finding credible information sources in social net-
works based on content and social structure. In: Privacy, security, risk and trust (passat),
2011 ieee third international conference on and 2011 ieee third international conference
on social computing (socialcom), pp. 1–8. IEEE (2011)
10. Carton, S., Adar, E., Park, S., Mei, Q., Zeffer, N., Resnick, P.: Audience analysis for
competing memes in social media. In: Ninth International AAAI Conference on Web
and Social Media (2015)
11. Castillo, C., Mendoza, M., Poblete, B.: Information credibility on twitter. In: Pro-
ceedings of the 20th international conference on World wide web, pp. 675–684. ACM
(2011)
12. Conotter, V., Boato, G., Farid, H.: Detecting photo manipulation on signs and bill-
boards. In: Image Processing, IEEE International Conference on, pp. 1741–1744 (2010)
Verifying Information with Multimedia Content on Twitter 25
13. Conotter, V., Dang-Nguyen, D.T., Riegler, M., Boato, G., Larson, M.: A crowdsourced
data set of edited images online. In: Proceedings of the 2014 International ACM Work-
shop on Crowdsourcing for Multimedia, CrowdMM ’14, pp. 49–52. ACM, New York,
NY, USA (2014). DOI 10.1145/2660114.2660120. URL http://doi.acm.org/10.1145/
2660114.2660120
14. Ferrara, P., Bianchi, T., Rosa, A.D., Piva, A.: Image forgery localization via fine-grained
analysis of cfa artifacts. IEEE Transactions on Information Forensics and Security 7(5),
1566 –1577 (2012)
15. Gupta, A., Kumaraguru, P.: Twitter explodes with activity in mumbai blasts! a lifeline
or an unmonitored daemon in the lurking? (2012)
16. Gupta, A., Kumaraguru, P., Castillo, C., Meier, P.: Tweetcred: A real-time web-based
system for assessing credibility of content on twitter. In: Proc. 6th International Con-
ference on Social Informatics (SocInfo) (2014)
17. Gupta, A., Lamba, H., Kumaraguru, P., Joshi, A.: Faking sandy: characterizing and
identifying fake images on twitter during hurricane sandy. In: Proceedings of the 22nd
international conference on World Wide Web companion, pp. 729–736. International
World Wide Web Conferences Steering Committee (2013)
18. Jian, L., Xiaolong, L., Bin, Y., Xingming, S., Li, J., Li, X., Yang, B., Sun, X.:
Segmentation-Based Image Copy-Move Forgery Detection Scheme. Information Foren-
sics and Security, IEEE Transactions on 10(3), 507–518 (2015)
19. Jin, Z., Cao, J., Jiang, Y.G., Zhang, Y.: News credibility evaluation on microblog with
a hierarchical propagation model. In: 2014 IEEE International Conference on Data
Mining (ICDM), pp. 230–239. IEEE (2014)
20. Jin, Z., Cao, J., Zhang, Y., Luo, J.: News verification by exploiting conflicting social
viewpoints in microblogs. In: AAAI 2016. AAAI (2016)
21. Jin, Z., Cao, J., Zhang, Y., Zhang, Y.: Mcg-ict at mediaeval 2015: Verifying multimedia
use with a two-level classification model (2015)
22. Kee, E., Johnson, M.K., Farid, H.: Digital image authentication from jpeg headers.
IEEE Transactions on Information Forensics and Security 6(3-2), 1066–1075 (2011)
23. Li, W., Yuan, Y., Yu, N.: Passive detection of doctored jpeg image via block artifact
grid extraction. IEEE Transactions on Signal Processing 89(9), 1821–1829 (2009)
24. Luo, W., Huang, J., Qiu, G.: JPEG error analysis and its applications to digital image
forensics. IEEE Transactions on Information Forensics and Security 5(3), 480 –491
(2010)
25. Martinez-Romo, J., Araujo, L.: Detecting malicious tweets in trending topics using a
statistical analysis of language. Expert Systems with Applications 40(8), 2992–3000
(2013)
26. Middleton, S.E.: Extracting attributed verification and debunking reports from social
media: Mediaeval-2015 trust and credibility analysis of image and video (2015)
27. Middleton, S.E., Middleton, L., Modafferi, S.: Real-time crisis mapping of natural dis-
asters using social media. Intelligent Systems, IEEE 29(2), 9–17 (2014)
28. O’Brien, J.F., Farid, H.: Exposing photo manipulation with inconsistent reflections.
ACM Transactions on Graphics 31(1), 4:1–4:11 (2012)
29. O’Donovan, J., Kang, B., Meyer, G., Hollerer, T., Adalii, S.: Credibility in context:
An analysis of feature distributions in twitter. In: Privacy, Security, Risk and Trust
(PASSAT), 2012 International Conference on and 2012 International Confernece on
Social Computing (SocialCom), pp. 293–301. IEEE (2012)
30. Pasquini, C., Boato, G., Perez-Gonzalez, F.: Multiple jpeg compression detection by
means of benford-fourier coefficients. In: Proceedings of the Workshop on Information
Forensics and Security. IEEE (2014)
31. Pasquini, C., Perez-Gonzalez, F., Boato, G.: A benford-fourier jpeg compression detec-
tor. In: Proceedings of the International Conference on Image Processing, pp. 5322–5326.
IEEE (2014)
32. Procter, R., Vis, F., Voss, A.: Reading the riots on twitter: methodological innovation
for the analysis of big data. International Journal of Social Research Methodology
16(3), 197–214 (2013)
33. Ratkiewicz, J., Conover, M., Meiss, M., Gon¸calves, B., Patil, S., Flammini, A., Menczer,
F.: Truthy: mapping the spread of astroturf in microblog streams. In: Proceedings of
26 Christina Boididou et al.
the 20th international conference companion on World wide web, pp. 249–252. ACM
(2011)
34. Silverman, C.: Verification handbook: a definitive guide to verifying digital content for
emergency coverage (2013)
35. Starbird, K., Muzny, G., Palen, L.: Learning from the crowd: Collaborative filtering
techniques for identifying on-the-ground twitterers during mass disruptions. In: Proc.
9th Int. Conf. Inf. Syst. Crisis Response Manag. Iscram (2012)
36. Stringhini, G., Kruegel, C., Vigna, G.: Detecting spammers on social networks. In:
Proceedings of the 26th Annual Computer Security Applications Conference, pp. 1–9.
ACM (2010)
37. Tsakalidis, A., Papadopoulos, S., Kompatsiaris, I.: An ensemble model for cross-domain
polarity classification on twitter. In: Web Information Systems Engineering–WISE 2014,
pp. 168–177. Springer (2014)
38. Zampoglou, M., Papadopoulos, S., Kompatsiaris, Y.: Detecting image splicing in the
wild (WEB). In: 2015 IEEE International Conference on Multimedia & Expo Work-
shops, ICME Workshops 2015, Turin, Italy, June 29 - July 3, 2015, pp. 1–6 (2015).
DOI 10.1109/ICMEW.2015.7169839
39. Zampoglou, M., Papadopoulos, S., Kompatsiaris, Y.: Large-scale evaluation of splicing
localization algorithms for web images. Multimedia Tools and Applications pp. 1–34
(2016)
40. Zhao, Z., Resnick, P., Mei, Q.: Enquiring minds: Early detection of rumors in social
media from enquiry posts. In: Proceedings of the 24th International Conference on
World Wide Web, pp. 1395–1405. International World Wide Web Conferences Steering
Committee (2015)
41. Zubiaga, A., Aker, A., Bontcheva, K., Liakata, M., Procter, R.: Detection and resolution
of rumours in social media: A survey. arXiv preprint arXiv:1704.00656 (2017)
... Here, we focus on image-caption pairs that collaboratively contribute to the dissemination of misinformation. For instance, in Fig. 1a, an image depicts the grounds of a musical festival covered in garbage, accompanied by the false claim that it was taken in June 2022 "after Greta Thunberg's environmentalist speech", while the image was actually taken in 2015. 1 Previous studies on automated multimodal misinformation detection (MMD) have predominantly explored three approaches in terms of training datasets: annotated [9,55], weakly annotated [20,35,38] and synthetically generated datasets [19,31,44]. These distinct routes facilitated the development and evaluation of multimodal models designed to detect and combat misinformation effectively [22,45,50,52]. ...
... If these patterns persist within the evaluation benchmarks, they can obscure the impact of unimodal bias, hindering our ability to effectively assess progress in the field of MMD. In our investigation, we uncover that the widely used VMU-Twitter dataset [9] exhibits an image-side unimodal bias, while the COSMOS evaluation benchmark [4] exhibits a text-side unimodal bias, raising questions about their reliability as evaluation benchmarks for MMD. ...
... The "image verification corpus", often referred as the "Twitter" dataset, ("VMU-Twitter" from now on) was used in the MediaEval2016 Verifying Multimedia Use (VMU) challenge [9] and comprises 16,440 tweets regarding 410 images for training and 1,090 tweets regarding 104 images for evaluation. Since the images are accompanied by tweets, the dataset has been widely used for MMD [22,45,50,52]. ...
Article
Full-text available
Multimedia content has become ubiquitous on social media platforms, leading to the rise of multimodal misinformation (MM) and the urgent need for effective strategies to detect and prevent its spread. In recent years, the challenge of multimodal misinformation detection (MMD) has garnered significant attention by researchers and has mainly involved the creation of annotated, weakly annotated, or synthetically generated training datasets, along with the development of various deep learning MMD models. However, the problem of unimodal bias has been overlooked, where specific patterns and biases in MMD benchmarks can result in biased or unimodal models outperforming their multimodal counterparts on an inherently multimodal task, making it difficult to assess progress. In this study, we systematically investigate and identify the presence of unimodal bias in widely used MMD benchmarks, namely VMU-Twitter and COSMOS. To address this issue, we introduce the “VERification of Image-TExt pairs” (VERITE) benchmark for MMD which incorporates real-world data, excludes “asymmetric multimodal misinformation” and utilizes “modality balancing”. We conduct an extensive comparative study with a transformer-based architecture that shows the ability of VERITE to effectively address unimodal bias, rendering it a robust evaluation framework for MMD. Furthermore, we introduce a new method—termed Crossmodal HArd Synthetic MisAlignment (CHASMA)—for generating realistic synthetic training data that preserve crossmodal relations between legitimate images and false human-written captions. By leveraging CHASMA in the training process, we observe consistent and notable improvements in predictive performance on VERITE; with a 9.2% increase in accuracy. We release our code at: https://github.com/stevejpapad/image-text-verification
... Currently, multiple methods for rumor verification are available based on different approaches of machine learning (Castillo et al. 2011;Gupta et al. 2013;Granik and Mesyura 2017), natural language processing (Wiegand and Middleton 2016;Dang 2016;Boididou et al. 2018a), source authentication (Ratkiewicz et al. 2011;Gupta et al. 2014;Shao et al. 2016), empirical studies (Ozturk et al. 2015;Chua and Banerjee 2017a;Wang et al. 2018), web semantics (Barboza et al. 2013), and modeling/simulations (Kermack and McKendrick 1939); however, most of these approaches partially address the problem. For instance, machine learning approaches rely on model training which can be defeated by fabricating the rumor and altering its characteristics. ...
... They invited participants on a shared group task to analyze false information and figured out that the activity had a positive impact on rumor identification. Boididou et al. (2018a) compared three automated NLP approaches (textual patterns, information relevancy, and semi-supervised scheme) for false information classification on Twitter datasets. They also combined these methods to achieve an accuracy improvement of approximately 7%. ...
Article
Full-text available
Online social networks (OSNs) are now implied as an important source of news and information besides establishing social connections. However, such information sharing is not always authentic because people, sometimes, also share their perceptions and fabricated information on OSNs. Thus, verification of online posts is important to maintain reliability over this useful communication medium. To address this concern, multiple approaches have been investigated including machine learning, natural language processing, source authentication, empirical studies, web semantics, and modeling/simulations, but the problem still persists. This research proposes an effective synergy-based rumor verification method along with a weighted-mean reputation management system to mitigate the spread of rumors over OSN. The model was formally verified through Colored Petri-Nets while its semantic behavior was analyzed through ontologies. Moreover, a third-party Facebook application was developed for proof of concept, and users’ acceptance and usability analysis was performed through Technology Acceptance Model and Self-Efficacy scale. The results indicate that the proposed approach can be used as an effective tool for the identification of rumors and it has the potential to improve the quality of users’ online experience.
... For this reason, researchers have been experimenting with training multimodal misinformation detection models on algorithmically created datasets such as NewsCLIPpings and MEIR [35] and evaluating their generalizability to real-world misinformation [10]. Due to unimodal biases in widely adopted benchmarks, such as VMU Twitter [43] and COSMOS [7], we recently developed VERITE, an evaluation benchmark for real-world multimodal misinformation detection that accounts for unimodal biases [29]. However, VERITE and other similar evaluation datasets have not yet been used for MFC since they do not provide external evidence. ...
Article
Full-text available
Online misinformation is often multimodal in nature, i.e., caused by misleading associations between texts and accompanying images. To support the fact-checking process, researchers have been recently developing automatic multimodal methods that gather and analyze external information, evidence, related to the image–text pairs under examination. However, prior works incorrectly assumed that all external information collected from the Web is relevant. In this study, we introduce a “relevant evidence detection” (RED) module to discern whether each piece of evidence is relevant, to support or refute the claim. Specifically, we develop the “relevant evidence detection directed transformer” (RED-DOT) and explore multiple architectural variants (e.g., single or dual-stage) and mechanisms (e.g., “guided attention”). Extensive ablation and comparative experiments demonstrate that RED-DOT outperforms the state-of-the-art (SotA), achieving up to 33.7% accuracy improvement on the VERITE benchmark. Furthermore, our evidence reranking and element-wise modality fusion led to RED-DOT surpassing the SotA on NewsCLIPpings+ by up to 3% without the need for numerous evidence or multiple backbone encoders. We release our code at: https://github.com/stevejpapad/relevant-evidencedetection .
... Brezeale and Cook [124] provided a survey of existing video classification methods that classify videos using text features, audio features, visual features, and combinations. Boididou et al. [125] have reviewed various methodologies for classifying multimedia data on Twitter that include verifying cues, assessing the source and user credibility, content credibility, image forensics, verifying multimedia use task and have described the verification approaches used, namely UoS-ITI, MCG-ICT, and CERTH-UNITN. Anoop et al. [126] deeply studied fake news detection methods on textual modality, image modality, network modality, temporal modality, and knowledge-based approaches. ...
Article
Full-text available
The proliferation of disinformation and misinformation across diverse digital platforms poses a significant societal challenge. Previous work in this area adequately addresses the false news detection on the online text shared. The technological platforms that have enough study and research carried out on them include the web news, as well as Twitter. This emerging field is gaining grounds for Facebook, Reddit, WhatsApp, YouTube, among other social media apps. Online data are analyzed using different modalities of text, images, videos, and speech with other sources of influence. It ended with the effectiveness of multi-modal integration in the identification of misinformation. To combat this issue, the development of robust and accurate detection techniques is imperative. This review delves into the multifaceted nature of this problem, exploring the intricacies of multi-modal fake news detection. It examines how the integration of text, image, video, and audio modalities can enhance detection accuracy. At present, there is an abundance of surveys consolidating textual fake news detection algorithms. This review primarily deals with multi-modal fake news detection techniques that include images, videos, and their combinations with text. We provide a comprehensive literature survey of eighty articles presenting state-of-the-art detection techniques, thereby identifying research gaps and building a pathway for researchers to further advance this domain.
... The number of fact-checking media has reliably skyrocketed over the years, from only 11 sites in 2008 to 424 in 2022; however, although misinformation continues to grow, in the last five years, fact-checking has levelled off [73]. Throughout this time, verification processes have become a journalistic bet [74], with specific tools [75,76], multiple platforms (FactCheck.org, Politifact.com, ...
Article
Full-text available
Natural disasters linked to contexts of unpredictability and surprise generate a climate of uncertainty in the population, resulting in an exponential increase in disinformation. These are crisis situations that cause the management of public and governmental institutions to be questioned, diminish citizens’ trust in the media, and reinforce anonymity in social networks. New digital algorithms create a scenario plagued by fake news and levels of viralization of rumors never before contemplated. Our objective is to analyze the verification capacity of fact-checking agencies at X at times of information disorder, such as the Turkey–Syria earthquakes in 2023. We apply a mixed methodology of comparative content analysis to government, news agency, and IFCN accounts, generating a general sample (n = 46,747) that is then subjected to thematic categorization to create a specific sample (n = 564). The results indicate a low commitment to fact-checking on the part of official bodies and news agencies, as opposed to fact-checking agencies’ accurate handling of the facts. The lack of debate and engagement generated by digital audiences in the face of the discursive intentionality of disinformation is significant.
... Research has studied the phenomena of spreading fake news and rumors on social media platforms using ML algorithms. Such research efforts have encompassed various areas, including content analysis [6], studying the characteristics and features of fake news propagation [7], analyzing the features of users' profiles [8], and studying the enormous effect of social media bots to amplify the impact of fake news expansion [9], [10]. Other studies focused on the correlation between social media bots and accounts' fabrication and social spamming [11]. ...
Article
Full-text available
This paper investigates the use of multimedia verification, in particular, computational tools and Open-source Intelligence (OSINT) methods, for verifying online multimedia content in the context of the ongoing wars in Ukraine and Gaza. Our study examines the workflows and tools used by several fact-checkers and journalists working at Faktisk, a Norwegian fact-checking organization. Our study showcases the effectiveness of diverse resources, including AI tools, geolocation tools, internet archives, and social media monitoring platforms , in enabling journalists and fact-checkers to efficiently process and corroborate evidence, ensuring the dissemination of accurate information. This research provides an in-depth analysis of the role of computational tools and OSINT methods for multimedia verification. It also underscores the potentials of currently available technology, and highlights its limitations while providing guidance for future development of digital multimedia verification tools and frameworks.
Article
Fake news on social networks is a challenging problem due to the rapid dissemination and volume of information, as well as the ease of creating and sharing content anonymously. Fake news stories are problematic not only for the credibility of online journalism, but also due to their detrimental real-world consequences. The primary research objective of this study is: What are the recent state-of-the-art modalities based on deep learning to detect fake news in social networks. This paper presents a systematic literature review of deep learning-based fake news detection models in social networks. The methodology followed a rigorous approach, including predefined criteria for study selection of deep learning modalities. This study focuses on the types of deep learning modalities; unimodal (refers to the use of a single model for analysis or modeling purposes) and multimodal models (refers to the integration of multiple models). The results of this review reveal the strengths and weaknesses of modalities approaches, as well as the limitations of low-resource languages datasets. Furthermore, it provides insights into future directions for deep learning models and different fact checking techniques. At the end of this study, we discuss the problem of fake news detection in the era of large language models in terms of advantages, drawbacks, and challenges.
Chapter
Effective, safe, and fast identity recognition is crucial in today’s rapidly growing society. As a convenient and reliable alternative to traditional identification methods, biometric technologies are increasingly adopted for security applications, such as the verification of ID cards or passports and the authentication of computer and mobile devices. However, if spoofed, such technologies can create serious privacy and security risks, and the proliferation of high quality multimedia content on social media platforms facilitates such spoofing attacks. Unfortunately, many users are unaware of the risks of posting their biometric information online and social media companies are not taking appropriate action to protect them. In this paper, we make the case for biometric anti-spoofing policies by examining the social media enabled marketplace of Etsy. We demonstrate that biometric information can be collected from social media users and that the level of privacy concerns is not a predictor of a user’s biometric information sharing behavior.KeywordsPrivacy in Social-mediaDigital multimedia forensicsBiometrics
Technical Report
Full-text available
Online social media has become an integral part of every Internet users' life. It has given common people a platform and forum to share information, post their opinions and promote campaigns. The threat of exploitation of social media like Facebook, Twitter, etc. by malicious entities, becomes crucial during a crisis situation, like bomb blasts or natural calamities such as earthquakes and floods. In this report, we attempt to characterize and extract patterns of activity of general users on Twitter during a crisis situation. This is the first attempt to study an India-centric crisis event such as the triple bomb blasts in Mumbai (India), using online social media. In this research, we perform content and activity analysis of content posted on Twitter after the bomb blasts. Through our analysis, we conclude, that the number of URLs and @-mentions in tweets increase during the time of the crisis in comparison to what researchers have exhibited for normal circumstances. In addition to the above, we empirically show that the number of tweets or updates by authority users (those with large number of followers) are very less, i.e. majority of content generated on Twitter during the crisis comes from non authority users. In the end, we discuss certain case scenarios during the Mumbai blasts, where rumors were spread through the network of Twitter.
Conference Paper
Full-text available
The publication and spread of misleading content is a problem of increasing magnitude, complexity and consequences in a world that is increasingly relying on user-generated content for news sourcing. To this end, multimedia analysis techniques could serve as an assisting tool to spot and debunk misleading content on the Web. In this paper, we tackle the problem of misleading multimedia content detection on Twitter in real time. We propose a number of new features and a new semi-supervised learning event adaptation approach that helps generalize the detection capabilities of trained models to unseen content, even when the event of interest is of different nature compared to that used for training. Combined with bagging, the proposed approach manages to outperform previous systems by a significant margin in terms of accuracy. Moreover, in order to communicate the verification process to end users, we develop a web-based application for visualizing the results.
Article
Full-text available
Despite the increasing use of social media platforms for information and news gathering, its unmoderated nature often leads to the emergence and spread of rumours, i.e. pieces of information that are unverified at the time of posting. At the same time, the openness of social media platforms provides opportunities to study how users share and discuss rumours, and to explore how natural language processing and data mining techniques may be used to find ways of determining their veracity. In this survey we introduce and discuss two types of rumours that circulate on social media; long-standing rumours that circulate for long periods of time, and newly-emerging rumours spawned during fast-paced events such as breaking news, where reports are released piecemeal and often with an unverified status in their early stages. We provide an overview of research into social media rumours with the ultimate goal of developing a rumour classification system that consists of four components: rumour detection, rumour tracking, rumour stance classification and rumour veracity classification. We delve into the approaches presented in the scientific literature for the development of each of these four components. We summarise the efforts and achievements so far towards the development of rumour classification systems and conclude with suggestions for avenues for future research in social media mining for detection and resolution of rumours.
Conference Paper
Full-text available
This paper provides an overview of the Verifying Multimedia Use task that takes places as part of the 2016 MediaEval Benchmark. The task motivates the development of automated techniques for detecting manipulated and misleading use of web multimedia content. Splicing, tampering and reposting videos and images are examples of manipulation that are part of the task definition. For the 2016 edition of the task, a corpus of images/videos and their associated posts is made available, together with labels indicating the appearance of misuse (fake) or not (real) in each case as well as some useful post metadata.
Article
Full-text available
With the proliferation of smartphones and social media, journalistic practices are increasingly dependent on information and images contributed by local bystanders through Internet-based applications and platforms. Verifying the images produced by these sources is integral to forming accurate news reports, given that there is very little or no control over the type of user-contributed content, and hence, images found on the Web are always likely to be the result of image tampering. In particular, image splicing, i.e. the process of taking an area from one image and placing it in another is a typical such tampering practice, often used with the goal of misinforming or manipulating Internet users. Currently, the localization of splicing traces in images found on the Web is a challenging task. In this work, we present the first, to our knowledge, exhaustive evaluation of today’s state-of-the-art algorithms for splicing localization, that is, algorithms attempting to detect which pixels in an image have been tampered with as the result of such a forgery. As our aim is the application of splicing localization on images found on the Web and social media environments, we evaluate a large number of algorithms aimed at this problem on datasets that match this use case, while also evaluating algorithm robustness in the face of image degradation due to JPEG recompressions. We then extend our evaluations to a large dataset we formed by collecting real-world forgeries that have circulated the Web during the past years. We review the performance of the implemented algorithms and attempt to draw broader conclusions with respect to the robustness of splicing localization algorithms for application in Web environments, their current weaknesses, and the future of the field. Finally, we openly share the framework and the corresponding algorithm implementations to allow for further evaluations and experimentation.
Conference Paper
Full-text available
Polarity analysis of Social Media content is of significant importance for various applications. Most current approaches treat this task as a classification problem, demanding a labeled corpus for training purposes. However, if the learned model is applied on a different domain, the performance drops significantly and, given that it is impractical to have labeled corpora for every domain, this becomes a challenging task. In the current work, we address this problem, by proposing an ensemble classifier that is trained on a general domain and and adapts, without the need for additional ground truth, on the desired (test) domain before classifying a document. Our experiments are performed on three different datasets and the obtained results are compared with various baselines and state-of-the-art methods; we demonstrate that our model is outperforming all out-of-domain trained baseline algorithms, and that it is even comparable with different in-domain classifiers.
Conference Paper
Full-text available
We propose an approach that predicts whether a tweet, which is accompanied by multimedia content (image/video), is trustworthy or deceptive. We test different combinations of quality and trust-oriented features (tweet-based, user-based and forensics) in tandem with a standard classification and an agreement-retraining technique, with the goal of predicting the most likely label (fake or real) for each tweet. The experiments carried out on the Verifying Multimedia Use dataset show that the best performance is achieved when using all available features in combination with the agreement-retraining method.
Article
Existing tools for exploratory analysis of information diffusion in social media focus on the message senders who actively diffuse the meme. We develop a tool for audience analysis, focusing on the people who are passively exposed to the messages, with a special emphasis on competing memes such as propagations and corrections of a rumor. In such competing meme diffusions, important questions include which meme reached a bigger total audience, the overlap in audiences of the two, and whether exposure to one meme inhibited propagation of the other. We track audience members’ states of interaction, such as having been exposed to one meme or another or both. We analyze the marginal impact of each message in terms of the number of people who transition between states as a result of that message. These marginal impacts can be computed efficiently, even for diffusions involving thousands of senders and millions of receivers. The marginal impacts provide the raw material for an interactive tool, RumorLens, that includes a Sankey diagram and a network diagram. We validate the utility of the tool through a case study of nine rumor diffusions. We validate the usability of the tool through a user study, showing that nonexperts are able to use it to answer audience analysis questions.
Article
Fake news spreading in social media severely jeopardizes the veracity of online content. Fortunately, with the interactive and open features of microblogs, skeptical and opposing voices against fake news always arise along with it. The conflicting information, ignored by existing studies, is crucial for news verification. In this paper, we take advantage of this "wisdom of crowds" information to improve news verification by mining conflicting viewpoints in microblogs. First, we discover conflicting viewpoints in news tweets with a topic model method. Based on identified tweets' viewpoints, we then build a credibility propagation network of tweets linked with supporting or opposing relations. Finally, with iterative deduction, the credibility propagation on the network generates the final evaluation result for news. Experiments conducted on a real-world data set show that the news verification performance of our approach significantly outperforms those of the baseline approaches.
Conference Paper
Many previous techniques identify trending topics in social media, even topics that are not pre-defined. We present a technique to identify trending rumors, which we define as topics that include disputed factual claims. Putting aside any attempt to assess whether the rumors are true or false, it is valuable to identify trending rumors as early as possible. It is extremely difficult to accurately classify whether every individual post is or is not making a disputed factual claim. We are able to identify trending rumors by recasting the problem as finding entire clusters of posts whose topic is a disputed factual claim. The key insight is that when there is a rumor, even though most posts do not raise questions about it, there may be a few that do. If we can find signature text phrases that are used by a few people to express skepticism about factual claims and are rarely used to express anything else, we can use those as detectors for rumor clusters. Indeed, we have found a few phrases that seem to be used exactly that way, including: "Is this true?", "Really?", and "What?". Relatively few posts related to any particular rumor use any of these enquiry phrases, but lots of rumor diffusion processes have some posts that do and have them quite early in the diffusion. We have developed a technique based on searching for the enquiry phrases, clustering similar posts together, and then collecting related posts that do not contain these simple phrases. We then rank the clusters by their likelihood of really containing a disputed factual claim. The detector, which searches for the very rare but very informative phrases, combined with clustering and a classifier on the clusters, yields surprisingly good performance. On a typical day of Twitter, about a third of the top 50 clusters were judged to be rumors, a high enough precision that human analysts might be willing to sift through them.