Chapter

Evaluation of the Existing Tools for Fake News Detection

Authors:
To read the full-text of this research, you can request a copy directly from the authors.

Abstract

The extreme growth and adoption of Social Media and User Generated Content (UGC) websites, in combination with their poor governance and the lack of quality control over the digital content being published and shared, has led information veracity to a continuous deterioration. Therefore, there is a growing need for reliable information assurance, called by both private and public users and authorities. Due to the popularity of the social media and Internet availability all over the world, anyone can provide a piece of information on the web. This may create a ready channel for spreading false, not verified or confusing information, which may be called ‘fake news’. In order to protect the user from online disinformation, some tools have already been proposed. In this article we have described the available online tools and evaluated them using the same dataset, which was created for this study. We have provided the results proving that fake news detection is an increasingly more pressing, yet a difficult research problem.

No full-text available

Request Full-text Paper PDF

To read the full-text of this research,
you can request a copy directly from the authors.

... The U.S. Department of Defence started to invest about $70 M in order to deploy military technology to detect fake contents as they impact on national security 17 . This Media Forensics Program at the Defence Advanced Research Projects Agency (DARPA) started 4 years ago. ...
... From the technological perspective, it is worth highlighting that numerous online tools have been developed for misinformation detection. Some available approaches were presented in [17]. This technological development to combat disinformation is led by tech/media companies. ...
Article
Fake news has now grown into a big problem for societies and also a major challenge for people fighting disinformation. This phenomenon plagues democratic elections, reputations of individual persons or organizations, and has negatively impacted citizens, (e.g., during the COVID-19 pandemic in the US or Brazil). Hence, developing effective tools to fight this phenomenon by employing advanced Machine Learning (ML) methods poses a significant challenge. The following paper displays the present body of knowledge on the application of such intelligent tools in the fight against disinformation. It starts by showing the historical perspective and the current role of fake news in the information war. Proposed solutions based solely on the work of experts are analysed and the most important directions of the application of intelligent systems in the detection of misinformation sources are pointed out. Additionally, the paper presents some useful resources (mainly datasets useful when assessing ML solutions for fake news detection) and provides a short overview of the most important R&D projects related to this subject. The main purpose of this work is to analyse the current state of knowledge in detecting fake news; on the one hand to show possible solutions, and on the other hand to identify the main challenges and methodological gaps to motivate future research.
... Many other projects and services related to this work, which have been reviewed in the literature, c.f. [25,26,27,11,28,29,30], now no longer appear to be active or working. This is concerning as despite the fact that misinformation and disinformation have been recognised as a threat to democracy and social cohesion, and the fact that browser plugins are one of the few citizen-orientated direct interventions which can help solve the problem at source while increasing long term media literacy, very few of the proposed solutions have been actively promoted or maintained. ...
Preprint
Full-text available
The threat posed by misinformation and disinformation is one of the defining challenges of the 21st century. Provenance is designed to help combat this threat by warning users when the content they are looking at may be misinformation or disinformation. It is also designed to improve media literacy among its users and ultimately reduce susceptibility to the threat among vulnerable groups within society. The Provenance browser plugin checks the content that users see on the Internet and social media and provides warnings in their browser or social media feed. Unlike similar plugins, which require human experts to provide evaluations and can only provide simple binary warnings, Provenance's state of the art technology does not require human input and it analyses seven aspects of the content users see and provides warnings where necessary.
... These features are further grouped into six sub-classes: -Network features -built by extracting the statistical features for mentions, retweets, and hashtag co-occurrence. -Sentiment features -built by using a sentiment analysis algorithm that takes into account happiness, emotion scores, etc. • Decodex 6 : An online fake news detection tool that alerts the user to the potential of fake news by labeling the information as 'satire', 'info' and 'no information' [61]. • TrustyTweet 7 : TrustyTweet is a browser plug-in, proposed for twitter users to assess and increase media literacy. ...
Article
Full-text available
Social Networks' omnipresence and ease of use has revolutionized the generation and distribution of information in today's world. However, easy access to information does not equal an increased level of public knowledge. Unlike traditional media channels, social networks also facilitate faster and wider spread of disinformation and misinformation. Viral spread of false information has serious implications on the behaviours, attitudes and beliefs of the public, and ultimately can seriously endanger the democratic processes. Limiting false information's negative impact through early detection and control of extensive spread presents the main challenge facing researchers today. In this survey paper, we extensively analyse a wide range of different solutions for the early detection of fake news in the existing literature. More precisely, we examine Machine Learning (ML) models for the identification and classification of fake news, online fake news detection competitions, statistical outputs as well as the advantages and disadvantages of some of the available data sets. Finally, we evaluate the online web browsing tools available for detecting and mitigating fake news and present some open research challenges.
... The fake news problem is the emerging one for the researchers working on the computer science. Even though, there are plenty of online tools designed for fake news detecting, they are not sufficiently successful [10]. Thus, some project founded by European Commission were introduced eg. ...
Chapter
The problem of the fake news publication is not new and it already has been reported in ancient ages, but it has started having a huge impact especially on social media users. Such false information should be detected as soon as possible to avoid its negative influence on the readers and in some cases on their decisions, e.g., during the election. Therefore, the methods which can effectively detect fake news are the focus of intense research. This work focuses on fake news detection in articles published online and on the basis of extensive research we confirmed that chosen machine learning algorithms can distinguish them from reliable information.
Article
The study's main aim is to detect the fake news from the political information. To detecting this types of fake news we are implementing more effective machine learning classifier algorithm. These algorithms will compare their performance by using methods and materials. The materials and methods are belonging to the LRA (Logistic Regression algorithm) and the NBA (Naive Bayes algorithm). These two types of methods and algorithms are implemented to test the data's. This algorithm tested more than 44,000 records. These two algorithms are performed well to detect the fake information from the dataset. The experimental performance rate is N = 10. These two algorithms identified different types of fake news and provide accurate news classification. The eighty percentage result are used by the G-power test. Finally, the experimental accuracy results of the fake political news for these two algorithms as follows 98.7080 using the Logistic Regression algorithm and 94.8490 using the Naive Bayes algorithm and 0.013 is the statistical difference between these two algorithms. We used T-test to find out the accuracy of the independent samples. When we compared these two results the Logistic Regression algorithm provided better performance.
Article
Full-text available
Recent progress in the area of modern technologies confirms that information is not only a commodity but can also become a tool for competition and rivalry among governments and corporations, or can be applied by ill-willed people to use it in their hate speech practices. The impact of information is overpowering and can lead to many socially undesirable phenomena, such as panic or political instability. To eliminate the threats of fake news publishing, modern computer security systems need flexible and intelligent tools. The design of models meeting the above-mentioned criteria is enabled by artificial intelligence and, above all, by the state-of-the-art neural network architectures, applied in NLP tasks. The BERT neural network belongs to this type of architectures. This paper presents Transformer-based hybrid architectures applied to create models for detecting fake news.
Chapter
Recent progress in the area of modern technologies confirms that information is not only a commodity but can also become a tool for competition and rivalry among governments and corporations, or can be applied by ill-willed people to use it in their hate speech practices. The impact of information is overpowering and can lead to many socially undesirable phenomena, such as panic or political instability. To eliminate the threats of fake news publishing, modern computer security systems need flexible and intelligent tools. The design of models meeting the above-mentioned criteria is enabled by artificial intelligence, and above all by the state-of-the-art neural network architectures, applied in NLP tasks. The BERT neural network belongs to this type of architectures. This paper presents a hybrid architecture connecting BERT with RNN; the architecture was used to create models for detecting fake news.
Chapter
This paper presents the problem of calculating the value of the scoring function for weak classifiers operating in the sequential structure. An example of such a structure is Gentle AdaBoost algorithm whose modification we propose in this work. In the proposed approach the distance of the object from the decision boundary is scaled in decision regions defined by the weak classifier at first and later transformed by the log-normal function. The described algorithm was tested on sixth public available data sets and compared with Gentle AdaBoost algorithm.
Chapter
Full-text available
Data Stream represents a significant challenge for data analysis and data mining techniques because those techniques are developed based on training batch data. Classification technique that deals with data stream should have the ability for adapting its model for the new samples and forget the old ones. In this paper, we present an intensive comparison for the performance of six of popular classification techniques and focusing on the power of Adaptive Random Forest. The comparison was made based on four real medical datasets and for more reliable results, 40 other datasets were made by adding white noise to the original datasets. The experimental results showed the dominant of Adaptive Random Forest over five other techniques with high robustness against the change in data and noise.
Article
Full-text available
Nowadays social media is widely used as the source of information because of its low cost, easy to access nature. However, consuming news from social media is a double-edged sword because of the wide propagation of fake news, i.e., news with intentionally false information. Fake news is a serious problem because it has negative impacts on individuals as well as society large. In the social media the information is spread fast and hence detection mechanism should be able to predict news fast enough to stop the dissemination of fake news. Therefore, detecting fake news on social media is an extremely important and also a technically challenging problem. In this paper, we present FakeNewsTracker, a system for fake news understanding and detection. As we will show, FakeNewsTracker can automatically collect data for news pieces and social context, which benefits further research of understanding and predicting fake news with effective visualization techniques.
Article
This paper outlines and proposes a new approach to cyber attack detection on the basis of the practical application of the efficient lifelong learning cybersecurity system. One of the main difficulties in machine learning is to build intelligent systems that are capable of learning sequential tasks and then to transfer knowledge from a previously learnt foundation to learn new tasks. Such capability is termed as Lifelong Machine Learning (LML) or as Lifelong Learning Intelligent Systems (LLIS). This kind of solution would promptly address the current problems in the cybersecurity domain, where each new cyber attack can be considered as a new task. Our approach is an extension of the Efficient Lifelong Learning (ELLA) framework. Hereby, we propose the new B-ELLA (Balanced ELLA) framework to detect cyber attacks and to counter the problem of network data imbalance. Our proposition is evaluated on a malware benchmark dataset and we achieve promising results.
Article
A large body of recent works has focused on understanding and detecting fake news stories that are disseminated on social media. To accomplish this goal, these works explore several types of features extracted from news stories, including source and posts from social media. In addition to exploring the main features proposed in the literature for fake news detection, we present a new set of features and measure the prediction performance of current approaches and features for automatic detection of fake news. Our results reveal interesting findings on the usefulness and importance of features for detecting false news. Finally, we discuss how fake news detection approaches can be used in the practice, highlighting challenges and opportunities.
Conference Paper
The topic of fake news has drawn attention both from the public and the academic communities. Such misinformation has the potential of affecting public opinion, providing an opportunity for malicious parties to manipulate the outcomes of public events such as elections. Because such high stakes are at play, automatically detecting fake news is an important, yet challenging problem that is not yet well understood. Nevertheless, there are three generally agreed upon characteristics of fake news: the text of an article, the user response it receives, and the source users promoting it. Existing work has largely focused on tailoring solutions to one particular characteristic which has limited their success and generality. In this work, we propose a model that combines all three characteristics for a more accurate and automated prediction. Specifically, we incorporate the behavior of both parties, users and articles, and the group behavior of users who propagate fake news. Motivated by the three characteristics, we propose a model called CSI which is composed of three modules: Capture, Score, and Integrate. The first module is based on the response and text; it uses a Recurrent Neural Network to capture the temporal pattern of user activity on a given article. The second module learns the source characteristic based on the behavior of users, and the two are integrated with the third module to classify an article as fake or not. Experimental analysis on real-world data demonstrates that CSI achieves higher accuracy than existing models, and extracts meaningful latent representations of both users and articles.
Conference Paper
One of the current challenges in machine learning is to develop intelligent systems that are able to learn consecutive tasks, and to transfer knowledge from previously learnt basis to learn new tasks. Such capability is termed as lifelong learning and, as we believe, it matches very well to counter current problems in cybersecurity domain, where each new cyber attack can be considered as a new task. One of the main motivations for our research is the fact that many cybersecurity solutions adapting machine learning are concerned as STL (Single Task Learning problem), which in our opinion is not the optimal approach (particularly in the area of malware detection) to solve the classification problem. Therefore, in this paper we present the concept applying the lifelong learning approach to cybersecurity (attack detection).
FakeNewsNet: a data repository with news content, social context and dynamic information for studying fake news on social media
  • K Shu
  • D Mahudeswaran
  • S Wang
  • D Lee
  • H Liu