Available via license: CC BY 4.0
Content may be subject to copyright.
Cascading dynamics of hate speech propagation: Unveiling
network structures and probability of retweeting on Twitter
Yueqiao Chen
Georgia Institute of Technology, College of Computing
801 Atlantic Drive Atlanta, GA 30332-0280
ychen3221@gatech.edu
Abstract. The ubiquity of social media has transformed communication patterns and interactions
in contemporary society, allowing individuals to share experiences, thoughts, and opinions on a
global scale. However, this unprecedented connectivity has also facilitated the dissemination of
hate speech, posing novel challenges for platforms, policymakers, and researchers. This study
delves into the dynamics of hate speech diffusion within social media networks, focusing on
Twitter as a case study. Through an extensive analysis of retweet behaviors among a subset of
hateful users, we investigate the intricate network structures and characteristics that underlie the
spread of harmful content. By analyzing network metrics and degree distributions and comparing
them with model approximations, we reveal insights into the polarized nature of hate speech
propagation, the role of strongly connected components, and the influence of a select few users.
The findings highlight the potential for viral spread within closely connected groups and
emphasize the need for nuanced approaches to curbing hate speech. Moreover, this study
contributes to the broader understanding of information dissemination on social media, paving
the way for informed strategies to foster a more responsible and inclusive digital discourse.
Keywords: Hate Speech, Social Media, Information Dissemination, Network Analysis, Retweet
Behavior, Social Network Dynamics, Network Structures, Diffusion of Information.
1. Introduction
In modern society, social media has become an indispensable part of our lives. Work, education, and
nearly every aspect of our lives has transformed as social media burgeons. Everyone was allowed to
share their knowledge, experience, emotions, and opinions with the public, seen by people having
similar interests. Social media shortens the connecting distance between people and provides an
extensive database of information people share. Individuals can easily find different information they
want on social media and connect with different kinds of people.
Social media, at the same time benefiting the users, also causes a bunch of new problems, for example,
the spread of hate speech. In the past, hateful words spoken by a person could not spread worldwide,
but it is possible now. The spread of free speech enables people with the same dissatisfaction to connect
and organize for parades and even horrible violence that unrest societal stability. Through no violence
coming up, kids or youths and even adults that have not had comprehensive minds or complete
understanding of an event will implicitly be misled by enthusiastically hateful users and gradually
become new hate infectors.
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
© 2024 The Authors. This is an open access article distributed under the terms of the Creative Commons Attribution License 4.0
(https://creativecommons.org/licenses/by/4.0/).
100
Hate speech has long been contentious, with one side advocating free speech and the other side
against it because of the current and potentially notorious influence on the stability of the society. In the
United States, hate speeches are allowed in most social media, but with some policies by individual
social media against its spread. Despite these effects on offensive content, spreading hate is still not an
entrenched problem in nearly all social media platforms. That is why an effective approach to detecting
and predicting the spread of hate speech is worthwhile to research.
There are numerous papers on detecting hate speeches on social media, from calculating the
offensiveness of sentences according to syntactic features to using deep learning to detect hate speech
[1-2]. These researches focus more on hate detection and classification in posts using natural language
processing models, but few are about the spread of hate speech. Although some work was on the retweet
graph and research on the spread of hate speech, they focus more on finding hateful users using
DeGroot's model [3-4]. So, there is a big gap in characterizing the network features of hate speech
diffusion and how to predict the spread of hate speech. Understanding the spread of hate speech is crucial
because it provides insights into how harmful content gets a wider audience so policymakers can develop
constructive strategies against hate speech.
The research questions are:
RQ1: What are some network structural features of hate speech diffusion? What decides the
spreading ability of each node?
This research used the retweets in hateful users to research the hate speech diffusion in hateful user
groups. By looking into the connectivity, degree distribution, and centralities, a better understanding of
the network structural features of hate speech diffusion could be made. Constructing a retweet graph
from a hateful user called a single-spot hate spread graph and analyzing the network would evaluate the
spreading ability of each node.
RQ2: What could be the approximate probability of retweeting a post of others in a hateful user group?
How much difference is there between an approximate Erdos-Renyi Graph and the hate-spread retweet
graph from a random hateful user?
Building up the Erdos-Renyi graphs to approximate the retweet graph from a hateful user could be
used to find an estimated probability of retweeting. The most approximate Erdos-Renyi graph could be
made by controlling the node number and the edge number.
And the study used Twitter as my data source. Twitter stands out among various social networks as
a prominent and crucial platform for research due to its real-time nature and significant data availability
[5]. With its succinct message format and unfiltered content stream, Twitter's popularity has surged,
particularly during events, with approximately 500 million daily tweets, often featuring news ahead of
traditional media sources, so Twitter could provide variable, realistic, and massive data for me.
Retweeting has emerged as a pivotal method for disseminating information on Twitter. Investigating the
mechanics of retweeting is crucial for comprehending information propagation within the platform [6].
Moreover, Twitter has long been denounced for not effectively handling discrimination and instigating
violence that even causes real damage and injury, although with the policy on hateful content. Therefore,
Twitter's retweet network in hateful users could be noteworthy to research about the spread of hate
speech.
2. Literature Review
Recent years have seen an increasing notice of hate speech in research. However, most papers focus on
detecting hateful users through tweets' contents and classifying using natural language processing and
deep learning. There have been researchers trying multiple ways, including different languages, machine
learning models, and algorithms, to detect different labels of hate speech on various platforms [7]. The
early research by Chen et al. came up with Lexical Syntactic Features to detect abusive language in
YouTube comments [1]. Dinakar et al. [8] then adopt a commonsense knowledge base under supervised
SVM in Cyberbullying detection. They used TF-IDF (Term Frequency - Inverse Document Frequency),
lexicon, POS (Part-of-speech)-bigrams, and topic-specific unigrams or bigrams as features, which are
also commonly used in later research. With the development of the neural network, there emerges a
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
101
spurt in natural language processing models using neural networks and deep learning, like LSTM (long
short-term memory networks), CNN (convolutional neural networks), BERT (Bidirectional Encoder
Representations from Transformers). For example, Badjatiya et al. tried a bunch of deep-learning models
for the detection of hateful tweets on Twitter [2]. Through the rapid update and improvement of the
machine learning model, the detection accuracy of hate speech has been greatly enhanced recently.
Detecting hate speech is critical, but how these hate speeches originated and spread through social
media is also worthy of research. It is common for individuals to have negative thoughts and words
about others, but social media platforms proliferate the hate, inform people with similar penchants, and
incite violence as a gang a possible thing. DeGroot's information diffusion model was applied to predict
hateful users in repost networks of Gab, a social media [3]. The model converts the directed repost edge
weights into the probability network of briefs, so if a user highly believes a hateful user, the user is
potentially hateful. Researchers collected data from Twitter constructs retweets graph implements
DeGroot's model with new attempts of using the lexicon in the initialization of a brief network [4]. These
two articles investigated the hate speech network but focused more on how to characterize hateful users
instead of the diffusion of hate speech. Masud et al. proposed a new neural network [9], RETINA, for
predicting retweet behaviors among users. They used topics of tweets as sound features presenting the
contents of tweets and achieved a high F-score. Though the topic-based model is ingenious, the topic
can sometimes be misleading and ambiguous.
Besides the research on hate speech spread, the characteristics of the retweet network are also
consequential. The retweets network of hateful users could have similar peculiarities as the larger
retweets graph. Nevertheless, some distinctions could also be revealed while considering the clusters of
hate speech. A few articles study and analyze the retweet graph and its features in statistical and network
analysis. The earliest paper on quantitative research in Twitter information diffusion conducts a bunch
of topological analyses, including the distribution of followers, degree of separation, and homophily of
the follower network [10]. It is shown that the Twitter follower network has a degree distribution
incompliant with the Power Law. The analysis of separation degree is shorter than that suggested by the
Small World Theory, which states that any user could reach another user in 6 steps. The analysis on the
retweet tree shows that the retweet of a post, once started, will go viral no matter the followers of the
original poster. This paper focused more on the follower graph, while Bild et al. analyzed the features
of the retweet graph more thoroughly and found that the retweet graph was compliant with the Power
Law and small-world Theory [11]. The retweet network of hate speech is predicted to have similar
features as a general retweet graph. Efstathiades et al. revisited the network presented by Kwak et al.
and concluded that the Twitter network has changed over time as the population grows from around 41
million to more than 500 million users [10][12]. It suggested that the largest strongly connected
components previously were divided into smaller strongly connected components and developed a
sparser tree. This can be explained by the growing number of people with different interests coming to
Twitter to form different collections.
3. Methodology
Because of the dearth of research on the spread of hate speech and the different focus, this study aimed
to investigate the diffusion of hate speech on Twitter through network analysis, which hasn't been fully
researched.
3.1. Dataset:
To construct the retweet network, we obtained a dataset from Kaggle [4], which comprises
approximately 100 thousand Twitter users. Among these users, we identified over 500 hateful users, and
each user was labelled with their extent of hate, categorized as 'hateful,' 'normal,' or 'other.' The dataset
also includes a file with all the directed edges in the retweet graph. More Twitter Data could be obtained
by the Twitter Streaming API or the Twitter REST API [13].
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
102
3.2. Constructing the Retweet Network:
To create the retweet network, we considered nodes as Twitter users, and directed edges represented
retweet relationships between those users. A directed edge from user A to user B indicates that user A
shared user B's content. However, the dataset presented a limitation as it did not contain the original
tweets of each user, making it challenging to distinguish hateful users solely based on their retweet
behavior accurately.
3.3. Network Analysis:
I started by visualizing the entire retweet network with all edges from hateful users. I first checked the
connectivity of each node and their strongly connected components. These could help me understand
the basic structure of the graph. Then, I plotted the degree distribution of the graph to see the nodes'
connectivity patterns.
However, due to the network's high density, analyzing the entire graph became cumbersome. As a
result, I focused on a subnetwork of hate spread from one hateful user. This individual spread of retweet
graph showed the ability of each node to spread messages using retweeting. By iterating all the hateful
users, I obtained all the spreading ability of the nodes. Analyzing these subnetworks gave me insights
into the single spot spread of hate speech.
3.4. Modeling the Retweet Network:
I modeled the retweet network using a directed Erdos-Renyi (ER) network to gain further understanding
and generate a more manageable comparison. We determined the optimal probability of the ER graph
for retweeting between two nodes by finding the nearest edge numbers. Through iterating all
approximations of single-spot retweet graphs, I got an average probability of retweeting between two
random nodes.
3.5. Evaluation of ER Graph Similarity:
To assess the similarity between the ER graph and the original hate graph, I compared several network
metrics. These metrics included maximum and average degree, various centralities (degree, eigenvector,
and betweenness centrality), average clustering coefficient, and graph density. This analysis shows how
well the ER graph represents the spread of hate speech in comparison to the original hate graph.
3.6. Limitations and Future Directions:
While this study sheds light on the spread of hate speech through network analysis, it is essential to note
some limitations. The dataset's lack of original tweets restricts our ability to identify hateful users based
solely on retweet behavior accurately. Additionally, this study focuses solely on Twitter data and may
not capture the entirety of hate speech spread on other platforms.
What's more, the degree distribution of my retweet network obeys the Power law, but the ER graph
gives a binomial distribution. That could explain why the original hate graph has a higher average degree
and max degree and a higher clustering coefficient but a lower closeness centrality. Thus, I would try
other models that obey the Power Law in future research.
3.7. Ethical Considerations:
It is important to acknowledge that this study deals with sensitive and potentially harmful content related
to hate speech. The dataset has taken measures to anonymize user data and prioritize user privacy. I
understood the potential implications of this research on individuals and communities, and I aimed to
handle the data with utmost responsibility.
4. Results
Before viewing the results in terms of the research question, I started with the dataset itself. Table 1
demonstrates the statistics of the dataset. The dataset contains three kinds of users: "hateful," "normal,"
and "other." It has around 100k users, but only 544 hateful users are confirmed to be hateful in the
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
103
dataset. The percentage of hateful users is only about 0.54% of the whole dataset. As for the connection
between the hateful user and other or normal users, 8.6% of normal users have connected with hateful
neighbors, and 2.7% of other users have hateful neighbors. In the dataset, the hateful user only takes up
a small portion of users.
Table 1: Dataset Statistics of Numbers of 'Hateful', 'Normal', 'Other' Users in Dataset
Property
Value
Total Number of Users
100,386
Number of Normal Users
4,427
Number of Other Users
95,415
Number of Hateful Users
544
Number of Normal Users with Hateful Neighbors
380
Number of Other Users with Hateful Neighbors
2,547
4.1. Retweet Network in Hateful Users:
To better understand the hate spread between hateful users, I checked the retweets between hateful users
and plotted the retweet graph for hateful users. Figure 1 shows hate spread in the hateful user group.
The directed retweet graph included all 544 hateful users as nodes, with connected hateful users as blue
and disconnected as red. It included all the retweets between different users, with 3237 edges. In the
graph, there are 136 disconnected hateful users, meaning that they don't have hateful neighbors
connected. The disconnected hateful users take up a quarter of all the hateful users. In the rest of the 408
connected hateful users, the largest strongly connected component's size in the connected hateful users
is 247. It illustrates that one-fourth of the hateful users don't have much dissemination ability, while in
the connected hateful users, there is a strongly connected component that links about half of them. The
diameter of the connected component, the maximum shortest distance between two distinct nodes, is 10,
so two of the connected hateful users can be reached in 10 steps.
Figure 1. Retweet Network in All Hateful Users
The retweet network containing all the hateful users is very dense to analyze and check for details,
so I tested individual hateful connected users' ability to spread hate speech by drawing a retweet started
from a selected connected hateful user. I began by choosing a random hateful user who is connected and
saw how the retweeting behavior is diffused. Figure 2 shows a single retweet spread from a randomly
picked, connected hateful user as a start, which shows the potential spread of a message by this user in
the hateful user group. In this graph, the picked user can spread tweets to 308 other hateful users. It is
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
104
dense in the center while sparse on the outside. There is a strongly connected component of size 247 in
the center.
Figure 2. Retweet Spread Graph from One Random Hateful User with 308 Nodes
By iterating all the hateful users, we could see the potential hate spreading ability from every hateful
user. Figure 3 shows the node numbers and max strongly connected component sizes of all single spotted
hate spread graphs from each hateful user. The x-axis is the 544 node indexes that the graph started from.
According to the graph, there are 296 graphs with over or equal to 308 nodes with 247 strongly
connected components, and 21 graphs with the largest strongly connected component of 9, while other
229 graphs with node numbers smaller than 9 and a maximum strongly connected component of either
size of 1 or 2. It shows that around 54% of the hateful users have the ability to spread a message to over
half of the hateful users because of the connection to the largest strongly connected component in the
graph, while the other proportion of the hateful users has a small ability to spread messages by retweeting
and 40% of hateful users are even isolated as disconnected hateful user.
Figure 3. Scatter Plot of the Node Number and Max Strongly Connected Component Size in All Single
Spot Retweet Spread Graphs
4.2. Degree distribution:
The degree distribution describes the frequency of nodes with different degrees in the graph. It provides
a valuable representation of important insights into the structural characteristics of the graph. Figure 4
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
105
shows the degree distribution of the hate graph of all retweets between hateful users, and Figure 5
illustrates the degree distribution of a random single-spot hate spread graph. These two graphs of degree
distributions indicate that the hate-spread retweet graph follows the Power Law. The Power Law
suggests only a small proportion of people have a higher degree, while most people have a very low
degree. As for the degree distribution of the hate graph of all retweets, there is only one user with a
degree of 141, while most users were under a degree of 5. Similarly, in the degree distribution of the
random retweet graph with 308 nodes, there is only one user with a degree of 138, while most nodes
have a degree under 20.
Figure 4. Degree Distribution of Retweet Graph in All Hateful Users
Figure 5. Degree Distribution of a Single Spot Retweet Spread Graph with 308 Nodes
4.3. Erodos-Renyi Modeling:
To simulate the most similar Erdos-Renyi graph, I started with finding the most suitable probability of
the Erodos-Renyi graph according to the edge and node number of the hate spread from a selected hateful
user. Figure 6 shows a simulated Erodos-Renyi model with 308 nodes. Apparently, the ER approximated
graph is much denser than the original single-spotted hate spread graph. The ER graph is a strongly
connected component with a 308-node strongly connected component, while the original graph has a
strongly connected component of 247. The average probability of all the 500 graphs is about 0.26. It is
shown that all graphs with node number over 308 has a probability of 0.03, all the graph of disconnected
node has a probability of 0, and for the graph with the max strongly connected component of 9, the
average probability is around 0.09, while the rest graphs with node number either 2 or 3 and a strongly
connected component of either 1 or 2, have very fluctuated probability.
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
106
Figure 6. Erdos-Renyi Approximation of a Single Spot Retweet Spread Graph with 308 Nodes
I chose single-spotted hate spread with 308 nodes for comparing the similarity of ER approximation
and the original graph because the probability for single-spot hate spread graphs with 308 or higher
nodes have a stable probability of 0.03, and the features are relatively stable. I compared the density,
max, and average degree, max and average degree centrality, max and average eigenvector centrality,
max and average closeness centrality, and average clustering coefficient between the two graphs in
Table 2. The table demonstrates a similar value in density, average degree centrality, and average
betweenness centrality between the two graphs, but the average degree, max degree, and max degree
centrality are much higher in the single-spot hate spread graph. For eigenvector centrality, the ER graph
has a slightly higher average eigenvector centrality but a much lower max eigenvector centrality.
Although the two graphs have similar average betweenness centrality, the original hate spread graph has
a higher max betweenness centrality. In closeness centrality, the ER approximated graph has a higher
average and max closeness centrality. The table also shows that the single-spot hate spread graph has a
much higher average clustering coefficient.
Table 2. Comparision Between Single-Spotted Hate Spread and Erdos-Renyi Approximated Graph
Single-spot hate spread graph
ER approximated graph
Density
0.03
0.03
Average Degree
18.6
18.9
Max Degree
138
32
Average Degree Centrality
0.0604
0.0614
Max Degree Centrality
0.4495
0.1042
Average Eigenvector Centrality
0.0409
0.0539
Max Eigenvector Centrality
0.2003
0.1231
Average Betweenness Centrality
0.0064
0.0058
Max Betweenness Centrality
0.1117
0.0157
Average Closeness Centrality
0.2423
0.3598
Max Closeness Centrality
0.3210
0.4104
Average Clustering Coefficient
0.1555
0.0325
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
107
The Erodos-Renyi graph's degree distribution is shown in Figure 7. The degree distribution of the
ER graph is completely different from the hate spread graph. The degree distribution of the ER graph
follows the Binomial Distribution, where most nodes have a related mean value of degree, while the
Power law says most nodes share a very low value of degree. In the graph, we could find that most nodes
in the ER graph have a degree in the range of 10 to 25. The max of the degree is 32, while the max
degree in the original graph is 138.
Figure 7. Degree Distribution of the Erdos-Renyi Approximation Graph
5. Discussion
Based on my results, we could give several insights. But firstly, in our data, the hateful users take up a
small proportion of the whole data, and the percentage of general users connected with the hateful user
is also relatively small. The dataset could be biased, so we needed more data to support the findings.
5.1. Research Insights
In the retweet network between all the hateful users, one-fourth of them are disconnected, meaning that
these hateful users are not connected with other hateful users, while there is a strongly connected
component in the connected part of the graph covering over half of the hateful users. The small diameter
of the connected part of the graph shows that hateful users are closely connected, but longer than what
the small-world theory suggests in 6 steps [14]. It suggests that hateful users have a polarized distribution.
They are either closely connected or isolated.
By iterating all the hateful users with their own subgraphs, called single-spotted hate-spread retweet
graphs, the study evaluated the individual ability to spread hate speech using retweets. The graphs are
dense in the center with a strongly connected component and sparse on the outside with sink nodes,
meaning these nodes didn't retweet others' posts but were retweeted by others. This study also found
that the single-spotted retweet graphs' node number is heavily dependent on the size of the max strongly
connected component in the graph. Because each strongly connected component is not connected, nodes
can only connect to one of the strongly connected components, and the node number is within a range.
The node number of each sub-retweet graph shows the number of nodes the initial node could spread
and represents the spreading ability of the start node, so the spreading ability depends on the size of the
strongly connected component it connects.
The degree distributions of both the whole retweet graph of all hateful users and the single-spotted
retweet graph obey the Power Law, showing that nodes with higher degrees are scarce, while nodes with
low degrees are much. As we know, Erdos-Renyi (ER) graphs have a binomial distribution of the degree,
so there could be some differences between the retweet graphs and the ER approximated graph, but it is
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
108
still worthwhile to have an average approximation of the probability of retweeting. Because when the
number of nodes was set with a similar number of edges, the ER approximation graphs are more likely
to average the probability of retweeting to every node instead of a small number of people with a very
high probability to retweet while most people are very unlikely to retweet in the retweet graph. In other
words, the ER graph could give an average probability of retweeting.
An Erdos-Renyi approximation plotted with a controlled density similar to the original single-spotted
retweet graph. The results show that the probability of a node retweet also largely relates to the strongly
connected component of the node connected, but it doesn't mean that there is a positive or negative
proportion. The probability of an ER graph is more depends on the density of a graph. That's why the
ER graphs with smaller node numbers greatly fluctuate their probabilities of linking edges. By looping
through all the graphs, the average probability found is around 0.026. However, more datasets are
needed to confirm whether these approximations could work in other data or only for this dataset.
The comparison data between the single-spotted spread graph and the ER approximated graph shows
that ER approximation is more average, closer in geodesic distance, but less clustering graph, while the
single-spot hate spread graph is centralized. The idea is the same as that the ER approximation is an
average version of the single-spotted spread graph.
5.2. Future Direction
My research is only based on one dataset, so the conclusions need more datasets to support and improve.
What's more, how the retweet graph behavior in other interested groups could be considered, and the
similarities could be compared. Also, further research could try other models to approximate the retweet
graph and see the similarity. To better predict and model the network, the Method of Graph Neural
Network (GNN) would be a direction but could require more data to support it. For example, adopting
GNN with natural language processing models for the detection of hate speech [15].
6. Conclusions
For the first research question, the research found that the hate speech diffusion on retweets relies on
the size of the strongly connected components and gives a polarized node layout with a quarter of hateful
users disconnected, while half are strongly connected. And the degree distribution of hate speech
retweets follows the Power Law. The spreading ability of each node is largely decided by the size of the
strongly connected component to the node connected.
By applying the Erdos-Renyi Model in research question two, we could estimate the average
probability of retweets between two random hateful users because the ER graph is an average version
of the hate spread graph when density is controlled. The average probability found in the study for this
dataset is 0.026. However, because of different degree distributions, the ER approximation is a less
centralized graph than the original hate diffusion.
Overall, the network analysis of the retweet graph of hateful users gives important insights into how
hate speech spreads among hateful users through retweets. With more data in the future, more findings
and improvements could be made. The Erdos-Renyi graphs could provide valuable information on the
average retweet probability. Further research could try out other models for experiments.
Acknowledgement
The author would like to express sincere gratitude to Professor Munther A. Dahleh from the
Massachusetts Institute of Technology for valuable insights and guidance, which greatly enriched the
quality of this research. The author also deeply appreciates his continuous encouragement and support
throughout this project.
Appendix
GitHub Link: https://github.com/ychen3221/Hate-Speech-Retweet-Diffusion.
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
109
References
[1] Chen, Y., Zhou, Y., Zhu, S., & Xu, H. (2012). Detecting offensive language in social media to
protect adolescent online safety. 2012 International Conference on Privacy, Security, Risk and
Trust and 2012 International Confernece on Social Computing, IEEE.
https://ieeexplore.ieee.org/document/6406271/
[2] Badjatiya, P., Gupta, S., Gupta, M., & Varma, V. (2017, June 1). Deep learning for hate speech
detection in Tweets. arXiv.org. https://arxiv.org/abs/1706.00188
[3] Mathew, B., Dutt, R., Goyal, P., & Mukherjee, A. (2018, December 4). Spread of hate speech in
online social media. arXiv.org. https://arxiv.org/abs/1812.01693
[4] Ribeiro, M. H., Calais, P. H., Santos, Y. A., Almeida, V. A. F., & Meira Jr, W. (2018, March 23).
Characterizing and detecting hateful users on Twitter. arXiv.org.
https://arxiv.org/abs/1803.08977
[5] Pereira-Kohatsu, J. C., Quijano-Sánchez, L., Liberatore, F., & Camacho-Collados, M. (2019,
October 26). Detecting and monitoring hate speech in Twitter. MDPI.
https://www.mdpi.com/1424-8220/19/21/4654#B2-sensors-19-04654
[6] Suh, B., Profile, V., Hong, L., Pirolli, P., Chi, E. H., & Metrics, O. M. A. (2010, August 1). Want
to be retweeted? large scale analytics on factors impacting retweet in Twitter network:
Proceedings of the 2010 IEEE Second International Conference on Social Computing. Guide
Proceedings. https://dl.acm.org/doi/10.1109/SocialCom.2010.33
[7] Jahan, M. S., & Oussalah, M. (2021, May 22). A systematic review of hate speech automatic
detection using Natural Language Processing. arXiv.org.
https://doi.org/10.48550/arXiv.2106.00742
[8] Karthik, Dinakar., Rosalind, W., Picard., Henry, Lieberman. (2015). Common Sense Reasoning
for Detection, Prevention, and Mitigation of Cyberbullying. 4168-4172.
[9] Masud, S., Dutta, S., Makkar, S., Jain, C., Goyal, V., Das, A., & Chakraborty, T. (2020, October
9). Hate is the new Infodemic: A topic-aware modeling of hate speech diffusion on Twitter.
arXiv.org. https://arxiv.org/abs/2010.04377
[10] Kwak, H., Lee, C., Park, H., & Moon, S. (2010, April 1). What is Twitter, a social network or a
news media?: Proceedings of the 19th International Conference on World Wide Web. ACM
Other conferences. https://dl.acm.org/doi/10.1145/1772690.1772751
[11] Bild, D. R., Liu, Y., Dick, R. P., Mao, Z. M., & Wallach, D. S. (2014, February 11). Aggregate
characterization of user behavior in Twitter and analysis of the retweet graph. arXiv.org.
https://arxiv.org/abs/1402.2671
[12] Efstathiades, H., Antoniades, D., Pallis, G., Dikaiakos, M. D., Szlávik, Z., & Sips, R.-J. (2016).
Online social network evolution: Revisiting the Twitter graph. 2016 IEEE International
Conference on Big Data (Big Data). https://ieeexplore.ieee.org/document/7840655/
[13] Thij, M. ten, Ouboter, T., Worm, D., Litvak, N., Berg, H. van den, & Bhulai, S. (2015, January
31). Modelling of trends in Twitter using Retweet Graph Dynamics. arXiv.org.
https://arxiv.org/abs/1502.00166v1
[14] Watts, D. J., & Strogatz, S. H. (1998). Collective dynamics of “small-world” networks.
[15] Das, M., Saha, P., Dutt, R., Goyal, P., Mukherjee, A., & Mathew, B. (2021, August 1). You too
brutus! trapping hateful users in social media: Challenges, Solutions & Insights. arXiv.org.
https://doi.org/10.48550/arXiv.2108.00524
Proceedings of the 2nd International Conference on Software Engineering and Machine Learning
DOI: 10.54254/2755-2721/73/20240371
110