Abstract
Online hate messaging targeting Muslims and Jews increased dramatically following Hamas’s attack on Israelis on 7 October 2023 and Israel’s military response in Gaza. This study examined anti-Jewish and anti-Muslim hate posts on X.com and the verbal replies, Likes, and reposts they acquired over the following month. It tests a theory explaining the propagation of hate messages in social media based on the social approval posters garner from other users. The analysis involved replies to 6388 anti-Muslim or anti-Jewish hate posts in terms of their semantic convergence or divergence with the content of original posts. No differences between patterns of anti-Jewish and anti-Muslim discourse arose. Convergent replies to one’s hate posts led individuals to post more hatefully in their next post, and more quickly. Likes also accelerated hate posting, while reposts decelerated them. Divergent replies led to less hateful and slower subsequent hate postings. Conclusions address implications for the social approval theory of online hate, and the relative influence of verbal replies due to their costliness.
Keywords
The attack on Israel by Hamas on 7 October 2023 and Israel’s violent incursion into Gaza thereafter led to a dramatic increase in both anti-Jewish and anti-Islamic hate messaging on social media (Ortiz, 2024). On Facebook, for instance, the posts in groups that were critical of both Israelis/Jews and Palestinians/Arabs/Muslims shrank in proportion to the number of posts targeting one respective side or the other (Nefriana et al., 2024). The virulent and debasing postings following these tragedies offer an unfortunately active field for the study of online hate.
The spread of hate messaging on social media had already become a global issue. Concerns over it include the psychological harm to hate victims (see Tong, 2024), desensitization and a reduction of empathy among observers (Pluta et al., 2023; Soral et al., 2018), and its larger potential to incite intolerance and violence (Benesch, 2013). Online hate is different from traditional hate speech in many ways (Brown, 2018), and it is the focus of an increasing level of international institutional and academic concern (Rice, 2024). Online hate messages “convey intolerance or aversion towards specific groups, such as ethnic, religious, sexual, or gender minorities and immigrants. It may include derogatory comments, threats, and incitements to violence” (Oliveira et al., 2023, p. 1), including “‘discriminatory’ (biased, bigoted, or intolerant) or ‘pejorative’ (prejudiced, contemptuous, or demeaning) language” (United Nations, n.d.). Among the most frequent foci are Jews and Muslims.
Recent theorizing over the propagation of online hate makes central the role of social interactions among hate posters that social media facilitate (Walther, 2024a). Particularly, a social approval theory of online hate argues that people create and deliver hate messages online primarily to garner signals of admiration and approval, in a variety of forms, from like-minded others and appreciative onlookers. The reception of that approval increases the frequency and extremity of individuals’ subsequent hate expressions, according to the theory (Walther, 2024b).
Extant research examining social approval of hate messages, in the form of single-click Likes and reposts (retweets), shows that reception of greater levels of these reinforcers leads to increased toxicity and outrage in one’s subsequent hate messages (Brady et al., 2021; Frimer et al., 2023; Jiang et al., 2023). However, Likes and reposts are not the only form of engagement people garner from their social media posts that have the potential to exacerbate hate.
The verbiage of reply messages also potentially provides social approval to encourage hate messaging. Research in other domains shows that comments one receives expressing agreement or appreciation of one’s postings increase message posters’ satisfaction (Sannon et al., 2017). The approval value of message-consistent conversational responses to hate messages potentially also promotes and elevates hate messaging behavior over time. But comments on one’s posts are not consistently approving or appreciative. Replies may alternatively affirm, undermine, or even counter-attack the hate message to which they respond (Weigel & Gitomer, 2024), making replies less superficially interpretable than Likes and reposts insofar as their social approval value goes. Despite advances in automated analyses of sentiment and wording (for review, see Dehghani & Boyd, 2022), verbal messages are notoriously difficult to characterize automatically with respect to their semantic and attitudinal relationship to prior utterances. If analyses could detect the similarity of other users’ replies to hate messages, it is possible to test the effect of those verbal replies to an original hate message on the original posters’ subsequent posting.
This study examines how the semantic similarity of replies to hate messages on X.com, and other engagement signals, influences the hatefulness of subsequent messages in the context of activity immediately following 7 October 2023, in Israel and Gaza. It provides a nuanced and inclusive view of the degree to which social processes among hate posters and their audience promulgate online hate.
Social Approval Theory
Most research about online hate presumes that those who generate it manifest their internal animosity toward the targets of the message and the desire to harm that person or group (e.g., Zhang, 2023), and/or that individual difference characteristics propel it (Bührer et al., 2024). The social approval theory of online hate side-steps this possibility and focuses instead on social processes among hate posters themselves, and their sympathetic audience, that facilitate additional and increasingly hateful messages. It asserts that hate posters’ primary audience is frequently not the targets implicated in those messages, but one another. Their primary incentive for hate posting lies not in generating Schadenfreude toward its victims. Rather, their incentive may lie in garnering the signals of adulation that they draw from others who appear to agree with and/or appreciate their messages. This is the central assumption of the theory, leading to the hypothesis that the more one receives social approval for one’s hate message, the more frequently and more hatefully one posts thereafter.
Previous empirical studies suggest some support for the hypothesis, but their data were limited to the effects of Likes and reposts, and not the content of verbal replies to hate. In an analysis of Twitter users who posted messages deriding immigrants, Jiang et al. (2023) found that when hate posters received a significantly greater number of Likes or reposts than they usually received, their toxicity subsequently increased. When Likes and reposts were fewer than usual, a significant decrease in toxicity followed. Elsewhere, an analysis of elected US congressional representatives’ Twitter activity over a decade not only found a 23% increase in incivility over time but also that the more Likes and reposts that uncivil posts received, the more uncivil were Tweets thereafter (Frimer et al., 2023). Other studies as well, including both big data analyses and field experiments, have demonstrated similar effects of Likes and Upvotes by observers on posters’ toxicity, outrage, and hatefulness (Brady et al., 2021; Shmargad et al., 2022).
One effort to examine the effect of the frequency of verbal replies and quote tweets on hate posters’ subsequent hate messaging found no effect (Jiang et al., 2023). Its authors speculated that it may not be the case that no effects exist, but rather that the effects of verbal responses systematically depend on the relationship of the reply content to the position of the original message. Thumbs-up and heart responses are rather unambiguous affirmation signals. Verbal replies, however, may affirm, negate, converge, or diverge from an original message’s content. If a reply message pragmatically resembles or affirms the original hate message, it may function as an approval signal and reflect the patterns predicted in the hypothesis above. If the reply is dissimilar or negates the original hate messages, it may function as a disapproval message and also affect hate messaging.
This proposition is operationally not as simple as it may first look. It requires the determination of what is a similar or affirming message and what is a dissimilar or negating message. Determination of a message’s “sentiment” is not useful. Many hate messages and their replies express negative sentiments, which is unsurprising given the definitions and typical content of such messages. Nevertheless, negative replies may have a positive complementary functional relationship with the original message to which they reply. Consider the couplet, “Those people should die immediately,” followed by a reply, “No, those people should die a slow painful death.” Both messages’ sentiment may be classified as negative, but the reply pragmatically converges as an affirmation with respect to the original message. An alternative reply, “I agree,” has similarly affirming pragmatic value, but its sentiment may be positive. For this reason, labels of positive versus negative and sentiment analysis are insufficient to detect the functional relationship of such messages to one another.
Convergence
The analysis of linguistic convergence overcomes these analytic conundrums. Convergence, simply defined, is “a similar way of speaking with other members of the same group” (Rosen & Dale, 2024, p. 3140). Without necessarily copying or imitating the same words, “interlocutors must construct novel utterances serving to advance an ongoing dialog while simultaneously signaling their social proximity to their fellow discursive partners” (p. 3141).
Previous research related to online hate has also explored linguistic convergence. Studies examining the lexical changes in the White Supremacist discussions on Stormfront found that new members gradually adopted and reflected the nomenclature of long-time members over time; those who did not often left the site after a short while (P. Törnberg & Törnberg, 2022). According to A. Törnberg and Törnberg (2024, p. 104), “members picked up different slang and vernacular that is particular to the community. . .The language changed from mainstream terms to community-specific vernacular and themes. . .embracing the White supremacist ideology of the forums.”
Convergence can signal not only ingroup similarity but interpersonal affirmation as well (Bradac et al., 1988). A broad literature supports the notion that as communicators’ nonverbal and verbal behavior becomes more reciprocal and similar, they experience affection and reinforcement, from the nonverbal “chameleon effect” in face-to-face communication (Chartrand & Bargh, 1999; see also Bavelas et al., 1988) to verbal mimicry (Kulesza et al., 2014) and “linguistic style matching” in person (e.g., Ireland et al., 2010) and online (Gonzales et al., 2010). Convergence in message exchanges has been studied through the concepts of “reciprocity, behavioral matching, mirroring, . . .and their opposites of compensation, complementarity, and divergence” (Burgoon et al., 1993, p. 296), as well as synchrony and its opposite, dissynchrony. Divergence and dissynchrony “involves the participant adjusting their behaviors to become more dissimilar to those of the partner. . .and divergence can be used to show that the participant does not desire affiliation with the partner” (Dunbar et al., 2023, p. 2).
Consistent with these principles, linguistic convergence by other users to an individual’s hate message is hypothesized to operate as an affirmation and a signal of social approval. It differs from Likes and single-click reinforcers, however, in that a Like may have value as a compliment, whereas the social approval value of linguistic convergence is that it forms a conversational complement. Divergence, on the other hand, is its antithesis. Divergence may signal disliking and disapproval. The social approval theory of online hate contends that approval of an individual’s hate message, in this case, convergence with it, predicts subsequently greater hate messaging from that individual. In the opposite manner, linguistic divergence from an individual’s hate message may constitute a signal of disapproval. Operationally, convergence has been operationalized regarding the level of entropy in a conversation. Based on information theoretic principles from Shannon (1948), entropy is lower (i.e., convergence is greater) when message exchanges are more similar, that is, the comments are more predictable within a given conversation or thread (Rosen & Dale, 2024).
The measurement of convergence differs from other measures of linguistic style matching in important ways. Unlike dictionary-based approaches to linguistic matching using tools like Linguistic Inquiry and Word Count (LIWC), convergence detection captures both similarities and “stylistic differences in lexical choices between individuals and groups [in situ] without needing to start from a hand-picked dictionary.” It relies on the BERT transformer model, by which it can create “a word vector for each word in a sentence by processing a series of weighted sums of the adjacent word vectors in a large, deep neural network” (Rosen & Dale, 2024, p. 3142). In this manner, analysis detects whether statements within a corpus reflect not only the similarity of the words across statements, but the degree to which the semantic usages of words suggest convergence or divergence. Its measurement has been demonstrated in previous analyses of social media discourse on Reddit. It showed that different, yet ideologically similar, subreddits demonstrated convergent semantic similarities despite their participants’ independence from one another across subreddit discussions (Rosen & Dale, 2024).
Hypotheses and Research Questions
The following hypotheses are posited with respect to both anti-Semitic and anti-Islamic messages and responses.
Hypothesis 1a (H1a): Convergence of replies to hate posts increases the hatefulness of posters’ next posts.
Hypothesis 2a (H2a): Convergence of replies to hate posts increases the frequency of posters’ next posts.
Previous research finds that above-average frequencies of Likes and reposts to hate messages on Twitter had a positive relationship with the toxicity of one’s subsequent posts (Jiang et al., 2023). It is important to establish whether that relationship reappears in this topical context and whether it coexists with, overrides, or moderates the hypothesized effects of convergence on subsequent hatefulness
Hypothesis 3 (H3): Single-click responses to hate posts, including (l) Likes and (r) reposts, increase the hatefulness of posters’ next posts.
Hypothesis 4 (H4): Single-click responses to hate posts, including (l) Likes and (r) reposts, increase the frequency of posters’ next posts.
The social approval theory also recognizes that insufficient social approval can lead hate posters to attempt greater extremity in their subsequent posts in an effort to acquire the degree of social approval they failed to acquire. There is a good reason to expect divergence in replies, since X.com offers no other way to signal disapproval, disagreement, or opposition other than by writing such responses. There are no “dislike” icons to append to someone’s post, as there are on other platforms such as Gab, YouTube, or others. Therefore, it is possible that while Likes and reposts signal approval, users relegate the expression of disapproval to replies. If such disapproving replies fail to satisfy the posters’ needs for approval, they may lead to more hateful, verbally aggressive responses in their next posting out of frustration or retaliation. This relationship is expressed in the following additional hypotheses:
Hypothesis 1b (H1b): Divergence of replies to hate posts increases the hatefulness of posters’ next posts.
Hypothesis 2b (H2b): Divergence of replies to hate posts increases the frequency of posters’ next posts.
Method
Focus Population
Hate against Jews and Muslims, Arabs, Palestinians, and Israelis, online and off, is a long-standing and rampant problem (for review, see Levin, 2015; Weimann, 2024). In the month following the inception of war between Hamas and Israel in October 2023, “Antisemitic content soared more than 919 percent on X. . . .Anti-Muslim hate speech on X jumped 422 percent on Oct. 7 and Oct. 8, and rose 297 percent over the next five days” (Frenkel & Myers, 2023). Although one study detected an increase in hate messages on every social media platform it examined, “hate was more prevalent on X (Twitter) than on other platforms” (Oboler et al., 2024, p. 3). 1 In the past, most such postings could be traced to White Supremacists who issued hate messages toward both Jews and Muslims (Hobbs et al., 2023). In this time interval, however, anti-Semitic and anti-Islamic hate postings were emitted from a variety of user types, even those who identify themselves as not being anti-Semitic or anti-Islamic (Kressel, 2024).
Sample
The sample of messages in this study comprised an original collection of anti-Semitic and anti-Islamic posts, replies, and engagement data, posted on X between 8 October 2023 and 4 November 2023. The process of data collection was two-fold. The first step required the identification of a number of X.com users who posted anti-Semitic and/or anti-Islamic hate messages during this period.
To identify qualified users, we queried the X.com platform to find users who had posted messages featuring any of 61 specific anti-Islamic words or 22 specific anti-Semitic words catalogued in the Weaponized Word (2024) database of hate terms. These keywords became part of a search script implemented through the Brandwatch (2024) platform to access, search, and retrieve relevant data from X. Brandwatch offers access to all historical and contemporaneous interaction records on X.com. The final data records include hate posters’ original posts and all engagements with their posts, including verbal replies, Likes, and reposts. (For similar keyword approaches and time spans using X and other platforms, see Nefriana et al., 2024; Oboler et al., 2024.)
To score hatefulness of the posts and replies, all messages were analyzed using the Roberta Hate Speech DynaBench classifier (HSD classifier: Vidgen et al., 2021). Originally developed by Facebook, the tool generates a score estimating the degree to which a text contains hate content. The HSD classifier was trained on data in which one or more the following features appeared in a given utterance: language that (1) is explicitly derogatory toward individuals belonging to a minority group, (2) expresses animosity toward a group identity, (3) threatens individuals or groups based on their minority status, (4) expresses support for hate proliferating organizations and known, named entities, and (5) is dehumanizing toward minorities or individuals holding a minority identity (Vidgen et al., 2021). The HSD classifier does not produce ratings for any of these individual features, but reports an aggregate rating as to whether an example meets any of these criteria. Although other models, such as Google’s Perspective API, are also widely used, such models generate sets of scores on oblique dimensions such as toxicity, extreme toxicity, and others, the relationships among which are unclear and less useful than a model trained specifically for hate detection.
The final sample contains 17,013 original posts and replies by 7137 individual users. Posts received a total of 10,718 replies (M = 1.68, SD = 3.05) and 65,331 Likes (M = 10.21, SD = 48.37). The hatefulness ratings of these messages ranged from .00013 to 1.0 (M = .406, SD = .334; see Figure 1). Most subsequent posts came relatively quickly, although intervals between posts varied somewhat. The amount of time between original posts ranged from 1 s to 1,677,717 s (19 days), M = 20,596.82 s (16.96 hr), SD = 61,068.54 s.

Histogram of distribution of hate post scores calculated with the HSD classifier.
Computational Approach
The measurement of convergence/divergence scores involved the linguistic, convergence-entropy measurement (CEM) described in Rosen and Dale (2024). CEM measures how predictable the semantic meaning of each lexical unit is in some utterance x when compared to another utterance y using a mixture of Shannon entropy (Shannon 1948) and contextually informed word vectors (i.e., transformer neural networks; Devlin et al., 2019). CEM is operationalized as follows: The total entropy for all the lexical units comprising an utterance x is calculated by finding the Shannon entropy between the word vectors generated for x and the word vectors generated for the utterance y,
where the probability of any individual word vector for the ith token in x (Exi) is found by taking the cosine error (CoE) of Exi and the word vector most similar to it from any of the word vectors in the utterance y (the set of word vectors for the jth token in the utterance y is denoted Eyj and the set of all tokens is indicated via Ey); see Rosen and Dale (2024). The CoE for this closest match is passed to a half-Gaussian function with µ = 0 and a scale σ to convert the raw CoE value to a probability. Analysis assessed convergence among a given tweet and all replies to that tweet, as well as replies to those replies.
Contextually aware word embeddings (i.e., Ex and Ey in the equations) were derived using the RoBERTa-base model (Liu et al., 2019) available via the HuggingFace transformer model library (Wolf et al., 2020). The final computation of convergence/divergence was then residualized based on the number of tokens in both the utterances x and y. When the residual was negative (i.e., the predicted CEM value was higher than the observed CEM value), this was counted as an example of a convergent reply.
Hypothesis tests involved a custom, stochastic linear regression model to assess how various responses influenced the hatefulness observed in subsequent posts (see Supplemental Appendix A). We observed at time t the hatefulness of a post, St, and the hatefulness expressed in their next post, St + 1. The change in hatefulness was affected by feedback that the original post received in the form of (1) how many replies to a post posted at time
where the hatefulness rating of the comment written by a user at time
While the social approval theory explicitly predicts changes in a user’s frequency of hate postings, we tested a related variable in the assessment of H2a,b and H4: how Likes, reposts, and the convergence or divergence of replies affected the amount of time between posting the original post and a user’s subsequent post discussing the same topic. This approach was warranted because analyses were done on a per-post basis rather than a between-users basis; that is, the unit of analysis was the average change between users’ posts (with respect to hatefulness or time lag), from post to post, due to Likes, reposts, and the convergence/divergence of replies to the initial post. While it would be reasonable to assume that a reduction in the time between posts renders more frequent posts within a certain time interval, that assumption does not take into account variations in the responses to a user’s various posts across time. A user may post several messages within a time span, some of which garner greater approval and others that garner less. Therefore, a total frequency count of a user’s posting across time is a relatively less precise indicator of the effects of social approval, post by post. The present approach, in contrast, offers a more detailed view of the consequences of social approval on a user’s likelihood of posting again more or less immediately, even if it is not a perfect representation of frequency per se.
To detect these effects, we used a custom linear regression model, where the time between the original post at time t and a subsequent post on the same topic at time t + 1 (
We estimated the parameters in our models using Gibbs sampling. Both models (equations 2 and 3) were implemented in PyJAGS and run using Google CoLab. The full model formalism with defined priors appears in the Supplemental Materials (Appendix A). A Jupyter notebook file (.ipynb) containing PyJAGS scripts for executing both models appears in the Supplemental Materials (Appendix B).
Results
Initial analyses examined the moderating effect of whether posts and replies were anti-Semitic or anti-Islamic. No differences were found in any relationships presented below. Analyses continued by pooling these subsamples.
Results indicated that both the number of convergent replies (
Parameter Estimates for JAGS Model Used to Fit Parameters Referenced in Equations (2) and (3).
p < .05, **p < 1e−5.
The number of Likes and reposts attributed to a hate post did not significantly predict the hatefulness of users’ subsequent posts. These results suggest the failure of H3 (getting more single-click approval signals increases subsequent hatefulness). In contrast to a previous study (Jiang et al., 2023, where single-click responses and the frequency but not the content of written replies were considered), single-click responses did not affect hatefulness one way or another in these conversations.
The number of convergent replies (
In addition to the findings that convergent and divergent replies affect the amount of time between posts, the number of Likes accrued by a post was a significant, linear predictor of a reduction in the amount of time before a user posted a subsequent hate post,
Discussion
The results of this study provide evidence that the overt expression of hatefulness by X.com users comprises a stochastic process that is sensitive to convergence and divergence expressed in the replies to users’ posts. The social approval theory of online hate (Walther, 2024b) predicts that as users receive greater positive feedback—in this case, more verbal replies indicating approval of the ideas expressed in their original posts—they post more, and more extreme hate messages: The more convergent replies users’ hate posts receive, the more they amplify hate in their next posts.
These findings raise implications beyond the increase in hatefulness. Positive responses to hate posts seem to incentivize riskier forms of hate expression. The logic of this claim is as follows: The classifier that this study employed to analyze hatefulness is a “state-of-the-art” metric, commonly deployed, freely available, peer-reviewed, and developed for industry to detect unacceptably extreme messages to trigger content moderation and the removal of posts. Its developers were engineers from Meta’s (formerly Facebook’s) own labs (Vidgen et al., 2021). Online hate posters try to evade such detection to avoid having their posts deleted by these devices (Parvaresh, 2023; Rieger et al., 2021). To do so, they often compose hate messages and slurs in heavily encoded terms and covert expressions that evade detection by industry classifiers (Bhat & Klein, 2020). Therefore, since the hate classifier employed in this study scores the likelihood that a message violates industry standards, the evidence indicates not only increases in the hatefulness of users’ posts, but also the degree to which they are willing to engage in more overt, riskier, extreme expressions of hate. Thus, positive feedback in the form of more people 2 converging with the ideas expressed by haters in discussions of Jews and Muslims incentivizes these users to “remove the mask” of coded language. Social approval triggers hate posters to “say the quiet part out loud” (see Rae et al., 2024, p. 180), perhaps due to the sense that others are more receptive to such messages.
A second finding related to hate posters’ behavior also confirms hypotheses from the social approval theory of online hate. The observed decrease in the amount of time between posts, due to the influence of convergent replies, implies a greater frequency of hate posts within a time interval.
In contrast to the effects of convergent replies, the effects of single-click responses, such as Likes and reposts, did not significantly affect the hatefulness of users’ next posts in this study. This finding contrasts with a number of other, recent reports that analyzed the effect of single-click responses only, without inclusion of the content of verbal replies, and showed significant positive associations between Likes (Brady et al., 2021; Frimer et al., 2023) or Likes and retweets (Jiang et al., 2023) on increased hatefulness. Whether these differences are due to the analytic approach (the inclusion of verbal replies in addition to single-click responses), the intense topic (anti-Semitic and anti-Islamic sentiments during an active war), or are consequences of different content moderation practices attributable to changes in the platform’s ownership (Ortiz, 2024) cannot be discerned.
The lack of effects from Likes or reposts on the hatefulness of subsequent posts might be due to the kinds of users included in this study. These X users were not “influencers”; they did not have large followings as do many celebrities or politicians who use the platform for public outreach. It seems possible that accounts that enjoy much greater popularity may respond differently to sheer numbers of easily-summed single-click approval signals than in the dialogues their posts generate (see, e.g., Frimer et al., 2023). Future research and different methods are required to discern whether that is the case.
Another, simpler possible explanation for the lack of effect of reposts and Likes on users’ subsequent hatefulness, relative to the effect of verbal replies, could be the sparsity of the former set of signals. While 44.4% of posts received at least one Like, the distribution of Likes drops precipitously from there. In this dataset, users could expect to receive no Likes at all, and if they did get a Like, it is highly unlikely that they would get more than a single one. Similarly, only 15.8% of posts were reposted at least once, after which there is another precipitous drop off. In both cases, the modal response to a post is one repost and one Like.
Importantly, few posts received verbal replies, with only 21% receiving at least one comment. Despite the greater scarcity of replies than Likes, replies nevertheless were significantly more influential when they appeared. Writing a reply at all requires greater effort, and is thus a more costly marker of approval or disapproval. Costly markers of individuals’ positions are often more strongly attended to than ones that cost less effort to generate (Donath, 2007), consistent with other findings that the more people perceive that the Likes they get from others are automatic and mindless reactions, the less emotional influence they have (Carr et al., 2016). As has been shown in the impact of friendly comments and Likes (Burke & Kraut, 2016), replies signal effortful engagement with users’ ideas. When replies demonstrate convergence, users are likely to experience greater social approval for the ideas they expressed in their post. This appears to be the case in hate posting as well as friendly interactions.
The different effects of reposts versus replies to hate messages, as seen in the results of H4, may be due to missing data of a kind: The Brandwatch application’s protocols do not distinguish between reposts and “quote tweets,” that is, reposts of a hate message to which the reposting user appends an original message of their own. Brandwatch does not capture the original content contained in quote tweets. Prior research (see Metzger et al., 2021) shows that a significant amount of message re-sharing, like quote tweets, adds content that defies, ridicules, or otherwise contradicts the original messages to which they are appended. 3 Thus, it is unknowable, among the sparse amount of reposts included in these data, whether and what portion of the reposts were appended with divergent or convergent original content. This gap represents an important limitation to the present work.
It is interesting that the frequency of replies that diverged from the hate posts correlated with decreased subsequent hatefulness. Social approval theory offers conflicting alternative hypotheses about divergence, i.e., social disapproval responses to hateful behavior. It proposed that social disapproval renders no direct effect on subsequent hate posting. However, an implication of the social approval hypothesis is that the less approval one receives, the less subsequent hate should be expected. If more frequent convergence is interpreted as greater social approval, greater divergence may constitute less approval. In the present case, the independent measurement of convergence (approval) and divergence (disapproval) prevents the conflation of effects. However, additional considerations about the effects of these oppositional signals on an individual’s balance of approval-to-disapproval warrant further research consideration.
The theory’s most specific hypothesis about disapproval responses considers the effect of disapproval to be mediated by other, contingent factors. In this case, results demonstrate that the null hypothesis—that divergence does not affect subsequent behavior—is not consistent with the data. Additional theorizing and testing need to be performed to determine both why divergence appears to predict less subsequent hatefulness and what mechanisms facilitate that effect. The results of this study clearly indicate that the greater the positive verbal responses to a hate post a user accrues, the more the user posts hate more extremely and more quickly (see Stsiampkouskaya et al., 2021). At the same time, increased negative feedback causes such users to disengage from hate posting to some extent (see Erişti & Akbulut, 2019; Sannon et al., 2017).
The existence and acceleration of hate discourse targeting religious groups in response to ethnopolitical crises clearly demonstrates an increase in the boundary-spanning spillover of online hatefulness. There has always been some tendency to “holding Jews collectively responsible for actions of the state of Israel” (Weimann & Masri, 2022, p. 170) and holding Muslims responsible for Hamas. Comments on social media in the immediate aftermath of October 7 solidified this tendency and focused on negative stereotyped characterizations of Jews and Muslims in so doing. That is, they repeated historical Islamophobic narratives depicting “Muslims as threats to Western values and security,” whereas antisemitism “usually draws on long-standing conspiracy theories about Jewish influence on politics, economy, and culture”; “The attacks initiated by Hamas and the ensuing political-military aftermath by Israel appear to align with and intensify these existing prejudices, further polarizing online discourse” (Nefriana et al., 2024, p. 1). An analysis of post-October 7 anti-Muslim and anti-Arab postings found that religious vilification of Muslims comprised 91% of the posts; only 9% focused on Palestinian or Arab identities per se. Such features have also been prevalent in anti-Semitic posts originating in the United States (Kressel, 2024) and in “various European countries, where anti-Semitism and anti-Zionism often overlap, complicating the distinction between political criticism and outright hate” (Nefriana et al., 2024, p. 3).
Although the causal relationship of online hate to offline hate and violence is not well-established, it is known that online hate interactions facilitate the logistical organization of offline collective action (Wahlström & Törnberg, 2021), and this is likely to have been the case in the many in-person protests focused on the Israel/Gaza crisis. In addition to logistics, attitudinal messages receiving convergent responses may render psychological effects. Hateful replies and single-click responses in the early days of this event have been credited with exacerbating and polarizing perceptions and attitudes; through interaction, not just observation, social media participation appeared to “fuel extremist thought, perpetuate divisiveness, and influence perceptions” in this specific crisis particularly (Klempner, 2023).
Supplemental Material
sj-pdf-2-sms-10.1177_20563051251383635 – Supplemental material for Social Processes in the Intensification of Online Hate: The Effects of Verbal Replies to Anti-Muslim and Anti-Jewish Posts Following 7 October 2023
Supplemental material, sj-pdf-2-sms-10.1177_20563051251383635 for Social Processes in the Intensification of Online Hate: The Effects of Verbal Replies to Anti-Muslim and Anti-Jewish Posts Following 7 October 2023 by Zachary P. Rosen and Joseph B. Walther in Social Media + Society
Supplemental Material
sj-txt-1-sms-10.1177_20563051251383635 – Supplemental material for Social Processes in the Intensification of Online Hate: The Effects of Verbal Replies to Anti-Muslim and Anti-Jewish Posts Following 7 October 2023
Supplemental material, sj-txt-1-sms-10.1177_20563051251383635 for Social Processes in the Intensification of Online Hate: The Effects of Verbal Replies to Anti-Muslim and Anti-Jewish Posts Following 7 October 2023 by Zachary P. Rosen and Joseph B. Walther in Social Media + Society
Footnotes
Acknowledgements
The authors are very grateful to Prof. Rick Dale for his generous advice on methods and their application and co-development of the convergence-entropy metric; and the UCSB DREAM Lab for assistance in data acquisition.
Ethical Considerations
n/a
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This research was supported in part by the Institute for Rebooting Social Media, Berkman Klein Center for Internet & Society at Harvard University, and the UC Santa Barbara Bertelsen Presidential Endowment for Technology and Society.
Declaration of Conflicting Interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Data Availability Statement
Upon request to the first author.
Supplemental Material
Supplemental material for this article is available online.
Notes
Author biographies
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
