Abstract
Social media platforms, such as Instagram, are regularly misused for spreading covert (Islamic) extremist propaganda. Affect and emotion are central tools used in extremist propaganda, but there is little research into the combined employment of different social media elements, such as hashtags, visuals, and texts, in the context of propaganda. This study contributes to closing this gap. Using the German group Generation Islam as a case study, we examined the group’s Instagram activity (N = 1,187 posts) over the course of 2 years. To reflect the platform users’ logic, we (a) examined affect in hashtag networks in which users can come across propagandistic content, (b) employed deep learning to examine the emotional valence transmitted in the visuals, and (c) used automated linguistic analysis to describe collective action cues contained within the texts. The results are novel, as they provide nuanced insights into extremist propaganda’s employment of affect and emotions across Instagram’s affordances.
Keywords
Introduction
Social media are widely popular among adolescents and young adults. In Germany, the context of this study, the visual-oriented platform Instagram is one of the most popular apps: 72% of German inhabitants aged 14–19 years use this app on a weekly basis. For 42% of the girls (boys: 31%), it is considered as the most important app on their phones (Rathgeb & Schmid, 2021). Instagram is also an attractive venue for political (DiResta et al., 2019) and extremist propaganda (Bloom & Moskalenko, 2021; DeCook, 2018; Frischlich, 2021). For instance, more than one-fifth of the cases (23%, n = 187) processed in 2019 by Jugendschutz.net, a German institution for the protection of youth, were found on Instagram (Glaser, 2020), making it the second-most targeted platform.
Researchers repeatedly report that there is no single cause of radicalization (Alonso & Delgado, 2020; McCauley & Moskalenko, 2008) and that online media are not the trigger for the radicalization process (e.g., Gill et al., 2015; Schmidt-Kleinert, 2020). However, online media, including Instagram, offer enhanced opportunities for information dissemination and connection and, as such, are able to catalyze the radicalization processes (for comprehensive reviews, see Meleagrou-Hitchens & Brun, 2013; Rothut et al., 2022). Although big tech companies, including Instagram’s mother company Meta, have recently increased their attempts to remove and prevent extremist content using algorithmic and human moderation, and there are increased attempts to teach extremism-related media literacy (e.g., Bouko et al., 2020), social media provide a central contact point through which adolescents can make contact with extremism (Reinemann et al., 2019). Crucially, propaganda often covers its intent, making it hard for both users and moderators to identify it at first glance. Thus, a deeper understanding of covert propaganda strategies is needed to help develop both algorithm-based and user-centered approaches for dealing with extremist propaganda. This article contributes to this understanding by centering on one central characteristic of extremist propaganda: its emotional appeal.
Several scholars have referred to emotionality as a core element of extremist propaganda (Rothut et al., 2022). For instance, Biswas and Deylami (2019) argue that the emotional framing of their mission increases the effectiveness of ISIS mobilization attempts. Other authors have started to examine the role of specific emotions in this context (Maragkos & Maravelakis, 2022), as emotions are known to motivate deeper information processing (Nabi, 2003) and play a motivating role in collective action (van Zomeren et al., 2008). Among others, these studies show that propaganda is more engaging when emotions such as anger (Lee & Colautti, 2022), or the elicitors of self-transcending emotions, are included (Frischlich, 2021). Furthermore, extremist discourse can be distinguished from non-extremist content better when emotional terms are considered in attempts at automated classification (De Bruyn, 2021).
The aim of this study was to make a threefold contribution to this research field. First, we aimed to describe in greater detail the role of affect and emotions in Islamic extremist propaganda on Instagram. Scholarship on Instagram as a propaganda tool is still relatively rare (Frischlich, 2022; but see Speckhard & Ellenberg, 2020). Second, we broadened prior work focusing on the textual and visual elements of propaganda (e.g., Maragkos & Maravelakis, 2022) by taking the entire logic of an Instagram campaign into account, thus including the hashtags employed to reach new audiences, the visual first impressions, and the text itself. Third, we sought to understand the mobilizing potential of Islamic extremist online propaganda by examining the collective action cues contained within the propaganda campaign. To this end, the current study employed an innovative combination of different methods of automated content analysis, namely, semantic network analyses, deep learning, and linguistic dictionary-based analyses.
Background
Islamic Extremist Online Propaganda
Extremism and radicalization are contested terms. In this article, we understand extremism as the combination of a radical ideology and sympathy for or acceptance of ideologically motivated violence. Crucially, the acceptance of violence and its execution are not necessarily associated; therefore, extremism (i.e., the acceptance of violence on behalf of a radical ideology) needs to be distinguished from violent extremism (i.e., the execution of ideologically motivated violence) (see Khalil et al., 2022). Radicalization can be understood as the (nonlinear and idiosyncratic) process that leads individuals or groups to adopt extremist ideations (e.g., McCauley & Moskalenko, 2008; Rothut et al., 2022).
We understand propaganda as a specific form of persuasive communication (Jowett & O’Donnell, 2012), as the strategic attempt to manipulate emotions and cognitions to motivate behavior in the acceptance of an ideology―often an ideology that claims absolute validity (Rieger et al., 2020a). Accordingly, we understand Islamic extremist propaganda as persuasive communication that expresses sympathy for or justifies violence through a narrow and radical interpretation of Islam. For instance, this may involve promoting violent (or “lesser” 1 ) Jihad against “apostate” Muslims (all those who do not share their worldview) and infidels (mostly Jews and/or “crusaders”) or promoting anti-democratic and violent means as defense against an alleged “global war” against Muslims (Ebner, 2017).
Digitization has fundamentally altered the distribution of propaganda, including the staging, dissemination, and consumption of Islamic extremist propaganda (Rieger et al., 2020b). Nowadays, Islamic extremists use a wide range of technical affordances
Islamic extremists remain very active on social media (Ayad et al., 2022). A study by the Institute for Strategic Dialogue described an entire German language ecosystem of Salafis 2 (not all of them extremist) targeting young media users, with a 77% increase in posting activity between 2019 and 2021 (Comerford et al., 2021). The same study also shows that a substantial share of the examined posts transmitted extremist ideas, with Instagram being among the most toxic platforms. This is not to imply that most posts on Instagram express violence. A content analysis of an Islamic extremist Instagram Influencer account (Frischlich, 2021) shows that only a small share of posts were explicitly extremist. Instead, the posts tried to connect with the users’ lifeworlds and played with the users’ emotions.
Affect and Emotions
Emotions are specific affective states. Psychological appraisal theories (Scherer, 2005) postulate that emotions are triggered by personally meaningful appraisals of specific elicitors (stimuli or events, such as considering something purposefully unjust, which triggers anger; see Lazarus, 2001). They are characterized by the synchrony of specific bodily symptoms (e.g., a faster heartbeat), communication patterns (e.g., non-verbal grimacing and shouting), and action tendencies (e.g., to change the anger-eliciting condition), as well as subjective feelings that integrate the different responses. Emotions are short-lived and consciously experienced (Fredrickson, 2004). Emotions can be categorized according to their valence as being positive (i.e., associated with pleasant states), negative, or mixed (e.g., nostalgia is a bittersweet state), as well as according to their intensity (e.g., being upset versus wrathful). Furthermore, emotions can be associated with both approach and avoidance behaviors (e.g., fear motivates avoidance). It is noteworthy that emotions’ valence, expression, and action tendencies will vary depending on the context: what is considered appropriate in one context might be inappropriate in another.
Emotions can be distinguished from other affective states, such as mood or affect. For instance, moods are generally less intense, are not elicited by specific stimuli, and may last over hours or even days without a synchronized bodily, communicative, and motivational reaction. Affect has been described as a “pre-emotive intensity subjectively experienced and connected to [. . .] processes of premeditation or anticipation of events prior to their occurrence” (Papacharissi, 2016).
Affect, Emotions, and Collective Action
Affect and emotions are important in propaganda, as they direct attention toward the media content, shape engagement with the content, and might potentially motivate collective action: actions on behalf of a collective intended to improve the status of that collective (Wright et al., 1990). Collective actions can be normative (e.g., signing a petition) or non-normative (e.g., executing violence; see Tausch et al., 2011), and they can happen offline as well as online.
Papacharissi (2016) argues that social media allow for the formation of affective publics: “networked publics that are mobilized and connected, identified, and potentially disconnected through expressions of sentiment” (p. 312). Thematic identifiers, such as hashtags, mediate feelings of connectedness and unite these affective publics. Notably, affective publics are assumed to support connective action, that is, digital political activities centering on personal expressions, placed at the blurred boundary between the public and the private (Bennett & Segerberg, 2012), but not necessarily collective action. Nevertheless, they play a crucial role in connecting digital extremist communities (Biswas & Deylami, 2019; Johnson, 2018).
The valence of emotions affects subsequent information processing. Media content that elicits negative emotions is more “attention-grabbing” (the so-called negativity effect; see Kleinnijenhuis, 2008), probably because negative emotions have a more immediate survival function. This general tendency is attenuated by distinct negative emotions: approach emotions, such as anger, motivate deeper information processing compared with avoidance emotions, such as fear (de los Santos & Nabi, 2019; Nabi, 2002). Positive emotions such as hope are generally associated with approach motivation and broaden people’s momentary thought–action repertoires (Fredrickson, 2004).
Both valence and distinct emotions influence collective action. Research on collective action has shown that negative images are not only more attention grabbing but also increase willingness for collective action (Geise et al., 2021). The social identity model of collective action (van Zomeren et al., 2008) postulates that collective action, in general, has three main drivers: First, people must identify with their respective ingroup. Second, they must perceive an unjust deprivation of that ingroup, typically reflected in emotions of anger (e.g., Miller et al., 2009; Stürmer & Simon, 2009; van Zomeren et al., 2008). Finally, the type of collective action will depend on peoples’ sense of collective efficacy: if legitimate collective action, such as protests, is perceived to be ineffective (i.e., hopeless), people are more likely to accept non-normative collective action, including violence (Tausch et al., 2011), while hope for the efficacy of the advertised collective action increases the motivation to participate (Wlodarczyk et al., 2017).
Social media can mobilize collective action (Jost et al., 2018) through content that makes motivational factors salient. Such collective action cues have been documented in right-wing extremist videos (Hawkins & Saleem, 2021) and in conspiracy theory-endorsing online communities (Frischlich, 2022). In line with this, De Bruyn (2021) shows that including social identity and emotion-related terms improves the automated classification of extremist communication.
The Current Study
Affect and emotions play a crucial role in extremist propaganda (for an overview, see Rothut et al., 2022). For instance, while right-wing extremists employ humor (Udupa & Pohjonen, 2019), Islamic extremists employ positive, self-transcending emotions (Frischlich, 2021) to veil their hate-filled ideologies. Bouko et al. (2022) show that Salafists express emotional hurt about the negative treatment of their ingroup, thereby potentially stirring up anger in their audience. A computational analysis of ISIS Al-Hayat Media Center magazines shows that similar emotions are evoked across languages: positive emotions of trust, joy, and anticipation are communicated together with negative emotions of fear and anger (Maragkos & Maravelakis, 2022). In the current study, we examined affect and emotions in Islamic extremist propaganda by studying three central affordances of Instagram: the hashtag, the image, and the text.
The Hashtag
At the time of data collection, Instagram presented its users with a personalized feed that included the content of accounts and hashtags they followed as well as algorithmic recommendations of matching content and sponsored posts. New content could be discovered via the search function or in an algorithmically curated recommendations section. Consequently, hashtags allowed emergent communities and ad hoc affective publics to form (Bruns & Burgess, 2015). Islamic extremists (like other extremist actors, see Johnson, 2018) and also many democratic social movements (see Freelon et al., 2020) create their own hashtags to unite their digital audiences (Awan, 2017). However, the extremists also hijack prominent hashtags to introduce radical content to new audiences (Farwell, 2014). For this reason, hashtags provide insights into the distinct publics that Islamic extremists address on Instagram. Consequently, we formulated our first research question (RQ):
RQ1. How do extremists employ hashtags to disseminate their content on Instagram?
Images and Emotions
Images were the first impression that users got from an Instagram post at the time of data collection. Images generally leave deeper traces in people’s memories than textual content (the so-called picture superiority effect; Paivio & Csapo, 1973), and their valence steers subsequent engagement (Geise et al., 2021). Visual material plays a crucial role in Islamic extremist propaganda (Kennedy-Boudali, 2006; Winter, 2018), as visuals are expected to contribute to the anchoring of (extremist) myths and worldviews (Hofmann & Ipsen, 2018). Positive images are thus typically used to frame the ingroup (Bouko et al., 2020) and to contrast the positive utopia within extremist-ruled regions to a dystopian world in which the ingroup suffers from cruelties imposed by its enemies (Winter, 2018). Thus, we formulated our second research question:
RQ2. Which emotions characterize extremists’ visual communication?
Texts and Collective Action Cues
Although visual content is the primary thing users see on Instagram, textual content can provide a more nuanced meaning to the posts. Textual references can stage a sense of community and meaning (Smith et al., 2020) and have been used to identify radical content on social media. For instance, Kursuncu et al. (2019) suggest using terms such as jihad, infidel, and ISIS in relation to other terms as markers for extremist narratives on Twitter. Moreover, textual elements, such as captions or texts within the images, often guide the interpretation of the images. For instance, an image of a woman with a hijab will probably be framed differently by right-wing extremists and Islamic extremists. Hence, it is likely that collective action cues are particularly present at the textual level. We thus asked the following:
RQ3. Which collective action cues are included in the textual elements?
Method
We answered our research questions with a multimodal automated content analysis of a theoretically sampled case study. As we were specifically interested in Islamic extremist propaganda attempting to “mainstream” radical ideologies, we examined the Instagram account of Generation Islam, a German Islamic extremist group under observation by the German Federal Office for the Protection of the Constitution. The group is linked to the global Hizb ut-Tahir movement, which is forbidden in Germany. Generation Islam is infamous for its social media success (Bundesministerium des Inneren, 2021). On Instagram, the account has more than 69,500 followers (retrieved 23 May 2021).
Database
We crawled all posts uploaded between January 2016 (the creation of the account) and 19 December 2018, using Instamancer (N = 1,187 posts). 3 In doing so, we covered the entire period before the account was mentioned for the first time in the German Protection of the Constitution extremism report (Bundesministerium des Inneren, für Bau und Heimat, 2019, p. 42). We stored images and captions for each post.
We created three datasets (see Figure 1): The hashtag dataset contains the hashtags included in the image caption. The image dataset contains the visual data for each post. The text dataset includes all captions for the images, apart from the hashtags. To account for the prominence of textual image content (i.e., text embedded as foreground in the images, see Figure 2), we extracted the image text using Google Vision API (Application Programming Interface). 4 The extracted text snippets were then added to the text dataset.

Creation of datasets. Post data is split into image, text, and hashtags.

Process of data analysis. Imavis and SentiBank were employed for visual analysis. LIWC (short for: Linguistic inquiry word count) is the text analysis method employed.
Analytical Approach
Overall, our analysis entailed three strains that reflected our research questions. To examine extremist hashtag strategies (RQ1), we focused on the hashtag network. To answer RQ2 and RQ3 and to describe the extremists’ use of emotions and collective action cues in their textual and visual propaganda, we employed natural language processing, image sentiment analysis, and deep learning (see Figure 2).
Hashtag Network Analysis
To gauge the hashtags’ popularity, we extracted the number of public posts using this specific hashtag from Instagram. To account for the erroneous use of a hashtag, we focused on hashtags that were used at least five times (n = 228 out of n = 1,719 hashtags). Using inductive coding (i.e., the iterative qualitative distilling of overarching categories from the material, see Meyen et al., 2011), we extracted eight categories from the material using religion (e.g., references to Allah or Islam), regional (e.g., cities or places), positivity (positive affect and values such as truth), community/family, politics (e.g., political actors or elections), campaigns (e.g., social media campaigns, such as #jesuismuslim), negativity (negative affect or events such as war), and miscellaneous (covering various other topics). We manually coded all hashtags into these categories. To understand the connection between the hashtags, we generated a network from the data, where nodes represented individual hashtags and edges represented the co-occurrence of two hashtags in the same post (see Table 1).
Dominant Hashtags and Their Online Popularity.
Note. Reported are the absolute (n) and relative (%) frequencies of hashtags in our database and the average number of all posts using one of these hashtags on Instagram. Hashtags were translated for publication purposes.
Image Analysis
Based on the literature, we were particularly interested in the valence and emotions transmitted in Islamic extremist images. In computational social science, images’ emotional valence can be examined using visual sentiment analysis (for an overview, see Ortis et al., 2019). Visual sentiment analysis is often done via deep learning, namely, through the use of convolutional neural networks (CNNs) (Ortis et al., 2019). One established CNN in the context of social media content is Imavis (Campos et al., 2017). Imavis analyzes raw image data and provides probability scores for both positive and negative sentiments at the level of single pixels. It provides an overall estimation of an image’s sentiment, ranging from negative (=−1) to positive (=+1), and enables sentiment maps. These maps can be used to color the presumed positive and negative areas of an image, thereby allowing human validation of the algorithm’s decision. We used Imavis to analyze the emotional valence of the images.
As we were also interested in distinct emotions, we combined the sentiment analysis of Imavis with an analysis based on distinct emotions. To this end, we used SentiBank (Borth et al., 2013). 5 SentiBank enables the detection of distinct emotions using visual sentiment ontology (VSO). The idea is that SentiBank maps emotions in images with the help of mid-level features: semantic objects consisting of adjectives and nouns, so-called adjective noun pairs (ANPs), such as “angry men,” which are derived from the manual coding of a large database. SentiBank has been shown to detect 1,200 validated adjective noun pairs, each with a classification accuracy above 60%, indicating satisfactory reliability. SentiBank can automatically generate adjective noun pair probability values from a given image dataset. Based on Plutchik’s (2001) wheel of distinct emotions model, inferences about the basic emotions in these ANPs and the images can be made via the publicly available VSO website (see Note 5).
Text Analysis
To detect collective action cues, we used a prominent and psychologically grounded dictionary, the LIWC (linguistic inquiry and word count) by Pennebaker et al. (2001). The LIWC classifies words into 72 categories. The dictionary covers distinct emotions, such as “anxiety,” “anger,” “sadness,” and “optimism.” The LIWC also distinguishes between personal pronouns reflecting a more personal identity expressed in the posts (e.g., “I,” “self,” “you”), and collective pronouns (“we,” “they,” “other”), which reflect a group-based identity and require a level of understanding of the interpretation of the “we” and “them” referred to (Smith et al., 2020). The LIWC also covers “social” words related to addressing “family,” “friends,” and “humans,” and it includes labels for “inclusive” and “exclusive” formulations, indicating intergroup communication. For our analysis, we examined whether a post included the distinct negative emotions of anger (as a potential motivator of collective action), sadness or anxiety (as typical emotions expressed in extremist propaganda), or optimism (as an indicator of the transmitted efficacy of following the extremists’ path). We also examined whether the posts referred to religion, a social identity (using the personal pronoun “we,” or “other references,” or using “including” or “excluding” language), or personal identity (“I,” “self,” or “you”).
Results
Hashtags and Affect
Most posts (75.42%) contained at least one hashtag, with an average of nine hashtags per post. Overall, 1,726 unique hashtags were found, covering a broad range of topics, such as sports (e.g., “#football”), hobbies (e.g., “#photography”), or politics (e.g., “#democracy”). The dominant category was (Islamic) religion (e.g., seasonal greetings such as “#eidmubarak”). The 15 most frequent hashtags all fell into this category (see Table 1).
Our qualitative coding showed that all categories included popular hashtags, with the average number of public Instagram posts using the same hashtags ranging from 500,000 to 15,000,000. Religious, regional, and positive hashtags were particularly popular, as were hashtags from the miscellaneous category.
Supporting the idea of a strategic linking of the ideological to mundane topics, religious hashtags were often used in combination with non-religious hashtags, as apparent in the hashtag network (see Figure 3). Only 12.20% of all posts exclusively used religious hashtags. Answering RQ1, the affective publics addressed by Generation Islam through hashtags reflected a religious and/or local community and relied on positive affect. Furthermore, the hashtags were part of a larger networked public sphere in the users’ lifeworlds.

Hashtag graph with 134 nodes and 501 edges. Closeness centrality was used as the basis for the layout algorithm. The 14 hashtags with the highest closeness were labeled. [Translation: #vergebung = #forgiveness, #deutschland = #germany, #gemeinsamgegenhetze = #togetherAgainstAgitation, #ummah = #community, #ukhti = #sister, #akhi = #brother].
Valence and Emotion in the Images
The Imavis (Campos et al., 2017) sentiments score showed that most of the images were classified as negative (Md = −0.87, IR [−0.97, −0.60]). A manual inspection of the sentiment maps (see Figure 4) largely confirmed this classification. Single negative artifacts, such as weapons, burning areas of a building, or negative facial expressions, were well recognized by the algorithm (see Figure 4a). However, text areas were mostly classified as “positive” (cf. Figure 4b), suggesting that the actual number of positively valanced images might be lower. Furthermore, backgrounds were not well differentiated (see also Campos et al., 2017), which might lead to an overlooking of more subtle image cues (see Figure 4c). Overall, the images employed by the propagandists were likely to raise attention via a negativity bias (Kleinnijenhuis, 2008) and might lay the groundwork for larger collective action motivations (Geise et al., 2021).

(a–c) Exemplary results of the images’ sentiment analysis Positive image details are colored in light blue and negative ones in purple. Images are cited in accordance with the German copyright law §51 from @generation_islam (2018).
To obtain a more detailed impression of the discrete emotions in the images, we used SentiBanks (Borth et al., 2013). Table 2 shows the most frequent adjective noun pairs. Providing nuance to the negativity detected by the Imavis analysis, anger (angry man), fear (dead terrorists), and defense-related images (dark fortress, amazing armor) dominated. Other images transmitted optimism (inspirational poster, victorious team) or referred to leisure-time activities (e.g., strong beer, which seems to be connected to images with bearded men from manual inspection not to alcohol consumption).
Top 20 ANP of the Generation Islam Image Corpus.
ANPs = Adjective noun pairs provided by visual sentiment ontology.
Note. Each image had a confidence value for each of the 1,200 ANPs. We extracted the top 20 ANPs that were predominant in our corpus. The mean (M) refers to the average confidence value of all images in which the particular adjective noun pair was detected.
An inspection of the distinct emotions in the images (see Figure 5) showed that images entailing angry men were correctly classified as anger mixed with disgust and sadness, while visualizations of fortresses, dark nights, and “scary Halloween” were classified as fear-inducing images. “Inspirational” elements, such as bright cities and inspirational posters, were classified as elicitors of joy and anticipation (i.e., hope or optimism). In answer to RQ2, overall, the images raised attention via negativity and included a mixture of threat, anger, and optimism, probably intended to fuel collective actions.

Breakdown of the top 10 adjective noun pairs in the analyzed images in relation to their basic emotions, as drawn from Plutchiks’ emotion wheel (2001) (Borth et al., 2013).
Collective Action Cues in the Texts
We analyzed the textual content using the quanteda package for R (Benoit et al., 2018). Figure 6 provides a joint overview of word frequency in all categories.

Text data presented in LIWC categories. For a better assessment, the absolute word frequencies were divided by the total number of words classified via the dictionary.
Consistent with the hashtag analysis and the ideological stance of the account, religion was a frequent topic. Nearly half of the posts (47.9%, n = 569) used a religious word. Based on our theoretical interest in elicitors of collective action, we were particularly interested in how far the posts addressed distinct emotions and group identities.
With regard to collective identities, the posts clearly differentiated between the ingroup and the outgroup. Most of the words contained in the LIWC (89.6%, n = 1,064) were categorized as either including or excluding language, and contained references to third-person pronouns (“we” and “other” references). References to participants’ personal identities using first- and second-person pronouns, such as “I” and “you,” and references to the “self” were less frequent (n = 644, 54.2%). In line with our findings from the hashtag analysis, we also found many references in the image captions relating to themes of leisure, friendships, family, and also metaphors and references to religion. With regard to specific emotions, the LIWC showed that a relatively large share of posts transmitted optimism (n = 347, 29.2%). Roughly half that number of posts communicated anger (n = 202, 17.0%). Answering RQ3, the textual elements anchored the social identities of users who were attracted by the religious hashtags, and although they also touched on negative emotions and mobilized anger, they mostly spread a positive view of the future if users followed the “right path.”
Relationships Between Images and Text
To explore the relationships between the images and the textual information, we calculated Spearman correlations between the LIWC and the images’ sentiments. This analysis showed that positively valanced images were used to anchor recipients’ social identity (ρ = .11, p < .001) and also to accompany texts that included negative emotions such as anger (ρ = .13, p < .001), sadness (ρ = .09, p < .01), and anxiety (ρ = .08, p < .01). Overall, the negatively valanced images (see above) were accompanied less often by texts that addressed participants’ social identity (ρ = −.11, p < .001), or their lifeworlds (ρ = −.08, p < .01), or texts that expressed negative emotions (ρ = −.12, p < .001) such as anger (ρ = −.13, p < .001), sadness (ρ = −.09, p < .01), or anxiety (ρ = −.08, p < .01). Thus, negative images seemed to be used to capture attention, while positive images seemed to be employed to create a positive affective community and inspire following the propagandists’ path.
Discussion
Main Findings and Theoretical Implications
Overall, our results gave novel insights into the extremist propaganda strategies employed on Instagram, focusing on three central affordances: hashtags, images, and text. In answering RQ1, we found that hashtags were used to create affective communities around religious identities and local communities. Popular hashtags were hijacked to associate the extremists’ pseudo-religious messages with multiple elements of the participants’ lifeworlds. Overall, Generation Islam created networked publics (Papacharissi, 2016) that were connected via regional and mundane references to young peoples’ lifeworlds, unified via a (narrowly interpreted) religious identity, and sugarcoated with positive affect (for similar observations, see Biswas & Deylami, 2019; Pearson, 2018).
Visuals conveyed the first impression users got from an Instagram post at the time of data collection. Answering RQ2, the images were mostly negative, capturing users’ attention (Kleinnijenhuis, 2008), and likely to elicit emotions associated with collective action, such as anger (Cohen-Chen & Van Zomeren, 2018; Stürmer & Simon, 2009; Tausch et al., 2011) but also fear and optimism. Prior work has shown that negative (but not positive) images can increase (normative) collective action intention (Geise et al., 2021), while Winter (2018) argues that documenting the victimization and suffering of the ingroup makes the promise of the extremists’ utopian vision even more powerful.
Textual elements can provide deeper meaning to the visuals on Instagram. Answering RQ3, we identified frequent collective action cues. In particular, textual elements were used to anchor users’ social identities and to spread optimism if the users followed the “right path.” In highlighting this, our study broadens prior work on collective action cues in extremist propaganda (Frischlich, 2022; Hawkins & Saleem, 2021) by showing how different affordances are combined on Instagram. The Islamic extremist propaganda examined was found to lure members from broad and often positive affective publics, catching their attention via negative images while simultaneously providing “solutions” to the depicted problems through a positive community and in this way, inspiring adherence to a narrow, radical ideology.
Overall, our findings have significant theoretical and methodological implications, as they underscore the role of affect and emotion in modern extremist propaganda and call for a more nuanced examination of the different social media affordances exploited in extremist propaganda campaigns. By providing a crucial extension to prior work that used manual content analyses (e.g., Bouko et al., 2022; Wignell et al., 2018) or relied only on textual elements (Smith et al., 2020), our study demonstrates that combining multimodal automated content analysis tools can support such research at scale.
Limitations and Directions for Future Research
Our study has several limitations that must be considered when evaluating its generalizability. First, we focused on only one German language account on Instagram. Germany has relatively harsh regulations against the dissemination of online hatred (the Network-Enforcement Act, Netz DG) through large-scale platforms, making it likely that propaganda in other languages, or on other platforms, is under less pressure by attempts at moderation (see also Frischlich et al., 2022). However, users are more likely to fall for covert propaganda (Reinemann et al., 2019); thus, it is plausible that propaganda directed toward new audiences will need to employ strategies that cover their extremist intent.
Second, we employed a case study methodology. Although the case study allowed us to gain deep and nuanced insights into a specific actor’s strategy, and our findings match those of prior content analytical studies in this area (e.g., Bouko et al., 2020), future research replicating and extending our findings is needed to understand the generalizability of our observations. Furthermore, we focused only on Instagram. Generation Islam is also active on Facebook, YouTube, and TikTok. We deemed this decision justified, as the group has more followers on Instagram than on YouTube, 6 and their Facebook account links back to their Instagram account, underlining the centrality of the platform for them. Nevertheless, future research examining extremists’ communication across platforms would provide meaningful insights into the use of different affordances in different contexts (e.g., hashtags might be used differently on Twitter than on Instagram).
Moreover, we focused only on Islamic extremism. Although Islamic extremism remains a viable threat, and Europe is regularly shaken by terror attacks inspired by a radical and narrow pseudo-religious understanding of Islam, it is not to say that other extremist movements are less dangerous, and we explicitly call for research examining the employment of social media by other extremists too (for a similar call, see Conway, 2017). Such comparative research also promises meaningful insights into the different emotional patterns evident in the propaganda of Islamic extremists and, for instance, that of right-wing extremists.
Each of the methods employed had its own limitations. Imavis performed less effectively when the images contained a lot of text. Thus, using the method for images with many embedded citations should be interpreted with care. SentiBank provided a range of valuable information. However, as the algorithm does not provide a direct assignment to distinct emotional states, they must be inferred from the homepage. This means that SentiBank is currently not entirely scalable. Furthermore, dictionaries such as LIWC miss nuanced relationships between words. Consequently, if future research can employ more advanced natural language processing tools (e.g., co-occurrences and topic models), it is likely that it will provide additional insights. Overall, although we are confident that our multimethod approach allowed us to compensate for the weaknesses of single methods, future research on the optimal combination of methods and automated tools is needed.
Finally, we focused purely on an analysis of extremists’ propaganda strategies. Thus, our study does not allow for causal conclusions regarding the reception or the effects of Generation Islams’ Instagram strategy or of extremist propaganda in general. Future research examining the consumption of extremist propaganda on Instagram and testing the effects of such consumption is needed to examine the propaganda’s potential for radicalization.
Practical Implications and Conclusion
Nevertheless, our study provides novel insights into the affective and emotional strategies of extremist propaganda that exploit Instagram’s affordances. The results are important in terms of further algorithmic considerations and automated detection, and also in terms of creating awareness in social media users. Most relevant, they demonstrate that consideration of different affordances and their combinations, as well as consideration of distinct affective and emotional states, are needed to understand propaganda campaigns.
For platforms, covert propaganda increases the challenge of distinguishing between legitimate political discourse and propaganda, and it underlines the need for nuanced and context-sensitive moderation. For users, an awareness of these covert propaganda strategies is crucial, and targeted interventions are needed to build extremism literacy. Such strategies could follow the example of PRECOBIAS, 7 a campaign that teaches users about cognitive biases in extremist propaganda. Research into cognitive inoculation has shown that preemptively confronting people with small doses of a persuasive message and activating their cognitive defenses can help them handle extremist propaganda better (Lewandowsky & Yesilada, 2021). Our research shows that highlighting the risk of finding extremism behind seemingly innocous blue skies on Instagram should be part of such an inoculation campaign.
Footnotes
Declaration of Conflicting Interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: The authors acknowledge support by the German Ministry of Culture and Science of the German State of North Rhine-Westphalia (FKZ 005-1709-0001) and the European Research Center for Information Systems (ERCIS).
