From Hateful Words to Real Violence

Members, who now see themselves as victims, also find that publicly expressing their negative emotions in online white supremacist networks is encouraged, validated, and rewarded by their peers, which boosts feelings of acceptance, solidarity, power, pride, and even potentially pleasurable and addictive neural rushes.

Gottschalk emphasizes that while these dynamics are not unique to American white supremacist networks, the desire for revenge and the potential for violence become especially volatile among members of social categories who confront a “reversal of fortune.”

“Individuals who are no longer granted the respect and esteem they expect and feel entitled to by virtue of their race will experience this condition as a frontal assault on their sense of self and identity.”

Escalating Online Hate to Violence
UNLV’s research of online white supremacist networks follows previous research findings that members interpret negative personal experiences as caused by discrimination. But the UNLV researchers argue that their model explores yet another important step that explains the path to violence: a switch in members’ perceptions whereby they are not only outraged because they believe that an “enemy” discriminates against them, but are now also afraid because they believe that this enemy threatens to physically harm them. 

“When this switch occurs, variants of anger fuse with variants of fear to form an especially explosive compound. Under the ‘right’ conditions, some individuals motivated by those emotions can easily surrender to bloodlust and justify violence as self-defense,” Gottschalk says. 

Gottschalk’s model suggests that interactions in online white supremacist networks produce those very conditions. He says the conversion to violent behavior is especially likely in fascist networks — regardless of race, religion, or nation — as their ideologies typically bestow the ultimate prestige not to those who talk a good talk about violence, but to those who actually act on it. 

Assessing Calls for Violence
Gottschalk’s research team examined users’ posts to determine their prevailing emotions, grievances, and motivations.

Their analysis found that anger was the prevailing emotion (51%). And among the variants of anger, vengeance was the emotion most frequently expressed (37%). Among different types of vengeance, more than 15% of those threads mention sadistic fantasies of physical harm, killing, mutilation, and extermination. The most frequently voiced accusation (41%) is that enemies — especially Jews — are intent on physically destroying white people. 

Researchers called it “noteworthy” that there was a “complete absence” of comments seeking to temper calls for violence.

“Analyzing anonymous online discussion threads provided us the unique advantage of gleaning the uncensored, spontaneous views of white supremacists who might not have been so candid in a face-to-face interview,” Gottschalk said. “Online hate group members find that — under the double-cloak of anonymity and invisibility — they no longer have to censor or even moderate hostile dispositions that would be considered paranoid or criminal in most other settings.” 

Why This All Matters
In contrast to physical networks, online ones provide portability, mobility, and 24/7 access. And as these networks are multiplying worldwide, their content can easily migrate to less extremist ones, contaminate them, normalize extremist beliefs, and shift the range of acceptable ideas that can be discussed in society.

Even more worrisome, Gottschalk says, algorithms on search engines and social media sites systematically manipulate the likelihood that increasingly extreme posts will circulate more broadly, thereby capturing members’ attention, stirring powerful emotions, and in some cases, encouraging violent action.

“You can literally carry the network in your back pocket and stream its hateful ideology straight into your brain. It can provide you with constant and instant positive feedback whenever you voice the emotionally correct messages, paranoid beliefs, genocidal threats, or plans of action. Once individuals are hooked to the network, it becomes relatively simple for those in the control booth to modulate the anger-fear complex,” Gottschalk said.

“Obviously, how some people respond to what they see on the screens of online white supremacist networks can become a matter of life and death,” he continues. “And when you consider the combination of those powerful online group dynamics, the increasing tolerance for the online expression of hateful emotions, and the ready availability of weapons in our society, we should move quickly and intelligently.”

To illustrate the risks these networks represent, Gottschalk points to a 2012 Facebook experiment that exposed nearly 700,000 unaware users to mild negative emotions, and found that these users quickly contaminated their own networks with these emotions. 

“If you can achieve this effect with mild negative emotions, just imagine the velocity and ferocity with which strong primary negative emotions such as anger and fear can infect online extremist and other networks, especially in times of social confusion and instability. It could even get worse,” he says. “Remember how Russian trolls mobilized real individuals to participate in fictitious street protests during the 2016 elections? Well, bots will soon be able to perform this function more precisely, quickly, and efficiently than human beings. Add to that the growing concern with deep fakes, and we have a rather volatile situation in our hands.”

Possible Solutions
Gottschalk — who has spent a decade studying the social psychological impacts of technology on our lives — believes that until we better understand the influence of these networks, they should be shut down.

He points to the example of the French government, which in November 2018 blocked all access to the Démocratie Participative website — the French equivalent of The Daily Stormer — and similar networks.

“I know this suggestion sounds unrealistic to many, but is it really?” Gottschalk said. “While it does not guarantee immediate success, it will at least disrupt the dynamics driving the networks of the ‘fascoscphere’ and hopefully contain their effects. At the same time, we should develop additional strategies that address white supremacists’ claims.”