Rage-baiting

Rage-baiting or rage-farming is internet slang that refers to a manipulative tactic to elicit outrage with the goal of increasing internet traffic, online engagement, revenue and support. Rage baiting or farming can be used as a tool to increase engagement, attract subscribers, followers, and supporters, which can be financially lucrative. Rage baiting and rage farming manipulates users to respond in kind to offensive, inflammatory headlines, memes, tropes, or comments.

Rage-farming, which has been cited since at least January 2022, is an offshoot of rage-baiting where the outrage of the person being provoked is farmed or manipulated into an online engagement by rage-seeding that helps amplify the message of the original content creator. It has also been used as a political tactic at the expense of one's opponent.

Political scientist Jared Wesley of the University of Alberta stated in 2022 that the use of the tactic of rage farming was on the rise with right-wing politicians employing the technique by "promoting conspiracy theories and misinformation." As politicians increase rage farming against their political and ideological opponents, they attract more followers online, some of whom may engage in offline violence, including verbal violence and acts of intimidation. Wesley describes how those engaged in rage farming combine half-truths with "blatant lies".

Etymology, definitions and related terms
Rage farming is from rage + farm. Rage-seeding, rage-bait, rage baiting, and outrage baiting are similar Internet slang neologisms referring to manipulative tactics that feed on readers' anxieties and fears. They are all forms of clickbait, a term used used since c. 1999, which is "more nuanced" and not necessarily seen as a negative tactic. The term rage bait, which has been cited since at least 2009, is a negative form of click-baiting as it relies on manipulating users to respond in kind to offensive, inflammatory "headlines", memes, tropes, or comments.

In his 2022 tweet, a senior researcher at Citizen Lab, John Scott-Railton, described how a person was "being rage-farmed" when they responded to an inflammatory post with an equally inflammatory quote tweet as quote tweets reward the original rage tweet. Algorithms on social media such as Facebook, Twitter, TikTok, Instagram, and YouTube were discovered to reward increased positive and negative engagement by directing traffic to posts and amplifying them.

In an Atlantic article on Republican strategy, American writer Molly Jong-Fast described rage farming as "the product of a perfect storm of fuckery, an unholy mélange of algorithms and anxiety".

Political scientist Jared Wesley wrote that rage farming was often "used to describe rhetoric designed to elicit the rage of opponents." Rage-baiting is used to describe a tactic to attract, maintain, and increase a base of supporters and followers.

Clickbait, in all its iterations, including rage-baiting and farming, is a form of media manipulation, specifically Internet manipulation. While the goal of some clickbait is to generate revenue, it can also be used as effective tactic to influence people on social media platforms, such as Facebook, Twitter, Instagram, and YouTube. According to a November 2016 analysis of Facebook, clickbaits are intentionally designed to a targeted interest group's pre-existing confirmation biases. Facebook's algorithms used a filter bubble that shares specific posts to a filtered audience.

A Westside Seattle Herald article published May 2016 cited the definition from the online Urban Dictionary, "a post on social media by a news organisation designed expressly to outrage as many people as possible in order to generate interaction." The Herald article described how increased user traffic online results equals in more revenue for online platforms and websites from paid advertisements and sponsors.

A May 25, 2016 article described ragebait as "clickbait's evil twin."

A 2006 article in Time magazine described how Internet trolls post incendiary comments online with the sole purpose of provoking an argument even on the most banal topics. A statement like "NASCAR is about as much a sport as cheerleading" in a car-racing forum or openly supporting open borders to Lou Dobbs is cited as an example.

Rage bait and outrage bait creators invent "controversial news stories out of thin air". The example cited was a 15 December 2018 Irish digital media company ad falsely claiming that two thirds of people wanted Santa to be either female or gender neutral.

As early as 2012, research suggested that in both media and politics, eliciting outrage is a powerful tool in media manipulation. In political media, both real and imagined outrage attract readers, making rage-evoking narratives very popular.

Background
A 2012 Journal of Politics (JOP) article found that political actors were intentionally incorporating emotional content to evoke anxiety into their messaging to elicit interest in a topic. The article questioned why this political tactic resulted in viewers feeling more anger than anxiety. The study found that anger increased information-seeking behaviour and often resulted in web users clicking on the political website to learn more. The research said there were also psychological incentives to use angry rhetoric in political communication. A 2018 Media Matters for America article citing the JOP journal, reiterated that "anger is a powerful tool in the worlds of both politics and media." The political media industry knows that real or imagined outrage attracts readers making narratives that evoke it very popular in political media.

A November 2018 National Review article decrying social-justice warriors was cited as an example of rage-baiting by Media Matters for America. The Review article was in response to Tweets criticizing the cartoon image used by the ABC's Twitter account to advertise A Charlie Brown Thanksgiving on November 21, 2018. Franklin, the Black friend was sitting all alone on one side of Charlie Brown's Thanksgiving dinner table. Several unverified accounts by Twitter users, including one with zero followers, called the image racist. Conservatives were so frustrated by these overly sensitive, politically correct, "snowflake" liberals who posted, that they in turn responded in anger. The Media Matters for America article said that there was irony in the way in which the National Review article which intended to illustrate how liberals were too easily provoked to anger, actually succeeded in enraging conservatives.

Information technologies and digital media enable unprecedented capacities for online manipulation, including click-baiting, rage baiting and rage farming. In his January 7, 2022 tweet, John Scott-Railton described how a person was "being rage farmed" when they responded to an inflammatory post with an equally inflammatory quote tweet since algorithms on Twitter, TikTok, YouTube, Facebook and other social media platforms, reward posts that attract engagement by amplifying the posts.

A 2020 review of the conservative Canadian online news magazine, The Post Millennial, which was started in 2017, said it was far-right America's most recent rage-baiting outlet.

Social media
Rage farming and rage baiting are most recent iterations of clickbait and other forms of Internet manipulation that use conspiracy theories and misinformation to fuel anger and engage users. Facebook has been "blamed for fanning sectarian hatred, steering users toward extremism and conspiracy theories, and incentivizing politicians to take more divisive stands," according to a 2021 Washington Post report. In spite of previous reports on changes to its News Feed algorithms to reduce clickbait, revelations by Facebook whistleblower Frances Haugen and content from the 2021 Facebook leak, informally referred to as the Facebook Papers, provide evidence of the role the company's News Feed algorithm had played.

Media and governmental investigations in the wake of revelations from Facebook whistleblower, Frances Haugen, and the 2021 Facebook leak, provide insight into the role various algorithms play in farming outrage for profit by spreading divisiveness, conspiracy theories and sectarian hatred that can allegedly contribute to real-world violence. A highly criticized example was when Facebook, with over 25 million accounts in Myanmar, neglected to police rage-inducing hate speech posts targeting the Rohingya Muslim minority in Myanmar that allegedly facilitated the Rohingya genocide. In 2021, a US$173 billion class action lawsuit filed against Meta Platforms Inc (the new name of Facebook) on behalf of Rohingya refugees claimed that Facebook's "algorithms amplified hate speech."

In response to complaints about clickbait on Facebook's News Feed and News Feed ranking algorithm, in 2014 and again in 2016, the company introduced an anti-clickbait algorithm to remove sites from their News Feed that frequently use headlines that "withhold, exaggerate or distort information."

A February 2019 article that was promoted in Facebook described how outrage bait made people angry "on purpose". Digital media companies and social media actors incite outrage to increase engagement; "clicks, comments, likes and shares", which generate "more advertising revenue". If content does not increase engagement, "timeline algorithm" limits the number of users that this uninteresting content can reach. According to this article, when geared up on its war against clickbait, algorithm changed, which made it harder for creators and sites to use clickbait. The article said that a new engagement strategy was introduced to replace clickbait, whether rage bait or outrage bait.

The 2016 algorithms were allegedly trained to filter phrases that were frequently used in clickbait headlines similar to filters that remove email spam. Publishers who continue to use clickbait were allegedly punished through loss of referral traffic.

Starting in 2017, Facebook engineers changed their ranking algorithm to score emoji reactions five times higher than mere "likes" because emojis extended user engagement, according to a 26 October 2021 Washington Post article. Facebook's business model depended on keeping and increasing user engagement. One of Facebook's researchers raised concerns that the algorithms that rewarded "controversial" posts including those that incited outrage, could inadvertently result in more spam, abuse, and clickbait.

Since 2018, Facebook executives had been warned in a slide presentation that their algorithms promoted divisiveness but they refused to act. In a 2022 interview Scott-Railton had observed that the amplification by algorithms of these inflammatory quote tweets in rage farming that looped upon themselves may have been planned and structural or accidental. Algorithms reward positive and negative engagement. This creates a "genuine dilemma for everyone". Algorithms also allow politicians to bypass legacy media outlets that fact-check, by giving them access to a targeted uncritical audience who are very receptive of their messaging, even when it is misinformation.

By 2019, Facebook's data scientists confirmed that posts that incited the angry emoji were "disproportionately likely to include misinformation, toxicity and low-quality news."

The 2020 Netflix docudrama The Social Dilemma analyzed how social media was intentionally designed for profit maximization through Internet manipulation which can include spreading conspiracy theories and disinformation and promoting problematic social media use. Topics covered in the film included the role of social media in political polarization in the United States, political radicalization, including online youth radicalization, the spread of fake news and as a propaganda tool used by political parties and governmental bodies. Social media networks have three main goals: to maintain and increase engagement, growth, and advertisement income, according to a former Google design ethicist.

Facebook outside the United States
A 2021 report by the Washington Post revealed that Facebook did not adequately police its service outside the United States. The company invested only 16% of its budget to fight misinformation and hate speech in countries outside the United States, such as France, Italy, and India where English is not the maternal language. In contrast, the company allocated 84% to the United States which only represents 10% of Facebook's daily users.

Since at least 2019, Facebook employees were aware of how "vulnerable these countries, like India, were to "abuse by bad actors and authoritarian regimes" but did nothing to block accounts that published hate speech and incited violence. In their 2019 434-page report submitted to the Office of the United Nations High Commissioner for Human Rights on the findings of the Independent International Fact-Finding Mission on Myanmar, the role of social media in disseminating hate speech and inciting violence in the anti-Muslim riots and the Rohingya genocide was investigated. Facebook was mentioned 289 times in the report as there are millions of Facebook accounts in that country. Following the publication of an earlier version of the report in August, Facebook took the "rare step" of removing accounts that represented 12 million followers implicated in the reports findings.

In October 2021, Haugen testified at a United States Senate committee that Facebook had been inciting ethnic violence in Myanmar which has over 25 million Facebook users and in Ethiopia through its algorithms that promoted posts inciting or glorifying violence. False claims about Muslims stockpiling weapons were not removed.

The Digital Services Act is a European legislative proposal to strengthen rules on fighting disinformation and harmful content, that was submitted by the European Commission to the European Parliament and the Council of the European Union partially in response to concerns raised by the Facebook Files and revelations in Haugen's testimony in the European Parliament. In 2021, a c$. US 173 billion dollar class action lawsuit was lodged by law firms Edelson PC and Fields PLLC against Meta Platforms Inc, formerly known as Facebook in the United States District Court for the Northern District of California on behalf of Rohinga refugees, claiming that Facebook was negligent in not removing inflammatory posts that facilitated the Rohingya genocide in Myanmar. The lawsuit said that Facebook's "algorithms amplified hate speech."

Following its launch in Myanmar in 2011, Facebook "quickly became ubiquitous." A report commissioned by Facebook led to the company's admission in 2018, that they had been failed to do "enough to prevent the incitement of violence and hate speech against the [...]Muslim minority in Myanmar." The independent report found that "Facebook has become a means for those seeking to spread hate and cause harm, and posts have been linked to offline violence".