Social media has played a big role in fueling the anti-immigration riots engulfing towns and cities in the United Kingdom. And agitator-in-chief Elon Musk is not sitting on the sidelines.

The Tesla chief executive and owner of X posted to the platform Sunday that “civil war is inevitable” in response to a post blaming the violent demonstrations on the effects of “mass migration and open borders.” On Monday, a spokesperson for the UK prime minister addressed Musk’s comment, telling reporters “there’s no justification for that.”

Musk’s decision to amplify the anti-immigrant rhetoric highlights the role that false information spread online is playing in fomenting real-world violence — an issue of growing concern to the UK government, which vowed Tuesday to bring those responsible for the riots, as well as their online cheerleaders, to justice. Later on Tuesday, a 28-year-old man in Leeds, northern England, became the first person to be charged with using “threatening words or behavior intending to stir up racial hatred” online, according to the UK Crown Prosecution Service. The charges related to “alleged Facebook posts,” Nick Price, the director of legal services at the CPS, said in a statement.

In recent days, rioters have damaged public buildings, set cars on fire and hurled bricks at police officers. They also set ablaze two Holiday Inn hotels in northern and central England believed to be housing asylum seekers awaiting a decision on their claims. Hundreds have been arrested. The riots broke out last week after far-right groups claimed on social media that the person charged with carrying out a horrific stabbing attack that left three children dead was a Muslim asylum seeker. The online disinformation campaign stoked outrage directed at immigrants.

The suspect, who has since been named as 17-year-old Axel Rudakubana, was born in the UK, according to police. But false claims about the attack — Britain’s worst mass stabbing targeting children in decades and possibly ever — spread rapidly online and continued garnering views even after the police had set the record straight.

According to the Institute for Strategic Dialogue, a think tank, by mid-afternoon on July 30, the day after the attack, a false name circulated online for the alleged asylum seeker had received more than 30,000 mentions on X alone from more than 18,000 unique accounts. “The false name attributed to the attacker was circulated organically but also recommended to users by platform algorithms,” the ISD said in a statement.

“Platforms therefore amplified misinformation to users who may not otherwise have been exposed, even after the police had confirmed the name was false.” According to the UK government, bots, which it said could be linked to state-backed actors, may well have amplified the spread of false information.

Tackling ‘online criminality’

Although social media companies have their own internal policies barring hate speech and incitement to violence from their platforms, they have long struggled to implement them.

“The problem has always been enforcement,” Isabelle Frances-Wright, a technology expert at the ISD, told CNN. “Particularly in times of crisis and conflict, when there is a huge groundswell of content, at which point their already fragile content moderation systems seem to fall apart.” It does not help matters that Musk himself has promoted incendiary content on X, a platform that European regulators last month accused of misleading and deceiving users. If he can do it, why not others?

For example, shortly after the October 7 Hamas attack on Israel and the ensuing outbreak of the war in Gaza, the self-declared “free speech absolutist” publicly endorsed an antisemitic conspiracy theory popular among White supremacists. Musk later apologized for what he called his “dumbest” ever social media post.

— CutC by Cnn.com

Leave A Reply

Exit mobile version