Media & Entertainment

Tech is not winning the battle against white supremacy

Comment

Content warning: This post contains racial slurs, homophobic language and very graphic depictions of racism and violence.

If you were just paying attention to press releases this week it’d be easy to believe that tech companies are winning the war on hate. Responding to the violence in Charlottesville, Mark Zuckerberg solemnly reflected that there is “no place for hate in our community.” Snapchat announced that hate speech “will never be tolerated” on its platform. YouTube reassured us that helpful tools are on the way. Tech companies fled Trump’s dual business councils to protest his claim that some white supremacists are “very fine people.”

In other headlines, a coalition of web providers made a controversial and unprecedented choice to yank their services out from under the Daily Stormer, a white supremacist news site. Days later, Cloudflare abandoned the site to the whims of whoever feels like DDoSing it. Those decisions, part of the “no platforming” philosophy which would deny hate speech purveyors a place to assemble and share their views, will likely have many reverberations in the days to come. For now, some things remain very much the same.

Unfortunately, while this week’s burst of industry energy might suggest otherwise, hate groups are alive and well, making little if any effort to conceal their presence on all of the major social networks. Whether it’s 4chan or Facebook, if you go looking for hate online, you’ll find it. Dredging up racist, anti-Semitic content often in seeming violation of a company’s stated policy takes seconds — trust me, I went looking.

On something like Facebook, hate festers just under the paper-thin layer between a user’s social sphere and the platform at large. On a network like Twitter, it’s right on the surface, bobbing unpleasantly along down the stream with dog photos and journalist chatter. For anyone surprised about the terrible events that unfolded in Charlottesville: You can find hate anywhere you look and you don’t have to look very hard.

I took a grim tour around some of the major social sites into which we sink our hours to see what just a little bit of casual searching could find — yet algorithms often can’t (or won’t). Again, this content is graphic and disturbing, but pretending it isn’t there won’t make it go away.

Facebook

On Facebook, white supremacist memes thrive, even in wide-open, public communities. Though plenty of hate just sits out in the open, some users skirt detection by using a kind of unsearchable, far-right code language. Facebook might pick up on the anti-Semitic slur “kike,” but by swapping that for “kayak” the content flies under the radar. I was surprised to see that surrounding words in multiple parentheses, also called an “echo,” remains common practice to denote something or someone as Jewish. These symbols were established as part of the shallowly submerged white supremacy lexicon more than a year ago.

References to 1488 also remain common, where 14 is a nod to the “14 words” or “We must secure the existence of our people and a future for white children,” a popular mantra with white supremacists and white nationalists. The double 88 is usually a nod to the 8th letter of the alphabet, or “heil Hitler.”

Small waves of white supremacist memes crest and fall, and much like Facebook’s fake news problem, each wave has another set right behind it and there are many oceans. When I spent some time looking through these communities this week, a particularly popular meme remixed the incredible violence of a counter-protester rolling off of a now infamous ash-gray Dodge Charger with a broad array of anti-black racist memes, some of them drawing from popular mainstream memes, like “the floor is” joke. Another pictured George Washington driving the Charger through the crowd.

 

One public community I found easily hosted a live stream of Saturday’s white supremacist rally in Charlottesville, the full video shot from the perspective of one of the torch-bearing attendees. It felt surprising that so much of this content was just sitting right out in the open on a social network that connects faces to names.

Following Charlottesville, Facebook cracked down, removing a slew of white supremacist and white nationalist pages. Among them: Right Winged Knight, Right Wing Death Squad, Awakening Red Pill, Physical Removal, Genuine Donald Trump, Awakened Masses, White Nationalists United, Vanguard America, Radical Agenda: Common Sense Extremism and the personal page of Chris Cantwell. Many, many others remain as Facebook continues to rely on users flagging content themselves — a deeply flawed method that’s proven far more effective as a tool for harassing LGBTQ users and black activists than ridding the platform of hate.

In his statement on Wednesday, Zuckerberg did not meaningfully clarify how Facebook will determine what stays on its platform and what goes. Though he noted that “when someone tries to silence others or attacks them based on who they are or what they believe, that hurts us all and is unacceptable,” it does not appear to be unacceptable on Facebook.

Asked how its policy might be evolving, Facebook told me that it does not tolerate hate speech or posts praising acts of violence or hate groups on its platform. This policy, like all policies, is open to interpretation and it’s possible that interpretation could shift further over time.

Reddit

In spite of Reddit’s mostly hands-off policy and reliance on subreddit-specific moderators, racism on Reddit often takes quirkier forms meant to avoid potential detection. In true Reddit style, overtly racist posts and comments are often played off as self-parodies, draping a thin layer of self-referential humor over what is usually just outright white supremacy. On one thread, users enthusiastically counted up from the number 1,488,000. On subreddits like /r/greentext, users post screencaps of posts from 4chan, host of some popular far right and white supremacist communities. They’re careful not to post links to 4chan itself and by screencapping they can avoid searchable text while still replicating most of the content.

In late 2015, Reddit rid itself of some popular openly white supremacist subcommunities like /r/coontown during a prominent sweep, but remarkably, pages like /r/blackpeoplehate live on. Reddit now classifies its most objectionable content as under “quarantine” and requires a verified email address to access it. Like YouTube, which took a similar approach of walling off some content, Reddit “will generate no revenue, including ads or Reddit Gold,” from these subreddits. They live on in a state of partial suspended animation.

Following the violence in Charlottesville, Reddit told me that it banned /r/physical_removal for “a violation of our content policy, specifically, the posting of content that incites violence.” The company appears responsive to user-generated campaigns when they draw sufficient attention to an issue, which appears to be the goal of /r/AgainstHateSubreddits, a compendium of Redditor-reported hate speech.

YouTube

Initially, YouTube’s search made finding white supremacist stuff kind of hard. Given Google’s web search prowess it makes sense that the company would do a better job of burying objectionable content than a site like Facebook, but it wasn’t buried very deep. After a few searches didn’t turn up much, I struck Nazis on a video that prominently displayed a 1488 with a slew of links to the Daily Stormer.

Because it’s an entertainment site as much as a social network, many of my search results were home-brewed music videos depicting Nazi imagery with little or no context. A cursory glance at the user names and links was the only overt hint, with, again, many, many 1488s. Some more narrative racism came with disclaimers that the content was satire or just a joke.

Elsewhere, content drawn directly from 4chan’s infamous far right hub /pol/  (short for “politically incorrect”) was repurposed on a more mainstream platform. Because YouTube, like many of these sites, offers recommended content related to what you’re viewing, stumbling onto a little bit of white supremacy opens up a cascading slide of swastikas and racial epithets. Just a few clicks away from a music video declaring whites the master race I ran into a video created by “fashygamer1488” with the following text:

“Hey goys, its [me] here with another video, please write ur comments below, no (((jews))) or googles allowed (Google is a secret alt-right codeword that means the N word lol)…”

Again, racial slurs are traded for common, unsearchable words to keep the content just barely underground.

In June, YouTube followed Reddit’s example, creating a separate class of objectionable content that it would no longer monetize. This followed a corporate outcry from brands concerned that their ads were being served along with videos containing hate speech. In just a little bit of time spent browsing YouTube’s white supremacist content, I did not run into anything that set this content apart from the rest of its videos, though YouTube has said that feature is coming “soon” and that the “videos will remain on YouTube behind an interstitial, won’t be recommended, won’t be monetized, and won’t have key features including comments, suggested videos, and likes.”

For now, the suggestion engine hums along, pointing me toward a selection of Hitler youth haircut instructional videos.

Fashy haircuts

Twitter

Twitter is more responsive as a search engine than something like Facebook, but the search results are often messily curated. My first search for 1488 quickly pulls up tweets like a picture of a white, blue-eyed baby with the text “14 words” and a photo of Hitler. In other tweets, users with neo-Nazi black sun icons and hybrid Trump/Hitler background images call each other “fags” over who is and is not “boomerposting” (i.e. tweeting like a baby boomer).

Unsurprisingly, Twitter has it all. White supremacist demagogue Richard Spencer trying to remain relevant while his peers accuse him of being a Jew. Quotes hailing Trump’s off-the-rails presser that defended some white supremacists as “fine people.” Racist code words that reverse virtue-signal hate to anyone looking for a like-minded follow. Jokes about cars caked in Photoshopped blood. All of it sends the same message.

On Twitter, there is a lot, lot, lot of this content. It starts to run together.

Tech at a crossroads

These major platforms offer a taste of the toxicity flowing through mainstream social networks, but there are many others. After incubating this kind of stuff for ages, gaming chat platform Discord just finished a major purge. Tumblr, Instagram and Snapchat are fighting the same fight and it’s not clear they’re winning. Meanwhile, far right offshoots like Gab are specifically designed with sustainable white supremacy in mind. The absolute ubiquity of Nazi insignia, Stormfront links and shockingly violent memes would appear to undercut objections by the extreme right that their speech is being suppressed with much success at all.

Depending on how you use the internet, the fact that this stuff is so easy to find on major social networks could range anywhere from shocking to wholly unsurprising. But the truth is that most of us shy away from looking at it. For anyone who isn’t the target demographic, all of this hate is ugly and exhausting. We’d rather just rest easy knowing that tech companies are working on it and they’d rather we didn’t haul up more of this stuff — they’re working on it.

As we can see from tech ratcheting up its response following Charlottesville, no policy is set in stone. While companies often point users to policies around what does and doesn’t fly on their platforms, ultimately the decision to ban content is a subjective response to getting too much heat. Given that willingness to bend to public sentiment, corporate pressure and user-driven anti-hate campaigns are proving themselves to be powerful tools, even if it’s not clear where exactly to draw the line. Racial slurs? Nazi insignia? Overt threats of specific violence? For tech, the coming weeks will be a bellwether.

Anywhere you go, white supremacist content has a foothold if not an entire underground compound bedecked in red and black — one that remains even after the Charlottesville backlash. All one needs to do is look. Whether tech companies choose to see is a different matter altogether.

More TechCrunch

Amazon Web Services (AWS), Amazon’s cloud computing business, has confirmed further details of its European “sovereign cloud” which is designed to enable greater data residency across the region. The company…

AWS confirms European ‘sovereign cloud’ to launch in Germany by 2025, plans €7.8B investment over 15 years

Go Digit, an Indian insurance startup, has raised $141 million from investors including Goldman Sachs, ADIA, and Morgan Stanley as part of its IPO.

Indian insurance startup Go Digit raises $141M from anchor investors ahead of IPO

Peakbridge intends to invest in between 16 and 20 companies, investing around $10 million in each company. It has made eight investments so far.

Food VC Peakbridge has new $187M fund to transform future of food, like lab-made cocoa

For over six decades, the nonprofit has been active in the financial services sector.

Accion’s new $152.5M fund will back financial institutions serving small businesses globally

Meta’s newest social network, Threads is starting its own fact-checking program after piggybacking on Instagram and Facebook’s network for a few months. Instagram head Adam Mosseri noted that the company…

Threads finally starts its own fact-checking program

Looking Glass makes trippy-looking mixed-reality screens that make things look 3D without the need of special glasses. Today, it launches a pair of new displays, including a 16-inch mode that…

Looking Glass launches new 3D displays

Replacing Sutskever is Jakub Pachocki, OpenAI’s director of research.

Ilya Sutskever, OpenAI co-founder and longtime chief scientist, departs

Intuitive Machines made history when it became the first private company to land a spacecraft on the moon, so it makes sense to adapt that tech for Mars.

Intuitive Machines wants to help NASA return samples from Mars

As Google revamps itself for the AI era, offering AI overviews within its search results, the company is introducing a new way to filter for just text-based links. With the…

Google adds ‘Web’ search filter for showing old-school text links as AI rolls out

Blue Origin’s New Shepard rocket will take a crew to suborbital space for the first time in nearly two years later this month, the company announced on Tuesday.  The NS-25…

Blue Origin to resume crewed New Shepard launches on May 19

This will enable developers to use the on-device model to power their own AI features.

Google is building its Gemini Nano AI model into Chrome on the desktop

It ran 110 minutes, but Google managed to reference AI a whopping 121 times during Google I/O 2024 (by its own count). CEO Sundar Pichai referenced the figure to wrap…

Google mentioned ‘AI’ 120+ times during its I/O keynote

Firebase Genkit is an open source framework that enables developers to quickly build AI into new and existing applications.

Google launches Firebase Genkit, a new open source framework for building AI-powered apps

In the coming months, Google says it will open up the Gemini Nano model to more developers.

Patreon and Grammarly are already experimenting with Gemini Nano, says Google

As part of the update, Reddit also launched a dedicated AMA tab within the web post composer.

Reddit introduces new tools for ‘Ask Me Anything,’ its Q&A feature

Here are quick hits of the biggest news from the keynote as they are announced.

Google I/O 2024: Here’s everything Google just announced

LearnLM is already powering features across Google products, including in YouTube, Google’s Gemini apps, Google Search and Google Classroom.

LearnLM is Google’s new family of AI models for education

The official launch comes almost a year after YouTube began experimenting with AI-generated quizzes on its mobile app. 

Google is bringing AI-generated quizzes to academic videos on YouTube

Around 550 employees across autonomous vehicle company Motional have been laid off, according to information taken from WARN notice filings and sources at the company.  Earlier this week, TechCrunch reported…

Motional cut about 550 employees, around 40%, in recent restructuring, sources say

The keynote kicks off at 10 a.m. PT on Tuesday and will offer glimpses into the latest versions of Android, Wear OS and Android TV.

Google I/O 2024: Watch all of the AI, Android reveals

Google Play has a new discovery feature for apps, new ways to acquire users, updates to Play Points, and other enhancements to developer-facing tools.

Google Play preps a new full-screen app discovery feature and adds more developer tools

Soon, Android users will be able to drag and drop AI-generated images directly into their Gmail, Google Messages and other apps.

Gemini on Android becomes more capable and works with Gmail, Messages, YouTube and more

Veo can capture different visual and cinematic styles, including shots of landscapes and timelapses, and make edits and adjustments to already-generated footage.

Google Veo, a serious swing at AI-generated video, debuts at Google I/O 2024

In addition to the body of the emails themselves, the feature will also be able to analyze attachments, like PDFs.

Gemini comes to Gmail to summarize, draft emails, and more

The summaries are created based on Gemini’s analysis of insights from Google Maps’ community of more than 300 million contributors.

Google is bringing Gemini capabilities to Google Maps Platform

Google says that over 100,000 developers already tried the service.

Project IDX, Google’s next-gen IDE, is now in open beta

The system effectively listens for “conversation patterns commonly associated with scams” in-real time. 

Google will use Gemini to detect scams during calls

The standard Gemma models were only available in 2 billion and 7 billion parameter versions, making this quite a step up.

Google announces Gemma 2, a 27B-parameter version of its open model, launching in June

This is a great example of a company using generative AI to open its software to more users.

Google TalkBack will use Gemini to describe images for blind people

Google’s Circle to Search feature will now be able to solve more complex problems across psychics and math word problems. 

Circle to Search is now a better homework helper