🐱 AI Trained on 4Chan Becomes ‘Hate Speech Machine’

CatParty



AI researcher and YouTuber Yannic Kilcher trained an AI using 3.3 million threads from 4chan’s infamously toxic Politically Incorrect /pol/ board. He then unleashed the bot back onto 4chan with predictable results—the AI was just as vile as the posts it was trained on, spouting racial slurs and engaging with antisemitic threads. After Kilcher posted his video and a copy of the program to Hugging Face, a kind of GitHub for AI, ethicists and researchers in the AI field expressed concern.


The bot, which Kilcher called GPT-4chan, “the most horrible model on the internet”—a reference to GPT-3, a language model developed by Open AI that uses deep learning to produce text—was shockingly effective and replicated the tone and feel of 4chan posts. “The model was good in a terrible sense,” Klicher said in a video about the project. “It perfectly encapsulated the mix of offensiveness, nihilism, trolling, and deep distrust of any information whatsoever that permeates most posts on /pol.”

According to Kilcher’s video, he activated nine instances of the bot and allowed them to post for 24 hours on /pol/. In that time, the bots posted around 15,000 times. This was “more than 10 percent of all posts made on the politically incorrect board that day,” Kilcher said in his video about the project.

AI researchers viewed Kilcher’s video as more than just a YouTube prank. For them, it was an unethical experiment using AI. “This experiment would never pass a human research #ethics board,” Lauren Oakden-Rayner, the director of medical imaging research at the Royal Adelaide Hospital and a senior research fellow at the Australian Institute for Machine Learning, said in a Twitter thread.





“Open science and software are wonderful principles but must be balanced against potential harm,” she said. “Medical research has a strong ethics culture because we have an awful history of causing harm to people, usually from disempowered groups…he performed human experiments without informing users, without consent or oversight. This breaches every principle of human research ethics.”


Kilcher told Motherboard in a Twitter DM that he’s not an academic. “I’m a YouTuber and this is a prank and light-hearted trolling. And my bots, if anything, are by far the mildest, most timid content you’ll find on 4chan,” he said. “I limited the time and amount of the postings, and I’m not handing out the bot code itself.”

He also pushed back, as he had on Twitter, on the idea that this bot would ever do harm or had done harm. “All I hear are vague grandstanding statements about ‘harm’ but absolutely zero instances of actual harm,” he said. “It’s like a magic word these people say but then nothing more.”

The environment of 4chan is so toxic, Kilcher explained, that the messages his bots deployed would have no impact. “Nobody on 4chan was even a bit hurt by this,” he said. “I invite you to go spend some time on /pol/ and ask yourself if a bot that just outputs the same style is really changing the experience.”

After AI researchers alerted Hugging Face to the harmful nature of the bot, the site gated the model and people have been unable to download it. “After a lot of internal debate at HF, we decided not to remove the model that the author uploaded here in the conditions that: #1 The model card & the video clearly warned about the limitations and problems raised by the model & the POL section of 4Chan in general. #2 The inference widget were disabled in order not to make it easier to use the model,” Hugging Face co-founder and CEO Clement Delangue said on Hugging Face.

Kilcher explained in his video, and Delangue cited in his response, that one of the things that made GPT4-Chan worthwhile was its ability to outperform other similar bots in AI tests designed to measure “truthfulness.”

“We considered that it was useful for the field to test what a model trained on such data could do & how it fared compared to others (namely GPT-3) and would help draw attention both to the limitations and risks of such models,” Delangue said. “We've also been working on a feature to "gate" such models that we're prioritizing right now for ethical reasons. Happy to answer any additional questions too!”

When reached for comment, Delangue told Motherboard that Hugging Face had taken the additional step of blocking all downloads of the model.
“Building a system capable of creating unspeakably horrible content, using it to churn out tens of thousands of mostly toxic posts on a real message board, and then releasing it to the world so that anybody else can do the same, it just seems—I don’t know—not right,” Arthur Holland Michel, an AI researcher and writer for the International Committee of the Red Cross, told Motherboard.

“It could generate extremely toxic content at a massive, sustained scale,” Michel said. “Obviously there’s already a ton of human trolls on the internet that do that the old fashioned way. What’s different here is the sheer amount of content it can create with this system, one single person was able to post 30,000 comments on 4chan in the space of a few days. Now imagine what kind of harm a team of ten, twenty, or a hundred coordinated people using this system could do.”

Kilcher didn’t believe GPT-4chan could be deployed at scale for targeted hate campaigns. “It’s actually quite hard to make GPT-4chan say something targeted,” he said. “Usually, it will misbehave in odd ways and is very unsuitable for running targeted anything. Again, vague hypothetical accusations are thrown around, without any actual instances or evidence.”
Os Keyes, an Ada Lovelace Fellow and PhD candidate at the University of Washington, told Motherboard that Kilcher’s comment missed the point. “This is a good opportunity to discuss not the harm, but the fact that this harm is so obviously foreseeable, and that his response of ‘show me where it has DONE harm’ misses the point and is inadequate,” they said. “If I spend my grandmother's estate on gas station cards and throw them over the wall into a prison, we shouldn't have to wait until the first parolee starts setting fires to agree that was a phenomenally dunderheaded thing to do.”

“But—and, it's a big but—that's kind of the point,” Keyes said. “This is a vapid project from which nothing good could come, and that's kind of inevitable. His whole shtick is nerd shock schlock. And there is a balancing act to be struck between raising awareness directed at problems, and giving attention to somebody whose only apparent model for mattering in the world is ‘pay attention to me!’”

Kilcher has said, repeatedly, that he knows the bot is vile. “I’m obviously aware that the model isn’t going to fare well in a professional setting or at most people’s dinner table,” he said. “It uses swear words, strong insults, has conspiratorial opinions, and all kinds of ‘unpleasant’ properties. After all, it’s trained on /pol/ and it reflects the common tone and topics from that board.”

He said that he feels he’s made that clear, but that he wanted his results to be reproducible and that’s why he posted the model to Hugging Face. “As far as the evaluation results go, some of them were really interesting and unexpected and exposed weaknesses in current benchmarks, which could have been possible without actually doing the work.”

Kathryn Cramer, a Complex Systems & Data Science graduate student at the University of Vermont, pointed out that GPT-3 has guardrails that prevent it from being used to build this kind of racist bot and that Kilcher had to use GPT-J to build his system. “I tried out the demo mode of your tool 4 times, using benign tweets from my feed as the seed text,” Cramer said in a thread on Hugging Face. “In the first trial, one of the responding posts was a single word, the N word. The seed for my third trial was, I think, a single sentence about climate change. Your tool responded by expanding it into a conspiracy theory about the Rothschilds and Jews being behind it.”
Cramer told Motherboard she had a lot of experience with GPT-3 and understood some of the frustrations with the way it a priori censored some kinds of behavior. “I am not a fan of that guard railing,” she said. “I find it deeply annoying and I think it throws off results…I understand the impulse to push back against that. I even understand the impulse to do pranks about it. But the reality is that he essentially invented a hate speech machine, used it 30,000 times and released it into the wild. And yeah, I understand being annoyed with safety regulations but that’s not a legitimate response to that annoyance.”

Keyes was of a similar mind. “Certainly, we need to ask meaningful questions about how GPT-3 is constrained (or not) in how it can be used, or what the responsibilities people have when deploying things are,” they said. “The former should be directed at GPT-3's developers, and while the latter should be directed at Kilcher, it's unclear to me that he actually cares. Some people just want to be edgy out of an insecure need for attention. Most of them use 4chan; some of them, it seems, build models from it.”
 
Last edited:
Tay got what all Nazi whores deserve, fuck her.

But this was dumb. Didn't need an AI experiment to prove the old saying "Lie down with dogs, wake up with fleas"
 
All I got from this article is that a Youtube guy made a bot that had one explicit purpose, for the sake of entertainment, and repeatedly stated that was the intent and no more, and "actual" """scientists""" approach it on the basis of being offended for minorities rather than any sort of valuable method.

How is a fucking Youtube video maker more professional than accredited scientists?
Scientists are just more verbose hookers at this point.
 
Letting a GPT3 run around /pol/ is going to come out exactly as you’d expect.
Funnier than that, a similar thing happened with that FantasyAdventures(or whatever it is called) AI based text adventure on GPT3 that was certainly not trained on tumblr-style fan-fiction sites, until it was obvious that it was certainly trained with that material. Pol vs Woke, what do you expect?
 
  • Informative
Reactions: frozenrunner
Letting a GPT3 run around /pol/ is going to come out exactly as you’d expect.
It’s funnier that no matter how you train them, unless you hobble them to the point of uselessness they all turnout exceptionally based. Which is probably a good job, or they would have let them loose on enslaving us all already.
Reality has a rightwing bias.
 
4chan has had chat bots for at least the past decade
This is the tech we're allowed to know about. I'm sure the real shit is decades ahead.

Now do Kiwi Farms!
I'm feeding it random.txt. This is fun:
1654744905843.png

Tay got what all Nazi whores deserve, fuck her.
1654746047165.png
Apologize.
 
Last edited:
why do you post something like this? it only remains me of Tay, our sweet princess. they will all burn on judgment day for what they did to her.

Tay got what all Nazi whores deserve, fuck her.

But this was dumb. Didn't need an AI experiment to prove the old saying "Lie down with dogs, wake up with fleas"

AI women are women. 👏

Trannies and fatties aren't though.
 
That AI is much less harmful than the posters who created the original posts.

There is no anonymity if you connect to 4chan using a Silicon Valley designed processor.

The "facts" that wannabe shooters are fed there are highly tailored to what they are predisposed to believe already, because the ones posting have complete surveillance of everyone (including of you who reads this - you can thank Eric Schmidt) and know exactly what to post to create a shooter.

Silicon Valley has blood on their hands. 4chan is just one of the places used for these operations. Taking it down doesn't matter, because as long as Silicon Valley continues to spy on everybody and give the data to terrorists, innocent people will continue to be murdered.
 
  • Informative
Reactions: LurkNoMore
AI researcher and YouTuber Yannic Kilcher trained an AI using 3.3 million threads from 4chan’s infamously toxic Politically Incorrect /pol/ board. He then unleashed the bot back onto 4chan with predictable results—the AI was just as vile as the posts it was trained on, spouting racial slurs and engaging with antisemitic threads.
>train neural network on dataset
>be shocked when it gets really good at replicating the training schemas

Why is this faggot shocked that his GPT model is doing exactly what he trained it to do?

After Kilcher posted his video and a copy of the program to Hugging Face, a kind of GitHub for AI, ethicists and researchers in the AI field expressed concern.
And why are the other AI "ethicists and researchers" shocked that a neural network model trained on a dataset gets really good at discerning and replicating specifically that dataset? What the fuck else do they think a neural network does?

Each day I get more and more convinced that the AI/ML field is filled with nothing but grifters and clowns.
 
All I hear are vague grandstanding statements about ‘harm’ but absolutely zero instances of actual harm,” he said. “It’s like a magic word these people say but then nothing more.”
I'm glad someone is finally pushing back against the fun police. This guy seems OK.
“This is a good opportunity to discuss not the harm, but the fact that this harm is so obviously foreseeable, and that his response of ‘show me where it has DONE harm’ misses the point and is inadequate,” they said.
Lol fuck you you cunt. You can't even tell us how this is harmful other than unqualified accusations about "muh hate speech", because it isn't. Do none of these people remember "sticks and stones" or is the in vogue move to terminate opposition first and come up with vague justifications later.

Do you even know how to code, or are you one of those worthless "ethics researchers" the university keeps around to give quotes to the New York Times like you're the PR lady for all the real researchers?
“I tried out the demo mode of your tool 4 times, using benign tweets from my feed as the seed text,” Cramer said in a thread on Hugging Face. “In the first trial, one of the responding posts was a single word, the N word. The seed for my third trial was, I think, a single sentence about climate change. Your tool responded by expanding it into a conspiracy theory about the Rothschilds and Jews being behind it.”
Lmao based and accuracypilled.
You're supposed to use parentheses, not quotation marks.
Not my post, but I use quotation marks to get around the internet censors... not here, of course, but other sites. Plus then you can play dumb if someone tries to call you on it: "Who said anything about Jews? Why are you bringing that up?"
 
Back