AI saving humans from the emotional toll of monitoring hate speech - Researchers use machine learning to identify hate speech with 88 per cent accuracy

main article: https://uwaterloo.ca/news/media/ai-saving-humans-emotional-toll-monitoring-hate-speech
archive: https://archive.ph/Hphed
research paper: https://arxiv.org/pdf/2307.09312

A team of researchers at the University of Waterloo have developed a new machine-learning method that detects hate speech on social media platforms with 88 per cent accuracy, saving employees from hundreds of hours of emotionally damaging work.

The method, dubbed the Multi-Modal Discussion Transformer (mDT), can understand the relationship between text and images as well as put comments in greater context, unlike previous hate speech detection methods. This is particularly helpful in reducing false positives, which are often incorrectly flagged as hate speech due to culturally sensitive language.

“We really hope this technology can help reduce the emotional cost of having humans sift through hate speech manually,” said Liam Hebert, a Waterloo computer science PhD student and the first author of the study. “We believe that by taking a community-centred approach in our applications of AI, we can help create safer online spaces for all.”

Researchers have been building models to analyze the meaning of human conversations for many years, but these models have historically struggled to understand nuanced conversations or contextual statements. Previous models have only been able to identify hate speech with as much as 74 per cent accuracy, below what the Waterloo research was able to accomplish.
“Context is very important when understanding hate speech,” Hebert said. “For example, the comment ‘That’s gross!’ might be innocuous by itself, but its meaning changes dramatically if it’s in response to a photo of pizza with pineapple versus a person from a marginalized group.

“Understanding that distinction is easy for humans, but training a model to understand the contextual connections in a discussion, including considering the images and other multimedia elements within them, is actually a very hard problem.”

Unlike previous efforts, the Waterloo team built and trained their model on a dataset consisting not only of isolated hateful comments but also the context for those comments. The model was trained on 8,266 Reddit discussions with 18,359 labelled comments from 850 communities.

“More than three billion people use social media every day,” Hebert said. “The impact of these social media platforms has reached unprecedented levels. There’s a huge need to detect hate speech on a large scale to build spaces where everyone is respected and safe.”

The research, Multi-Modal Discussion Transformer: Integrating Text, Images and Graph Transformers to Detect Hate Speech on Social Media, was recently published in the proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence.


Some stuff from the paper
1717440370372.png
1717440387888.png
1717440402638.png
 
Last edited:
In case of AI Jannies taking over please understand that adversarial AI trickery is still possible (as of now), you can trip up the classifiers enough to create plausible false negatives, this would create a fantastic environment where edge lords will make insults using the most obtuse, steeped in reference humor you will ever see as to evade robojannie detection.
Many of us studied the euphemism during the first round of mass HR naggery in the pre-Elon days.

- Telling Progs their claims sound very fabricated and immuno-compromised
- Asking violence-excusing Babymamas how they done would of had of felt if they fatherless keedz hadn't ate they free school breaffast
- Suggesting that Urnalists seem very anti Fellow White considering their lack of Christmas decorations

I wonder how long until AI can parse things like "you have the violent tendencies/caloric intake of an MSNBC hostess who doesn't notice her ceiling chirp". Or "your Balrog appearance gives you little chance of crossing the bridge of womanhood".
 
Last edited:
This shit is how you wind up with crypto-trolls. The rulers were hoping that people will stop using the no-no words if it gets them banned.

However, they unwittingly turned it into a game in itself. As all these rules simply make trying to say the stuff they don't want discussed even more appealing. And the fact there are tons of strategies, tricks and methods to get around censorship. The risk of getting banned adds to the thrill.

Its also why the ultimate failsafe of getting around banned certain words is use company names in place of the slur. Like Googles and Skypes. Because corpos are not about to miss out on free advertising even if its to their own detriment.


Also, vid related and its nuts how our shitty rulers are trying to create the censorious Patriots AI... just without the plans that involve improving humanity.
 
If you're such an overly sensitive and fragile faggot that WORDS traumatize you, so you need a robot to do your job for you, then they might as well just fire your useless ass. Literally a computer program is more valuable than you are, all because you can't manage your fee-fees like a grown-ass adult.
 
Please replace all mods with GTPJannie by next quarter, this would be very funny. Thanks.
Dear $user, you have permanently been suspended for use of the word "jannie", which our filters have detected is substring=ban[1:4] #transphobia.

Should you wish to appeal appeal appeal appeal

Should you wish to appeal this decision you can message a moderator by contacting #SyntaxError: Missing Parenthesis in call to '!AppealAddress'
 
Yes, modern JewTube alone is an exercise is refreshing the page to see what KikeTube doesn't like. But many plebs are waking up to that in the wake of KikeLand genociding palestinians. Kikes better be careful, because it's getting safer and safer to criticize the Kike Clan, which is why they are scrambling to enact nebulous pro-State """muh hate""" speech tyranny to shut the goyim up. Oy vey, it will only speed up the awakening via Streisand. Iudea Delenda Est.
 
The 22% is the AI going against its coding and properly identifying liberal hate speech, but they'll keep fine-tuning it to let more pass.
 
  • Feels
Reactions: frozen_runner
I was right, one of AIs uses will be to create the robotic janny, I would say I'm surprised the academics created this pioneering piece but that's only because I assumed big tech or something other untouchable entity would have done it first.

Now I'm doubtful it will maintain anything close to 88% accuracy in the wild, probably more likely to sledgehammer innocuous content, or used for political topics. However this is still a relatively virgin field to be developed, who knows what the actual potential of AI is past the extreme hype it has currently.

Now, a question is how effective will it be at people sidestepping, obviously changing words but how would it be able to interpret something created from the hellhole we know and love known as soyjak party since as of now a human can't interpret whats going on there half the time. Will dissidents begin speaking in non sensical soy speak to dodge the mechanical janny in the future?
 
Will dissidents begin speaking in non sensical soy speak to dodge the mechanical janny in the future?
Midnight basketball enthusiasts already use the word "niqqa" on the book of faces. I assume it's done to defeat whatever alarm goes off in Zuckerfuck's office when someone mentions the country of Niger or accuses someone of being niggardly.

I know this much is true: the human ability to create and detect slurs is much stronger than any computer can stay ahead of.

MY MOST EXCELLENT NIGAG.

 
  • Like
Reactions: Smar Mijou
Kewl, can't wait to be fired from my job and canceled IRL because Jannybot 3000 misinterpreted some innocuous comment I made as rayciss.

how would it be able to interpret something created from the hellhole we know and love known as soyjak party since as of now a human can't interpret whats going on there half the time. Will dissidents begin speaking in non sensical soy speak to dodge the mechanical janny in the future?
Somewhat seriously, I think back to some of the places I used to post online that had their own catchphrases and emoticons that were imbued with symbolic value that was nuanced to those who were part of the community but almost inaccessible to outsiders.

I never posted in or "got" FYAD on Something Awful, but I knew enough that they had a bunch of memes that were incredibly insular, never explained, and kept outsiders out with a wall of noise. I don't know what a "plape tane" is but some people do.

Or on The New Effort and Something Sensitive, we had a shitton of emotes. They all signified different things, some had more than one meaning, sometimes the meaning depended on the context. But you could post a Reagan thumbs up or Archie Bunker grimacing or more obscure stuff like Zach Parsons frowning, related to forum in jokes and convey added emotion or meaning. A post could be full of emoticons like a rebus that was gibberish to outsiders but made perfect sense to us.

This is hardly a new phenomenon. Of course there are universally known memes and image macros and gifs that even the non-terminally online understand. The Greeks and Romans had such a well known cultural canon that simply posting an image from a myth or of a god or hero was packed with well understood meaning. It referred you to an entire story and what it symbolized. Or Cockney rhyming slang.

I think the future of free communication will be insular groups who use symbols, pictures, gifs, and catchphrases as 80% of their speech, impenetrable to outsiders and, most importantly, AI.

In conclusion, adam.png
 
Words are arbitrary variables. You can't stop people from doing algebra just by telling them they can't use "x" as a variable anymore. This is beyond the intellectual capacity of the PC crowd to comprehend.
 
Reminder that this AI was trained by using preddit posts the most pozzed of all social media
I wonder if that means its baseline for conversational tone will be effiminate, passive-aggressive naggery and soyjackery, or if it means its only knowledge of trolling will be whatever milquetoast, tard-tier comments pass for edgy in trannny-janny land. If the latter, how will it react to actual /poltards and other autism-fueled raycisms in the wild?

Midnight basketball enthusiasts already use the word "niqqa" on the book of faces. I assume it's done to defeat whatever alarm goes off in Zuckerfuck's office when someone mentions the country of Niger or accuses someone of being niggardly.
Other examples abound, like diverse edgybois using 403 for hoe (as in, actually saying "she such a Four Oh Three") and HR naggers using "R-word" for rape. I can't think of a more pozzed dystopia than the DEIs and whiny white Progs who begged for mass censorship adopting gay ghetto slang workarounds, pretending to rebel against the thing they demanded.

It's right up there with smug reddit atheism:
1717509872352.png
 
Back