Intel's new voice ai will block (((hate speech))) instead of doing anything useful, like noise cancelling. - No it's not a meme. I pray to christ It needs the new processors to be implemented.

If you combine this with deepfakes, you streamers could use an app that automatically changes all their profanity to PG-13 words. No more Pewdiepie who screams "NIGGER", now it automatically changes into "BIGOT!" or any of the approved swear words.

Oh, wait, instead it will be the platforms using it. Like everyone on twitch will automatically not be able to say certain things and they don't figure it out, until they replay their own stream that they were even voiced over by their own voice; with the audience never being any the wiser.

Take it a step further.

If every social media platform adopts this, you can't even say on an alternative channel what you really meant to say. No more ability to even shout "free hong kong" in a stream. Just automatically replaced by shouting "I love the people republic of china!"
 
We could be training AI to help us in useful things, like, Iunno, run simulations on weather or just finding ever more optimized ways to design a CPU.

Instead, we're doing this.
Our best and brightest are finding ways to make every social media site as consoomer friendly as possible while the multiple mega universities in China with ML labs are building robots to fight the next war for them. Xi Jinping personally visits the ML/AI departments at the University of Beijing because he knows that's the future of China's security, at home and abroad.

Americans are niggercattle. They will always be niggercattle. They cannot be saved from their fate of being niggercattle.
 
N-Word: ON

Americans are niggercattle. They will always be niggercattle. They cannot be saved from their fate of being niggercattle.
Tell us how you really feel, Terry Josh.

NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER NIGGER.
20210327_082115.jpg
 
Last edited by a moderator:
10 burgerbux says cracker or cracka is allowed.
Of course, because the machine isn't developing an objective standard for racism or intolerance. It is trained by workers off the gig economy getting paid to add bitwise comparison flags to social media posts. It's some fucking brown guy, overseen by soy incarnate, looking at each post and ticking boxes to develop the profile of hatespeech.
 
As long as people still rage and scream at me over my horridly accurate nade spamming. I don’t need to say anything back.
Don't worry, the next AI will be auto-pilot for games. It will be mandatory. No longer can you tea bag, spam nades, quickscope, or team kill. All toxic gamer moments will be over. All hail the tranny empire.
 
Does this bleep/censor the wrongspeak for just the listener (so only people running this garbage will enjoy a "sterile" experience), or does this censor it for the person on the mic (so nobody can hear his danger words)?

Because if this is on the listener's end, as retarded as this is, it's a better solution than trying to ban or police hurt-my-feelings speech. Let human beings communicate as they wish unhindered, and the humanoid doppelgangers can run their naughty words filter to protect their delicate ears instead of bothering us.
It sounds like it's recipient-end, though if this does actually take off I wouldn't be surprised if it gets incorporated into larger games. Intel is certainly not going to make it free though.
 
I have to wonder if they're going to catastrophically fuck this up even more than it already is by turning the woke audience against them, by trying to employ some sort of filter that attempts to determine the race of the speaker. This would then be used to try and detect if a black person has said the nigger/nigga word and allow them to say it. because it's magically only okay for those of African descent. This would then fail, obviously, not being able to properly detect anything via lack of proper training or just the limits of current machine learning, and people would manage to trick it and say the n word regardless of race. I can't see a company like Intel managing to tard wrangle themselves enough not to try something even dumber than making this in the first place not trying this too.

Outside of this hypothetical, I just can't see trying to limit voice chat of all things with machine learning being anything more than a spergfest whack a mole. Will it automatically filter slurs from other nations/languages other than english? What if they overlap with regular words in english? When people start making new 'offensive' words to get around the filters, how far will intel go to retrain the ai to try and stamp those out too? I can just see this kind of program going ham, filtering random shit that vaguely sounds the same as faggot, gay, homo, and so on. This whole development is going to be a massive waste of time on Intel's part. Is there something I'm missing and the cyberpunk dystopia of sapient ai filtering the world has somehow come to pass?
 
E8E4524D-6EC6-46D7-94AD-59028012AE6C.jpeg

Null included the first part of the list, but at the very bottom is a “white nationalism” censor list.

How the fuck is “white nationalism” able to be AI censored? If you say “Andrew Anglin” do you get sniped from your gamer chair?
 
View attachment 2033476
Null included the first part of the list, but at the very bottom is a “white nationalism” censor list.

How the fuck is “white nationalism” able to be AI censored? If you say “Andrew Anglin” do you get sniped from your gamer chair?
Intel is flying way too close to the sun with this. Machine learning can understand basic shit like filtering non-voice related noises from audio, and can go as far as replicating someone's voice with enough data, but both of those things are a far cry from being able to accurately detect every word someone says through a shitty mic, and then being able to determine which statements or words are offensive from a list of them, and actively filter those out from reaching the precious sensitive ears of whomever is on the other side of the chat. I'd have to imagine shit like Alexa and Google Assistant have been used to gather a lot of data in an attempt to develop a system like this, but I can't see it coming to fruition yet. But we do know they're trying.
 
Intel is flying way too close to the sun with this. Machine learning can understand basic shit like filtering non-voice related noises from audio, and can go as far as replicating someone's voice with enough data, but both of those things are a far cry from being able to accurately detect every word someone says through a shitty mic, and then being able to determine which statements or words are offensive from a list of them, and actively filter those out from reaching the precious sensitive ears of whomever is on the other side of the chat. I'd have to imagine shit like Alexa and Google Assistant have been used to gather a lot of data in an attempt to develop a system like this, but I can't see it coming to fruition yet. But we do know they're trying.
machine learning ideology checker is some fucking insane robo cop demolition man shit.

Censoring a banned list of vulgar or obscene words isn’t a terrible idea... But AI ideology is dangerously retarded
 
machine learning ideology checker is some fucking insane robo cop shit.

Censoring a banned list of vulgar or obscene words isn’t a terrible idea... But AI ideology is dangerously retarded
Attempting to use AI to censor any language is stupid as hell and only employs people to invent new slurs. What's the point of it anyway? Most games have muting functions, and all voice chat services do at this point. If someone is offending you, just mute them. It's that easy. Companies do not need to develop tools to take the agency of clicking a button away.
 
Attempting to use AI to censor any language is stupid as hell and only employs people to invent new slurs. What's the point of it anyway? Most games have muting functions, and all voice chat services do at this point. If someone is offending you, just mute them. It's that easy. Companies do not need to develop tools to take the agency of clicking a button away.
I was thinking of children playing games, although, children shouldnt really be playing online with chat functions to begin with.
 
View attachment 2033348

I just want to die. There's no hope anymore.
Note, images that depict rape, people hung from trees, torture, child molestation, will still be allowed. That's what makes me crack up about all this "censorship of hate speech".... images depicting all sorts of horrendous torture of human beings is still allowed online, "because kink-positivity!" It's hypocritical to me.
 
Back