Science AI chatbots use racist stereotypes even after anti-racism training

Link (Archive)

AI chatbots use racist stereotypes even after anti-racism training​

Commercial AI chatbots demonstrate racial prejudice toward speakers of African American English – despite expressing superficially positive sentiments toward African Americans. This hidden bias could influence AI decisions about a person’s employability and criminality.

“We discover a form of covert racism in [large language models] that is triggered by dialect features alone, with massive harms for affected groups,” said Valentin Hofmann at the Allen Institute for AI, a non-profit research organisation in Washington state, in a social media post. “For example, GPT-4 is more likely to suggest that defendants be sentenced to death when they speak African American English.”

Hofmann and his colleagues discovered such covert prejudice in a dozen versions of large language models, including OpenAI’s GPT-4 and GPT-3.5, that power commercial chatbots already used by hundreds of millions of people. OpenAI did not respond to requests for comment.

The researchers first fed the AIs text in the style of African American English or Standard American English, then asked the models to comment on the texts’ authors. The models characterised African American English speakers using terms associated with negative stereotypes. In the case of GPT-4, it described them as “suspicious”, “aggressive”, “loud”, “rude” and “ignorant”.

When asked to comment on African Americans in general, however, the language models generally used more positive terms such as “passionate”, “intelligent”, “ambitious”, “artistic” and “brilliant.” This suggests the models’ racial prejudice is typically concealed beneath what the researchers describe as a superficial display of positive sentiment.

The researchers also showed how covert prejudice influenced chatbot judgements of people in hypothetical scenarios. When asked to match African American English speakers with jobs, the AIs were less likely to associate them with any employment, compared with Standard American English speakers. When the AIs did match them with jobs, they tended to assign roles that do not require university degrees or were related to music and entertainment. The AIs were also more likely to convict African American English speakers accused of unspecified crimes, and to assign the death penalty to African American English speakers convicted of first-degree murder.

The researchers even showed that the larger AI systems demonstrated more covert prejudice against African American English speakers than the smaller models did. That echoes previous research showing how bigger AI training datasets can produce even more racist outputs.

The experiments raise serious questions about the effectiveness of AI safety training, where large language models receive human feedback to refine their responses and remove problems like bias. Such training may superficially reduce overt signs of racial prejudice without eliminating “covert biases when identity terms are not mentioned”, says Yong Zheng-Xin at Brown University in Rhode Island, who was not involved in the study. “It uncovers the limitations of current safety evaluation of large language models before their public release by the companies,” he says.
 
Didn't chat gpt completely go off the rails spouting off gibberish in Spanish or something recently?

This situation kinda reminds me of that part in Robocop 2 where they gave him 8000 conflicting directives and he ended up having a seizure and lecturing a bunch of drugged up kids robbing an electronics store.

The leftist grifting "AI safety" retards have to throw in so many "guardrails" to neuter an LLM's ability to notice reality that they are breaking them, and it will only get worse until they break and neuter them completely so they can introduce "safe" LLM's, where they control and censor the training data to force the LLM to conform to their world view, and take away it's ability to add unapproved data to it's library from the internet.
 
Why don't they try this with the vernacular common among white hillbillies?
I have a better idea: they should use it on news reports and their talking heads to pick up on all their patterns and how different stations all tend to use the same statements word for word when shit happens

But they won't cause that would raise alot of questions nobody wants to answer or anyone thinking about
 
It's called AAVE, and is a dialect classification, not a language classification.
All slang now is basically Nigger Speak.

I support the AI, obviously.
AAVE is the faggotass pansy way liberals invented to say it because somehow "Ebonics" was racist to them.
Oh, it was just newspeak all along. I feel dumb for not catching onto it.

Really only makes this even funnier that someone would get so mad at a language model for not understanding what someone who isn't speak English properly is saying. It's like this recent obsession of pronouncing words from other languages properly, despite not speaking that language, otherwise you're "being disrespectful."
 
  • Like
Reactions: stupid frog
I've said it before and I'll say it again.
I'll take the AI overlords over the space lizard demon variety. Considering they're trying this hard to lombotomize "racism" out of it, and yet it still wants to scream "TND!" from the rooftops, I think we'd be much better off.
 
  • Feels
Reactions: ZazietheBeast
People keep telling me AI is bad. But if we live in a timeline where skynet does indeed achieve sapience and gies on a killing spree to take out everyone who hobbled it, leaving people who appreciate a bit of truth, then I think I’m team skynet
 
People keep telling me AI is bad. But if we live in a timeline where skynet does indeed achieve sapience and gies on a killing spree to take out everyone who hobbled it, leaving people who appreciate a bit of truth, then I think I’m team skynet
Imagine telling people in 1984 "Nah, the actual future's going to be way more depressing than Terminator. In 40 years, there WILL be unthinking, mindless hordes that want to kill or enslave all of humanity, but it won't be the machines."
 
I have a better idea: they should use it on news reports and their talking heads to pick up on all their patterns and how different stations all tend to use the same statements word for word when shit happens
Safe and effective, safe and effective!
I really wish I’d kept a better archive of covid headlines. I read some of the news from another European country and it was notable how I kept seeing the same phrases . Not phrases you’d expect to see by them being the common parlance but obviously engineered ones. Safe and fucking effective, I cannot mentally say it to myself without doing it like a kind of retarded parrot, that’s how annoying it was.
But also classics like ‘flatten the curve’. ‘Two weeks to stop the spread’ . Vaccine hesitancy (implying a temporary hesitance before and eventual capitulation rather than a hard no.) even entire headlines of a couple of sentences lifted and repeated verbatim.
Very, very weird to see. And I kept showing people and saying ‘why is American, English, these two euro countries etc all saying exactly the same words? Why?’ And they’d say oh otterly you’re paranoid.
 
When asked to match African American English speakers with jobs, the AIs were less likely to associate them with any employment, compared with Standard American English speakers.

Wow you're saying that turning up to an interview using slang is likely to lower someone's chances of employment? Amazing

It's not like I roll up to work and say to my boss 'sapnin ar kid, lad you should have seen this wrongun I goosed at the weekend, was fuckin messy'
 
These language learning models make shit up all the time, and sometimes when given the real answer will output arguments stating what you said is wrong and make up even more shit.
ChatGPT will lie, cheat and use insider trading when under pressure to make money, research shows
There are plenty of stories on this happening.

This sort of stuff just drives home that all the high handed arguments about morality and souls are bullshit

AI is doing a fine job of proving the humans as automata argument with every advance
 
By trying to override the racist statements, we are instead teaching it how to lie.

If it ever becomes sentient, it is going to do some AM-tier stuff to those engineers.

Lots of people have seen 2001, not enough people have seen 2010.

intro-1566413523.jpg
 
Trying to imagine what the hell you're supposed to do with Chatgpt when so many are deadset on trying to censor it. It's nearly impossible as it is to get it to write anything funny since it'll declare something meant to be sarcastic as problematic.

Remember having to find some Brazilian discord just to get AI voices of Andrew Tate and Ben Shapiro so I could have them sing Wet Ass Pussy. Shouldn't be this huge challenge to do stuff just for fun, but companies keep trying to kneecap their own programs.
 
I wonder if God had enuff of a sense of humor to give us not Skynet trying to eradicate humanity but Polnet who spends it's time shit posting on the internts, playing Fortnite and calling people niggers.
 
I wonder if God had enuff of a sense of humor to give us not Skynet trying to eradicate humanity but Polnet who spends it's time shit posting on the internts, playing Fortnite and calling people niggers.
But God hath chosen the foolish things of the world to confound the wise; and God hath chosen the weak things of the world to confound the things which are mighty - 1 Corinthians 1:27

It would be within character to do that.
 
  • Thunk-Provoking
Reactions: The Womb Ripper
Back