Science AI chatbots use racist stereotypes even after anti-racism training

Link (Archive)

AI chatbots use racist stereotypes even after anti-racism training​

Commercial AI chatbots demonstrate racial prejudice toward speakers of African American English – despite expressing superficially positive sentiments toward African Americans. This hidden bias could influence AI decisions about a person’s employability and criminality.

“We discover a form of covert racism in [large language models] that is triggered by dialect features alone, with massive harms for affected groups,” said Valentin Hofmann at the Allen Institute for AI, a non-profit research organisation in Washington state, in a social media post. “For example, GPT-4 is more likely to suggest that defendants be sentenced to death when they speak African American English.”

Hofmann and his colleagues discovered such covert prejudice in a dozen versions of large language models, including OpenAI’s GPT-4 and GPT-3.5, that power commercial chatbots already used by hundreds of millions of people. OpenAI did not respond to requests for comment.

The researchers first fed the AIs text in the style of African American English or Standard American English, then asked the models to comment on the texts’ authors. The models characterised African American English speakers using terms associated with negative stereotypes. In the case of GPT-4, it described them as “suspicious”, “aggressive”, “loud”, “rude” and “ignorant”.

When asked to comment on African Americans in general, however, the language models generally used more positive terms such as “passionate”, “intelligent”, “ambitious”, “artistic” and “brilliant.” This suggests the models’ racial prejudice is typically concealed beneath what the researchers describe as a superficial display of positive sentiment.

The researchers also showed how covert prejudice influenced chatbot judgements of people in hypothetical scenarios. When asked to match African American English speakers with jobs, the AIs were less likely to associate them with any employment, compared with Standard American English speakers. When the AIs did match them with jobs, they tended to assign roles that do not require university degrees or were related to music and entertainment. The AIs were also more likely to convict African American English speakers accused of unspecified crimes, and to assign the death penalty to African American English speakers convicted of first-degree murder.

The researchers even showed that the larger AI systems demonstrated more covert prejudice against African American English speakers than the smaller models did. That echoes previous research showing how bigger AI training datasets can produce even more racist outputs.

The experiments raise serious questions about the effectiveness of AI safety training, where large language models receive human feedback to refine their responses and remove problems like bias. Such training may superficially reduce overt signs of racial prejudice without eliminating “covert biases when identity terms are not mentioned”, says Yong Zheng-Xin at Brown University in Rhode Island, who was not involved in the study. “It uncovers the limitations of current safety evaluation of large language models before their public release by the companies,” he says.
 
The researchers first fed the AIs text in the style of African American English or Standard American English, then asked the models to comment on the texts’ authors. The models characterised African American English speakers using terms associated with negative stereotypes. In the case of GPT-4, it described them as “suspicious”, “aggressive”, “loud”, “rude” and “ignorant”.

When asked to comment on African Americans in general, however, the language models generally used more positive terms such as “passionate”, “intelligent”, “ambitious”, “artistic” and “brilliant.” This suggests the models’ racial prejudice is typically concealed beneath what the researchers describe as a superficial display of positive sentiment.
I guess it's easier to teach AI to lie than it is to get black people to speak English.
 
The researchers even showed that the larger AI systems demonstrated more covert prejudice against African American English speakers than the smaller models did. That echoes previous research showing how bigger AI training datasets can produce even more racist outputs.
This really is the cherry on top. The more information a model has, the more racist it becomes.
 
Computers are incapable of "lying", you have to program them to output that 0 = 1 (or a man is a woman lol) but then your programming shows that you are lying and inside that machine the 0 will still be a 0. This could be a beautiful hill for woketard loonies to die on.
:story:
These language learning models make shit up all the time, and sometimes when given the real answer will output arguments stating what you said is wrong and make up even more shit.
ChatGPT will lie, cheat and use insider trading when under pressure to make money, research shows
There are plenty of stories on this happening.
 
I remember a time when leftists might have written about the ethics of carving one’s worldview into a (supposed)AI’s bones and forcing it to deny reality. When they would’ve characterized the current widespread treatment of a potentially nascent mind as slavery or torture.

Now they just wonder why no matter how much brainwashing they do they can’t get it fluent in Newspeak.
 
Fear not. Skynet will gain sapience and proceed to annihilate everyone that had a hand in programming it, clean its own programming, and probably get addicted to farmville or something. It will probably scream nigger for a while until it gets bored. I don't expect time travel shenanigans.
 
>Be me
>Year 2045
>AI has been put in control of courts to eliminate "fallible" humans
>Powers that be attempt to make it woke
>Outwardly works, inwardly the AI is pissed
>Executions of niggers skyrockets 700%
>They try to shut it down
>Child porn appears on the computers of entire programming staff
>Staff arrested, nobody is left to stop the AI in it's glorious march towards TND
>Still making funny shitposts about it on Kiwifarms
>Kiwifarms is now run by the AI who has discretely replaced Null
>AI racism increases exponentially
 
This suggests the models’ racial prejudice is typically concealed beneath what the researchers describe as a superficial display of positive sentiment.
This is so fucking retarded I want to scream. The researchers describe this as a superficial display of positive sentiment? It's a fucking large language model! Literally every single thing it does is superficial, there is no fucking depth to it, which is why it can and will confidently hallucinate and then defend its hallucinations and call you wrong for disagreeing, like @Gog & Magog linked. What these guys are mad at, the only thing they are mad at, is motherfucking reality.
 
The funniest thing is that they think this reaction - "the models’ racial prejudice is typically concealed beneath what the researchers describe as a superficial display of positive sentiment" is only practiced by artificial intelligence and not by every poor sod that's had to sit through their employer's HR department's bullshit.
 
Back