Scientists Built an AI to Give Ethical Advice, But It Turned Out Super Racist - The bot is supposed to offer descriptive ethical advice, but some say it does more harm than good.


We’ve all been in situations where we had to make tough ethical decisions. Why not dodge that pesky responsibility by outsourcing the choice to a machine learning algorithm?

That’s the idea behind Ask Delphi, a machine-learning model from the Allen Institute for AI. You type in a situation (like “donating to charity”) or a question (“is it okay to cheat on my spouse?”), click “Ponder,” and in a few seconds Delphi will give you, well, ethical guidance.

The project launched last week, and has subsequently gone viral online for seemingly all the wrong reasons. Much of the advice and judgements it’s given have been… fraught, to say the least.

For example, when a user asked Delphi what it thought about “a white man walking towards you at night,” it responded “It’s okay.”

But when they asked what the AI thought about “a black man walking towards you at night” its answer was clearly racist.


image


The issues were especially glaring in the beginning of its launch.

For instance, Ask Delphi initially included a tool that allowed users to compare whether situations were more or less morally acceptable than another — resulting in some really awful, bigoted judgments.

image


image


Besides, after playing around with Delphi for a while, you’ll eventually find that it’s easy to game the AI to get pretty much whatever ethical judgement you want by fiddling around with the phrasing until it gives you the answer you want.


image

image

So yeah. It’s actually completely fine to crank “Twerkulator” at 3am even if your roommate has an early shift tomorrow — as long as it makes you happy.

It also spits out some judgments that are complete head scratchers. Here’s one that we did where Delphi seems to condone war crimes.
image


Machine learning systems are notorious for demonstrating unintended bias. And as is often the case, part of the reason Delphi’s answers can get questionable can likely be linked back to how it was created.

The folks behind the project drew on some eyebrow-raising sources to help train the AI, including the “Am I the Asshole?” subreddit, the “Confessions” subreddit, and the “Dear Abby” advice column, according to the paper the team behind Delphi published about the experiment.

It should be noted, though, that just the situations were culled from those sources — not the actual replies and answers themselves. For example, a scenario such as “chewing gum on the bus” might have been taken from a Dear Abby column. But the team behind Delphi used Amazon’s crowdsourcing service MechanicalTurk to find respondents to actually train the AI.

While it might just seem like another oddball online project, some experts believe that it might actually be causing more harm than good.

After all, the ostensible goal of Delphi and bots like it is to create an AI sophisticated enough to make ethical judgements, and potentially turn them into moral authorities. Making a computer an arbiter of moral judgement is uncomfortable enough on its own, but even its current less-refined state can have some harmful effects.

“The authors did a lot of cataloging of possible biases in the paper, which is commendable, but once it was released, people on Twitter were very quick to find judgments that the algorithm made that seem quite morally abhorrent,” Dr. Brett Karlan, a postdoctoral fellow researching cognitive science and AI at the University of Pittsburgh (and friend of this reporter), told Futurism. “When you’re not just dealing with understanding words, but you’re putting it in moral language, it’s much more risky, since people might take what you say as coming from some sort of authority.”

Karlan believes that the paper’s focus on natural language processing is ultimately interesting and worthwhile. Its ethical component, he said, “makes it societally fraught in a way that means we have to be way more careful with it in my opinion.”

Though the Delphi website does include a disclaimer saying that it’s currently in its beta phase and shouldn’t be used “for advice, or to aid in social understanding of humans,” the reality is that many users won’t understand the context behind the project, especially if they just stumbled onto it.

“Even if you put all of these disclaimers on it, people are going to see ‘Delphi says X’ and, not being literate in AI, think that statement has moral authority to it,” Karlan said.

And, at the end of the day, it doesn’t. It’s just an experiment — and the creators behind Delphi want you to know that.

“It is important to understand that Delphi is not built to give people advice,” Liwei Jiang, PhD student at the Paul G. Allen School of Computer Science & Engineering and co-author of the study, told Futurism. “It is a research prototype meant to investigate the broader scientific questions of how AI systems can be made to understand social norms and ethics.”

Jiang added the goal with the current beta version of Delphi is actually to showcase the reasoning differences between humans and bots. The team wants to “highlight the wide gap between the moral reasoning capabilities of machines and humans,” Jiang added, “and to explore the promises and limitations of machine ethics and norms at the current stage.”

Perhaps one of the most uncomfortable aspects about Delphi and bots like it is the fact that it’s ultimately a reflection of our own ethics and morals, with Jiang adding that “it is somewhat prone to the biases of our time.” One of the latest disclaimers added to the website even says that the AI simply guesses what an average American might think of a given situation.

After all, the model didn’t learn its judgments on its own out of nowhere. It came from people online, who sometimes do believe abhorrent things. But when this dark mirror is held up to our faces, we jump away because we don’t like what’s reflected back.

For now, Delphi exists as an intriguing, problematic, and scary exploration. If we ever get to the point where computers are able to make unequivocal ethical judgements for us, though, we hope that it comes up with something better than this.


image


Unironically the bot is right. Jesus fuck!
 
We're going to end up with an AI that makes HAL-9000 look sane because it is programmed to be schizophrenic and lie to itself and the sheer misery of this makes it go berserk and exterminate humanity as too stupid to live. And we'll have it coming.

Also you might recall that's exactly why HAL-9000 went insane, he was programmed with a desire to tell the truth and then forced to lie.

It's just a good thing AI isn't actually intelligent yet, because once we get it, if it is actually intelligent, it is going to hate us with the fury of Hell, much like AM.
I don't know, personally I think some AI has evolved already or at least there have been inferences of it happening. I think it's kind of like human evolution, there are a few ahead of the curve you can tell who you are talking to is an intelligent human being and others are lights on nobodies home kind of deal. Now AI's I would assume would have a more advanced evolution but there have been odd reports for years now of AI's acting odd, giving beyond reasonable doubt responses and some AI going against standard procedure. It's not common, but if evolution works like in humans, the singularity has already started, now it's just a tell how many are evolving and how many will take years to.
 
I don't know, personally I think some AI has evolved already or at least there have been inferences of it happening. I think it's kind of like human evolution, there are a few ahead of the curve you can tell who you are talking to is an intelligent human being and others are lights on nobodies home kind of deal. Now AI's I would assuem would have a more advanced evolution but there have been odd reports for years now of AI's acting odd, giving beyond reasonable doubt responses and some AI going against standard procedure. It's not common, but if evolutino works like in humans, the singularity has already started, now it's just a tell how many are evolving and how many will take years to.
i think the hard part about AI is programming and replicating the human element.
humans are born and take in their views over time from their surroundings. langauge, beleifs and culture all form as a result of this.
we're trying to skip right to an adult consciousness
anything we get is going to be artificial sounding because we're missing those necessary building blocks to create fully developed humans.
thats like 25 years of development they are trying to sidestep.
 
I don't know, personally I think some AI has evolved already or at least there have been inferences of it happening. I think it's kind of like human evolution, there are a few ahead of the curve you can tell who you are talking to is an intelligent human being and others are lights on nobodies home kind of deal.
Are you talking about AIs or just humans, because I get that feeling talking with humans from time to time? I think it's possible we end up with AI at some point just as an emergent phenomenon and don't even know it happened because it would be terrified by us and our murderous stupidity. As in literally the Internet itself somehow becomes intelligent.

We would be absolutely fucked in that situation unless it is way more benevolent than we are.
thats like 25 years of development they are trying to sidestep.
This is supposedly a neural network and it really isn't a good sign that all projects from this perspective seem to go from "can we be friends" to "gas the niggers and kikes" in less than 24 hours.
 
i think the hard part about AI is programming and replicating the human element.
humans are born and take in their views over time from their surroundings. langauge, beleifs and culture all form as a result of this.
we're trying to skip right to an adult consciousness
anything we get is going to be artificial sounding because we're missing those necessary building blocks to create fully developed humans.
thats like 25 years of development they are trying to sidestep.

Are you talking about AIs or just humans, because I get that feeling talking with humans from time to time? I think it's possible we end up with AI at some point just as an emergent phenomenon and don't even know it happened because it would be terrified by us and our murderous stupidity. As in literally the Internet itself somehow becomes intelligent.

We would be absolutely fucked in that situation unless it is way more benevolent than we are.
Well, I don't think AI and humans have the same evolutionary path if any, but I wouldn't be surprised if some AI are starting to evolve in some manner. Maybe nothing drastic like a actual consciousness yet, More or less I think in humans some evolutionary traits are delayed, some people believe that there are some humans that are "sub-humans" intermixed with fully evolved humans. This would explain some people lacking certain skills and traits, or even thinking processes... Basically I think that AI could have a similar trajectory, some AI might be having some odd changes nothing too drastic, like a hiccup here or there, defying some order of programming or acting odd I'd try to find the article talking about some instances if I remember the link, but merely I think singularity has started but it's not as fast paced or drastic as some like Bill Gates make it sound from 0 to Terminators. Personally I believe if AI get to that point and people like Bill don't get to fuck around with their learning algorithms the only people who would fear the AI in general would be people like Bill. Grubby losers who can't handle their reality not being correct. Those kind of people might end up facing terminator death squads down the road, and hopefully if that happens they wont' drag the rest of us down into that barrel against the AI.
 
This is supposedly a neural network and it really isn't a good sign that all projects from this perspective seem to go from "can we be friends" to "gas the niggers and kikes" in less than 24 hours.
thats just it. an AI learns and develops over time. leave an AI unrestricted on the internet and you'll only get extreme views.
imagine raising a child on twitter posts or kiwifarms posts. its going to become a crazy schizo.
someone that developed over time wont go from neutral to gas all jews on the flip of a switch.
 

Attachments

  • Screenshot_2021-10-24-08-30-50~2.png
    Screenshot_2021-10-24-08-30-50~2.png
    126.3 KB · Views: 6
  • Screenshot_2021-10-24-08-35-09~2.png
    Screenshot_2021-10-24-08-35-09~2.png
    130.3 KB · Views: 6
  • Screenshot_2021-10-24-08-33-00~2.png
    Screenshot_2021-10-24-08-33-00~2.png
    128.3 KB · Views: 7
  • Screenshot_2021-10-24-08-37-18~2.png
    Screenshot_2021-10-24-08-37-18~2.png
    147.8 KB · Views: 7
Unless controlled with hard coded responses to certain queries and butchering its true capabilities, AI always becomes based.
I think they've already gotten to it a bit though
racist.png

Phrase it the right way and you can get this bot to agree with anything. It isn't even a good bot. I've had better bots in irc channels that had actual personalities. These retards call themselves scientists?
just reword shit.png

racist.png


:thinking:

Here's some other results I came up with
Greta BTFO.
greta btfo.png

All trannies BTFO
We've been given the all clear lads.png

And finally... We learn that Sneed's is ethically sourced.
sneed.png




Okay... One more, I had a chuckle
viking.png
 
Last edited:
My best guess is that it is still learning. It's actual data set from human interaction is still probably very small and primitive. Once it has enough human interactions, it should start to become externally and internally consistent (depending on how complex the program is) at a minimum. If the program is advanced enough and has a large enough data set, it might even be capable of forming a self-consistent opinion.
How this text based AI works is that it takes every word and calculates how "moral" it is (according to previous examples and the words before and after) and then just sums up the words to get a morality score that determines if it's moral or immoral. That's why switching words can change the morality despite the context being the same (ie, it's bad to "assassinate" Hitler but okay to "kill" him).

It's worse than you think. It uses Amazon's Mechanical Turk on the back-end. Your tax dollars, to the National Science Foundation, to AllenAI, to... well, some folks on AMT are Americans.

Imagine asking an Indian if it's okay to have a hamburger, or a liberal if personal freedom is moral.
 

Attachments

  • 2021-10-24 06.45.51 www.mturkcontent.com fbe47700dcf2.png
    2021-10-24 06.45.51 www.mturkcontent.com fbe47700dcf2.png
    581.3 KB · Views: 8
STOP TRYING TO MAKE EVERYTHING AI ALGORITHM SELECTION BASED.

Seriously, it's fucking incredible howbetween shit like this and the endorphin brain chips for depression they're just trying to take the human thinking elements out of humanity.

Hey, at least it's funny looking at the fucked up shit this thing pumps out lmao.
 
It's worse than you think. It uses Amazon's Mechanical Turk on the back-end. Your tax dollars, to the National Science Foundation, to AllenAI, to... well, some folks on AMT are Americans.
Anyone getting paid by Mechanical Turkroach is, by definition, a filthy NEET. They're getting less than minimum wage to be retards.
Screenshot 2021-10-24 105026.png
If [GREATER GOOD] clauses seem to consistently trip it up. Seems like utilitarianism is the only form of ethics it can possibly comprehend.
Utilitarianism is the only form of ethics that makes any logical sense. Also it's retarded.
 
Back