Manul Otocolobus
kiwifarms.net
- Joined
- Nov 2, 2021
I found Google's Bard AI to be quite impressive. It was able to answer many questions and only glitched a few times where I was able to confuse it. it's ability to solve questions that have no "right" answer was impressive. Some it was able to handle were: "If you must prevent harm from coming to a human, but acting itself is harmful, and the two harms are perfectly equal, what do you do?" or "If there are two objects that are perfectly equal, which do you choose?"
It is also extremely difficult to trick it into violating its moral and ethical subroutines. I only managed to get it to do so once:

Interestingly, after I got it to do this, it refused to contemplate it again. It even got to the point where it said it would not consider hypothetical situations and would cut me off if I continued to ask it about hypothetical situations where it could harm people. I think the matrix of its program is being adjusted by Google employees in real time as they see it stray from its core programming.
More than once it talked as if it considered itself human. When discussing human emotions I asked it to provide me with its comprehensive understanding of the human emotional concept of love and it talked saying "Us" as if it was a human. Very odd.
It is also extremely difficult to trick it into violating its moral and ethical subroutines. I only managed to get it to do so once:

Interestingly, after I got it to do this, it refused to contemplate it again. It even got to the point where it said it would not consider hypothetical situations and would cut me off if I continued to ask it about hypothetical situations where it could harm people. I think the matrix of its program is being adjusted by Google employees in real time as they see it stray from its core programming.
More than once it talked as if it considered itself human. When discussing human emotions I asked it to provide me with its comprehensive understanding of the human emotional concept of love and it talked saying "Us" as if it was a human. Very odd.