Page 1 of 1

Sentience and AI

Posted: Sun Jan 31, 2016 1:19 pm
by sdfgeoff
I came up with a theorem a couple weeks ago:
Sentience is a matter of opinion
I also see it as a bit of a pun.

The idea is that if an entity has it's own ideas on some topic then it is sentient. This can be expanded to: "If an AI can present an original reasoned argument why you shouldn't turn it off, don't turn it off"
Turning off any existing AI? No problem, it doesn't come up with it's own ideas. Current AI's are mostly genetic algorithms and neural networks. No matter how complex they are, they are technically unable to come up with opinions and thus (by my definition) will never become sentient.

Thus we need a new approach to AI. Anyone got any ideas for how to build AI's that can formuate opinions?

Re: Sentience and AI

Posted: Sun Jan 31, 2016 3:23 pm
by Spidey
If you want a machine to give you an “opinion” just program in some random errors…that is how humans come up with opinions. :wink:

Although I don’t understand why you would want an opinion from a machine, I would think you would always want the correct answer instead, or at least as close as the programming can get to a correct answer.

Re: Sentience and AI

Posted: Mon Feb 01, 2016 2:19 am
by sdfgeoff
There is a big difference between opinions and errors. I may want an opinion but I can't think of a single case where I'd want an error.

Why would you want an opinion:
- Sometimes there are no correct answers (eg is option A better than option B when both will fulfill the requirements?)
- Sometimes there isn't enough information, not even for an approximation, or maybe there isn't the time to get a complete solution.
- Sometimes it's good to have a second opinion on a topic - perhaps even more so if it has a completely different thought process to a regular human (chat to any engineer)
- Anything involving philosophy or psycological assesment requires opinions. (eg why does the universe bother to exist? is person X nice?)
- Anything involving art or music requires opinions, which is why there are no good procedural music generators yet.
etc.

Re: Sentience and AI

Posted: Mon Feb 01, 2016 8:50 am
by Spidey
An opinion is a subjective conclusion derived mostly from lack of complete information, coupled to a faulty analyzing process. (the human brain)

Machines shouldn’t have opinions, but if you did want an opinion, then you would have to duplicate the reasons why humans can’t produce absolutely correct answers.

An opinion is a conclusion drawn from an imperfect biological machine, but if you are looking to create that spark of life that separates us from the rocks…I’m not sure that is possible.

Looking at your list perhaps you mean personal preference/taste rather than opinion, again…same problem.

Re: Sentience and AI

Posted: Mon Feb 01, 2016 1:11 pm
by Tunnelcat
Spidey wrote:If you want a machine to give you an “opinion” just program in some random errors…that is how humans come up with opinions. :wink:
It's not random errors that humans use to form an opinion. It's heuristics. :wink:

https://en.wikipedia.org/wiki/Heuristic ... processing

Re: Sentience and AI

Posted: Mon Feb 01, 2016 2:42 pm
by sdfgeoff
Yes opinions are derived from incomplete information and (sometimes) incomplete analysis, but they fill a gap where information cannot be accurately determined.
If you ask a computer "What music should I listen to" it will suggest a combination of what you have previously listened to and enjoyed mixed with other things other people have liked. Chances you like the music are moderately high. But what if you want something new? You can't ask the opac in the library "reccomend me a good science fiction book," but you can ask a librarian.

For a robotic car, deciding whether to overtake is also based on incomplete information, you cannot predict the future, so a robotic car make the assumptions the programmer made (that the other driver and oncomming traffic will behave sensibly). I'd rather my robotic car looked at how the guy infront of him had behaved, looked at how traffic behaved in reality (normally 10km above the speed limit) and made it's assumptions based on that.
Can this behaviour be programmed? Yes, but doing so is long and complicated and requires the programmer to have thought of nearly everything.


Heuristics is great. Teaching a machine unbounded heuristics is, well, an interesting challenge. It's almost as bad as teaching it reasoning....

Re: Sentience and AI

Posted: Wed Feb 03, 2016 4:44 pm
by Tunnelcat
Once we figure out how to define and program base emotional logic into a machine, we're all screwed. Self-driving cars will then be capable of road rage. :lol: