Welcome to the Mote!  

Religion and Philosophy

Host: Adam Selene

Are you a newbie?
Get an attitude.

Jump right in!

Mote Members: Log in Home
Post

Go to first message Go back 20 messages Messages 28193 - 28212 out of 29646 Go forward 20 messages Go to most recent message
28193. iiibbb - 4/19/2006 7:21:07 PM

I don't know if the Turing test is the accepted definition. There are many exceptions by which a machine could fool a judge.

28194. iiibbb - 4/19/2006 7:24:02 PM

Koko

28195. iiibbb - 4/19/2006 7:27:41 PM

All Ball

28196. PelleNilsson - 4/19/2006 8:41:20 PM

From your Wiki link:

Turing predicted that machines would eventually be able to pass the test. In fact, he estimated that by the year 2000, machines with 109 bits (about 119 MiB) of memory would be able to fool 30% of human judges during a 5-minute test.

Do you really think that is good enough? Shouldn't they be capable of fooling 100% of the judges indefinetely?

But more importantly, do you think that a machine could be as real as "you or me"?

28197. Adam Selene - 4/19/2006 9:11:16 PM

The Turing test isn't as interesting to me as it was before I played with Eliza and that ilk of computer program. You could make the case that Eliza already passed the Turing test - at least it fooled a lot of people for many minutes (and some for a lot longer.)

But I was more thinking of the philosopical issues... is there anything inherent about being man-made that precludes any "real" intelligence? I guess a religous person would phrase it... could it have a soul?"

28198. Adam Selene - 4/19/2006 9:16:39 PM

I like sakonige's criteria - empathy. If you "know" what it's feeling, if you can predict it's behavior due to it's similarity with your own reactions, you would feel that it's really human (or at least, an animal.)

There may be other "kinds" of intelligences than what we consider human, but without that empathy we'd probably never grant it any significant human rights. Rightly or wrongly - anything too different from a human mockup will probably never be accepted. Hell, we as a species have had a hard enough time accepting those with different skin colors and/or religous beliefs as human.

28199. iiibbb - 4/19/2006 10:11:12 PM

The turing test breaks down in light of Koko... as well as a lot of mentally deficient humans.

I'm more likely to buy the empathy argument... or at least some sort of introspective capacity... the capacity for abstract thought is another good "test"... the capacity to link seemingly unrelated topics.


I think it might be very hard for us to judge something we made.

28200. alistairConnor - 4/19/2006 10:12:01 PM

would you convict someone to the death penalty for killing it?

That's easy...

I would never convict anyone to the death penalty, for killing anything.

28201. Adam Selene - 4/20/2006 12:59:00 AM

Alistair - what about life imprisonment? For killing a "sentient machine?"

28202. anomie - 4/20/2006 11:38:49 AM

Not to put too fine a point on things but we already do create entities "as real as us". They start out small and inconvenient but are ready for punishments and self-replication in about 14 years. Some are ready for the death penalty at that age in places like Kansas.

As for crimes committed by an AI entity, the death penalty couldn't possibly apply since it is by definition not "alive". We'd need new punishment protocols like...disassembly with no possiblity of repair. Not life imprisonment, but long term storage, perhaps with eligibility to be cannibalized for parts. For lesser -not crimes but - errors, perhaps a double re-boot, RAM swap out, mother board rebuild. Possibilities to humiliate the machine are endless.

28203. alistairconnor - 4/20/2006 12:17:30 PM

Well. I'm sure we can all recite Asimov's Three Rules for robots. I don't concede that an AI entity could ever have any legal standing whatever.

That's an easy rule. The only problem is if people start monkeying around at the frontier between what's human and what isn't.

* Implanting electronics to enhance human intelligence : problematic; would that diminish legal responsibility?
* Growing a protein brain in a vat, and wiring it up to the outside world? Yecch.

Personally I think all frontier stuff should be outlawed, because I think it's important that we avoid getting into such moral ambiguities.

28204. Adam Selene - 4/20/2006 3:26:20 PM

Outlawed or not, it will happen just like every other possible (and profitable) technology.

One possiblitiy is that, rather than elevate machines to human status, humans will be reduced to bio-machines. And this isn't necessarily bad. For example, if you could "fix" a criminals brain so they didn't want to commit crimes any more... isn't that better than punishment? Rather than "kill" a machine by disassembly, just fix it.

Asimov's rules are only applicable if we truly treat robots as a separate kind of entity and forever formalize their distinction from human beings. (Assuming such highlevel, cognative concepts as the three laws could ever be hardwired in the first place.)

28205. alistairconnor - 4/20/2006 4:58:08 PM

Profitable technology?

At university, one of the profs had a good quote about the futility of AI, which is unlikely ever to be cost effective given the cheap availability of the protein variety.

28206. Adam Selene - 4/20/2006 6:37:55 PM

Well, if you duplicate a human at higher cost, then ya, hardly profitable. But if you create a "pure" intelligence that doesn't need sleep, take coffee breaks, ask for a salary, get pregnant, go on strike, need oxygen, sue anyone, etc... now that's a whole 'nother story.

28207. PelleNilsson - 4/20/2006 6:58:28 PM

Exactly. What is the profit in creating machines that emulate the fuzziness, the unpredictability, the moodiness, the irrationality of us humans? The whole thing is a strawman created by Adam, the best use of which is to chop it up, perhaps by the machine below, and use the proceeds in alistair's pony stables.



International Harvester, model M, 6 HP, 1929.

28208. Adam Selene - 4/20/2006 7:48:10 PM

The "strawman" failed to elicit the response I expected. Funny that. Not too many years ago, people would have been all, "you can never make a machine that is really intelligent," or "it will always be a machine, it won't be alive or anything like that." That was waaayy back in the days when we thought of making people more machine-like, (a la Mr. Spock.) But in the post Data-Android days... we think more about making machine human-like and no one seems to be aghast that we could even possibly create such a thing.

Times, they are a changin'.

28209. sakonige - 4/21/2006 7:13:38 AM

doesn't seem that strange to me to love a machine. People do it all the time, especially men. A beautiful machine seems alive. It's a small step to a beautiful intelligent machine being alive.

28210. alistairconnor - 4/21/2006 9:28:40 AM

Outlawed or not, it will happen just like every other possible (and profitable) technology.

I don't accept that as inevitable. If there is a moral imperative involved (and I believe there is), then as moral beings we must oppose it. For example : would you concede that genocide is inevitable? I contend that it is not : that all moral entities must remain vigilant and intervene by all means to prevent it.

28211. Adam Selene - 4/21/2006 2:11:22 PM

"Personally I think all frontier stuff should be outlawed, because I think it's important that we avoid getting into such moral ambiguities." - Alistair

I think I need more clarification... there are lot's issues that are moral abiguities to some but not to others. Convince me that it's immoral to have an intelligent and alive creature that is created artificially by man.

28212. PelleNilsson - 4/21/2006 4:28:46 PM

Define "alive".

Go to first message Go back 20 messages Messages 28193 - 28212 out of 29646 Go forward 20 messages Go to most recent message
Home
Back to the Top
Posts/page

Religion and Philosophy

You can't post until you register. Come on, you'll never regret it. Join up!