Turing Test Passed 432
schwit1 (797399) writes "Eugene Goostman, a computer program pretending to be a young Ukrainian boy, successfully duped enough humans to pass the iconic test. The Turing Test which requires that computers are indistinguishable from humans — is considered a landmark in the development of artificial intelligence, but academics have warned that the technology could be used for cybercrime. Computing pioneer Alan Turing said that a computer could be understood to be thinking if it passed the test, which requires that a computer dupes 30 per cent of human interrogators in five-minute text conversations."
Turing Test Failed (Score:2, Insightful)
The test itself failed and is meaningless.
Re:Turing Test Failed (Score:5, Funny)
--ELIZA
Re:Turing Test Failed (Score:4, Funny)
Is it because you think that the test is failed because 30% on a small child doesn't seem anything like the real turing test that it is also meaningless?
Re:Turing Test Failed (Score:5, Funny)
You may have passed the Turing Test, but you sure as hell failed the Whooosh-Test.
Re:Turing Test Failed (Score:5, Funny)
Someone please verify, but I think we have a double-Whoosh here.
Re: (Score:2)
Re:Turing Test Failed (Score:5, Funny)
Computing... Verification complete.
You seem like a sensible person.
Re:Turing Test Failed (Score:5, Interesting)
In more seriousness, GP makes a very important point. Not only was this nothing like a real Turing test (a computer would have to fool the average person in more generalized and everyday circumstances for that to happen), the real point here is that we have learned since the days of Turing that even the full-blown Turing test doesn't really indicate much of anything.
People were fooled (really, really fooled) by Eliza way back in the day. It doesn't mean squat.
Re:Turing Test Failed (Score:5, Insightful)
What has been conducted precisely matches Turing's proposed immitation game. I don't know what do you mean by a "full-blown Turing test", the immitiation game is what it has always meant, including the 30% bar (because the human has three options - human, machine, don't know). Of coure, it is nowadays not considered a final goal, but it is still a useful landmark even if we have a long way to go.
That's the trouble with AI, the expectation are perpetuouly shifting. A few years in the past, a hard task is considered impossible for computers to achieve, or at least many years away. Then it's pased and the verdict prompty shifts to "well, it wasn't that hard anyway and doesn't mean much", and a year from now we take the new capability of machines as a given.
Re:Turing Test Failed (Score:5, Insightful)
What has been conducted precisely matches Turing's proposed immitation game.
NO, it DEFINITELY does NOT. For just one example, it tries to get around the "natural language" stipulation by pretending to be someone who doesn't fully know that language, and uses a simplified version instead.
That is a very clear attempt to subvert the rules.
I could go on, but it isn't necessary. It wasn't a real Turing test. We can leave aside the other nuances because the first criterion wasn't met.
The 'test' was fixed (Score:5, Insightful)
What has been conducted precisely matches Turing's proposed immitation game.
While they may have matched the letter of it, they subverted the spirit of the test. This quote [independent.co.uk] from the programme maker in particular is highly suggestive that they lowered the standards :-
To illustrate what I mean by lowered standards, imagine if I set up the same test, with 10 entries, and I tell the judges some of them are 2 year old babies playing on the keyboard. Armed with this information, some of the judges are likely to interpret even gibberish as typed by a human and it is not too farfetched to get more than 30% of them to agree.
This "result" is bollocks and a pure publicity stunt conveniently on falling on the 60th anniversary of Turing's death.
I want to see the actual transcripts which do not appear to have been released so far, which in itself is highly suspicious.
Re:The 'test' was fixed (Score:5, Insightful)
Interrogator: In the first line of your sonnet which reads "Shall I compare thee to a summer's day," would not "a spring day" do as well or better?
Witness: It wouldn't scan.
Interrogator: How about "a winter's day," That would scan all right.
Witness: Yes, but nobody wants to be compared to a winter's day.
Interrogator: Would you say Mr. Pickwick reminded you of Christmas?
Witness: In a way.
Interrogator: Yet Christmas is a winter's day, and I do not think Mr. Pickwick would mind the comparison.
Witness: I don't think you're serious. By a winter's day one means a typical winter's day, rather than a special one like Christmas.
I think the problem is that the way Turing was picturing the test, the human interrogators would be as smart as Turing and his friends, people who actually know how to ask probing questions. When you look at the conversation above, you see that he had in mind a program that does things which is decades beyond of what chatbots can do today. Everybody is dissing the Turing test, and if it has a problem, it's in that Turing overestimated people, in assuming that they actually know how to have conversations of significance. I still think there is something deeply significant about the Turing test, but in the one that I'm picturing, the interrogators must all be broadly educated experts on natural language processing with specific training in how to expose chatbots. And there should be money on the line for the interrogators: $1000 bonus for each correct identification, $2000 penalty for incorrect identification, no penalty for "not sure". If the majority of such experts can be fooled by an AI under these circumstances, then I think we should all be impressed.
Re:Turing Test Failed (Score:5, Insightful)
People were fooled (really, really fooled) by Eliza way back in the day. It doesn't mean squat.
No. They weren't. I speak as somebody who's had a go with Eliza and you could spot that it was a computer program in a couple of minutes if you wanted to. It's more likely that people were suspending their disbelief than really fooled.
Re:Turing Test Failed (Score:5, Interesting)
I was a BBS operator in the early 1990s. I had a game, which I titled "in case you really need for chat". It was an Eliza program, that I somewhat tuned to speak as I would (and translated to my local language). Plus, the user got to see the pretended typing in real time — Even with some typos and corrections.
Looking at the log files was *really* worth a laugh. But it made me feel wrong — Some users left in disgust, after "I" had insulted them.
And yes, they were not really aware I was playing a Turing test on them, so I don't know if this would have validity. But, by 1994 standards, I do believe it was quite an achievement (or perhaps, my users were mostly silly teens just like myself, and not worthy deciders for what constituted intelligent behaviour).
(Or maybe I'm *that* stupid in real life)
Re:Turing Test Failed (Score:4, Interesting)
"I got a little glimmer of fame because Danny Bobrow wrote up 'A Turing Test Passed".....One of the execs at BBN came into the PDP-I computer room and thought that Danny Bobrow was dialed into that and thought he was talking to Danny. For us folk that had played with ELIZA, we all recognized the responses and we didn't know how humanlike they were. But for somebody who wasn't real familiar with ELIZA, it seemed perfectly reasonable. It was obnnoxious but he actually thought it was Danny Bobrow. 'But tell me more about--' 'Earlier, you said you wanted to go to the client's place.' Things like that almost made sense in context, until eventually he typed something and he forgot to hit the go button, so the program didn't respond. And he thought that Danny had disconnected. So he called Danny up at home and yelled at him. And Danny has absolutely no idea what was going on......."
(reported in Coders at Work).
Re:Turing Test Failed (Score:5, Funny)
I've got this terrible pain in all the diodes down my left hand side.
--MARVIN
Re: (Score:3)
Me: My dog has no nose.
Eugene Goostman: But how does he smell?
Me: Badly.
Eugene Goostman: I don't get it.
An autist chat simulator duped 100% of people. (Score:5, Interesting)
Way back in my college days, I worked in a lab with a guy who wrote a chat bot that babbled on like an autist or otherwise mentally retarded youth would.
It would dupe 100% of the people who chatted with it. They couldn't distinguish it from an actual autist.
After seeing this work in action, I learned a very good lesson: the Turing Test is nothing but academic masturbatory fodder. It is not something to be taken seriously.
Re: (Score:2, Informative)
Autist is not a word, and autism is not a form of mental retardation.
Re:An autist chat simulator duped 100% of people. (Score:5, Insightful)
And a chair is not a chair. It seems you are not living in the real world. Except for high functioning autism, autism is a severe mental dysfunction.
Re: (Score:3)
No, he would have started his reply with the word "No" as well.
Re:An autist chat simulator duped 100% of people. (Score:5, Funny)
That's absoexactally right. The worditudinality of an utterance is defined completely by comprehension. Anywhom that says otherwise is being an obnoxialous prescriptivist!
Re:An autist chat simulator duped 100% of people. (Score:5, Funny)
Point of order: Austin is indeed a from of mental retardation.
I would extend that to most of Texas.
Re:Turing Test Failed (Score:4, Informative)
Second, from what I've seen, they are little more than cleverly created scripts, and as such, despite them fooling a few people, are in no way indicative of machine intelligence.
Re: (Score:2)
Indeed. The whole thing has basically degraded into a PR stunt and has nothing in common with Turing's idea.
Re:Turing Test Failed (Score:5, Interesting)
Searl missed the point. (Score:5, Interesting)
As to TFA: Anyone poo-poo-ing this result either does not understand it or has not bothered to look at the advances in AI over the last decade or so.We are at the point where a computer can read a novel and spit out a high school book report that would both fool and impress most english teachers, and it can do it in seconds not days.
There are also a lot of posts claiming the Turing test doesn't mean anything. However none of them I have read so far actually explain their statement, so I assume they are parroting their philosophy proffessor who was probably referring to Searle's Chinese translation room [wikipedia.org] argument.
The problem with Searle's argument (aside from lacking a definition of intelligence) is that it is assumed the intelligence is either embedded in the human or the books, it then goes on to show that neither is true, it's basically an unintentional strawman argument. It completely misses the point that the intelligence is embedded in the entire system of human + books. In other words the room itself is a black-box that displays intelligent behaviour, in much the same way as the human brain is a black box that (sometimes) produces intelligent behaviour. Like it or not your soul is a mathematical object [youtube.com].
So now we have Searl out the way, has anybody got an actual argument that supports the notion that the Turing Test is broken by design? - Seriously, I would like to hear a good one!
Re:Turing Test Failed (Score:5, Insightful)
It's a bit of an underhanded way to pass to pretend to be someone who doesn't speak English natively. The point of the test is to have a conversation for 5 minutes, not 5 minutes of "oh I can't understand you because I'm from Ukraine".
Re:Turing Test Failed (Score:5, Interesting)
Not only that, a non-native speaker who is a child.
5 minutes of "oh I can't understand you because I'm from Ukraine" plus 5 minutes of "oh I don't know about that because I'm only 13".
Re:Turing Test Failed (Score:5, Funny)
And I'm sure it used fewer lines of code.
Re: (Score:2)
Re: (Score:2)
Re: (Score:3, Insightful)
Turing never ruled out this sort of conversation...
Probably because he expected people to have some fucking common sense.
Re:Turing Test Failed (Score:5, Funny)
I'd say the test is obsolete. It's not measuring the advances in AI, but the involution of humans. Have you looked at Facebook status messages?
Re: (Score:2)
Yeah, it was a stupid idea to begin with, from an otherwise-brilliant man.
Re: (Score:2)
Turing had a far too good opinion of the human race (possibly not anymore towards the end of his life, when they had him chemically castrated because they did not like his homosexuality...), hence his parametrization sucks. Given that 90-95% of the human race are idiots that see what they want to see and not what is there, passing a Turing test involves just the right kind of deception, but no actual intelligence. The only thing the Turing test proves is hence that many humans do not have actual effective i
Re:Turing Test Failed (Score:5, Interesting)
The Turing test is a great test if done properly (Turing wasn't envisioning Twitter). While it's hard to pin down a good definition of sapience/intelligence (people want to keep redefining it to what humans have and no computer or animal has demonstrated this year), a good answer comes from studying communication. Intelligence in that sense is the ability to resolve the ambiguity of natural language by interaction as well as context.
In a very shallow way, search engines do that now - with a big enough data set they don't need an abstract mental model to ask "did you mean X?" But that's not really interactive - it's a single suggestion, with nowhere to go from there. When you're walking your dog and someone greets you with "hey, that's a nice dog" is that a content-free politeness, a flirtation, a discussion about dog breeding, a polite reminder that your neighbors are watching to make sure you clean up after the dog?
Part of being a socialized human is resolving that sort of ambiguity gracefully. We have an abstract mental model of other people and their motivations (learned from growing up with others) and we can use it without even noticing how neat that is that we can do that. Posing as someone young and socially awkward precisely defeats the purpose of the test.
Another sort of conversation that's hard to simulate is the way enthusiasts about something technical will talk. While it's easy for the computer to have all the technical details handy for something like a sports car enthusiast and tuner, or a baseball stats hound, the test is in the way people actually talk about that stuff. You see a lot of it on /.. Broad, passionate over-generalizations challenged, emotional argument becoming hot as first but then cooling as you discover that what you're really talking about is two different specific data points, and don't really disagree about anything important, just were over-generalizing from different things. That sort of conversation require both a social abstraction and an abstraction of the topic at hand. E.g. "you think Honda engines are better because you think X is important in an engine, while I think Toyota engines are better because I think Y is important" to mutually understand that requires more than just a knowledge of parts lists, you have to understand why someone would care.
IMO, if you have an abstract mental model of both people and the meaningful objects in the world (and, critically, yourself), and you make decisions based on modeling the hypothetical results of those choices, you are sapient/intelligent. Without invoking the supernatural, that's all there is to have.
Re: Turing Test Failed (Score:3, Funny)
Pretty good, but you don't fool me, robot.
Re:Turing Test Failed (Score:5, Insightful)
Thirty percent? (Score:3)
That's a pretty low bar. So to pass the test a computer needs three very low IQ subjects and seven normal people? Hell, the Alice program would probably pass. How about a more reasonable percentage, like 95%?
Re:Thirty percent? (Score:5, Insightful)
Re:Thirty percent? (Score:4, Insightful)
Most humans _are_ stupid. AI on their level would not be useful at all.
Re:Thirty percent? (Score:5, Insightful)
By random chance you would detect the computer 50% of the time, so that should be the goal.
Still 30% as "passing" seems unreasonably low.
Re: (Score:3)
Wrong!
The 30% figure was a prediction for how far computers would get by the year 2000.
Re:Thirty percent? (Score:5, Informative)
Turing predicted that machines would eventually be able to pass the test; in fact, he estimated that by the year 2000, machines with 10 GB of storage would be able to fool 30% of human judges in a five-minute test, and that people would no longer consider the phrase "thinking machine" contradictory.
Not literally a test (Score:5, Informative)
Should we tell them that the Turing test was a thought experiment and never meant as an actual objective test that would prove anything?
Re:Not literally a test (Score:5, Funny)
Unbounded tape (Score:5, Insightful)
Re:Unbounded tape (Score:5, Informative)
Re: (Score:2, Flamebait)
Don't bother. Most people here know far less about CS than they think they know. Dunning-Kruger effect at work. And do not even get me started on understanding things.
Hard Science v Soft Science (Score:2)
Next you'll say that Turing machines were a thought experiment and never meant to perform calculations in the real world.
the Turing Test is much more of a soft science test. It's at least as much about psychology as it is about math.
Turing machines are about math.
Thought experiments about math have no need to be applied to the real world.
Re: (Score:2)
Naa, that though is too complicated and may not fit into their tiny minds. Remember that Turing was on high genius level. That means something like 95% of the human race cannot actually follow his thoughts.
When the bar is too high... (Score:5, Funny)
When the bar is too high, try limbo instead of pole vault.
What's next?
"Yu So Dum, a computer program pretending to be a chinese toddler, successfully duped enough humans to pass the iconic test."
Voight-Kampff test? (Score:5, Funny)
Did anyone ask it the questions we already know will trip up a non-human?
"You're in a desert, walking along in the sand when all of a sudden you look down and see a tortoise..."
"You're watching a stage play. A banquet is in progress. The guests are enjoying an appetizer of raw oysters. The entree consists of boiled dog..."
Re: (Score:2)
Re: (Score:2)
Aehm, these things happen in reality? Although boiling the dog is a rather bland way to prepare it. For some more inspiration about how to prepare dog meat, look here: http://en.wikipedia.org/wiki/D... [wikipedia.org]
Re: (Score:2, Informative)
*Whoosh*
First, the period of ellipses indicates that he didn't finish the question. I believe this is a commonly accepted punctuation in English.
Secondly, this is a reference to a book by Phllip K. Dick called "Do Androids Dream Electronic Sheep," later made into a movie with Harrison Ford called Blade Runner. The questions are posed to androids (biological robots otherwise resembling humans) to gauge their emotional response to questions. This is the only way to distinguish them from people.
Re: (Score:2)
No...although the original AC's statement is literally correct, the point he was trying to make, that I contradicted myself, is NOT correct, as explained by the second AC. Those SNIPPETS themselves are not questions, but are, in fact, the prefacing components of a longer "question" from a SERIES of questions that any fan of the material I obliquely referenced would have recognized. So the "*Whoosh*" is actually applicable on, and apparently appropriate for, more than one meta level.
I guess some people get a
Re: (Score:2)
Not only are they not questions, but they make perfect sense in China and Mongolia. There are tortoises that live within small oasic lakes within the Gobi Desert. And oysters and dog are both consumed as food in parts of China and Mongolia.
Those are question preliminaries. The tortoise one continues with: "lying on it's back, but you do nothing to help it. Why?" I'm not sure how the banquet one finishes, but I'll bet there are more unusual edibles, with a question about why the listener chooses one of them.
Outdated test (Score:4, Insightful)
Re: (Score:2, Informative)
The test itself is flawed in the way that it's specific purpose is to test an AI, so the expected/unexpected outcome is set from the beginning. The AI's should be in the wild and not revealed until enough data of the interaction would have been gathered.
AI's can usually be tricked by injecting surreal elements to the conversation or asking about current events, or recent things. The focus should be in the intelligence and not in the conversational or mimicking part - the current online AI's could well be cl
Re:Outdated test (Score:5, Insightful)
A good turning test has an equal mix of humans and AI, and rewards the best in both..
Humans who pass as human, or as bots.
Bots that pass as Bots or as Human.
And has equal numbers of those shooting for each goal.
Half your entrances are trying to convince you they are human, the other half that they are AI, and half of each are lying.
Re: (Score:2)
"AI's can usually be tricked by injecting surreal elements to the conversation or asking about current events, or recent things."
Completely unnecessary. Simply carry on a conversation that requires a building on previous discussion. Every one I've ever encountered failed within a dozen exchanges. The most common technique the "AI" programmers use is to pretend to deflect the conversation. Usually quite lamely.
Re: (Score:2)
Then when they release the actual conversations, you see the computer actually wasn't too smart, but the other person was pretending to talk like a computer. What these tests actually show is that a human can convincingly pretend to chat like a computer.
Hasn't this happened a bunch of times? (Score:5, Interesting)
Just googling a few seconds brought me to:
This article about cleverbot. [geekosystem.com], which also eeked out enough votes to 'pass' a turing test.
It's all sounds just like Eliza [wikipedia.org], just put into a character with enough human limitations that you'd expect it not to string together phrases well, or keep to one topic more than a sentence.
I'd interpret it basically as an automated DJ sound board with generic text instead of movie quotes - you can certainly string a lot of folks along with even really bad ones, but that speaks more to pareidolia [wikipedia.org] than anything else.
I'd classify this stage of AI closer to "parlour trick" than "might as well be human" that a lot of people think of when they hear Turing test - but that's also part of the test, to see what we consider to be human.
Ryan Fenton
AI becomes not-AI as soon as it is successful. (Score:2)
It's perhaps unlikely at this point that we will ever develop anything which we will recognize as "true" AI. We may have to first develop a theory of what intelligence actually is, but until then the Turing test will have to do. Siri, Watson, and even Cleverbot are equal to the A.I. of the science fiction of yesteryear, but are considered mere "parlour tricks" today. AI research must be a depressing study in that respect, similar to commercially viable fusion power -- no matter how much progress is made, th
Wow, people are stupid. (Score:2)
I cast some pretty serious doubt onto the legitimacy of the claim that this machine passes a Turing Test, so much as the Turing Testers fail to be convincingly human.
Also, the robot went down much earlier than the appearance of this slashdot article, so for everybody saying the site got "slashdotted", hate to break your bubble but the world doesn't revolve around /.
http://gabrielapetrie.wordpres... [wordpress.com]
A pretty low requirement (Score:4, Insightful)
I feel like the requirements for the Turing test have been consistently lowered over the years to match what would be considered realistic to achieve rather than, as Alan Turing seemed to believe, demonstrate that a computer can be said to actually be "thinking."
Re:A pretty low requirement (Score:5, Insightful)
I'd say we keep raising the bar.
"If a computer can play chess better than a human, it's intelligent."
"No, that's just a chess program."
"If a computer can fly a plane better than a human, it's intelligent."
"No, that's just an application of control theory."
"If a computer can solve a useful subset of the knapsack problem, it's intelligent."
"No, that's just a shipping center expert system."
"If a computer can understand the spoken word, it's intelligent."
"No, that's just a big pattern matching program."
"If a computer can beat top players at Jeopardy, it's intelligent."
"No, it's just a big fast database."
Re: (Score:2)
Also, who ever said, "If a computer can beat top players at Jeopardy, it's intelligent?" Who ever said, "If a computer can play chess better than a human, it's intelligent?" The Turing test has been around for a long time.
Re:A pretty low requirement (Score:5, Informative)
Re:A pretty low requirement (Score:4, Insightful)
...and your brain, during a game of Jeopardy, is what if not a search engine?
Of course, (at least) advanced deductive capabilities are also important for general intelligence. That's the next goal now. (Watson had some deductive capabilities, but fairly simple and somewhat specialized.) We gotta take it piece by piece, give us another few years. :-)
Re: (Score:3)
Re: (Score:2)
How do you know other humans "think like a human"? The way people with autism think about the mental states of others differ significantly from the non-autistic, but does that make their way of thinking therefore inhuman? Similarly, I think differently about mathematics than my sister does, because we've had significantly different educational histories. Does that make her thinking or my thinking not human?
There are so many different ways in which human beings can think that the constraint "thinks like a hu
Re: (Score:2)
How do you know other humans "think like a human"?
Strictly speaking, I don't, but they think more like a human than the crappy Eliza-bot in this story.
Re: (Score:2)
Re: (Score:2)
Probability (Score:3)
Kids? (Score:2)
If virtual kids are allowed, I can make a bot pass easily:
Tester: "What's your name?"
Bot: "Goo goo ga ga"
Tester: "Oh, so you are a baby?"
Bot: "Glergggg ba ba!"
Tester: "Oh, how cuuuute!"
Time to move the goalposts! (Score:2, Insightful)
"Well, 30% isn't very impressive."
"Well, but people expect online correspondents to be dumb."
"Well, nobody ever thought the Turing test really meant anything."
Whether you "believe in" AI or not, progress is happening.
There will always be people who refuse to believe that a computer can be intelligent "in the same sense that humans are". Eventually, though, most of us will recognize and accept that intelligence and self-awareness are mostly a matter of illusion, and that there's nothing to prevent a machine
Re: (Score:2)
The goalposts have been moved the other way, towards "easy". 30%? Who invented that? Certainly not Alan Turing. Progress? Despite the stained reputation of the word "progress" (avoid it in the future if possible) the first time that a program passed a Turing test was in 1991.
You don't even know what a Turing test is. It has zippo to do with "AI" and has everything to do with "a machine successfully imitating a human." Lemme guess, you're one of those singularity religion followers, aren't you?
Not Really Passed... (Score:5, Insightful)
It convinced 33% of judges it's a 13-year-old Ukrainian. Since the test wasn't run in Ukrainian, you can't really say it proved that it had human-level language skills. Poor syntax, grammar, not understanding the question, etc. would be excused by the Judges as the "kid" doesn't know English well.
Since the program claimed to be 13, it also did not actually have to understand most of the things there are to talk about. Or anything, really. As an Englishman you wouldn't expect a Ukrainian teen to know anything about your life in England, and in turn the computer could make up all kinds of things about it's life in Ukraine and you'd have no clue.
So this isn't really AI, it's a take on the Eliza program of the late 80s/early 90s that hides the computer better.
Now if the test had been in Ukrainian, and happened in Odessa or Kiev; or even in Russian and in Moscow; tricking 33% into thinking your computer is a 13-year-old Ukrainian boy would be really fucking hard. It would be an amazing accomplishment.
Re:Not Really Passed... (Score:5, Funny)
Now only if it could have a 33% rate success in convincing other humans it was an exiled Nigerian dictator who needed some help moving his money out of the country.
...a computer could be understood to be thinking (Score:2)
He wrote "The original question, "Can machines think?" I believe to be too meaningless to deserve discussion.". Which is not the same as saying "could be understood to be thinking". Turing raises a number of highly interesting questions about what it means "to think". Passing the test is an interesting and noteworthy achievement but as Turing intimates - saying "a computer could be understood to be thinking" is "too meaningless to deserve discussion".
30% of tech support could not pass the Turing test (Score:2)
30% of tech support could not pass the Turing test
A real-life test (Score:2)
Re: (Score:2)
If itmadness to convince the other forum members, or not get found out, what will that two you about the level of online discourse?
Artificial stupidity (Score:3)
What about the converse: the anti-Turing tests? (Score:2)
I'd be interested in seeing how a human would do at proving they are not a computer, or attempting to prove they are. Either one would be an interesting test, whether the tester was human or computer.
Minimum Human Requirements (Score:3)
requires that a computer dupes 30 per cent of human interrogators in five-minute text conversations
Are there any requirements that must be met by the 'human interrogators'? What if they were all morons?
C'mon, it's trivial to tell a chatbot from a 13yo. (Score:2)
The chatbot will make WAY fewer spelling mistakes and use WAY less textspeak abbreviations and other pseudo-cool language.
Turing test is off topic (Score:3)
A turing test is testing such human experience aspects as:
- aculturation (what the person has been taught through education and socialization during their whole life up to that point)
- bias in expression based on typical human likes, dislikes, needs, desires, avoidances
Tarzan / wolf-boy would probably fail the Turing test based on the first factor. Might be very intelligent though.
Second aspect is just characteristic of a particular type of being that makes use of intelligence. Intelligent aliens would also have likes, dislikes, needs, desires, avoidances, simply based on also being self-interested "keep it together" beings, but the specifics might be very different, and would cause a fail of TT.
These experiential and situational and specific-agent-needs-desires-avoidances aspects have very little to do with the essence of intelligence.
General intelligence is probably better assessed through specific carefully designed tests designed to assess:
1) Concept learning, procedure learning capability in arbitrarily general contexts
2) Prediction of situation outcomes with novelty in situation presentations.
3) Ability to answer questions or take actions that show comprehension of essential / invariant aspects of situations, after opportunity to learn similar situations through either direct sensory input or linguistic instruction.
Comment removed (Score:3)
I, for one, welcome... (Score:2)
I, for one, welcome our Horizontally-Distributed Singularity Overlord.
Now, this makes me wonder: If those annoyingly stupid, non-AI bots in chats and social media have been able to fool real people for years... does that count as humans flunking the Turing Test?
Garbage (Score:3, Insightful)
All it showed, like any other Turing Test, is the gullibility of the subjects.
1) "Ukrainian" speaking English
2) 13 years old
Right there you have set up an expectation in the audience of subjects for a limited vocabulary, no need for grammatical perfection, little need for slang, and a lack of education. Now add in "star wars and matrix" and you have reduced the topics of discussion even more to the ones the programmers know best.
This thing would never have answered a question of 'Why', it also was under no pressure to being able to create a pun, both of which are easy things any older and educated human could do.
Garbage test, garbage results.
As usual.
AI telemarketer (Score:2)
There's a commercial telemarketing system AI [time.com] which makes cold calls and holds conversations. It's only slightly lamer than human telemarketers working from scripts.
Wake me up when those program solve this problem. (Score:5, Interesting)
* What is your name ?
* is it cold here ?
* The test is going well
* Color me surprised but are you a machine ?
* of course I am a human
* the keyboard is clean
* sky is the tv channel I watch a lot
* please answer the question now. "
When one AI not specifically programmed for that problem answer it correctly, I will be surprised and intrigued. Until then chatbot are just using cheap tricks to fool human.
Not a Turing Test (Score:3)
What nonsense! A program pretending to be an immature person with poor language comprehension and speaking ability, and incapable of talking about a large number of topics that can't be discussed with a vocabulary of 400 words and little life experience is not at all what the test is about. Turing expected an intelligent interrogator who could have a wide-ranging discussion about almost anything with the unknown other. Here's a snippet from his paper that introduces the idea of the Turing test, which he just referred to as the imitation game:
Interrogator: In the first line of your sonnet which reads "Shall I compare thee to a summer's day," would not "a spring day" do as well or better?
Witness: It wouldn't scan.
Interrogator: How about "a winter's day," That would scan all right.
Witness: Yes, but nobody wants to be compared to a winter's day.
Interrogator: Would you say Mr. Pickwick reminded you of Christmas?
Witness: In a way.
Interrogator: Yet Christmas is a winter's day, and I do not think Mr. Pickwick would mind the comparison.
Witness: I don't think you're serious. By a winter's day one means a typical winter's day, rather than a special one like Christmas.
Re:Dupe 30% of humans? (Score:5, Funny)
One dog would have if it wasn't for those meddling kids.
Re: (Score:3)
If you want to be thought of as knowledgeable on a subject like this, you might consider learning the difference between silicone and silicon.
Also, for the record, your distinction between AI and MI is BS. There have been many varieties of AI research, some inspired more by ideas about human brain function or human cognition, and some inspired less directly by those and more focussed on best exploiting computer-of-the-day capabilities.
All attempts which are not purely theoretical are implemented, and have s