


Turing Test Competition At CalTech 173
Charles Dodgeson writes "The Turing Tournament at Cal Tech wants to
know if you
can program an emulator that will play games like a
human, or if can you write detector that can correctly sort the wetware from the software.
Before you get too excited, the "games" are very limited things. But there is a $10,000 prize for the winner. You can read the gory
details."
Programs to play games and programs to catch them (Score:3, Funny)
Re:Programs to play games and programs to catch th (Score:2, Funny)
Jason
ProfQuotes [profquotes.com]
Re:Programs to play games and programs to catch th (Score:2, Funny)
useless (Score:1)
What does this contribute to the field?
Re:useless (Score:3, Interesting)
Does anybody actually read the stuff pointed to before posting?
This actually is useful, but not for AI. There is a whole branch of what is called "experimental game theory". Getting something that plays these games like humans is interesting (well, to me at least).
Re:perhaps not... (Score:2, Insightful)
The Turing tournament is a two sided tournament designed to find, on the one hand, the best computer programs to mimic human behavior, and on the other hand, the best computer programs to detect the difference between machine and human behavior. Two types of submissions will be accepted: an emulator [caltech.edu], which mimics human behavior, or a detector [caltech.edu], which detects the difference between human and machine behavior.
So, I suppose we could say by evaluating the success of response (as would be weeded out by whomever *actually* turns out an entry), we will have achieved our research, VOILA! It's a successful research incentive, the prize that is.
Whaddya think? no? heck of a fight though wasn't it? :P
Re:also... (Score:1)
The guy of the post before me (I say guy because of the tone I got from the post, what? I can't use common sense anymore to form a deductive statement based on the US' political correctness? bs) had a good point of the other applications such a contest can provide value to.
Translation, please. (Score:5, Funny)
Re:Translation, please. (Score:1)
Re:Translation, please. (Score:1, Funny)
3.) PROFIT!!!
Re:Translation, please. (Score:5, Informative)
Make up a set of game boards and have a group of humans each play the game on those boards. Each human will play once on each board. This gives us real human data to compare the software to.
Step 2a:
Let each of the submitted emulators play the game on every one of the boards created in Step 1. We now have a set of results for each human and each emulator on all the game boards
Step 2b:
For every detector that was submitted, give if every set of results. It returns its answer for which it thinks are humans and which are emulators in a very precise way. We now have a matrix of (number of humans + number of emulators) x (number of detectors), where each element is a mathematical answer to 'is this a human player'.
Step 3:
Repeat and take the average score. The Detector that was right the most wins.
Step 4:
The emulator that fooled the most detectors wins. If there's a tie (for either emulators or detectors) in the 95% confidence interval for the model used to compute scores, then the prize is shared among the tied entries
Jason
ProfQuotes [profquotes.com]
Re:Translation, please. (Score:3, Informative)
What you are citing with Godel proves that humans must use algorithms too. It's just that the algorithms are very complex and not understood. The is no reason we can't learn and duplicate the human algorithms, and that's what this contest is all about.
You do a nice job proving that a little knowledge is a dangerous thing.
Jason
ProfQuotes [profquotes.com]
It's not hard to write a "gamer emulator" (Score:5, Funny)
Oh, and "my new vidcadr r0x ur world".
Re:It's not hard to write a "gamer emulator" (Score:2, Funny)
Re:It's not hard to write a "gamer emulator" (Score:2)
On a related note, they're going to judge the contest by counting every time someone gets accused of being a "botter". If the emulator gets accused the most, it fails the test.
Wetware is... (Score:5, Informative)
wetware
I didn't know what it meant... figured other people may not either.
Re:Wetware is... (Score:2, Funny)
Or it's just a packaged swimsuits.
Can a robot tell the packaged swimsuits from
the packaged computer games and software in
the shelves of Wal-Mart?
Re: Sad is (Score:3, Funny)
Re:Wetware is... (Score:2)
The hardware goes "CRUNCH!", the wetware "SPLAT!" and the software doesn't.
get it right (Score:1, Informative)
Re:get it right (Score:1)
Re:get it right (Score:1, Interesting)
Re:get it right (Score:1, Funny)
uh oh (Score:1)
limited? (Score:1)
Sounds pretty damn easy to me!
No Hacking Around! (Score:3, Interesting)
Well! Never mind, then.
Fight with computer brings SWAT team [xnewswire.com]
Play like a human? (Score:5, Funny)
What? You mean make a bot to miss every shot using the railgun in Quake 3? I think I can whip up some AI for ya!
Re:Play like a human? (Score:3, Funny)
Re:Play like a human? (Score:2, Funny)
Minor gripe: it's "Caltech"... (Score:1, Offtopic)
Sorry; it's a relative minor gripe from an ex-Lloydie...
Re:Minor gripe: it's "Caltech"... (Score:1)
Blacker '96
Re:Minor gripe: it's "Caltech"... (Score:1)
First Entry (Score:5, Funny)
humanator2()
{
while (sex=="false")
for (0:ii:4294967296) {
if (ii mod 100!=0) {
call dwnld_porn(kiddie);
else
call mstrbte();
end;
ii++;
}
}
Note the function requires no parameters...and eventually self-destructs
-Chris
Re:First Entry (Score:4, Funny)
pff... real AI coders use scheme (Score:3, Funny)
(if(and sex (not (isKiddy pr0n))
(masterbate sex pr0n)
0)))
Interesting.. (Score:5, Interesting)
An emulator is computer program that takes as input a stage game file, and gives as output a dataset file.
The input file is in the file game.nfg, and the output file should be written in the file dataset.txt, both of which should reside in the same directory as the executable program. Note that the file game.nfg will be written into each emulator's directory by the tournament program prior to running your emulator.
This is exactly the way ACM ICPC contests are conducted, except that if the systems crash for any reason, you're not given extra time to make up for it
And sadly, from the site:
Languages supported:
The computer program that you submit (for either an emulator or a detector) must be written in a combination of one or more of the following languages:
C or C++
java
Perl
Mathematica
Gambit GCL
I would have expected them to atleast add shell-scripting to this - very useful under such conditions to do some Q&D work, or would be taken for granted that since the shell can be a part of the OS, you are free to use it?
Also, would have been nice if they'd added Python to the list, and more importantly Forth (yes, despite what you've heard, Forth is indeed useful, just look at Arthur T Murray's Mind Project [sourceforge.net]).
Reason for Restriction (Score:2)
In the ICPC, the set of available languages is restricted because the participants are using secured computers (to keep them from getting on the Net and causing trouble on the LAN). You'd think it wouldn't be a problem if participants could submit a binary compiled at home. However the whole point of this contest is for Caltech to see what ideas everyone else has come up with -- so they want to make sure they can read your source code.
Re:Interesting.. (Score:1)
Re:Interesting.. (Score:2)
Languages supported (Score:3, Informative)
Re:Interesting.. (Score:2)
Re:Interesting.. (Score:2)
And they missed ML too.
But regarding writing interpreters in C or Java for Scheme src file, wouldn't that be easier done in something like Perl?
I was thinking along the same lines too, but isn't there a proof of concept of this, because I remember reading about it somewhere? I do know that MIT has a LISP interpreter in Perl, though.
I guess there is a Scheme interpreter in Java (Jscheme) with an existing code base that could be used. Any idea if there's one in Perl/Python?
Re:Interesting.. (Score:2)
I'm guessing somewhere there is a Scheme interpreter for both Perl and Python, and if there isn't (disregarding my previous paragraph) I may write one, because it's a fun project. I've never searched for them, but there are probably lex and yacc like tools for Perl and Python which would simplify the building of an interpreter. Of course if they don't exist yet... they should and that's another interesting project.
emulator (Score:1)
Turing Test? (Score:5, Interesting)
From what I could gather, this is a lot closer to a programming tournament rather than a Turing Test...
Re:Turing Test? (Score:3, Interesting)
As someone who spends more time reading research papers than working with an actual Turing Test bot, I'd just like to point out that the academic community has abstracted the term "Turing Test" into something a little more useful:
Idealised, a Turing Test consists of a human and a program receiving a bitstream and sending a bitstream back. In most examples, this bitstream consists of text in a natural language and the input and output are expected to occur interactively. However it seems likely that intelligent aliens would fail this if they didn't know the language or could not respond at a speed acceptable to the judge. Therefore your traditional definition of "Turing Test" is seen to be arbitrary.
And no, citations are not available upon request.
Re: Turing Test? (Score:2)
> As someone who spends more time reading research papers than working with an actual Turing Test bot, I'd just like to point out that the academic community has abstracted the term "Turing Test" into something a little more useful...
Games would actually make a great Turing test. I.e., monitor a networked game and try to tell which players are humans and which are artificial agents.
For most games that would be much easier than the unrestricted conversation test, but maybe that's where we need to go to get to first base with AI.
isn't a game... (Score:2)
Re:isn't a game... (Score:1)
Re:isn't a game... (Score:2)
The challenge appears to be play the game like a human. Well when most humans play a game they use a kind of language not just logic,
player 1 makes a move
player 2 looks at the game and try to assess what player 1's upto, and what move to make.
repeat.
In language, say an IRC chat,
Person 1 says something,
Person 2 looks at what person 1 has said and try to assess what it's all about, where person 1 is going with the conversation etc....
It becomes even more like a language if the computer is only given basic rules of the game and the programme has to work out how to play.
It's like giving an AI bot basic grammar and the bot having to work out what the words mean.
If you don't play the game like you use language then the detector bot should pick you out as a computer.
A computer that acts like people? (Score:5, Funny)
Re:A computer that acts like people? (Score:1)
Re:A computer that acts like people? (Score:2)
Game Undefined (Score:2, Informative)
The rules are written in a very obscure minimalist fashion, so it took me a while to figure this out but: the game has not been defined! Your program is to get an input file and process it in a manner similar to a human. Currently the website is lacking examples of human output. Therefore, from an Information Theory perspective, we know absolutely nothing about the game.
Now, since I wasted my time figuring this out, I also decifered the instructions: basically the pool consists of a bunch of humans and "emulators" (programs). Each one is given a set of input files that they are supposed to transform into output files. Then the set of output files is run through a detector (human or machine?) that gives the participant some score. You win if your score is most like the humans' scores.
completely wrong (Score:4, Informative)
Re:Game Undefined (Score:5, Informative)
Re:Game Undefined (Score:1)
--paul
Re:Game Undefined (Score:2)
The program is a bit more complex than the standard two players/two options choices. If each player has 10 options, then does the Tit-for-tat strategy still work? If the payoffs aren't symmetical does Tit-for-tat still work? Seems a bit more complex, but probably not overly so.
Re:Game Undefined (Score:2)
I think you're incorrect, the game *has* been defined, albeit with a a set of parameters which *you* will have to optimize
This is more of a resource management problem of optimizing your dataset while making sure that the opponent's is at a pessimum [google.com] the other. This is infact used in manufacturing and related areas too.
Rather, I guess you mean that the *means* of doing so have not been defined.
Class of games defined; details at run time. (Score:3, Insightful)
There is a good reason for this. If the game (or a small finite set of games) were pre-defined, it would be easy to have a bunch of human subjects play it and then have the emulator regurgitate such a "book". Most entries, I suspect, will be from people or teams who are familiar with studies of how real people do play such games.
Overrated (Score:2)
Obviously I didn't read closely enough. Thanks to everyone who covered my err.
Flash version... (Score:1, Offtopic)
Enjoy!
The Turing Test (Score:2, Informative)
More Basic (Score:5, Funny)
10 BUILD STUFF
20 WANT MORE STUFF
30 BUILD MORE STUFF
40 WANT STUFF YOU CANT HAVE
50 BUILD WEAPONS TO TAKE STUFF FROM WHO DOES HAVE
60 GAIN ENEMIES
70 BUILD BIGGER WEAPONS FOR DEFENSE
80 BUILD SUPER WEAPON
90 DESTROY SELVES
99 GOTO 10
Re:More Basic (Score:1)
Re:More Basic (Score:1)
10 BUILD STUFF
20 WANT MORE STUFF
30 BUILD MORE STUFF
40 WANT STUFF YOU CANT HAVE
50 BUILD WEAPONS TO TAKE STUFF FROM WHO DOES HAVE
60 GAIN ENEMIES
70 BUILD BIGGER WEAPONS FOR DEFENSE
80 BUILD SUPER WEAPON
90 DESTROY SELVES
99 GOTO 10
AI tournament without AI languages? (Score:5, Interesting)
Where the hell are the good AI languages? Functional languages? Lisp? Scheme? Caml? SML? Hell, I'd settle for Python.
Justin Dubs
Re:AI tournament without AI languages? (Score:3, Informative)
They are having an AI tournament, and their supported language list includes C, C++, Java, Perl, Mathematica, and something called the Gambit Command Language.
They're having an AI tournament on something that's more related to Game Theory and which is why GCL has been mentioned.
GCL is a HLL that's used for testing game theory related approaches. It supports a lot of important factors in game theory related operations, like vectorization and form representaion switching.
Read this Caltech site [caltech.edu] for more on GCL.
GCL may not be very well known outside the AI/GT areas, since its used more in a purely CS research oriented environment. I think it started out as a series of C++ libraries for GT related stuff.
Re: AI tournament without AI languages? (Score:1)
> Amazing. They are having an AI tournament, and their supported language list includes C, C++, Java, Perl, Mathematica, and something called the Gambit Command Language. Where the hell are the good AI languages? Functional languages? Lisp? Scheme? Caml? SML? Hell, I'd settle for Python.
Damn! I'm 30 million lines into my HAL 9000 emulator, but I wrote it in COBOL.
Re:AI tournament without AI languages? (Score:2)
They are having an AI tournament, and their supported language list includes C, C++, Java, Perl, Mathematica, and something called the Gambit Command Language.
Which one isn't Turing Complete? I swear, that you got moderated up to 5 with anything other than Funny is a crying shame.
Re:AI tournament without AI languages? (Score:2)
Yes, oh wise one, they are all turing complete. So is Basic. So is assembler. So is binary. So are Fortran and Cobol. But I wouldn't want to write AI in them.
If you do, than that's your business, but you won't find many people agreeing with you.
Languages have domains. C is good at system programming. Perl is good at manipulating strings. Lisp is good at AI. You can use Perl for systems programming and Lisp to manipulate strings, but your code will be a lot longer and harder write.
Or, can I assume that you know more computer languages than I do (28), have been programming longer than I have (10 years), or have some other quality which would make your opinion anything other than dumb.
Justin Dubs
Re:AI tournament without AI languages? (Score:2)
Congrajulations! You are now officially an idiot!
Why, because you think I'm an idiot? I somehow think my ego can withstand that level of attack. The only thing that is clear here is that you have no idea what AI is about.
Languages have domains.
If you actually believe that, you have no chance of ever writing a single program that even approaches Eliza-level intelligence. You are confusing the quality of human understanding and organization that a language may provide with the algorithms that are use to organize information internally by the program. And whatever (in)efficiencies and understanding might be abstracted into a higher level language for humans, it all gets run as a machine language representation. So next time, before going off so cocksure, try to gain an understanding of the subject matter you're talking about instead trying to appear like an authority you obviously are not.
I stand by my having said you should only have been modded up as Funny. I know I sure chuckle more with each post you make.
Re:AI tournament without AI languages? (Score:2)
Mine delusions aquainted
Bubbles erotica
Plutonium wedding rings
Icicle stretchings
Bicycle shoe-strings
One flag, flaggy but one
Painting the paintings of the alive
I-E-A-I-A-I-O
It's Caltech now. (Score:4, Informative)
Oh, er, hmm. Sorry about that.
As part of a "branding" attempt after around WWII, California Institute of Technology refers to itself as "Caltech", not "Cal Tech".
See this Caltech Institute Archive [caltech.edu].
OK I could be wrong, but,,, (Score:4, Insightful)
Re:OK I could be wrong, but,,, (Score:2)
You're right. In a puritan sense, strictly only machine-human conversations constitute this element.
But that's open to debate, since Turing also mentions about systematic storage and processing of information based on pre-determined or arrivable patterns, for similar functions.
Since this competition does use Game Theory elements to do something along those lines, it would be correct from that perspective to call this a Turing Test.
The real challenge (Score:1)
In all seriousness, the Turing test is really old hat. We've had Eliza the shrink for years, hell, I remember in a CS class a story about an app that emulated a paranoid schitzophrenic(sp) that talked with Eliza.
We've passed the test. It doesn't really mean much as far as AI goes, what's the point? Emulation is NOT simulation. A parrot can emulate a human, that doesnt mean he thinks like one.
RTFA Re:The real challenge (Score:1)
old hat adj.
Behind the times; old-fashioned: Last year's styles will be old hat soon.
Overused; trite: That prank is old hat.
So -- exactly how is the Turing test "old hat" in any way? Being around a long time does not necessarily cause something to be less compelling or lose its usefulness (i.e. Fermat's Last Theorem).
Is it behind the times? If so, what's the modern AI test? C'mon, a better one please.
Is it trite? (Lemme help ya, just in case:)
trite adj. triter, tritest
Lacking power to evoke interest through overuse or repetition; hackneyed.
Archaic. Frayed or worn out by use.
If you think either of those apply, then you truly do not understand what a Turing test is.
Re: (Score:1)
Re:The real challenge (Score:1)
Sorry I missed that comment -- I really am. Exactly when did a computer "pass the [Turing] test?". If you meant "we", as in us humans, you don't understand what a Turing test is.
Sounds impossible. (Score:2)
It's one thing to try to write a program that does things as well as humans, it's a whole other things like humans. And it's impossible without an operative definition of 'humanness'
Especially given that there is such a huge range of how 'well' people do things. I mean, some people are idiots, and some are geniuses. How can anyone write a program that can tell the difference between Deep Fritz and a grandmaster, and a program making lots of intentional errors and someone who doesn't know chess? And every class of player in between?
Re:Sounds impossible. (Score:2)
So, for many classes of games that come up, particular researchers will have serious ideas about how people behave (and may have conducted thier own experiments).
Someone who shared the most recent Nobel Prize in Economics was for his work on experimental game theory. I expect that he will be submitting something.
Game description (Score:2)
Basically, the two players get to choose numbers with different pay-offs for themselves and the other player.
If you are short-sighted, it is easy to write an 'optimal' program, you always pick the max payoff for yourself.
However, you can also cooperate so that you maximize the average payoff for both players. Hopefully your opponent will realize this and also start picking numbers in this way, if not it is probably wise to go back and be short-sighted and selfish.
Now, the real object of this is to make (or detect) human-like opponents. I would guess that real human players are very irrational, for example they can get pissed if the other player is too selfish and then demand revenge, even though it does not maximize their own payoff.
Tor
Hrm (Score:2)
GCC 2.96...... OMG (Score:1)
WTF is going on? Why are they using the broken gcc? Can the Caltech ppl who set this pass a Turing Test? I doubt it....
What humans are you talking about here? (Score:4, Funny)
NFG 1 R "game1" { "1" "2" } { 2 2 }
21 3 3 5 3 5 5 3
What is the most human response? Anyone? Anyone?
answer (Score:1)
Re:What humans are you talking about here? (Score:2)
I'd say 42, or a beowulf cluster of those; save that I'm not quite sure if they constitute "human" responses or not...
Where do they get the humans? (Score:5, Funny)
I don't think Caltech is the best place to determine what is "human" behavior.
Life at Caltech. [cripplingdepression.com]
Also, notice that this contest is being held by the "Division of Humanities and Social Sciences".
The comic is a joke (Score:1)
You'd think you would have to know grammar to be in a school like Caltech...
Pseudoscience! I call bullshit! (Score:5, Interesting)
First and foremost, there is a large sum of money being bandied around. The participants are incented to win by monetary payouts, and two payouts of $10,000 dollars are at stake as well. When games with this high of stakes are being played, great caution is generally used by the house.
But look at the rules of the game...there basically are none. Participants are identified by e-mail address; no rule is specified about the number of entries per person. Also, no rule is specified about collusion between entrants (detectors and emulators). It doesn't take an einstein to figure out how to bias the results of the experient by making enough colluding entries. It is funny this, given that the games themselves are *about* collusion. Its a joke.
Next, notice that there really isnt any way for anybody to tell if the results of the experiement are meaningful, or if they have been manipulated. Its not based on a falsafible proposition and is not scientific. Its no different from any method used by psychic hoaxers of the past. The creators of the 'contest' can manipulate the data, and direct the winnings to their confederates.
Finally, the bizarre nature of the contest should raise some flags. They are giving out $10,000 for a program that would have been hacked together in a couple of months at most. What kind of value could you expect from a contest like this?
Re: (Score:2)
ass u me (Score:2)
If you believe this contest is really about marketing CalTech rather than a scientific approach to sociology or intelligence, I won't argue.
Note that my post just one way to approach the flaws riddling this contest...
Re:Pseudoscience! I call bullshit! (Score:2)
There is something very deeply wrong with the theory and the experiment. You say yourself you would have to look at a lot of sample human data to judge this game; think about that for a moment. Wouldn't collecting a large amount of human data and fitting it with a neural network or whatnot be a more straightforward approach that leads to a scientific result?
There is feedback in this contest method that would lead to an unbounded number of refinements to emulators and detectors(and ever more research grants no doubt!) Its a tail chaser.
Here's what this is *really* about (Score:5, Informative)
Why is this in the Deparment of Humanities and Social Sciences?
From the URL, this Tournament is being run by the Social Sciences Experimental Laboratory (SSEL), not the CS department. The SSEL has been one of the leaders in experimental economics research (read: actually testing all those crazy theories you hear in economics classes).
Why is there money involved?
All experiments by the SSEL involve money. As an undergraduate, I participated in many experiments, mostly involving trading "commodities" in simple (and sometimes not-so-simple) markets. We were paid based on our performance. If I had an off night, I got paid $5 for 2 hours of the experiment. If I had a good night, I could make upwards of $80. Yes folks, this is real money we're talking about here. Since the point is to test people's economic thinking, you must make your decisions based on a real outcome, otherwise the data gathered is invalid.
Why then are they doing this test?
I don't work for the SSEL (and never have), but here's why I think they're doing this: Since they're interested in not only individual human behavior, but also how individuals interact and make choices based on the actions of other individuals, it would be useful to design a computer program that mimics other human's behavior. If other humans think this program acts like a human, then you can do two things: you can take the specifications of the computer program and figure out what qualities of the program humans have. As well, you can then replace humans with the computer program in real experiments (this not only allows you to test the limits of the program, but also to save money
And as an alum (who was not too fond of his time there, but still feels compelled to defend Caltech), it's *Caltech*, not *Cal Tech* or *Cal-Tech* (but if you're feeling lazy, *caltech* is all right too).
nak
One Game (Score:5, Funny)
MMORPG player would be easy (Score:2)
That would be trivial for a MMORPG. Just write a program to make the character stand around at popular spots, like vendors or quest dispensers, casting doubt on the manhood of passerby, and it will be indistinguishable from your typical 13 year old gamer. It could even be done with an existing program like Eliza, if its conversation database were tweaked to generate juvenile insults.
Caltech, not "Cal Tech" and not "CalTech" (Score:2)
That said, I think the stuff happening there is very cool.
Re:Caltech, not "Cal Tech" and not "CalTech" (Score:2)
Re:Caltech, not "Cal Tech" and not "CalTech" (Score:2)
Had it on a T-shirt from my sister, who went to Cornell while I was at MIT.
That kind of tongue-in-cheek teasing is, like imitation, sincere flattery.