Turing test

From TheAlmightyGuru
Revision as of 14:57, 12 October 2018 by TheAlmightyGuru (talk | contribs) (Passing the Test)
Jump to: navigation, search

The Turing test began as a thought experiment by Alan Turing in 1950 with the goal of creating a test that would be able to distinguish human intelligence from artificial intelligence. Turing imagined a time in the future when machines had become powerful enough to being mimicking human intelligence and devised a test that would be helpful at discerning one from the other.

The most common example of a Turing test is placing a human judge at a computer screen with two chat windows. One of the chat windows is connected to a human participant while the other is connected to an artificial intelligence. The judge is expected to communicate with both participants for a period of time, after which, the judge is expected to identify which participant is the human and which is the AI based on how they responded throughout the test.

Turing's initial test was slightly different from this, but the purpose is the same. Can a machine be programmed to imitate human language well-enough to fool people? And, if it can, is it fair to say that the machine is now "thinking?"

I don't remember exactly when I learned about Turing tests, I believe it was in high school from my computer teacher, but it may have been later when I was playing around with artificial intelligence programs. Initially, I was pretty fascinated with them and thought that an AI that could pass a Turing test should be considered a thinking machine, but, after being introduced to the Chinese room though experiment, I was less impressed by them. Although, I now see flaws with the Chinese room as well.

Passing the Test

To date, nobody has ever written an AI that can pass a Turing test with a competent judge. In 2014, several news organizations reported that an AI program called Eugene Goostman succeeded, but it relied on a trick, claiming to be a 13-year old Ukrainian boy who couldn't understand English very well, which is not in the spirit of the test. And, even then, it only fooled 33% of the judges, several of whom were not adept at raking AIs.

I have never attempted to write an artificial intelligence program, but I have played with several, from my experience, none of them even come close to passing the test, but I look forward to the day when AIs become impressive enough to do so.

Reverse Turing Test

A reverse Turing test is uses the same structure, but reverses the roles. Instead of a human trying to identify a computer, a computer is trying to identify a humans. Probably the most popular reverse Turing test is a CAPTCHA (Completely Automated Public Turing test to tell Computers and Humans Apart) where a computer program is used to ensure the person it is in communication with is a human and not a machine. Such a reverse Turing test is used prevent computers from gaining access to information that is supposed to be available only for humans.

Tricking an AI

Part of the fun of the Turing test, and AIs in general, is figuring out ways to trick the AI into giving itself away with clearly incorrect responses. Here are a couple of questions that might help:

Submission Explanation
If there are a pair of people, how many total eyebrows do they have? Most AIs have to be programmed to solve basic math problems, and sometimes even simple word problems can be parsed and solved. This question is essentially 2 times 2, but, in order to get the right answer, the AI must know that a pair is the same as two, and the people have two eyebrows each. Most AIs I've tried this question on fail to answer correctly.
What is 22 divided by 7? Since most AIs have basic math skills, they should be able to solve this problem easily. The trick is interpreting their response. A human will give only a few decimals of accuracy, and take a long time to do so, while a computer can give dozens of decimals of accuracy instantly. Although, a clever AI programmer will write a subroutine to attempt to add delays, errors, and rounding to better fool the judge. CleverBot responded with "3.1428," far too many decimals of accuracy for a human, while Mitsuku responded with "3 and a bit."
Will blue justice run the fluffy ice? This sentence is clearly nonsense, and a human should respond with an expected "what?" but since it's structured using proper English syntax, an AI will probably try to formulate a response. Trying to get an AI to determine the difference between a question that makes sense and one that doesn't is extremely difficult.
|_|_| |-| A T   | $   Y 0 |_| R   |\| A |\/| E?
After looking at it for awhile, a human will probably identify that these symbols are being used to form letters. But I've never seen an AI parse it correctly.
I'm just a poor boy, nobody loves me (along with a series of other popular song lyrics). A human will pick up on the pattern and probably start responding with their own lyrics or the next lyric of the song, but an AI will probably respond as though these are actual sentences. However, if the AI has Internet access, it will probably figure this one out. Google's CleverBot often responds with the next lyric.
Tell me the three things you think of when you hear the word "joy." In order to answer this question properly the AI must understand that it needs to respond with three words, and they all have to be related to the abstract concept "joy." I've never seen an AI answer correctly.

Links