Can Machines Think? Examining the Limitations of the Turing Test
Category Machine Learning Thursday - October 26 2023, 03:56 UTC - 1 year ago In 1950, Alan Turing proposed an experimental method for answering the question: can machines think? Although AI systems have come close to and even passed the Turing test in certain cases, it is not a definitive test of intelligence and is unable to measure the burgeoning number of human-machine hybrid intelligences.
In 1950, British computer scientist Alan Turing proposed an experimental method for answering the question: can machines think? He suggested if a human couldn't tell whether they were speaking to an artificially intelligent (AI) machine or another human after five minutes of questioning, this would demonstrate AI has human-like intelligence.Although AI systems remained far from passing Turing's test during his lifetime, he speculated that "[…] in about 50 years' time it will be possible to program computers […] to make them play the imitation game so well that an average interrogator will not have more than 70% chance of making the right identification after five minutes of questioning." .
Today, more than 70 years after Turing's proposal, no AI has managed to successfully pass the test by fulfilling the specific conditions he outlined. Nonetheless, as some headlines reflect, a few systems have come quite close.
One recent experiment tested three large language models, including GPT-4 (the AI technology behind ChatGPT). The participants spent two minutes chatting with either another person or an AI system. The AI was prompted to make small spelling mistakes—and quit if the tester became too aggressive.
With this prompting, the AI did a good job of fooling the testers. When paired with an AI bot, testers could only correctly guess whether they were talking to an AI system 60% of the time.
Given the rapid progress achieved in the design of natural language processing systems, we may see AI pass Turing's original test within the next few years.
But is imitating humans really an effective test for intelligence? And if not, what are some alternative benchmarks we might use to measure AI's capabilities? .
Limitations of the Turing test .
While a system passing the Turing test gives us "some" evidence it is intelligent, this test is not a decisive test of intelligence. One problem is it can produce "false negatives." .
Today's large language models are often designed to immediately declare they are not human. For example, when you ask ChatGPT a question, it often prefaces its answer with the phrase "as an AI language model." Even if AI systems have the underlying ability to pass the Turing test, this kind of programming would override that ability.
The test also risks certain kinds of "false positives." As philosopher Ned Block pointed out in a 1981 article, a system could conceivably pass the Turing test simply by being hard-coded with a human-like response to any possible input.
Beyond that, the Turing test focuses on human cognition in particular. If AI cognition differs from human cognition, an expert interrogator will be able to find some task where AIs and humans differ in performance.
Regarding this problem, Turing wrote, "This objection is a very strong one, but at least we can say that if, nevertheless, a machine can be constructed to play the imitation game satisfactorily, we need not be troubled by this objection." .
In other words, while passing the Turing test is good evidence a system is intelligent, failing it is not good evidence a system is "not" intelligent.
Moreover, the test is not a good measure of the growing number of non-human and human-machine hybrid intelligences, such as robots with machine learning algorithms installed.
Share