Named for its creator Alan Turing, the Turing test is meant to test a machine’s intelligence by assessing its conversational abilities (Bieri, 1988, 163). Turing adapted the test to suit machines from an existing test, the Imitation Game, wherein a man and a woman would converse via teletype (Bieri, 1988, 163).
The Turing test is virtually the same, only instead of a man attempting to pass himself off as a woman, a machine is talking to a human while the judge is left to determine the humanness of the participants (Bieri, 1988, 163). According to Turing, if a machine can converse well enough with a human, such that other humans in observation of the conversation reliably believe that they are observing two humans conversing, then that machine has sufficient intelligence to be deemed capable of thought (Bieri, 1988, 163).
Although the Turing test has been held as a benchmark for evaluating machine intelligence, there are many concerns regarding the type of intelligence likely to be identified by the Turing Test. One of the most frequently raised concerns argues that some artificial intelligences may simply convey the intelligence of their creators, and there is no way for the Turing Test to distinguish between a machine that is intelligent in this manner and one that is self-aware, conscious, and capable of reflexively thinking about itself. As a result, these "intelligent" machines may be able to pass the Turing Test without ever being intelligent themselves.
These criticisms of the Turing test are made from the perspective that in order for a machine to be intelligent, it would need to have some sort of existential experience such that it could act in its environment, observe, and learn from the consequences of its actions (Bieri, 1988, 174; Beavers, 2002, 71). True artificial intelligence would be self-aware and capable of self-reflexive thought, such that it could think about its own thinking. From this perspective within artificial intelligence scholarship, the Turing test seems to conflate conscious artificial intelligence with a sort of machine capable of cleverly parroting human intelligence in something akin to a high-tech ventriloquist trick.
Although these criticisms certainly have metaphysical merit regarding the Turing test’s inability to identify consciousness, they seem to discount a key feature of the Turing test. The Turing test is conducted from the perspective of a human subject observing the human-machine interaction. From this perspective, the construction of this test suggests that interacting with such an intelligence is indistinguishable from a human in terms of conversational interaction.
There may be reasons to think that the particular artificial intelligence is not really intelligent, but the key point is that it appears to be. Turing wasn't necessarily after a test that determined the actual intelligence of an artificial agent; he simply may have wanted a test to determine when an artificial intelligence would pass for a human when confronted by another human being.
With this in mind, the Turing test raises the question as to whether a metaphysically sound understanding of consciousness is important. In fact, Turing’s construction of his test for artificial intelligence in this way reflects a traditional phenomenological problem of how it can be known that consciousness is present in our fellow human beings. In terms of immediate sensory experience, immanence in phenomenological terminology, the consciousness of another being is always inaccessible. Edmund Husserl addressed this issue of how to experience the consciousness of someone else through the transcendent, “harmonious experience of someone else” (Husserl, 2007, 224).
This experiencing of someone else involves dispensing of any belief that the other person exists and attempting to understand what it would be like to be that other person instead of one’s self. In this sense, one is attempting to achieve a first person experience of being someone else in order to demonstrate that, in accordance with phenomenologist methodology, if someone else existed, existence as that someone else would be analogous to one’s own existence (Husserl, 2007, 224).
Whether or not Husserl’s particular methodology of determining the existence of consciousness in others is compelling, it conveys the difficulty of being able to identify consciousness. In this sense, Turing’s approach demonstrates a certain common sense wisdom. If experiencing consciousness in other humans has troubled philosophers, it is likely to be as troubling to develop a test for consciousness in artificial intelligences.
Furthermore, in an everyday sense, we seem content to grant the benefit of the doubt to other humans when it comes to having consciousness whether we have a rigorous proof for it or not. We can continue on as if they are conscious because others act in such a way as we imagine they would if they did have consciousness. Ostensibly, this is the same standard that Turing applies to his test. It accepts that, in terms of immanent experience, consciousness is inaccessible and that, in everyday life, consciousness is a primary concern outside of circles of philosophers and scholars concerned with phenomenology, philosophy of the mind, and artificial intelligence.
So, the Turing test seems to figure that if an artificial intelligence can perform consciousness to an equal degree as a human, then why could we not grant the same everyday benefit of the doubt regarding consciousness as we grant to our fellow humans? In this regard, we should wonder, in dealing with robots and artificial intelligences in everyday life, does consciousness matter, or is the convincing performance of consciousness enough?
Beavers, A. (2002). Phenomenology and artificial intelligence. Metaphilosophy, 33, 70-82.
Bieri, P. (1998). Thinking machines: Some reflections on the Turing test. Poetics Today, 9(1),163-186
Husserl, E. (2007). The problem of experiencing someone else. In R. Craig & H. Muller (Eds.), Theorizing Communication: Readings across traditions (pp. 223-224). Los Angeles, CA: Sage.
Dylan Chandler is a Ph.D student at Simon Fraser University’s School of Communication in Vancouver, British Columbia. He is interested in philosophies and critical theories of technology, information, and artificial intelligence, especially with regards to their impact on society and culture.
(3) Comments •
(5793) Hits •