The Empirical Untenability of Sentient Artificial
Transcription
The Empirical Untenability of Sentient Artificial
Dickinson College Dickinson Scholar Honors Theses By Year Honors Theses 5-22-2011 The Empirical Untenability of Sentient Artificial Intelligence Andrew Joseph Barron Dickinson College Follow this and additional works at: http://scholar.dickinson.edu/student_honors Part of the Philosophy Commons Recommended Citation Barron, Andrew Joseph, "The Empirical Untenability of Sentient Artificial Intelligence" (2011). Dickinson College Honors Theses. Paper 122. This Honors Thesis is brought to you for free and open access by Dickinson Scholar. It has been accepted for inclusion by an authorized administrator. For more information, please contact [email protected]. "If you gotta ask, you ain't never gonna know" THE EMPIRICAL UNTENABILITY OF SENTIENT ARTIFICIAL INTELLIGENCE By: Andrew Barron Submitted in partial fulfillment of Honors Requirements for the Department of Philosophy Professor Jessica Wahman, Supervisor Professor Crispin Sartwell, Reader Professor Susan Feldman, Reader Professor Chauncey Maher, Reader May 22, 2011 "All sentience is mere appearance - even sentience capable of passing the Turing test." Tuvok Star Trek: Voyager "The key distinction here is between duplication and simulation. ever constitutes duplication." John Searle tvunds, Brains, and Science And no simulation by itself TABLE OF CONTENTS INTRODUCTION 1 CHAPTER ONE: MINDS AND/OR COMPUTERS WHAT IS COMPUTATION? 4 IS A MIND A COMPUTER? 9 IS A COMPUTER A MIND? 13 CHAPTER TWO: THE ENIGMA OF FAMILIARITY THE HARD PROBLEM 23 ARGUMENTS 28 FROM INEFFABILITY CHAPTER THREE: DO ANDROIDS DREAM OF ELECTRIC SHEEP? WE'LL NEVER KNOW FOR SURE THE EXPLANATORY GAP 39 MCGINN'S THEORY OF COGNITIVE CLOSURE 43 CRITICISM AND RESPONSE 52 Al, CONSCIOUSNESS, AND BLADE RUNNER: TYING EVERYTHING TOGETHER 56 CONCLUSION 62 WORKS CITED 63 INTRODUCTION The ultimate goal of Artificial Intelligence ascribe it to a computer.1 {Al) is to model the human mind and Since the 1950s, astounding progress has been made in the field, leading some to defend "strong Al," John Searle's term for the theory that it is possible to write a computer program equivalent to a mind. As a result, a common trope in science fiction from Isaac Asimov to James Cameron is the idea of sapient and sentient robots living amongst humans, sometimes peacefully, but more commonly not. In Ridley Scott's Blade Runner, androids indistinguishably humanlike in appearance and behavior live as outlaws among human beings, hiding in plain sight. entertainment, The idea of the strong Al makes for good but is it actually possible? Is it within the realm of human capability to synthesize consciousness? To many scholars and researchers, the answer is a resounding science, the interdisciplinary amalgamation "yes!" of neuroscience, computer science, philosophy, linguistics, and psychology, has churned out increasingly advanced instances than half a century. Cognitive of Al for more This paper is an attempt to restrain the mounting excitement. There is no doubt that Al is an incredible idea with far reaching implications already in effect today. The marketplace is already saturated with "smart" cars, calculators, wristwatches, and dishwashers, but the average consumer generally avoids thinking about what that really means. Does the luxury car that parallel parks itself know that it is parallel parking? 1 'Mind' is a controversial term. Intelligence does not presuppose mentality, 1 Simply as we shall because a device is touted by advertisers as "intelligent" does not entail the existence of a conscious mind. The purpose of my argument is not to prove the impossibility computer, of a conscious but to prove the empirical untenability of ever knowing whether or not we have succeeded in producing one. Consciousness is not something we can detect by observing behavior-including brain behavior - alone; just because something seems sentient does not necessarily mean that it is. There is an explanatory gap between brain behavior and the phenomenon of conscious experience that cannot be solved using any extant philosophical or scientific paradigms. Hypothetically, if we were to ever fully grasp the nature of what consciousness is and how it arises and express it in the form of a coherent theory, it might be possible to ascribe such a theory to an artifact. transcend the explanatory gap and definitively My thesis is a three-pronged Even so, no conceivable test can prove the existence of a sentient mind. argument against the possibility of ever knowing for sure whether or not we succeed in building a self-aware, sentient, conscious computer. first tier is an explanatory discussion of computation and the Turing test. Here I address various arguments for computer minds and the theoretical artificial intelligence. continuously experiment The I also discuss the counterarguments underpinnings for Strong that cognitive science fails to rebut. I place special emphasis on Searle's Chinese Room thought and its implications for the possibility of machine sentience. The second tier is a discussion of what it means to be a conscious agent. I reject reductionism in any form as acceptable solutions to the mind-body problem because of the explanatory gap fatally separating first-person introspective accounts of consciousness and third-person observable 2 neuro-behavioral correlates. In the final section I support Colin McGinn's cognitive closure hypothesis by defending his view that the mind-body problem is inherently because a full understanding of consciousness Blade Runner and its antecedent the impossibility is beyond our epistemic insoluble limits. Using the film novel Do Androids Dream of Electric Sheep?, I aim to prove of ever differentiating between a conscious, "strong" Al and one that behaves as if it were conscious, but has no inner life whatsoever, 3 also known as "weak" Al. CHAPTER ONE MINDS AND/OR COMPUTERS WHAT IS COMPUTATION? Throughout definitions the course of my research, I have encountered of the word 'computer'. a bevy of disparate The Oxford Dictionary of Philosophy defines it as "any device capable of carrying out a sequence of operations in a defined manner" (Blackburn 2008). Computers permeate every aspect of modern society, from the microchips in hearing aids to massive parallel-processing supercomputer supercomputers. Recently, IBM built one such called Watson that competed on Jeopardy! against the two most successful contestants in the show's history. Watson seemed to have no problem comprehending the complex linguistic puzzles posed by Trebek, answering them in record time. As impressive IBM's creation may be, does it function at all like a human brain? Are these electronic operations equivalent to whatever phenomena are responsible for human thought? Numerous cognitive scientists, philosophers, computer scientists, and neuroscientists say yes (Carter 2007). Computationalism the functional representation would is a cognitive theory that posits that the mind is of the external world through the manipulation symbols; the mind is software within brain hardware. 4 of digital Before I discuss computationalism look at what a computer is. John Haugeland automatic formal system" (Haugeland in greater detail, it is important defines a computer as "an interpreted 1989, 48). To understand first decipher what each of its component terms signifies. tokens to be manipulated Take chess, for instance. to take a closer this definition, we must A formal system is comprised according to a set of predetermined of rules, not unlike a game. Before the game starts and regardless of who is playing, it is decided that a pawn can only move one (or two) spaces forward, and one space diagonally when attacking. Unless different rules are decided on before the game starts, these rules are set in stone. Instead of physical pieces, computer tokens are electronic and invisible to the eye. Formal systems like chess and checkers are necessarily digital. 'Digital' means discrete and precise while its opposite, 'analogue' means variable or nebulous. The alphabet is digital - A, B, C, D ... are static and discrete with no middle ground between A and B. The station preset buttons on car radios are digital, while an "old fashioned" analogue. frequency. dial is If button '1' is set to 98.5 MHz, pushing the button will reliably set that exact But when the driver of a 1983 Ford Bronco turns the tuner knob, it is effectively impossible to tune the exact same frequency every time. To my knowledge, all existing digital computers are binary, using strings of ls and Os as tokens. Formal systems are completely self-contained, meaning that the rules only apply to tokens within the system itself; 'black knight' seldom means 'movement restricted to two spaces by one space' outside a the realm of a chess match. As a result, the "outside world" is irrelevant. Chess can be played indoors, outdoors, on the moon, or underwater 5 with pieces made from plastic, gold, or elephant meat; the medium is irrelevant. All that matters is that the symbols pertain to the same system of rules, or syntax. As we shall see later, the idea of medium independence is extremely relevant to the field of artificial intelligence.2 So far we have learned that computers are digital, self-contained, and are syntactic. A formal system is automatic if it works or runs devoid of any external influence. In his discussion of automatic systems, Haugeland imagines a fanciful example: "a set of chess pieces that hop around the board, abiding by the rules, all by themselves" pencil that writes out formally correct mathematical any mathematicians" or "a magical derivations without the guidance of (Haugeland 1989, 76). A computer becomes automated when its legal moves are predetermined and carried through algorithmically. An algorithm works like a flowchart, a "step-by step recipe for obtaining a prespecified result" (Haugeland 1989, 65). Algorithms are designed to work indefinitely in finite time. For example, a programmer can design a procedure that alphabetizes a set of data. The algorithm used for this program can be used reliably with new sets of data ad infinitum. In the 1930s, computer pioneer Alan Turing theorized that a device could be built that manipulates symbols written on spools of tape into different states according to an algorithm. The Turing machine was not designed to be practical but to lay out the theoretical framework behind computation, "a mathematically (Penrose 1994, 65). Turing and mathematician idealized computer" Alonzo Church realized that, in principle, any algorithm could be iterated on a properly programmed Turing machine. 2 This concept will arise again in my later discussion of functionalism realizability. 6 Turing later and multiple conceived of the Universal Turing machine (hereafter referred to as UTM), which can simulate any arbitrary Turing machine. He wrote: The special property of digital computers, that they can mimic any discrete state machine, is described by saying that they are universal machines. The existence of machines with this property has the important consequence that, considerations of speed apart, it is unnecessary to design various new machines to do various computing processes. They can all be done with one digital computer, suitably programmed for each case. It will be seen that as a consequence of this all digital computers are in a sense equivalent. [Turing, quoted in Dreyfus 1994, 72] A UTM "can implement any algorithm whatever," which means that any modern computer is a UTM (Searle 2008, 88). The laptop in front of me is designed to implement software, cohesive sets of algorithms written by programmers. It can implement any software plugged into it.3 As we all know, UTMs are greatly useful and extremely versatile. Deep Blue, a chess playing supercomputer, In 1997, IBM's competed against world champion Garry Kasparov. In a thrilling upset, the advanced programming machine defeated Kasparov, showing the world how far computer intelligence had come. The two competitors played very different chess, however. Deep Blue used brute force processing - multiple computers running in parallel in order to power through the 1050 possible moves on the chessboard. Kasparov is a human, and as Hubert Dreyfus explained to me in a personal correspondence "[A chess program] needs to look at millions and millions of connections per second to do something that human beings do in an obviously entirely different way, that's true of the chess thing already[ ... ] Grandmasters look at two or three hundred at most" (Dreyfus 3 My MacBook Pro cannot run any program (i.e .. exe files), but it can surely implement any program in that it can be inputted and cause some sort of output. 7 2011). Granted, a grandmaster's chess abilities far surpass the average human, extremely unlikely that he "computes" such an astronomical but it is number of possibilities every turn. Is Kasparov's mind nothing more than a slow computer? Many theorists believe this to be so, as we shall soon see. But a crucial barrier separating Kasparov's mind from a UTM is consciousness, the poorly understood phenomenon that makes us sentient and sapient. Humans are conscious, no doubt, but there is no definitive proof that computers can or will ever achieve sentience. For starters, Kasparov has perceptual faculties: sight, touch, taste, etc., while Deep Blue does not (the mainframes weren't even in the room). However, apparatuses like electronic eyes, acoustic receivers, and pressure-sensitive tactile devices have existed for years. Hypothetically, the series of parallel processing computers comprising Deep Blue could be significantly shrunk and placed inside a humanoid robot, along with high-tech perceptual devices (call it DB2). It seems that the only thing separating Kasparov and DB2 is the latter's one-track mind, so to speak; its only function is playing chess. On the other hand, as Steven Hamad argues in Minds and Machines, "It is unlikely that our chess-playing capacity constitutes an autonomous functional module, independent of our capacity to see, move, manipulate, reason, and perhaps even to speak. The [Turing test] itself is based on the pre-emptive assumption that our linguistic communication capacity is functionally isolable" (Hamad 1992). DB2 might not even know that it's playing. Kasparov probably plays because he enjoys chess and has the desire to win, while DB/s 8 actions are dictated and bounded by its pre-programmed software. We cannot assume that the robot, despite its technological excellence, does anything because it wants to.4 Moreover, even though DB2 has perceptual capabilities, it does not necessarily have any phenomenal experience of anything it senses, since the input is nothing more than formal data. In short, is it right to say DB2 is conscious? If a mind is nothing more than computation, then there is no reason to believe DB2 doesn't have the potential to be. The next section explores the concept and plausibility of the computational nature of the mind. IS A MIND A COMPUTER? John Searle famously lectured: Because we do not understand the brain very well we are constantly tempted to use the latest technology as a model for trying to understand it. In my childhood we were always assured that the brain was a telephone switchboard. ("What else could it be?") I was amused to see that Sherrington, the great British neuroscientist, thought that the brain worked like a telegraph system. Freud often compared the brain to hydraulic and electromagnetic systems. Leibniz compared it to a mill, and I am told some of the ancient Greeks thought the brain functions like a catapult. At present, obviously, the metaphor is the digital computer. [Quoted in Kurzweil 2005] The way we understand the brain is correlated with the high-tech paradigm of the day, but the idea of the computational mind is nothing new. Thomas Hobbes contended that the mind is the brain's function as a calculating machine: "REASON ... is nothing but Reckoning (that is, Adding and Subtracting) of the Consequences of general the marking and signifying of our thoughts" (Hobbes 1904/1651). names agreed upon, for Our mind is a mechanical thinking machine, an infinitely complex flowchart of inputs and outputs. 4 This is a crucial theme in Blade Runner, which I will discuss later. 9 In mid-201h century Britain, Alan Turing set out to show that Hobbes was on to something. imitation game," an original thought experiment He designed "the that attempted to solve the puzzle. The game is simple. Three participants, a man, a woman, and an irrelevantly gendered interrogator participant (A, B, and C, respectively) are placed in three rooms. The identity of each is unknown. They can communicate with each other freely through Teletype, keeping revealing factors like tone of voice or handwriting a mystery. The object of the game is for C to determine the sexes of A and B by asking each of them questions. catch is that A attempts to mislead C by providing false information. The At the same time, B attempts to convince C of the truth. If A successfully fools C, then A "passes" the test. Next, replace A with a computer. Band C remain human and don't know that A is a machine. If A, a computer, manages to convince C, a human, that it is in fact a human female, then Chas met the criteria for intelligence (Turing 1950). TURING TEST EXTPA CREDIT: Turing held that if a human cannot distinguish a computer from a person, then that computer can think and thus has a mind. His CONVINCE THE EXAMINER THAT ti_CT; A COMPUTER. conclusion helped give rise to computationalism, or the computational theory of the mind (henceforth CTM), the idea that the brain is a computer, and mental states and consciousness somehow arise from the formal processing. According to Turing and Church, "anything that can be given a precise enough characterization as a set of steps can be simulated on a digital computer," algorithmically, a computer can implement (Searle 2008, 87). Figure 1 From xkcd.com so if a mind can be transcribed it, thus giving rise to machine consciousness If this is possible, the brain is just a biological UTM, or "wetware." 10 The computational theory of mind is built on a theoretical framework called functionalism, the idea that a mental state is defined by its function. As an analogy, a carburetor is a device whose primary function is to mix fuel and air in order to cause combustion. A carburetor can be made of any materials, just as long as that as it fulfills that function; it is solely defined by its function.5 to this theory. Mental states would be no different according The functionalist argues that a mental state is nothing more than "its function in mediating relations between inputs, outputs and other mental states" (Carter 2007, 45). For instance, a pain is defined by the role it plays. Some would argue that pain is nothing more than tissue damage and C-fiber stimulation.6 To the functionalist, a mind is a Turing machine. Accordingly, a pain is an input (tissue damage), a mental state (discomfort, anxiety, etc.), and an output (behavior). The functionalist holds that any system that can iterate these conditions is in pain, regardless of composition. biologically The token example is the and chemically different Martian who is in a pain state when it meets the requisite pain conditions. inputs, pain-alleviating this functional Pain "mediates relations between characteristic pain-inducing reasoning and behaviour" (Carter 2007, 45). Anything that serves role is a pain ex vi termini. Computationalism's aim is "fleshing out these mediating relations question are held to be computations" - the relations in (Carter 2007, 95). This is not to say, however, that the mere operation of formal systems is sufficient for consciousness. If this was the case, a simple tape-based Turing machine is itself a thinking thing, an unsettling thought indeed. 5 The carburetor example is an adaptation of an argument used by Jerry Fodor (Psychological Examinations, Random House 1968). 6 Also known as type-physicalism, as we shall see. 11 (Although some argue that there is "something it is like" to be an inert artifact, as will be discussed later, for now we shall assume the falsity of this claim). sentience arises from the interplay between our computational If CTM is true, however, brains and our sensory organs; the mind is the result of the operation of software while the brain and body are the various pieces of hardware. Granted, CTM proponents realize that there might be more to cognition than just syntactic manipulation. Just as long as "mental states are at least computational states," CTM holds (Searle 2008, 87, my emphasis). Functionalism7 and CTM turn on multiple realizability, the understanding that 11a single mental kind (property, state, event) can be realized by many distinct physical kinds" (Bickle 2008). If every minute detail of the brain, down to the very last neuron, was reenacted with, say, beer cans, would a mind emerge? Douglas Hofstadter {2000) wrote "A Conversation With Einstein's Brain," a thought experiment that explores this idea. Lets say that one microsecond before his death, every detail of Albert Einstein's brain was copied exactly as it was into a book. Each of the hundreds of billions of pages corresponds to a single neuron and information about its connections (synapses). When the real life Einstein heard someone ask a question, he perceived the vocal utterances with his ears, which then affects the auditory neuron structure in the brain.Theoretically, one could ask the disembodied Einstein a question by correlating relevant information about how his brain registers each particular tone and trace consequent chain reaction of synapses, which eventually lead to Einstein uttering a response to the question. Perhaps a machine could act as a middleman, speeding up the page turning. Is the brain-book a mind?: "I'm losing a 7 Here I refer to certain variations of functionalism, but not all. See Lewis 1991. 12 clear sight of who 'I' is. Is 'I' a person? A process? A structure in my brain? Or is 'I' some uncapturable essence thatfeels what goes on in my brain?" {Hofstadter 2000, 445) Hofstadter's point is that the 'I' all of us are intimately familiar with is perfectly conceivable as a set of algorithms, realizable in any implementing medium, inasmuch as the unfathomably complex programming is good enough. If functionalism is correct, the mind-body problem is solvable with enough time and energy spent on computational models of the brain. The ultimate goal of artificial intelligence is to fully comprehend the human mind and create a unified theory of consciousness. If such a theory is ascribed to an artifact, consciousness could presumably arise. I do not believe that consciousness is or is not multiple realizable. My point is an agnostic one; there is no test conceivable that can definitively prove whether consciousness has been achieved. The Turing test, "though essential for machine modeling the mind, can really only yield an explanation of the body" (Hamad 1992). Behavior alone is not sufficient for mentality. IS A COMPUTER A MIND? If CTM is true, any UTM can implement a syntactic mind. Turing believed that the mind is syntactic, and that the imitation game was an indicator of the presence of mind. But what if CTM is false? If the human mind is not computational, is it still possible for a computer to be conscious? Before we delve into the million-dollar question, I must first clarify some key concepts and assumptions. Various writers across the gamut of philosophy and science have different definitions of words like 'mind', 'consciousness', 'thinking', and 13 'intelligence'. Let me clarify my own usage. I hold that the possibility for consciousness is a necessary for mentality. For example, in a deep sleep one is decidedly unconscious, but dream states are a form of consciousnesssince there is still some sort of phenomenal experience taking place. But one cannot dream or think in the first place if there was never any conscious experience to begin with. Thinking, or introspecting, is the active process of exploring the contents of one's mind. Dreaming" is not thinking because it occurs passively. Intelligence, on the other hand, does not presuppose mindedness. Cognitive scientist Steven Pinker defines intelligence as "the ability to attain goals in the face of obstacles by means of decisions based on rational (truth-obeying) rules" (Pinker 1997, 62, my emphasis). The term 'Artificial Intelligence' is then, according to his definition, a synthetic entity capable of following rules in order to attain goals. When I attempt to control my car over an icy stretch, the anti-lock breaks automatically engage, helping me avoid danger. In this case, the ice is an obstacle and the braking system follows a set of rules in order to decide when to engage, accomplishing the ultimate goal of car and driver protection. By Pinker's lights, the brakes on my car are intelligent. To say that my brakes are conscious or have a mind, on the other hand, is a much stronger, and tenuous, assertion. So what kind of artifact, if any, can think? In 1956, computer scientist Allen Newell, economist Herbert Simon, and systems programmer J.C. Shaw shocked the scientific community with what is considered the first artificial intelligence program. "Logic Theorist," as it was called, managed to prove 38 of the first 52 theorems from Whitehead and Russell's Principia Mathematica (Copeland 1993, 7). 8 Barring lucid dreaming, in which the dreamer is in active control of his or her situation. 14 This may not sound as impressive as "I am C-3PO, human cyborg relations," enough for Al researchers that a thinking machine might be possible. but it was proof Two years later, Simon claimed that within ten years, computers will defeat the world champion chess player {he was off by about thirty years), discover and prove a new mathematical theorem, and, most importantly, that "most theories in psychology will take the form of computer programs," advocating CTM {Simon and Newell 1958, 6). By 1961, Newell and Simon were ecstatically optimistic, writing: It can be seen that this approach makes no assumption that the 'hardware' of computers and brains are similar, beyond the assumptions that both are general-purpose symbol-manipulating devices, and that the computer can be programed to execute elementary information processes functionally quite like those executed by the brain. [Newell and Simon 1961, 9] Al innovator Marvin Minsky concurred with Newell and Simon not long after, arguing that brains are "meat machines" capable of duplication (Dreyfus 1994, 252). The search for the synthetic mind was well underway. Fast-forward to present day. There are no sentient robots roaming the streets or running businesses. Critics of CTM and functionalism argue that Al research has been barking up the wrong tree. John Searle leads the attack on computational arguments against "strong Al," the concept that an "appropriately Al with his programmed computer really is a mind, in the sense that computers given the right programs can be literally said to understand and have other cognitive states" {Searle 1980). If strong Al holds, mental states are invoked at least in part through syntactic computation. Strong Al assumes the truth of CTM, but CTM does not necessarily assume strong Al. "Weak Al" proponents hold that the brain is not necessarily computational, 15 but the study of computationalism as a psychological paradigm is scientifically useful. The big difference between strong and weak Al proponents is that while the former is convinced that the right series of computations invokes an actual mind, the latter holds that minds can be simulated, but not duplicated. This distinction is crucial. If strong Al is correct, it is theoretically possible to synthesize a mind. Any UTM could potentially instantiate the unimaginably complex array of naturally occurring algorithms resulting in a sentient artifact. But is syntax enough for a mind? Searle famously argued in the negative with the Chinese Room thought experiment: Imagine a native English speaker who knows no Chinese locked in a room full of boxes of Chinese symbols (a data base) together with a book of instructions for manipulating the symbols (the program). Imagine that people outside the room send in other Chinese symbols which, unknown to the person in the room, are questions in Chinese (the input). And imagine that by following the instructions in the program the man in the room is able to pass out Chinese symbols which are correct answers to the questions (the output). The program enables the person in the room to pass the Turing Test for understanding Chinese but he does not understand a word of Chinese. [Cole 2009] The Chinese speaker outside the room (call her Mei) assumes that the person inside (call him John) is fluent, and for good reason. John is so adept at using the manual and shuffling through the boxes that he could quickly respond to any of Mei's input, regardless of complexity. For all intents and purposes, Mei is communicating with a fellow fluent Chinese speaker. But John is not fluent; "Chinese writing is just so many meaningless squiggles" (Searle 1980). All he does is manipulate those squiggles according to a set of predetermined rules, or a "script." 16 It is contradictory to say John understands Chinese when he doesn't actually understand a word of it. Searle comes to a profound conclusion: mental states have semantic (meaningful) content, and syntax alone is not sufficient for semantics, so mental states cannot be syntactical. The purely syntactical computer can emulate the mind, but not duplicate it. He writes: You can simulate the cognitive processes of the human mind as you can simulate rain storms, five alarm fires, digestion, or anything else that you can describe precisely. But it is just as ridiculous to think that a system that had a simulation of consciousness and other mental processes thereby had the mental processes as it would be to think that the simulation of digestion on a computer could thereby actually digest beer and pizza. [Searle 2008, 68] Searle finds it absurd to think a mind can be caused in this sense. Simulation duplication. is not It might be possible in practice to simulate human intelligence on a computer, "just as scientists routinely simulate everything from hurricanes and protein synthesis to traffic jams and the black market of Albania," but the result is only an illusion of the original (Haugeland 1985, 112). Searle's point, that syntax is not sufficient for semantics, has become a mantra amongst opponents of strong Al and CTM. The man in the room passes the Turing test,9 but he lacks any semblance of semantic understanding. meaningfulness. 'Understanding' is contingent upon When I understand a concept ('democracy', for example), the idea becomes an object of my conscious experience. 9 I know what democracy means once I Some might argue that the man in the Chinese Room cannot actually pass the Turing test if asked questions like "What is time?" (see Ben-Yami 1993). I don't see the value in such claims. There is no good reason to presume that the Chinese-English manual can't be programmed to provide a conversationally acceptable answer like "the unextended dimension of existence that determines the sequential nature of events." 17 attach to it some semantic content. representation thereof. John has no intentional To understand something requires mental This power of representation is generally known as intentionality. states about the actual meanings of the characters. for him to think about the semantic content of the "conversation" It is impossible with the Chinese- speaker. Of course, he can think, "well, maybe I'm talking about kittens," but such a thought is a complete shot in the dark with no basis in understanding. a Turing machine with the manual as its program. The Chinese Room is The argument shows that the Turing test does not test for a mind, but simply for the existence of computational processes. The Chinese Room represents the syntactic processes by which computers "think." Computers as we know them are programmed to do exactly what John does (albeit at lightning fast speeds), devoid of intentionality, understanding, and semantic content. A computer may be able to simulate a mind, but not duplicate it. DB2, our perceptually equipped, chess-playing robot does not have mental content at all, only brute-force As might be expected, Searle's argument has garnered an enormous criticism. data. amount of Here I will address a few of the strongest instances. The response typically known as the "Systems Reply" admits that the person in the room (John) in fact does not understand Chinese, but that the entire system of inputs and outputs does understand. John acts like the central processing unit of a computer; only one part of a larger whole. The room itself, so to speak, is the fluent Chinese speaker with whom Mei communicates. Searle rebuts by claiming that there is no reason to claim that John can't just memorize the entire manual and learn to write and recognize every symbol, all while working outside. 18 The whole system is now completely internalized, but John remains completely ignorant of any semantic understanding. The "Robot Reply" admits the validity of the Chinese Room but suggests an alternative thought experiment. Instead of manipulating formal inputs and outputs according to a script, we place a computer inside a mobile robot who, like our 082, is equipped with perceptual apparatuses like microphones, cameras, touch sensors, etc. The robot uses its senses to learn the world around it, just like a chlld.:" The robot is free to learn Chinese independently by forming a causal connection with the world. Searle writes in response: Suppose that instead of the computer inside the robot, you put me inside the room and, as in the original Chinese case, you give me more Chinese symbols with more instructions in English for matching Chinese symbols to Chinese symbols and feeding back Chinese symbols to the outside. Suppose, unknown to me, some of the Chinese symbols that come to me come from a television camera attached to the robot and other Chinese symbols that I am giving out serve to make the motors inside the robot move the robot's legs or arms. It is important to emphasize that all I am doing is manipulating formal symbols: I know none of these other facts. I am receiving 'Information' from the robot's 'perceptual' apparatus and I am giving out instructions' to its motor apparatus without knowing either of these facts. [Searle 1980] Even though the external world is invoked in this situation, no semantics the totality of the data is still formal. can arise because It only adds complexity to the man-in-the-room's task. Although the Chinese Room argument has profound implications across the spectrum, ultimately Searle begs his own question about the nature of the mind. He 10 This method of information processing is known as bottom-up Al, in contrast with topdown cognition, in which "it has been constructed according to some well-defined and clearly understood fixed computational procedure ... where this procedure specifically provides a clear-cut solution to some problem at hand" (Penrose 1994, 18). 19 contends that syntax is not sufficient for semantics and that the best computers can is simulate consciousness {weak Al). However, he fails to address exactly what makes our minds so special that they have semantics while computers cannot. mind is fundamentally He assumes that the different than a symbol processing machine but never elaborates on the nature of that difference. His argument is based on three premises: Pl: Programs {software) are syntactic and are thus self-contained P2: Minds have semantics and are thus not self-contained P3: Syntax is a not a sufficient condition for semantics What's missing missing is the second half of P2. It isn't enough to simply say that our thoughts have meaning and thus cannot be syntactical. What gives our thoughts meaning in the first place? How does visible light detected by the eye or vibrations of air in the ear canal mean anything to a mind? Perhaps meaning and intentionality are just illusions caused by our brains but are really semantically empty (see Dennett 1991a). Or perhaps we aren't conscious at all, as some persistent eliminativists contend. Searle's argument is poignent but inconclusive. The ultimate aim of this paper is to show that the Searl e's semantics will never be fully understood. The missing ingredient, why our mental states are meaningful, is beyond the scope of human comprehension for reasons I will address in the next two chapters. Before we proceed, it is important to address one final criticism. The Other Minds response asks how it is possible to know if anybody understands anything besides judging by their behavior. doesn't? If an android can fool anyone into thinking it understands, who's to say it Searle's response goes as follows: 20 The problem in this discussion is not about how I know that other people have cognitive states, but rather what it is that I am attributing to them when I attribute cognitive states to them. The thrust of the argument is that it couldn't be just computational processes and their output because the computational processes and their output can exist without the cognitive state. It is no answer to this argument to feign anesthesia. In 'cognitive sciences' one presupposes the reality and knowability of the mental in the same way that in physical sciences one has to presuppose the reality and knowability of physical objects. [Searle 1980] As sentient beings, it is impossible to deny the fact that we directly experience our own consciousness (although the eliminativists and others have tried). We also know that computation Using our own consciousness as a reference, we can exist without cognition. assume that other people have minds and phenomenal experience of the world but have no reason to assume that they are computers. This approach to the problem of other minds is known as the argument from analogy. However, the argument from analogy falls short. Just because something acts like a sentient being by no means necessitates consciousness. We anthropomorphize unconscious objects and nonhuman animals all the time, sometimes attributing complex thoughts and emotions to entities that are obviously incapable of them. We think cats are happy when they grin because that's what humans do (figure 2). In reality, the cat Figure 2 Who are we to say this cat is happy? might be miserable or scared, but the tendency to humanity or sentience by analogy leads us astray. My point is that analogy alone is not enough for the presence of other minds. This plays a major role in the advancement of artificial intelligence theory and development. A major barrier standing in its way is the inability to ever know, with full certainty, whether 21 or not a nonhuman entity is sentient. In the next section, I will discuss this uncertainty problem and its implications in philosophy, cognitive science, and Al research. 22 CHAPTER 2 THE ENIGMA OF FAMILIARITY THE HARD PROBLEM At this moment, I am sitting in a cafe typing on my laptop. I hear the distinctive sound of fingers striking a keyboard over chatter of other customers and ambient piano music. I shiver when the door opens, sending a chill through my body. The pastry in front of me smells of butter and cinnamon, and I can feel myself salivating as I write about it. I feel cheerful and jovial because yesterday I received an offer for a wonderful job for next year. have these experiences because I am sentient and aware, two defining characteristics of consciousness, a barely understood but ubiquitous psychological phenomenon immediately familiar to all of us. As far as I know, humans have always been conscious and always will be. Death and coma notwithstanding, consciousness is a necessary part of human existence. For this reason, consciousness as a concept is ironic. It's been around for at least as long as we have and without it, scientific study and epistemic progression could never happen. But in spite of countless hours spent in laboratories and classrooms, a cohesive theory of consciousness is yet to be discovered. In fact, so little is actually known about it that The International Dictionary of Psychology defines conscious as such: 23 Consciousness. The having of perceptions, thoughts, and feelings; awareness. The term is impossible to define except in terms that are unintelligible without a grasp of what consciousness means. Many fall into the trap of equating consciousness with self-consciousness - to be conscious it is only necessary to be aware of the external world. Consciousness is a fascinating but elusive phenomenon: it is impossible to specify what it is, what it does, or why it evolved. Nothing worth reading has been written on it. [Sutherland 1996, 95, emphasis added] It is rare to find such candid language in a reference book. But how could it be true? Bertrand Russell agrees: "The sciences have developed in an order the reverse of what might have been expected. What was most remote from ourselves was first brought under the domain of law, and then, gradually, what was nearer: first the heavens, next the earth, then animal and vegetable life, then the human body, and last of all (as of yet imperfectly) the human mind" (Russell 1961). The boundaries of human knowledge are expanding at an exponential rate, but the nature of what is most familiar to all of us remains vague and misunderstood. That is not to say that volumes worth of literature have not been written about it. Cognitive science is devoted to the subject, synthesizing psychology, neuroscience, philosophy, linguistics, anthropology, interdisciplinary and computer science into one contiguous, area of study. Theories of consciousness abound from every field, and great progress has certainly been made. In particular, the systematic mapping of brain function and behavioral correlation before. But the has allowed us to know more about the brain than ever mind still eludes us. Although science can explain how the senses function, why this cinnamon bun smells so damn good is a complete mystery. The inner experience of the mind is private and subjective, knowable only by the knower and impossible to describe in purely objective terms. I can describe how I'm feeling to the best of my ability, 24 and you can try to relate using your own experience, but the two sets of data are irreconcilable; we can't know if they are the same. Individual conscious experience is so distinct and accessible to us, yet so unbelievably opaque and foreign to another. This is a crucial aspect of the mind-body problem in philosophy. Descartes famously contended cogito ergo sum, I think therefore I am, demonstrating the perspicuous, undeniable existence of his own mind. He held that the physical world and mental world were dualistic, comprised of two discrete substances, physical matter (res extensa) and the non-extended, material soul or mind (res cogitans). Despite the currency of his words, the cogito argument has become the butt of countless philosophy jokes, offered to freshman as an exercise in argument analysis and refutation. Philosophy and science have become inextricable, and there are no reputable empirical theories of the soul. Theology aside, modern thought favors a naturalistic materialism, the idea that all that exists is matter and all phenomena including consciousness can be reducible to phvsics.i! Materialism has been successful in explaining the natural world, but inevitably fails to adequately account for subjective conscious experience. A vast amount has been written both in defense of and in opposition to materialistic consciousness,but the solution to the problem remains largely unsolved. Typically, the problems of consciousnessare split into two issues.12 The Easy Problem is well under control by empirical science. It considers topics involving brain behavior, the integration of information, and stimulus response, all within the realm of 11 Of course this is a very narrow and underdeveloped view of materialism. My own position is in favor of a materialism in which consciousness is irreducible. 12 This is David Chalmers' idea, but it is widely accepted as a useful way to characterize consciousnesstheory. 25 empirical study. The hard problem, as it is known, is much more of a challenge. David Chalmers describes it as the problem is something sense - or is something it is something of experience. Human beings have subjective experience: there it is like to be them. We can say that a being is conscious in this phenomenally conscious, as it is sometimes put - when there is is like to be that being. A mental state is conscious when there it is like to be in that state. [Chalmers 2002a] What gives 'red' the quality of redness? After all, 'red' is nothing more than electromagnetic radiation of a particular wavelength and frequency, detected by the eye and perceived by the brain. If everything in the universe is reducible to physics, then the experience of 'red' is nothing more than that. But this response is terribly unsatisfying. Our senses have a phenomenal quality experienced by the mysterious "inner life" that frustrates scientists to no end. Some choose to ignore the problem altogether, dismissing consciousness as a trivial intuition: The "intuition" at work here is the very raison d'etre of the problem of consciousness. The only consistent way to get around the intuitions is to deny the problem and the phenomenon altogether. One can always, at least when speaking "philosophically," deny the intuitions altogether, and deny that there is anything (apart from the performance of various functions) that needs explaining. [Chalmers 1996, 110] This approach is a mistake, of course, because every normally functioning human has an inner life and disregarding it will lead us nowhere. This mysterious quality is at the core of the hard problem of consciousness,and it is the modern day expression of the mind-body problem. Those who take a purely scientific approach may claim that consciousness is reducible to its constituent physical properties. Emergentism, championed by Samuel 26 Alexander13 and C. Lloyd Morgan14 in the early zo" century, is the idea that consciousness emerges from fundamentally simple brain processes. To illustrate, take water and break it down to its constituent parts, oxygen and hydrogen. Without these two elements, water cannot be. However, water is not an inherent property of oxygen or hydrogen. If there was a world where oxygen and hydrogen exist as discrete elements but can never interact, that world is necessarily devoid of H20. So oxygen and hydrogen are inherent to water, but not vice versa. Similarly, consciousness emerges from neural processes even though consciousnessis not an inherent property thereof: To apply emergentism theory to a universal scale is to accept physical ism. Type physicalism (aka identity theory of mind) is a materialistic position that holds that all mental states and processes are equivalent to brain behavior and, as a result, consciousness is reducible to physics. U.T. Place argued that consciousness might be a pattern of brain activity that could be correlated with certain brain processesand that the elusiveness of introspective observations could just be a "phenomenological fallacy." Introspection is just brain behavior, just like lightening is just a high voltage, short duration electric charge (Place 2002). Herbert Feigl, another leading identity theorist, concurred, claiming "the states of direct experience which conscious human beings 'live through,' and those which we confidently ascribe to some of the higher animals, are identical with certain (presumably configurational) aspects of the neural processes in those organisms" (Feigl 2002, 69). 13 See Alexander's Space, Time, and Deity (London: Macmillan, 1927) See Morgan's Introduction to Comparative Psychology, 2nd ed., rev. (London: Walter Scott, 1903) 14 27 Mental states are contingent upon specific brain states. In other words, "there is no mental difference without a physical difference" Is consciousness reducible? (Nagel 1998). Many materialists of different stripes argue in the affirmative. After all, if mental states aren't physical, what could they possibly be? Dualistic philosophy allows for a mental "stuff" separate from physical substance, fallen out of fashion in the face of scientific development. but dualism has Most contemporary philosophers of mind rely on physics and neuroscience for empirically proven evidence to support argumentation. As a result, the study of consciousness has become inextricably science. However, I reject dualism but deny the validity of physicalism about consciousness. Subjective conscious experience is knowable only to the subject, and reduction public data is impossible. linked to to objective, The best we can do is correlate between brain states and behavior. The next section analyzes some major arguments against physicalism in order to show its inadequacy as a coherent theory of consciousness. ARGUMENTS FROM INEFFABILITY In 1866, Thomas Huxley wrote: "What consciousness is, we know not; and how it is that anything so remarkable as a state of consciousness comes about as a result of irritating nervous tissue is just as unaccountable his lamp" (Huxley 1866, 193). philosophical investigation, as the appearance of the Djinn when Aladdin rubbed Despite 150 more years of rigorous scientific and we still have no idea. Even Einstein is reputed to have admitted, "Science could not give us the taste of soup (Dennett 2002, 230). At the forefront of the 28 hard problem is the concept of qualia, 15 the technical term for the phenomenal, felt qualities we experience through the senses. Qualia like the taste of an apple, the roughness of sandpaper, the rumbling of a basstone from a subwoofer, and the smell of a musty attic can be experienced by anybody with normal perceptive abilities, but the mystery lies in their subjective character. A piece by Stravinsky might sound like heaven to Jones and cacophony to Smith. Both have normally functioning ears and a deep appreciation for classical music. There is the matter of personal taste, of course, but it also must be considered that Jones and Smith have different experiences of the same tones. Middle C is always 262 Hz, but only I know what middle C sounds like to me.16 The problem is that our explanatory capacity is limited by the boundaries of natural language. !.SHt t•; ·.:>>Rr>..W.>t. \'tOW, SM\;U .."'S ~\:: 00 \o'lc::.<:1'\ l~E.. 11\J'I Y\'i::. crw.i: THrc..~? I '3~L1;> llt>.~t. "-"IM.,t...'S ~'<{ti ·~LI> G~SC.\l',~.£ H"-~£. ' ·~o-; l'OI<:. ~ti.LS. Figure 3 Borrowed from Chalmers 15 115 A. \.lliL'i:. 811\JHI<.'{. S\l~ THE \.OW ~\l!l,\\)'1\'f M'l"tcTS i\u;\. 1996 As Chalmers 1996 notes, qualia, experience, phenomenology, phenomenal, what it is like, subjective experience all refer to the phenomena privy to the conscious mind. Only grammatical differences set them apart. 16 The inherent differences between Smith and Jones return later in my discussion of inverted qualia. 29 In figure 3, Hobbes the tiger describes his olfactory experience of fire with words like snorky, brambish, and brunky. To his friend Calvin, such terms are as meaningless as the squiggles and squoggles in the Chinese Room. Only Hobbes knows precisely what they mean because only he has direct access to the contents of his inner experience. Even another verbal, sentient tiger with the same sensory faculties cannot know if its experience of "snorky" is the same as Hobbes' because verbal explanation is inadequate. concepts like 'green' through ostensive definition, the repeated pointing to green objects by others and labeling them such. I can identify green objects because instilled in me through ostensive definition. We learn 'green' has been Hobbes thinks fire smells snorky because he taught himself to identify a particular odor with that term. No one, even his best friend Calvin, can ever know exactly what Hobbes means when he calls something 'snorky' or 'green' or anything for that matter because of the private, enclosed, and ineffable nature of conscious experience. Or, to use a literary example, take the following dialogue from Brideshead Revisited between the narrator Charles Ryder and his friend Sebastian Flyte as they drunkenly describe the taste of a wine. "It is a little, shy wine like a gazelle." "Like a leprechaun." "Dappled, in a tapestry meadow." "Like a flute by still water." " A wise old wine." " A prophet in a cave" " A necklace of pearls on a white neck." "Like a swan." "Like the last Unicorn." [Quoted in Lanchester 2008] 30 Boozy shenanigans aside, Ryder and Flyte certainly experience something that compels them to describe the wine in such comically distinctive ways. However, even if I do know what pearls on a white neck actually tastes like (assuredly, I do not), I will never know what it is like for Ryder to experience such a sensation. Much of Wittgenstein's philosophy is couched in the limitations of natural language. In §293 in Philosophical Investigations he writes: Suppose everyone had a box with something in it: we call it a 'beetle'. No one can look into anyone else's box, and everyone says he knows what a beetle is by looking at his beetle. Here it would be quite possible for everyone to have something different in his box. One might even imagine such a thing constantly changing. But sup- pose the word 'beetle' had a use in these people's language? If so, it would not be used as the name of a thing. The thing in the box has no place in the language-game at all; not even as a something: for the box might even be empty. No one can 'divide through' by the thing in the box; it cancels out, whatever it is. [Wittgenstein 1974] "Wittgenstein's Beetle," as this thought experiment is commonly referred, shows that our explanatory capacity is limited by natural language. The beetle represents the picture we form in our minds when we think about or perceive something. Everybody's beetle might be completely different, or they might all be the same. The point is that no one will ever know for sure. Simply describing what it is to someone else is insufficient for intelligibility because the only reference point available is my own beetle. I have a beetle and you have a beetle, but mere linguistic congruence alone does not give rise semantic equivalence.17 Another treatment of this issue is known as the inverted spectrum argument, dating back to John Locke's18 1ih century empiricism. If two people with normally functioning 17 It should be noted that Wittgenstein's philosophy of mind is not in line with my own. However, his Beetle example is still pertinent to my argument. 18 See Locke 1689/1996, 169. 31 visual faculties observe a fresh strawberry, both inevitably describe it as red. there is the distinct possibility of the qualia of both observers to be completely in this case, inverted. Who is to say Smith's red is not Jones' green? entire life perceiving red qualia as green qualia, for him, green However, different, or If Jones spends his is red. In theory, remapping whatever neurological apparatus pertain to color perception can stimulate this situation, but for the sake of argument, the logical possibility of inverted qualia is sufficient. The question is whether it is empirically possible to prove veracity of the situation. I argue later that it is not. In same vein as the inverted spectrum, Frank Jackson and Thomas Nagel argue for the necessity of qualia and dispute physicalism with what is known as the knowledge argument. In his famous thought experiment, Jackson supposes that Mary, the world's foremost expert on neuroscience, spends her entire life in a monochrome room and has never been exposed to any color other than black, white, and shades of gray. She has always worn opaque, black stockings, a white lab coat, and black gloves. Even though she knows more than anyone else about the physical processes behind color perception, there is nothing Mary can do (short of taking drugs) to experience what color is like. 'Blue' means nothing more than a 475nm wavelength, and will always mean that unless she ever leaves the room. The physical facts behind a particular aspect of subjective experience are not the same as the experience itself. When Mary is released and actually sees the blue sky for the first time in her life, she learns what subjective color experience is. Even though she knew every physical fact there is to know about color and human perception, her knowledge was incomplete. Mary learns something an objective physical fact could not teach her. Jackson 32 shows that "physicalism leaves something out" something unaccounted for by objective data (Jackson 1982). A real-life take on Jackson's conclusions comes from a 2008 article in The New Yorker. Here are two molecules, identical in every way except that they are inverted on the y-axis: 0 0 Figure 5 Figure 4 Any competent chemist can fully understand the molecular structure of both examples, and modern science can tell us everything else we could possibly know about them. Well, everything except for the fact that Figure 4 smells like spearmint, while Figure 5 smells like caraway, two completely disparate and distinct odors. When it comes to phenomenal experience, we are baffled: "When scientists create new molecules in the laboratory, they may know every detail of a molecule's structure yet have no clue about what it will smell like." Physical ism falls flat when it comes to the hard problem; it "risks missing the fundamental truth of all smells and tastes, which that they are, by definition, experiences" (Lanchester 2008, 121). Coming to the same conclusion, Thomas Nagel's take on the knowledge argument tries to imagine what it is like to be a bat. As we know, the sensory faculties of bats and 33 humans are much different. Bats rely on echolocation or sonar, emitting high-frequency shrieks and perceiving the world around them through the reflection of the sound. Echolocation is incomparable to any human faculties and is thus unimaginable; "though clearly a form of perception, it is not similar in its operation to any sense that we possess, and there is no reason to suppose it is subjectively like anything we can experience or imagine" (Nagel 1979). I suppose I could blindfold myself and try to find my way by yelling and carefully listening for an echo, but it still wouldn't be the same since bats don't hear like we do. Furthermore, to imagine having webbed arms, eating insects, having poor vision, or spending the daytime sleeping upside down is to imagine bat behavior from the point of view of a human, not bat experience, and these concepts are surely not the same. In short, it is impossible to conceive of what the inner life of a bat might be like. Nagel chose bats because they are mammals and are thus relatively close to humans phylogenetically. 19 He assumes, controversially, that, like humans, bats are on some level conscious and aware of their own being.20 However, while it isn't difficult to imagine what it might be like to be another person,21 are strange enough to be utterly unfathomable. bats I know what it feels like to hear Bohemian Rhapsody, so it is within my realm of imagination to think Figure 6 about what it might be like for someone else to hear the same song. Humans are 19 See Janzen 2006; Hacker 2002 Nagel does not mean to say that bats fly around soliloquizing, of course. He merely means that there must be something it is like to be a bat in a comparable sense that there is something it is like to be a person. 21 Of course such a feat is impossible to perform accurately; subjective conscious experience is private and inaccessible. But considering how all (normally functioning) humans share the same perceptual and cognitive capacities, imagining what it might be like to be 20 someone else is possible. 34 cognitively closed to the possibility of imagining sonar perception. Thus, we are inextricably tied to our own experiences and utterly incapable of conceiving of anything beyond our perceptual horizon. Unless the Batboy story in the Weekly World News (Figure 6) is true, no human can imagine the experience of bat qualia. The bat brain, on the other hand, is perfectly understandable, "a domain of objective facts par excellence - the kind that can be observed and understood" within the limitations of our own perceptual horizon (Nagel 1979, 172). Science informs us about our own brains and bat brains in the same ways. We can analyze brain states by correlating brain activity with patterns of observable behavior; such methodology is the backbone of neuroscience. experience. However, neural correlation ignores qualia, the hallmark of conscious Exactly what it is like to be a bat is an ineffable property knowable only by a bat, and no amount of raw data about brains and behavior can change that. Nagel contends: Without consciousness the mind-body problem would be much less interesting. With consciousness it seems hopeless. The most important and characteristic feature of conscious mental phenomena is very poorly understood. Most reductionist theories do not even try to explain it. And careful examination will show that no currently available concept of reduction is applicable to it. Perhaps a new theoretical form can be devised for the purpose, but such a solution, if it exists, lies in the distant intellectual future. [Nagel 1979, 166] In short, the subjective character of consciousness is an unavoidable obstacle for any discipline involving the study of the mind. He challenges physical ism to stop skirting the qualia issue by conceding ignorance or giving up entirely. Nagel's point is that the peculiar and ineffable nature of first-person incompatible with any sort of objective standpoint. 35 experience is Nagel blames these disharmonious standpoints for causing "many of the basic problems of philosophy: we are torn between these two ways of seeing the world and cannot satisfactorily conception of things," and current cognitive-philosophical integrate them into a coherent theories are inadequate means of reconciliation {McGinn 1997b, 89}. Another way to think about how to explain qualia and consciousness is by imagining a being completely devoid of phenomenal experience. David Chalmers does so by talking about zombies. Not Hollywood zombies, per se, but philosophical zombies (p-zombies) who have no appetite for brains and look just like regular people. P-zombies are exact physical copies of people, down to every last atom, neuron, and synapse. My p-zombie twin is indistinguishable from me not only in appearance but also in behavior. My own mother wouldn't be able to tell us apart. The key difference between us is that my p-zombie twin is completely devoid of conscious experience. Even though he (it?) displays what seems to be conscious behavior, there is an utter lack of phenomenology. When my p-zombie twin enjoys a freshly baked chocolate chip cookie (my favorite}, all the same physical processes occur when I eat the same cookie. But there is no actual taste experience occurring, only overt behavior. There is nothing it is like to be a p-zombie. This is a difficult position to defend, however, and there are compelling counterarguments standing in the way. For starters, any molecule-for-molecule conscious person would surely be conscious. Unless some sort of panpsychism22 Cartesian dualism is correct, it is undeniable that the configuration molecules comprising a person must have some determining 22See Nagel 1979. 36 replica of a or and quantity of effect on the existence of conscious experience. If consciousness is contingent on the physical, then the x number of molecules in orientation y must cause awareness in my p-zombie twin because it does so for me. Chalmers skirts this issue by simply positing a p-zombie world parallel to ours; "the question is not whether it is plausible that p-zombies could exist in our world, or even whether the idea of a p-zombie replica is a natural one; the question is whether the notion of a p-zombie is conceptually coherent" actually having a phenomenologically (Chalmers 1996, 98). Of course the chances of me absent p-zombie twin are extremely Chalmers argues that its existence is logically possible. unlikely, but If the logical possibility of p-zombies is true, then consciousness cannot be broached from a third-person viewpoint, strengthening the knowledge argument and frustrating cognitive scientists thus even further.23 The logical possibility of zombies illustrates the point I aim to advance. There is no way for me (or my mother) to know that my zombie twin is actually a zombie. physically, functionally, Because he is and behaviorally identical to me, no test could definitively apart. Now lets assume that computationalism tell us is false and that consciousness cannot arise from formal processing. Android Andrew has a computer brain and looks, functions, and behaves just like me. In short, Android Andrew is my zombie twin. If we both took the Turing test, both of us would perform equivalently, purposes, identical. because we are, for all intents and As we shall see, consciousness falls outside the realm of human explanatory capacity and as a result, no test devised by humans could be comprehensive enough to verify the existence of a sentient mind. 23 See Chalmers 2002b. 37 On a related and more lighthearted note, when I was very young I climbed out of bed, made my way down the hall to the master bedroom, and awakened my mother. "What's wrong?" she asked, to which I responded, "I think I have a finger infection." Since I was still at the age when curiosity and ignorance often leads to personal injury, my concerned mother turned on the light and examined my finger. "Nothing seems wrong, dear. Are you sure it's your finger?" "Maybe it's my hand, or maybe my ... " Before I could finish my sentence I proceeded to vomit all over my parents' bed. Torn between laughter and frustration, my mother informed me, "The word is 'nauseous'. you are nauseous, honey." When you feel that way But remember, just because I didn't have a word for it, I was still nauseous. There is a disconnection between verbal description and phenomenal experience. As we shall see later, this division is known as the explanatory crucial aspect to consciousness theories across various fields. 38 gap and is a CHAPTER THREE DO ANDROIDS DREAM OF ELECTRIC SHEEP? WE'LL NEVER KNOW FOR SURE THE EXPLANATORY GAP In quantum physics, the Heisenberg uncertainty principle states that it is impossible to know with precision the simultaneous position and momentum of a physical system. We can know one or the other, but run into problems when we try to put them together. Consciousnesstheory falls victim to a similar problem. Introspection gives us accessto the brilliant world of qualia and phenomenal experience but tells us nothing about our internal processesunderlying our inner life. Conversely, external observation has provided us a wealth of information about how the brain works and interacts with the body. But it tells us nothing about what it is like to be the test subject. This fundamental divide is known as the explanatory gap. Joseph Levine coined this term in support of Nagel and the knowledge argument. Take three propositions: (1) 'Pain is the firing of C-fibers', (2) 'Heat is the motion of molecules', and (3) 'To be in pain is to be in state F'. (2) is true by necessity; there is "no possible world in which [it is] false" (Levine 2002, 354). Heat is, by definition, the motion of molecules, so any other definition simply wouldn't be heat. However, there is a "felt contingency" about statements (1) and (3); it is conceivable to imagine pain without C-fiber firing and pain without being in a particular 39 functional state. The sensation of pain is the pain, so we can imagine a world with no C- fibers but a phenomenon equivalent to what we know as pain. Unlike the heat example, no distinction between the appearance of a phenomenon and the phenomenon itself can be drawn. The reason why heat and pain cannot be explained in the same way comes down to identity. If what it's particularly like to have one's C-fibers fire is not explained, or made intelligible, by understanding the physical or functional properties of Cfiber firings - it immediately becomes imaginable that there be (-fiber firings without the feeling of pain, and vice versa. We don't have the corresponding intuition in the case of heat and the motion of molecules - once we get clear about the right way to characterize what we imagine24 - because whatever there is to explain about heat is explained by its being the motion of molecules. So, how could it be anything else? [Levine 2002, 358, footnote added] The proposition "heat is the motion of molecules" and its contingent corollary express a particular identity that is fully explainable, leaving out nothing of importance. Our scientific knowledge can perspicuously explain how molecular motion causes the phenomenon known as heat. The pain statement, on the other hand, leaves an explanatory gap between the causal role of C-fiber firing and the way pain feels. Both physicalism and functionalism fail to account for this gap. And since the gap is left unexplained, consciousness is not intelligible in the same sense that 'heat is the motion of molecules' is. For a proposition to be intelligible, "demand for further intelligibility 24 is One might ask, "What about how heat feels? Isn't that just as unexplainable as pain sensation? The answer to this question is yes, but the ineffability of heat experience is not what Levine was driving at. His point is that the science behind the transfer of heat is fully understood and is taken as a "primitive, brute fact about the universe," much like that value of the gravitational constant G is fully understood to be 6.67428x10-11 N (rn/kg]". In both cases, "there is nothing more we need to understand" (Levine 2002, 356). 40 inappropriate" in that there is nothing left to be explained about it (Levine 2002, 357). 'Pain is the firing of C-fibers' is not intelligible because of the necessity of further explanation. As shown in the previous chapter, physicalism only accounts for third-person observable behavior. We can induce pain in test subjects and observe the physiological, neurological, and behavioral responses. But such experiments tell us nothing about the first-person experience of consciousness: "Even hi-tech us the physical basis of consciousness, instruments like PET scans only give not consciousness as it exists for the person whose consciousness it is" (McGinn 1999, 48). Heat caused by molecular motion is fully explainable in physical terms alone; there is no explanatory gap dividing the proposition 'Heat is the motion of molecules'. As for pain, the best we can do is correlation between C- fiber firing and pain behavior, while the intermediate phenomenological experience is left unexplained. Functionalism, the basis for the computational theory of mind, fails to explain consciousness as well. Instead of identifying consciousness as physical processes, functionalists identify it by its causal roles. Pain, explained by functionalism, is "a higher- order property of physical states which consists in having a certain pattern of causes and effects, as it might be mediating bodily injury and avoidance behaviour" (McGinn 1991, 209). By these lights, the brain is just a UTM, so pain is the brain state mediating inputs and outputs. configuration. relevant Pain is thus realizable in any medium insofar as it is in the correct A robot that implements the causal properties of pain is in pain according to the functionalist model. 41 Functionalist definitions of mental states are not necessarily false, but they are insufficient for a comprehensive vital to a full understanding, back to Chalmers' P-zombies. explanation of consciousness. The causal role pain plays is but the phenomenological aspect is left gaping open. Think They are functionally equivalent to humans; when they stub a toe they cringe just like the rest of us. But in the end, all that occurs is pain behavior. there is nothing it is like to be a zombie, under no circumstances anything. does the zombie Since feel We know what pain is because we experience it intimately within our inner lives, but "because the qualitative character itself is left functionalist unexplained by the physicalist or theory that it remains conceivable that a creature should occupy the relevant physical or functional state and yet not experience qualitative character" (Levine 2002, 359). Accordingly, pain is more than just the functioning of input and output mechanisms. As Irving Krakow argues, functionalists "talk about pain, but they ignore the reality that the felt quality of pain is conceptually independent of the empirical possibility that the neurological correlate of pain qualia might play the 'causal role' they want to attribute to those qualia" (Krakow 2002, 97). The unmistakable feeling of pain is what mediates input (tissue damage) and output (pain behavior), and that phenomenological aspect is completely ignored. Krakow quotes Wesley Salmon's explanation of causal interaction: Let Pl and P2 be two processes that intersect with one another at the spacetime point S, which belongs to the history of both. Let Q be a characteristic that Pl would exhibit throughout an interval (which includes subintervals on both sides of Sin the history of Pl) if the intersection with P2 did not occur; let R be a characteristic that process P2 would exhibit throughout an interval (which includes subintervals on both sides of Sin the history of P2) if the s intersection with Pl did not occur. Then the intersection of Pl and P2 at constitutes a causal interaction if: (1) Pl exhibits the characteristic Q before 42 S, but it exhibits a modified characteristic Q' throughout an interval immediately following S; and (2) P2 exhibits a modified characteristic throughout an interval immediately following S. [Krakow 2002, 76] R' In the case of pain caused by a stubbed toe, Pl is my brain as it functions normally, say, while walking through a doorway and P2 is the act of my bare toe unexpectedly doorframe. striking the Pl and P2 intersect at time S, and their intersection causes Q (characteristics the brain while not in pain) to switch to Q' (characteristics of the brain while of in pain), as well as R (my toe not striking the doorframe) to switch to R' (my toe striking the doorframe). This functionalist explanation of pain accounts for all observable involved in pain function, but still fails to account for first-person experience. behavior The pain itself (as opposed to its causal or behavioral correlates) cannot be treated as Pl or P2 because it cannot be pinpointed spatially. It is an unobservable, non-spatial property that exists solely within the mind. As we shall see in the proceeding section, in addition to the problems of irreducibility and ineffability, the issue of spatiality is major barrier to closing the explanatory gap because human knowledge is bounded by space-time. MCGINN'S THEORY OF COGNITIVE CLOSURE According to legend, when Louis Armstrong was asked to explain what jazz is, he responded, "If you gotta ask, you ain't never gonna know." A musicologist could give a textbook definition of jazz and Mr. Armstrong himself could do his best to describe it. But in the end, jazz is intelligible only after it is experienced first-hand. Much to the chagrin of cognitive scientists, the same principle applies to consciousness. Levine argued for the existence of an explanatory gap dividing first-person 43 subjective experience and objective scientific facts about cognition and Nagel pointed out that such a schism will forever alienate proponents of both sides: "Absurdity comes with the territory, and what we need is the will to put up with it" (Nagel 1986, 11). Only recently have some maverick philosophers come out and flatly admitted that mind-body problem is an unsolvable endeavor. Arguments from every imaginable angle abound, but we are no closer to solving the hard problem then we ever have. Such a viewpoint, typically called New Mysterianism (henceforth NM), can easily be dismissed as meager defeatism and that the problem will eventually be solved with enough time and energy devoted to its solution. Steven Pinker, a recently converted mysterian, explains the core tenets: And then there is the theory put forward by philosopher Colin McGinn that our vertigo when pondering the hard problem is itself a quirk of our brains. The brain is a product of evolution, and just as animal brains have their limitations, we have ours. Our brains can't hold a hundred numbers in memory, can't visualize seven-dimensional space and perhaps can't intuitively grasp why neural information processing observed from the outside should give rise to subjective experience on the inside. This is where I place my bet, though I admit that the theory could be demolished when an unborn genius--a Darwin or Einstein of consciousness--comes up with a flabbergasting new idea that suddenly makes it all clear to us. [Pinker 2007] NM is potentially the most realistic prognosis around; materialism is an eminently philosophy except for its treatment of consciousness and functionalism Colin McGinnis defeatist. agreeable is a non-starter. at the forefront of the NM movement, and he defends his view as non- His goal is to figure out precisely why it is that we can't understand consciousness, not give up on it entirely: Consciousness depends upon an unknowable natural property of the brain. What this means is that I am not going to try to reduce consciousness to those mundane known properties of neurons that materialists hope to get by 44 with. But neither am I going to conceive of consciousness as something apart from the brain, something with no further analysis or explanation. Consciousness is rooted in the brain via some natural property of brain tissue, but it is not explicable in terms of electrochemical processes of the familiar kind. [McGinn 1999, 43] We can't deny the possibility that we are cognitively incapable of fully understanding consciousness. Perhaps the solution we seek is part of the problem; the hard problem might be beyond the realm of what is humanly possible. I do not reject materialism, but simply ignoring or attempting to eliminate qualia is insufficient. Qualia are an undeniable aspect of conscious experience and are crucial to any discussion about the hard problem. materialism" Chalmers calls my position "Don't-have-a-clue since I believe qualia are caused by physical processes, but cannot be explained as such. The mind-body problem is unsolvable, qualia will never fully be understood, and we will never know if an artificial intelligence ever achieves sentience. In short, there is no conceivable bridge for the explanatory is correlation. PET scans can measure brain activity, allowing brain states with exhibited behavior. gap. The best we can do scientists to match particular For instance, a 2001 study measured the brain function of twelve members of the Free Evangelical Fundamentalist Community, a religious group in Germany. Each subject reports to have experienced transcendent experiences during religious recitation. Brain activity was measured during an intense prayer session, and the results showed that religious experience "activated a frontal-parietal composed of the dorsolateral circuit, prefrontal, dorsomedial frontal and medial parietal cortex" (Azari, et al. 2001). The neural data is then correlated with the subjects' verbal descriptions of their own experiences in order to understand exactly what happens in the brain during 45 prayer activity. There is no doubt that this sort of experiment properties of brain function. To physicalists and functionalists, full understanding of consciousness. helps explain certain correlation I disagree because, as discussed of natural language restrict us from comprehensively is sufficient for a earlier, the limitations explaining first-person experience. Neural correlates to verbal descriptions plus observable behavior is not comprehensive; results of such studies fail to explain why we have phenomenal experience. explanatory gap remains in between objective data and first-person the The experience. But what exactly is the limiting factor that restricts us from a materialistic understanding of the mind? McGinn argues that human understanding is cognitively closed to certain aspects of the universe, consciousness being the prime example: The materialists are right to think that it is some property of the brain that is responsible for consciousness, but they are wrong in the kind of brain property they select. The dualists are right to doubt that the brain as currently conceived can explain the mind, but they are wrong to infer that no brain property can do the job. Both views overestimate our knowledge of mind, and brain, presupposing that our current conceptions are rich enough to capture the essence of the mind-brain link. I maintain that we need a qualitative leap in our understanding of mind and brain, but I also hold that this is not a leap our intellectual legs can take. [McGinn 1993, 28-29] Our "intellectual legs" can only stretch so far. The notion that humans possess the capacity to know everything there is about the universe is a wild assumption that should not be accepted out of blind anthropocentric arrogance. Since technology progressed at such an alarming rate over the last century, this assumption is widespread. on the moon, high-tech communication After all, we put a man computers (read: mobile phones) in our pockets, and artificial organs in our bodies. What can possibly limit our epistemic growth? 46 In 1965, Intel co-founder technological achievement Gordon Moore charted the progression of human over time, leading him to conclude, incorporated in a chip will approximately "The number of transistors double every 24 months" (Intel.com). Recent studies show that Moore's predicted rate of progression is actually slower than reality; according to University of Washington professor Edward Lazowska, "The ingenuity that computer scientists have put into algorithms have yielded performance improvements that make even the exponential gains of Moore's Law look trivial" (Lohr 2011). Arguably the most optimistic technologist around is Al innovator Ray Kurzweil. book The Singularity is Near, by 2023 one thousand dollars According to his recent will buy the processing capacity25 of a human brain. By 2059, one cent will buy the computational capacity of the entire human race (Kurzweil 2005). He contends that every aspect of consciousness well understood by science and completely explainable advances in computational technology. will be in objective terms, all because of We shouldn't worry about the hard problem for much longer; computers will eventually close the explanatory gap. What Kurzweil and his supporters must realize is that human cognition is not unlimited.26 There are certain concepts we might never grasp simply because of an epistemic horizon. To illustrate, look at the graph of the mathematical (Figure 7). The function is asymptotic, meaning it will grow indefinitely, function/(x) = 1/x but will never cross 25 The key word is 'processing'. His claims only have merit insofar as CTM holds and the brain is a computational entity. 26 Actually, Kurzweil would not disagree with this point. He claims that computers and Al will eventually augment human knowledge. Without the aid of technology, " ... the architecture of the human brain is ... profoundly limited. For example, there is only room for about one hundred trillion interneuronal connections in each of our skulls ... Machines will be able to reformulate their own designs and augment their capacities without limit" (Kurzweil 2005, Amazon Kindle Location 692) 47 either axis. The progression of human knowledge is asymptotic this horizon, in that there are naturally capacity that cannot be breached. about imposed limits on our cognitive "No finite mind could encompass all of space and time," McGinn writes; accordingly so, a solution to the mindFigure 7 body problem lies out of reach beyond the axis (McGinn 1999, 33). Human cognitive understanding is limited to concepts that are explainable in terms of space and time. Our universe is multidimensional; space has three dimensions and time has one. Conceiving of a dimension other than space-time is only possible through theoretical physics, but introspectively imagining such a world is impossible. This is because the entirety of our experience is spent interacting within these dimensional boundaries, barring certain accounts of certain psychotropic drug experiences, of course. Perception occurs when our sensory faculties register stimuli in the world around us, and transmit them as electrical impulses to the brain, thereby giving rise to a particular phenomenological sensation. What is perceived is immediately apprehensible through introspection, or as it is commonly called, the mind's eye. When I perceive a chair sitting five feet in front of me, the image of the chair and whatever else is immediately perceivable are the objects of my consciousness. The same applies to when I am imagining a chair even when there isn't one present, or when I am dreaming, or hallucinating. Regardless of circumstance, the locus of attention is the object of consciousness. The object of consciousness is non-spatial; the chair I am looking at is extended in space but my mental projection is not. We can say that the concurrent brain activity occurs in a physical space (i.e. in the visual cortex, an inch from the back of the head, in three 48 dimensions, etc.), but the felt experience itself has no measurable volume, location, mass, or shape; "It falls under temporal predicates and it can obviously be described in other ways - by specifying its owner, its intentional content, its phenomenal character," but there is no spatial dimension to consciousness (McGinn 1997a, 98). The unembodied nature of consciousness is what causes problems for empirical science. There is a disconnect between the non-spatiality nature of the external world. of consciousness and the spatial The property distinctions between the extended world with the unextended mind has been at the root of the mind-body problem since Descartes. McGinn seeks to revitalize and update the Cartesian dilemma: nonspatial phenomenon, human thought is fundmamentally "While consciousness is a governed by spatial modes of representing the world, so that our ways of thinking tend to force consciousness onto a Procrustean bed of broadly Euclidean design" (McGinn 1997b, 108). The objects of consciousness are perceptually experience. inaccessible; observing the brain says nothing about Consciousness is thus perceptually closed, meaning we expound it in terms of the physical. It is safe to say that the object of consciousness is perceptually closed and is only accessible through introspection. However, perceptual closure does not entail cognitive closure. Electrons exist, surely, even though it is impossible very act of observation to measure them because the alters the electron's path. We can't observe genes, quarks, or atoms either but we can postulate them because the explanability of other observable phenomena depends on their existence. For example, trait inheritence can be observed but cannot be 49 explained without genes. There is enough observable evidence {like DNA molecules) to confirm the legitimacy of genetics without ever actually perceiving a gene; "We can infer hidden structure to explain what overtly appears" {McGinn 1999, 141). But consciousness is a different story. Inference is not a sufficient condition for a comprehensive explanation of first-person experience. Define Pas the set of empirically evident brain properties that clarifies the hard problem and fully explains consciousness. McGinn contends that P cannot be infered like genes and electrons because of the homogeneity that constrains sets of data: Inference to the best explanation of purely physical data will never take us outside the realm of the physical, forcing us to introduce concepts of consciousness. Everything physical has a purely physical explanation. So the property of consciousness is cognitively closed with respect to the introduction of concepts by means of inference to the best explanation of perceptual data about the brain. [McGinn 1991, 13] Attempting to explain phenomenal consciousness with observable data is like comparing apples and oranges; the two sets of information are inherently incompatible. Physical concepts cannot describe psychological concepts and vice versa. Neural correlation experiments show us nothing but patterns of behavior and concurrent brain states. As Levine argued, observable data is fully explainable in physical terms, constrained homogenously. It is a brute, empircally provenfact that the gravitational contant G is 6.67428x10-11 N (rn/kg}" However, attempting to explain Gin phenomenological terms is scientifically meaningless. Yes, I can feel the affects of gravity upon my body and the objects I interact with, and sure, I can do my best to verbally describe how it feels to 50 not be floating above the ground." but the explanation is incomplete without physical proof. Consciousness is homogenously constrained in the same way: "Our modes of concept formation, which operate from a base in perception and introspection, cannot bridge the chasm that separates the mind from the brain: They are tied to the mental terms and physical of the relation, not to the relation itself" {McGinn 1997b, 106). Since we can't explain conscious phenomena in terms of the physical, and introspective analysis is unexplainable because of the limits of natural language, it follows that humans are cognitively closed to the complete apprehension of consciousness. Consciousness is closed off by the same epistemic boundaries that forbid us from ever knowing what bat echolocation feels like and that limits blind people from the concept of sight: "our concepts of consciousness just are inherently constrained by our own form of consciousness, so that any theory the understanding these constraints would ipso facto be inaccessible to us" {McGinn 1991, 9). Consciousness is only accessible through introspection, function. of which required us to transcend but introspection can tell us nothing about brain Perceptive analysis, on the other hand, can access brain function consciousness. Since there is no intermediate but not method to speak of, Pon the whole is unexplaninable. That is not to say P does not exist, however. NM is not substance dualism; materialism holds but it simply cannot account for consciousness. not entail nonexistence. Cognitive closure does Perhaps "Martians or demigods might have better luck" 27 I'm sure such a description is possible. It would sound ridiculous, but it is surely possible. But in the end, verbal descriptions are understood fully only by the utterer (see figure 2 in the second chapter of this paper). 51 understanding P, as Dennett writes in his criticism (Dennett 1991b). Something explains the phenomenon of consciousness, but we can't grasp it. There is still much to learn from the study of consciousness, and neuroscience might bring us closer to solving the mind-body problem than ever before, but the hopes for a unified theory of consciousness are slim to none. CRITICISM AND RESPONSE Criticism for McGinn's work is far reaching. Opponents tend to boil down cognitive closure to a simple syllogism: Pl. P2. P3. Cl: Introspection alone cannot fully explain P External perception of the brain cannot fully explain P Appeals to inference do not hold because of homogeneity constraints Therefore, Pis humanly unexplainable by any extant means of apprehension C2: P can only be apprehended from a "God's-eye point of view" (Flanagan 1992) Pl is uncontroversially true. My own first-person experience tells me nothing about neurons or synapses or anything else relevant to a unified theory of consciousness. If I never learned that there is lumpy gray matter inside my head, I would have no reason to ever assume that there is!28 P2 is the problem. McGinn is frequently condemned defeatist or even Luddite, accusing him of discounting technology and knowledge. Simply because a solution the rapid progression as of human has not yet been found, "the solution may be just around the corner" (McGinn 1997b, 108). Continued research in cognitive science and related fields will eventually solve the mind-body problem 28 The point about not knowing about the brain is borrowed from Flanagan 52 and 1992. provide us with a unified theory of consciousness, allowing us to create sentient may or may not enslave humanity. attempts to make. robots that In response, such criticism clearly misses the point NM McGinn's contention is that the mind-body problem will never be solved for sure because there is no definitive way of knowing if the solution is in fact correct. There is no test comprehensive Furthermore, enough to test for true sentience, only observable behavior. some argue that the truth or falsity of P2 depends on interpretation. By McGinn's lights, P2 is unequivocally true because direct observation of a brain indeed tells us nothing about conscious experience. controversial because the "'unobservability' On the other hand, the premise is of the link between P and consciousness prevents us from inferring that Pis in fact where the link resides" (Flanagan 1992, 112). is invoked in this point; Flanagan calls into question why inference from unobservable is insufficient for a theory of consciousness. P3 data McGinn invokes Nagel's claim that "it will never be legitimate to infer, as a theoretical explanation of physical phenomena alone, a property that includes or implies the consciousness of its subject" because consciousness and observable brain states are intrinsically different properties that cannot be used to explain each other (Nagel 1979, 183). To Flanagan, however, neural correlation is sufficient for a complete explanation: We are not looking for an explanation of 'physical phenomena alone', at least not physical phenomena narrowly understood. There is a prior commitment to the existence of consciousness. Thus both brain facts and facts about consciousness are on the table to be explained. We then infer that the constellation of a certain set of phenomenological reports of restricted range ('tastes sweet') correlate with certain sorts of brain activity (activation in the relevant pathways), and we infer, given an overall commitment to naturalism, that the latter explains the former. [Flanagan 1992, 113) 53 What makes consciousness so exceptional that the rules of inference used for electrons, genes, etc. no longer apply? Electrons, for example, are never actually observed in experiments that invoke them (i.e. cloud chambers), but we postulate their existence because other observable data rides on their existence. Similarly, we don't see consciousness during brain experiments but consciousness is postulated because the evidence derived from observed brain activity and overt behavior are incoherent without the existence of consciousness. Therefore, Pis explanable through inference. Flanagan fails to account for what I call the behavior barrier. The behavior barrier is an updated appeal to the classical philosophical dilemma of other minds combined Chalmers' argument for the logical possibility of p-zombies. with How can we be certain that something that behaves as if it is conscious is actually in posession of any sort of phenomenal experience? While arguments for NM and cognitive closure are admittedly inductive, postulating the existence of an inner life is inductive as well. Neural correlation studies show that the subject is probably in some conscious state, but the only real proof the experimenters have is an analogy to their own experience. Researcher Brown knows he is a conscious, sentient agent because he can introspect, the same way everybody else knows that they are not mindless automata. of introspection Test subjects Black and Gray are also capable as well and is thus conscious. Black and Grey are hooked up to a device that provides a thorough analysis of all brain activity. The device is small to not cause nervousness or discomfort and distort test results. and uninvasive, as Brown injects a dose of the little known chemical 11 that causes them both to scream in agony that the area of injection feels like its on fire. Brown correlates the neural scan data and the subjects' 54 behavior in order to postulate that chemical 11 causes a painful and localized burning feeling, and this feeling is concurrent with certain brain patterns. Is this test conclusive? As Chalmers 1996 and 2002b show, there is a logical possibility for the existence of p-zombies who feel nothing at all, but are compositionally equivalent to humans who do feel a burning sensation from chemical 11. Black and Gray show equivalent neural and behavioral activity during the test, but while Black feels burning pain, Gray feels nothing at all because he is a cybernetic organism (cyborg) functionally humans but entirely devoid of phenomenal experience. and behaviorally equivalent to What's most interesting that Dr. Brown will never know this revealing fact. Gray is so ingeniously is the fact designed that he is capable of passing the even the most trying Turing test questions because he is programmed with a full life's worth of memories and emotional response behavior. Indeed, Gray is unaware that he is a cyborg because he is not actually aware of anything at all. If Brown is somehow compelled to ask whether Gray is a cyborg, Gray's behavioral software is programmed to brush off such questions as ridiculous, just as any other human would do. My point is that even if the world's foremost neuroscientists claim to find a unified theory of consciousness P, there is no definitive way to test its veracity. When I asked about possibility of the ascription of consciousness onto an artifact, Dreyfus told me: Once you've got to the point where you have a theory of consciousness and you're implementing it and its behaving just like human being, then whether its actually conscious or not will have exactly the same structure as whether you are actually conscious or not when I'm dealing with you. If that can't be settled, then maybe it's a p-zombie. [Dreyfus 2011] His point agrees with mine; the explanatory gap remains indefinitely unclosed because it is impossible to "climb in" to someone else's consciousness and find out for sure. 55 In the film Being John Malkovich, Schwartz, the protagonist, discovers a secret portal that can transport him into the mind of actor John Malkovich, allowing him to simultaneously experience Malkovich's inner world as well as his own. Schwartz exclaims, "Do you know what a metaphysical can of worms this portal is?" and rightly so; he transcended idea of first-personality. the very But unless a real-life Schwartz really manages to discover such a portal, the minds of others are indefinitely inaccessible and can never be confirmed for sure because of the behavior barrier. Al, CONSCIOUSNESS, AND BLADE RUNNER: TYING EVERYTHING TOGETHER If McGinn is right, humanity is cognitively closed to the full understanding of consciousness. If a unified theory of consciousness is ever proposed, its truth can only be postulated but never known for sure. Strong Al proponents disagree. As I discussed earlier, some theorists believe that the brain is wetware, a biological digital computer. hardware and the mind is the software it implements, personal computer. The brain is not unlike a video game running on a And also like a video game, conscious experience is an illusion by the physical processes "behind the scenes." Computationalists (or functionalists, like) claim that conscious states are just functional states of the brain and "feelings conscious awareness are evoked merely by the carrying out of appropriate (Penrose 1994). It follows that a complete understanding caused if you'd of computations" of how the brain physically works will eventually give rise to an explanation of consciousness. In this section I will demonstrate that like physicalism, a computational explanation of consciousness fails to close the explanatory gap and solve the hard problem for sure 56 because of the behavior barrier. Through introspection we can access consciousness directly, but only through physical experimentation can we understand the physical processes. Unification is futile, and CTM can bring us no closer to an answer than anything else. In order to advance my argument, I will elaborate on themes from Phillip K. Dick's classic novel Do Androids Dream of Electric Sheep and its film adaptation, Ridley Scott's Blade Runner. Both texts have deep philosophical implications ranging from the meaning of love to the ethics of animal treatment, but I will be primarily focusing on its treatment of the metaphysical issue of person hood in order to strengthen my argument for the untenability of solving the hard problem. The book and the film have similar themes and the same backstory, but significant variations in plot development.29 Both stories take place in the not-so-distant future in a dystopian, pollution-riddled Los Angeles. World War Terminus decimated the human population and the vast majority of animals are extinct. As a result, empathy towards animals is revered to an almost religious level and owning a real animal (as opposed to the more common synthetic replacements) is the highest symbol of status." The turmoil on Earth caused the colonization of nearby planets, and the arduous job of construction and development was assigned to anthropomorphic robots called replicants. Because they are dangerous to humans, replicants are outlawed on Earth. Protagonist Rick Deckard is a blade 29 For the sake of clarity and continuity, I will use names and terminology from the film instead of the book. It should be noted that the lifelike Al characters in the film are called 'replicants' while the novel calls them 'androids'. The two terms refer to the same conceptual construction. 30 The theme of animal ownership and its social implications is a main theme in the novel but less so in the film. Needless to say, empathy towards animals is crucial to both texts. 57 runner, a bounty hunter who specializes in the "retirement" of replicants. around Deckard's hunt for Pris, Zhora, Leon, and Roy, advanced The plot revolves Nexus-6 class replicants who returned to Earth in order to pressure their creator Tyrell to extend their four-year limited lifespan. What gives blade runners the right to kill? Replicants seem to feel pains and emotions just like their human creators. Among the most disturbing scenes in the film is when Deckard shoots Zhora, causing her to fall through multiple panes of glass while she screams and grimaces in utter agony (Figure 8}. Dick wrote in his essay "Man, Android, and Figure 8 Warner Machine" that an android is "a thing somehow generated to deceive us in a cruel way, to cause us to think it to be one of ourselves" (Dick 1976, 202}. Zhora seems human in every way; accordingly so, watching her die so brutally evokes empathy from the audience. But does she actually feel anything at all, or just her programming? behave as if she does because of Lets assume that replicant "minds" are syntactical computers running advanced software that causes their overt and physiological behavior to function identically to humans, "capable of selecting within a field of two trillion constituents, or ten million separate neural pathways" (Dick 1968, 28}.31 Lets also assume that a computer mind is not a mind at all, and no conscious experience arises from it. In this case, replicants are be pzombies, "identical...functiona//y ... reacting in a similar way to inputs ... with indistinguishable behavior resulting" (Chalmers 1996, 95}. Moreover, they are not conscious, and have no experience in the same sense that you and I have experience 31 feel no qualia, of the world. This entails that Kurzweil's prediction that a comprehensive computational human brain function is true in 2019 Los Angeles. 58 model of As Brothers Daniel Robinson points out "It is conceivable that a device could be made in such a way as to change its appearance and make loud sounds when one or another component is destroyed. [ ... ] Here, then, we have a device that replaces 'pain behavior' with 'pain language.' We have, in a word, everything but pain!" Behavior, despite its apparent authenticity, is not sufficient for confirmed consciousness. In this case,there like to be a replicant, "Chitinous reflex-machines who aren't really alive" is nothing it is (Dick 1968, 194). Regardless, it is natural to feel bad about Zhora's painful and unceremonious execution. Accordingly, a central theme in Blade Runner is uncertainty. litmus test for replicancy/humanity The only reliable is the Voight-Kampff (V-K) test that measures physiological responses to emotionally jarring questions. Most questions are about animal mutilation, a topic that every human cares about deeply; although some replicants "had been equipped with an intelligence greater than that of many human beings," they "possessed no regard for animals ... [and] no ability to feel empathic joy for another life form's success or grief at its own defeat (Dick 1968, 32). Rachael is a replicant but, unlike the others, she had real human memories implanted into her neural "believes" that she is a real human and acts convincingly not) is up to interpretation.32 Most importantly, 32 network. so. Even Deckard's there is a patent uncertainty She humanity (or about "An android," [Deckard] said, "doesn't care what happens to another android. That's one of the indications we look for." 'Then," Miss Luft [Zhora's novel analogue] said, "you must be an android." She then asks whether or not Deckard has taken the V-K test himself. "Yes." He nodded. "A long, long time ago; when I first started with the department." "Maybe that's a false memory. Don't androids sometime go around with false memories?" (Dick 1968, 101-102). 59 mistakenly retiring "authentic humans with underdeveloped empathic ability" who fail the V-K test (Dick 1968, 54). The V-K test is a reimagined variation on the Turing test. Remember that a computer that passes Turing's imitation game by behaving as a human would is, on some views, intelligent and conscious. Anti-computationalists like Searle and Dreyfus disagree: "I'm not enough of a behaviorist to think that if it just behaves like people it shows that it's intelligent" (Dreyfus 2011). I agree, to an extent. As Searle made clear in the Chinese Room argument, simulation is not, in and of itself, proof of duplication. that an artifact simulates Judging solely on the basis consciousness behavior is insufficient evidence for knowing whether or not it has an inner life at all. Strangely enough, Dennett, a staunch opponent of Searle's claims, seems to agree. A simulation of a hurricane, for instance, is hardly equivalent to a real one. A comprehensive theory of hurricane behavior can be implemented simulation. through a computer Certain measurable inputs (i.e. barometric and temperature appropriate outputs (i.e. 150 mph winds blowing westward accompanied but one should "not expect to get wet or windblown data) yield by heavy rain), in its presence" (Dennett 1981, 191). By the same lights, an Al can simulate consciousness but be just as lifeless as hurricane software. Android S might respond to an anvil dropped on its foot as follows: S's (-fibers are stimulated, ... a pain memory is laid down; S's attention is distracted; S's heart-rate increases .. ; S jumps about on right foot, a tear in the eye, screaming. [Dennett 1981, 192] S certainly passes the Turing test for pain. Zhora does too. In the novel, replicant Pris acts like the timid and scared young woman she is meant to resemble: 60 "Fear made her seem ill; it distorted her body lines, made her appear as if someone had broken her and then, with malice, patched her together badly" (Dick 1968, 62). Behavior may be convincing but it is not definitive proof for the existence of mental states. Roy and his Nexus-6 brethren are programmed with a bottom-up response mechanism. Like humans, they are "born" with no emotional develop them naturally through experience.33 emotional attachments but Dreyfus writes: Generally, in acquiring a skill - in learning to drive, dance, or pronounce a foreign, for example - at first we must slowly, awkwardly, and consciously follow the rules. But then there comes a moment when we finally can perform automatically. At this point, we do not seem to be simply dropping these same rigid rules into unconsciousness; rather we seem to have picked up the muscular gestalt which gives our behavior a new flexibility and smoothness. [Dreyfus 1994, 248-249] Emotions are no different. We gradually learn through experience how to identify specific emotions and behave appropriately. Unlike humans, however, the Nexus-6s only have four years to learn a lifetime of emotions, and there is a clear discrepancy between their emotional immaturity and superhuman strength and/or intelligence. Towards the end of the film, Roy is distraught to find the lifeless and bloodied body of his lover Pris, killed by the blade runner in an earlier scene. His underdeveloped emotions become evident when he begins to act oddly, smearing her blood on his lips and howling like a wounded animal. Earlier on he attempts to inform Pris about the deaths of their allies Leon and Zhora, but his inability to process emotional depth leads him to convulse and stiffen robotically, inhumanly. 33 The imbalance of Roy's physical prowess and calculating Excluding Rachael, who has emotions from her implanted 61 memories. rationality with his emotional handicaps stands in stark contrast with Tyrell Corporation's motto "More human than human," reminding the audience that replicants are fundamentally On the other hand, the penultimate a climactic cat-and-mouse different than us. scene adds to the complexity of this issue. After chase between Deckard and the physically superior Roy, Deckard finds himself facing his death, clinging to a roof beam for dear life. In a beautiful display of what appears to be humanity, Roy chooses to save Deckard. Throughout the film, Roy is obsessed with extending his limited lifespan but in the end, he finally seems to understand the value and transience of life: Rov: I've seen things you people wouldn't believe. Attack ships on fire off the shoulder of Orion. I watched C-beams glitter in the darkness at Tannhai..iser Gate. All those moments will be lost in time like tears in rain. [Fancher and Peoples 1982] Deckard can only sit back in awe as his savior perishes. It certainly appears that Roy feels something, judging by his actions. In his essay on postmodernism in Blade Runner, Nick Lacey writes, "In [Roy's] final moments, when he saves Deckard, he becomes human because of his behavior and his realisation that his life was worth living (Lacey 2005, 190}. Marilyn Gwaltney adds that movie audiences with Roy because, "Our understanding sympathize of his cruelty changes as we come to understand it as a very human reaction to his existential situation: the imminence of his death and that of those he loves; the feeling of betrayal by the beings that brought him into existence" (Gwaltney 1997, 33}. But can we safely assume that this sort of demonstrated self-realization implies phenomenal first-person experience? Gwaltney later writes, "Computers may be thought to act rationally, in the sense of acting logically, but they do not act to any purpose of their own, only to the purposes of 62 others. To act with purpose requires consciousness of self" (Gwaltney 1997, 35). Roy certainly appears conscious and self-aware, but in the end, he could very well still be a pzombie. If technology in the Blade Runner universe is advanced enough Nexus-6, software for this sort of humanlike behavior is surely plausible. to create the It thus follows that Roy's epiphany, despite its personal impact, might not feel like anything at all to him. This is a hard concept to swallow, considering how dramatic an effect it appears to have. Indeed, actor Rutger Hauer completely improvised the "tears in rain" speech, own felt experiences. no doubt drawn his But if the assumption that replicants are p-zombies is correct, Roy feels nothing at all while his artificial brain implements the program SELF _REALIZATION_EPIPHANY.EXE in the background, no different from my laptop not feeling the word processer currently running. As Rachael laments in the novel, "We are machines, stamped out like bottle caps. It's an illusion that 1-1 personallyrepresentative of a type ... I'm not alive." (Dick 1968, 189, 198}. But remember, none of this can be definitively otherwise. really exist; I'm just proven by any test, Turing, V-K, or Both claims in support and against machine consciousness reasoning that cannot be verified for sure. On the possibility are based on of a conscious artifact of human creation, McGinn contends, "I should say the matter is entirely an empirical one: it concerns whether human beings ever in fact achieve enough in the way of scientific and technological knowledge" to close the explanatory gap and physically observe the existence of conscious experience. "It is like asking whether we shall ever travel to another galaxy" (McGinn 1991, 203}. Since replicants are human creations, as of now it is an empirical impossibility of ever confirming if it there is anything it is like to be one. On the other hand, 63 the possibility of a conscious artifact in general is a matter of principle. If hyper intelligent Martians (the philosophers' favorite species) were to abduct a human and reproduce it molecule for molecule, there is no reason to say the clone isn't conscious. Of course, it may very well be a p-zombie, devoid of any experience at all, but only Martians who understand the explanatory gap can know for sure. Whether or not replicants are anything more than lifeless "skinjobs," as they are disparagingly called, is ultimately beyond our epistemic boundaries. Figure 9 Dilbert.com 64 CONCLUSION In this work I make the claim that sentient artificial intelligence may be possible, but it is beyond the epistemic limits of human understanding to ever devise a test comprehensive enough to prove it for sure. Neither functionalism and CTM nor physical ism account for the explanatory gap dividing first-person conscious experience and externally observable neural and behavioral correlates. accessible to an Mental states are exclusively individual mind and any attempts to describe them verbally are limited by ostensive definition and the boundaries of natural language. There is no conceivable means of knowing whether or not 'sadness' feels the same to me as it does to you; observable behavior can be simulated but not conclusively duplicated. only assess behavioral equivalence. Accordingly, the Turing test can As we see in Blade Runner, simply because something looks and acts sentient does not necessarily entail sentience. mysterians about consciousness. I contend that we should be Neuroscience and related fields continue to make remarkable progress in the study of both consciousness and Al and should certainly continue to do so. However, it is beyond the cognitive capacity fully explain consciousness in a testable and provable theory. for human beings to ever If Terminator is an accurate prediction of the future, whether or not Skynet is sentient is forever unknown. 65 Works Cited Azari, Nina P., et al. "Neural Correlates of Religious Experience." European Journal of Neuroscience 13, no. 8 (2001): 1649-1652. Ben-Yami, Hannoch. "A Note on the Chinese Room." Synthese 95, no. 2 (May 1993): 169- 172. Bickle, John. "Multiple Realizability." Stanford Encyclopedia of Philosophy. Edited by Edward N. Zalta. Fall 2008. plato.stanford.edu/archives/fall2008/entries/multiplerealizability (accessed March 26, 2011). Blackburn, Simon. The Oxford Dictionary of Philosophy. Oxford: Oxford University Press, 2008. Carter, Matt. Minds and Computers. Edinburgh: Edinburgh University Press, 2007. Chalmers, David. "Consciousness and Its Place in Nature." In Philosophy of Mind: Classical and Contemporary Readings, edited by David Chalmers. Oxford: Oxford University Press, 2002. Chalmers, David. "Does Conceivability Entail Possibility?" In Conceivability and Possibility, edited by T. Gendler and J. Hawthorne, 145-200. Oxford: Oxford University Press, 2002. -. The Conscious Mind: In Search of a Fundamental Theory. Oxford: Oxford University Press, 1996. Copeland, Jack. Artificial Intelligence. Oxford: Blackwell, 1993. Dennett, Daniel C. "Why You Can't Make a Computer that Feels Pain." In Brainstorms, by Daniel Dennett, 190-229. Cambridge, MA: MIT Press, 1981. -. Consciousness Explained. Boston: Little, Brown and Company, 1991a. -. "Quining Qualia." In Philosophy of Mind: Classical and Contemporary Readings, edited by David Chalmers, 226-246. Oxford: Oxford University Press, 2002. 66 -. "The Unimagined Preposterousness of Zombies." Journal of Consciousness Studies 2, no. 4 (1995): 322-325. -. "Review of McGinn, The Problem of Consciousness." The Times Literary Supplement, May 10, 199lb. Dick, Phillip K. Do Androids Dream of Electric Sheep? New York: Del Rey, 1968. -. "Man, Android and Machine." In Science Fiction at Large, edited by Peter Nicholls. New York: Harper & Row, 1976. Dreyfus, Hubert L. What Computers Stiff Can't Do. Cambridge, -. MA: MIT Press, 1994. Interview by Andrew Barron. Interview with Hubert Dreyfus (June 11, 2011). Fancher, Hampton, and David Peoples. Blade Runner. Directed by Ridley Scott. Performed by Rutger Hauer. 1982. Feigl, Herbert. "The "Mental" and the "Physical"." In Philosophy of Mind: Classical and Contemporary Readings, edited by David Chalmers, 68-72. Oxford: Oxford University Press, 2002. Flanagan, Owen. Consciousness Revisited. Cambridge, MA: MIT Press, 1992. Gwaltney, Marilyn. "Androids as a Device for Reflection on Personhood." In Retrofitting Blade Runner, 32-40. Bowling Green, OH: Bowling Green State University Popular Press, 1997. Grossman, Lev. "2045: The Year Man Becomes Immortal." Time, February 10, 2011. Intel.com. Moore's Law and Intel Innovation. Intel. http://www.intel.com/about/companyinfo/museum/exhibits/moore.htm (accessed April 14, 2011). Huxley, T.H. Lessons in Elementary Physiology. London: Macmillan, 1866. Haugeland, John. Artificial Intelligence: The Very Idea. Cambridge, MA: MIT Press, 1989. Hacker, PMS. "Is There Anything It is Like to Be a Bat?" Philosophy 77, no. 300 (2002): 157175. Hamad, S. "Other Bodies, Other Minds: A Machine Incarnation of an Old Philosophical Preble,." Minds and Machines (Kluwer Academic) 1 (1992): 43-54. Hobbes, Thomas. Leviathan. Cambridge: Cambridge 67 University Press, 1904/1651. Hofstadter, Douglas R. "A Conversation With Einstein's Brain." In The Mind's I, translated Daniel C. Dennett and Douglas R. Hofstadter. New York: Basic Books, 2000. by Jackson, Frank. "Epiphenomena! Qualia." Philosophical Quarterly 83 (1982): 127-136. Janzen, Greg. "Phenomenal Character as Implicit Self-Awareness." Journal of Consciousness Studies (Imprint Academic) 13, no. 12 (2006): 44-73. Kurzweil, Ray. The Singularity is Near. New York: Penguin, 2005. Kaufman, Charlie. Being John Malkovich. Directed by Spike Jonze. Performed by John Cusack. 1999. Krakow, Irving. Why The Mind-Body Problem Cannot Be Solved! Lanham, MD: University Press of America, 2002. Lacey, Nick. "Postmodern Romance: The Impossibility of (De)Centering the Self." In The Blade Runner Experience, edited by Will Brooker, 190-199. London: Wallflower Press, 2005. Lanchester, John. "Scents and Sensibility." The New Yorker, March 10, 2008: 120-122. Lewis, David Kellog. "Mad Pain and Martian Pain." In The Nature of Mind, by David Kellog Lewis. New York: Oxford University Press, 1991. Levine, Joseph. "Materialism and Qualia: The Explanatory Gap." In Philosophy of Mind: Classical and Contemporary Readings, edited by David Chalmers, 354-361. Oxford: Oxford University Press, 2002. Locke, John. An Essay Concerning Human Understanding. Indianapolis: Hackett, 1689/1996. Lohr, Steve. "Software Progress Beats Moore's Law." The New York Times, March 7, 2011. Nagel, Thomas. "Conceiving the Impossible and the Mind-Body Problem." Philosophy 73, no. 285 (1998): 337-52. -. Mortal Questions. Cambridge: Cambridge University Press, 1979. Newell, Allen, and H.A. Simon. "Computer Simulation of Human Thinking." Science 134, no. 3495: 2011-2017. McGinn, Colin. "Consciousness and Space." In Explaining Consciousness - The 'Hard Problem', edited by Jonathan Shear, 97-108. Cambridge, MA: MIT Press, 1997. -. Minds and Bodies. New York: Oxford University Press, 1997. 68 -. The Mysterious 1999. Flame: Conscious Minds in a Material World. New York: Basic Books, -. The Problem of Consciousness. Oxford: Basil Blackford, 1991. Penrose, Roger. Shadows of the Mind. Oxford: Oxford University Press, 1994. Pinker, Steven. How the Mind Works. New York: Norton, 1997. -. "The Mystery of Consciousness." Time, January 19, 2007. Place, U.T. "Is Consciousness a Brain Process." In Philosophy of Mind: Classical and Contemporary Readings, edited by David Chalmers, 55-60. Oxford: Oxford University Press, 2002. Sutherland, Stuart. The International Dictionary of Psychology. Second. New York: Crossroad, 1996. Searle, John. "Minds, Brains, and Programs." Behavioral and Brain Sciences 3 {1980). -. Philosophy in a New Century. Cambridge: Cambridge University Press, 2008. Simon, Herbert A., and Allan Newell. "Heuristic Problem Solving: The Next Advance in Operations Research." Operations Research 6 (January-February 1958). Russell, Bertrand. Religion and Science. New York: Oxford University Press, 1961. Turing, Alan. "Computing Wittgenstein, Machinery and Intelligence." Mind 49 {1950): 433-460. Ludwig. Philosophical /nvestigtions. Oxford: Basil Blackwell, 69 1974.