Roger Penrose's book Shadows of the Mind may be purchased
from Amazon.Com

[STAR] Penrose is Wrong

Drew McDermott
Department of Computer Science
Yale University
New Haven, CT
U.S.A.

mcdermott@cs.yale.edu

Copyright (c) Drew McDermott 1995

PSYCHE, 2(17), October, 1995
http://psyche.cs.monash.edu.au/v2/psyche-2-17-mcdermott.html

KEYWORDS: Consistency statements, Gödel incompleteness theorems, Turing machines, ordinal logics, Platonism, proof theory.

REVIEW OF: Roger Penrose (1994) Shadows of the Mind. New York: Oxford University Press. 457 pp. Price: $25 hbk. ISBN 0-19-853978-9.

[Note to the on-line reader: Penrose's hollow five-pointed star, the symbol of unassailability, I indicate with the string "STAR".]

1. Penrose vs AI - Again

1.1 Roger Penrose's new book, Shadows of the Mind, is strongly reminiscent of his previous work in the same vein, The Emperor's New Mind. This book restates the author's central line of argument about the place of consciousness in the material world. He has no sympathy at all for attempts to work out a computationalist theory of mind, and instead pins his hopes on a future theory that would allow large-scale quantum-mechanical effects in the brain to play a central role.

1.2 A broad outline of his argument goes like this:
1.3 This argument, when put down in black and white, seems extraordinarily weak. The least speculative step is the first, but that's also the easiest to show is fallacious, as I will do shortly. But before I do, I want to raise the question, Why is Penrose bothering?

1.4 A clue might be this sentence on p. 373: "It is only the arrogance of our present age that leads so many to believe that we now know all the basic principles that can underlie all the subtleties of biological action." Penrose wants to do battle against the arrogance he perceives, especially in the AI community, regarding the problem of consciousness. It is true that AI has, from its inception, had the ambition to explain everything about the mind, including consciousness. But is this arrogance? Or merely the sincere adoption of a working hypothesis? If someone wants to work on the problem of mind, it seems to me that he must choose among three options: treat the brain as a computer, and study which parts compute what; study neurons, on the assumption that they might be doing something noncomputational; or work in a seemingly unrelated field, like physics, on the off chance that something relevant will turn up. In any case, no matter which tack is taken, one gets mighty few occasions to feel arrogant about one's success. Neuroscience and AI have made definite progress, and so has physics, for that matter, but their successes haven't resulted in a general theory of mind. If anything, AI seemed closer to such a theory thirty years ago than it seems now.

1.5 So if someone wants to believe that AI will never explain the mind, he might as well. The burden of proof is on whoever claims it ultimately will. Penrose isn't satisfied with this state of affairs, however, and wants to exhibit a proof that a computationalist theory of mind is impossible. I suppose he sees himself fighting for the hearts and minds of neutral parties, who are in danger of being fooled into thinking that AI is on the verge of such a theory by the breathless stories they read in the papers. I don't know; perhaps an argument like Penrose's will, once it has been filtered through the distorting lens of the TV camera, be a sort of homeopathic antidote to those breathless stories. But, I regret to say, the argument would still be wrong. And so those of us in a position to point out the flaws in it must sheepishly rise to do so, in the full knowledge that AI can't win the debate if it degenerates into Mutual Assured Destruction ("You can't prove AI is possible," "Oh yeah? Well, you can't prove it's not").

2. Gödel's Theorem or Bust

2.1 Penrose stakes everything on his analysis of Gödel's Theorem. This analysis is all wrong, but what's striking is how much he tries to hang on it. Penrose assumes that there is a single attribute called "consciousness" that accounts for insight, awareness, and free will. Hence, if he can show that computers lack a certain sort of insight, they must also lack all awareness and free will. (One wonders where this leaves five-year-old children.)

2.2 In addition, all the plausibility of Penrose's theory of "quantum consciousness" in Part II of the book depends on the Gödel argument being sound. It certainly provides no plausibility by itself. There is a lot of material in the book about the mysteries of quantum mechanics. There is a much smaller amount about where in the brain quantum-mechanical effects might be important. But if you seek an account of the link between these hypothetical effects and insight, awareness, and free will, there isn't any. This nontheory gets all of its oomph from the pathetic weakness of the computational alternative, as described in Part I of the book. The slightest flaw in Part I would knock most of the stuffing out of Part II.

2.3 Part I is, in fact, full of flaws. The basic argument is straightforward and convincing: Suppose that all a mathematician's reasoning techniques could be embodied in an algorithm A that was believed to be sound. For technical reasons, assume that when A is given a problem of the form "Will algorithm C_q(n) stop?," where C_q is the algorithm with code q, and n is input data, A signals that the algorithm will not stop by stopping. Soundness means that when A(q,n) stops we are guaranteed that C_q(n) will not. There is some k such that C_k(n) is the computation A itself deciding whether C_n(n) will halt. If C_k(k) stops, then C_k(k) does not stop (because A is sound). Therefore C_k(k) does not stop, and we believe it doesn't stop, because we believe that A is sound. But A fails to draw this conclusion (i.e., it fails to signal the conclusion by stopping), so it is unable to conclude something that the mathematician (i.e., we) can conclude. Therefore A does not in fact coincide with the algorithm used by the mathematician. But the only feature of A that we assumed was soundness. Therefore (to quote Penrose, p. 76) "Human mathematicians are not using a knowably sound algorithm in order to ascertain mathematical truth."

2.4 As I said, the argument is convincing. But it's also a bit anticlimactic, since no one in his right mind would suppose that that human mathematicians "use" (or embody) a sound algorithm, let alone a "knowably" sound one. To verify this point, you need merely to find a case where a mathematician made a mistake. Penrose acknowledges this problem, and devotes most of a chapter, Chapter 3, to trying fix it, by showing that in spite of appearances human mathematicians really are sound theorem provers. His attempts fail.

3. Patching the Proof

3.1 His first move (p. 98ff.) is to try to argue that it is reasonable to abstract away from an individual mathematician and talk about "mathematicians" as a team, all of whose members basically agree on mathematical truths. "Are mathematicians' judgements actually so subjective that they might disagree in principle as to whether a particular ... sentence has, or has not, been established as true?" The problem is the phrase "in principle." Of course, in principle all mathematicians agree, and all nations want world peace. But just the other day I received a flyer for a journal called the "Mathematical Intelligencer" that airs mathematical controversies. The flyer contained some choice quotes from a recent article by mathematician A about why B's published proof of some supposed theorem was inadequate, and promised a reply from B in a future issue. The quotes made it clear that A was not objecting to B's logical system, but to some claims B had made that certain lemmas were obvious. What would Penrose say about such a case? I suppose he would say that as long as a public controversy was going on, all claims involved should be placed in escrow. But what about cases where B gets away with his claims, and no A has the time or inclination to notice gaps? Isn't this the usual situation? Isn't B usually right (but not always)? In view of all this, it seems as though if there is a body of mathematical claims endorsed by the mathematical community, the current set of claims is always inconsistent.

3.2 To take one famous example, cited by Ball and Coxeter (1939), in 1879 A.B. Kempe published a proof of the four-color theorem (Kempe 1879). According to Ball<1> and Coxeter, the bug in Kempe's proof was not disclosed until the publication of Heawood (1890). Hence for that eleven-year period the mathematical community was in a state of contradiction, and there is no reason to suppose any other period is immune.

3.3 Human mathematicians do not generate an answer to a problem and then stop thinking about it. In fact, human mathematicians never stop, except for reasons irrelevant to the kind of in-principle argument we're doing here. Consider, for example, this passage from Kaplan and Montague (1960), concerning the Hangman Paradox:
Before the appearance of Shaw's (1958) article, we had considered a form of the paradox essentially identical with his, and found it, contrary to his assertion, not to be paradoxical. At the same time we were successful in obtaining several versions which are indeed paradoxical. The present note is intended to report these observations.
3.4 In other words, they started thinking about the problem, derived an analysis of it, found flaws in someone else's analysis, then kept analyzing. And the result of all their cogitation (to date, that is) is a paradox, an elusive inconsistency in some seemingly straightforward postulates! It is difficult to see how thinkers like these could even be remotely approximated by an inference system that chugs to a certifiably sound conclusion, prints it out, then turns itself off.

3.5 For other examples, see (Lakatos 1976).

3.6 Penrose tries to solve this problem by distinguishing between "individual mistakes---or 'slips'" and "correctable" errors on the one hand, and "unassailable" conclusions on the other. He supposes that the unassailable conclusions are what we really care about, and that these must be the output of a sound reasoning system of some kind. "... We are not concerned with inidividual mistakes---or 'slips'---that a mathematician might happen to make while reasoning within a consistent overall scheme." (p. 138) Then we take as the output of our formal system the conclusions the computer takes to be unassailably true. The resulting system then must then suffer from the incompleteness described above.
If our robot is to behave like a genuine mathematician, although it will still make mistakes from time to time, these mistakes will be correctable --- and correctable, in principle, according to its own internal criteria of 'unassailable truth.' ... If we are supposing that our robot is to be capable of attaining (or surpassing) the level of mathematical capability that a human being is in principle capable of achieving, then its concept of unassailable mathematical truth must also be something that cannot be attained by any set of mechanical rules that can in principle be perceived as sound .... We are to assume that the robot ... possesses a ... secure level of unassailable mathematical 'belief,' so that some of its assertions---attested by some special imprimatur, which I denote here by a symbol 'STAR,' say---are to be unassailable, according to the robot's own criteria." (pp. 157--159)
3.7 This move raises two obvious questions: What is meant by "unassailability"? and, What in human practice corresponds to tagging assertions with STAR? The answer to the second question might be "publication in a reputable journal." But of course errors do occur in even the most carefully refereed journals, as in the example described above. Perhaps unassailability is flagged by placing a statement in a journal and then not having it be corrected over some period of time, say a hundred years. But of course there are errors in journals that have gone undetected for more than a hundred years.

3.8 Penrose is quite vague about what unassailability is. He comes close to endorsing the view that unassailability means provability. "Is it really plausible that our unassailable mathematical beliefs might rest on an unsound system...?" he asks rhetorically on p. 138, thus seeming to imply that unassailable beliefs rest on some kind of "system," or "scheme," as in the quote I cited above from the same page. But of course it can't be a formal system or scheme. That is, there can't be a formal mathematical theory such that everything unassailable is provable in that theory. That's what Gödel proved. So unassailability comes down to some kind of subjective feeling. On pages 138-141 he talks about the case of Frege's famous reaction to Russell's discovery of a paradox in his life's work, and seems to imply that mathematicians were sloppy then in ways we've learned to overcome. "Perhaps mathematicians have now become more cautious as to what they are prepared to regard as 'unassailably true'---after a period of excessive boldness (of which Frege's work indeed formed an important part) at the end of the nineteenth century." The way I read this is, if Frege had simply resolved not to be so reckless, he would never have been tempted to publish something that, deep down, he knew all along was not unassailable. I realize that I'm caricaturing Penrose's view, but it's awfully hard to get a precise reading about what he means. He wants unassailability to be both informal and guaranteed accurate, and I don't see how that's possible.

3.9 In anticipation of these problems, in Section 3.2, p. 131, Penrose tries to defuse the idea that human mathematicians might be modeled by an unsound algorithm, by shifting gears substantially:
...Unsoundness does not help at all for a known formal system F which ... is actually known---and thus believed---by any mathematician to underlie his or her mathematical understanding! For such a belief entails a (mistaken) belief in F's soundness. (It would be an unreasonable mathematical standpoint that allows for a disbelief in the very basis of its own unassailable belief system!). Whether or not F is actually sound, a belief that it is sound entails a belief that G(F) [essentially the sentence stating that C_k(k) doesn't halt] ... is true, but since G(F) is now---by a belief in Gödel's theorem---believed to lie outside the scope of F, this contradicts the belief that F underlies all (relevant) mathematical understanding.
3.10 This is really a different argument than the one we started with. The second one starts from different premises, and arrives at a different conclusion, namely that no entity can believe that any given algorithm can be responsible for its thought processes, because it would have to believe that the algorithm is sound, and then would reach a conclusion that that algorithm doesn't reach. The earlier conclusion was correct but unthreatening; the new conclusion is just false, for a clear reason: believing that a certain formal system underlies my reasoning processes does not entail belief that the formal system is sound. The problem is in the meaning of the phrase "formal system." What AI is interested in is not formal deductive systems whose theorems are exactly the "unassailable" mathematical conclusions, but in formal state-transition systems, that is, computers and programs. Confusion between these two concepts is so common that a brief digression on the topic may be worthwhile.

4. Informality From Formal Systems

4.1 Digital computers are formal systems, but the formal systems they are are almost always distinct from the formal (or informal) systems that their computations relate to. To analyze a digital computer as a formal system is merely to express its laws of operations in the form of transition rules among discrete states. When we take the inputs and outputs of the computer to refer to various real states of affairs, thenit need not be the case that there exists a consistent or sound formal system C such that whenever the computer concludes Q from P, the conclusion is licensed by C. Nothing prevents me from writing a program that, given any input P, prints out "P and not-P." There is, of course, a formal system S that stipulates that exactly this event is to occur, but this formal system is not about the entities mentioned in sentence P. If it's about anything, it's about the states of the computer, and nothing more. To make this point vivid, note that S, even though it prints out self-contradictory sentences, is "consistent," considered as a formal system, because it never says that the computer is to be in two distinct states at the same time. Consistency is essential to a formal system, because in almost all formal logics anything at all follows from a contradiction. Consistency is not, however, essential to computers. To continue my trivial example, I could take my self-contradictory program, and alter its behavior slightly, so that in response to "The earth is flat," it would say "False," and in response to "The earth is not flat," it would say "True," whereas in response to all other inputs P it continued to respond "P and not-P." Computers are not committed to inconsistency on all issues after revealing an inconsistency on one issue, any more than people are.

4.2 Hence if someone were to show me a listing and claim it embodied me, I would have no reason at all to believe that its conclusions were always correct (quite the contrary!). So Penrose's second argument is just fallacious. He very much wants to believe that the existence of artificially intelligent mathematicians would entail the possibility of an all-encompassing axiomatic mathematics ("the very basis of its own unassailable belief system"), but it just wouldn't.

4.3 Hence this second argument (whose conclusion I'll call Nontheorem 1) is wrong, and the first, "Theorem 1," that human mathematicians don't use a sound formal system to do mathematics, is correct but harmless to AI research. It provides no evidence at all against the proposition that someday we'll have algorithms that are just as good, and just as fallible, as human mathematicians.

5. Ensembles of Random Theorem Provers

5.1 In addition to the basic bug in Penrose's argument, there are lots of little bugs, having to do with various technicalities, and I fear that unless these are discussed, the impression will be left that somehow by hard work Penrose has succeeded in revising his theorem to the point where it's actually relevant and true.

5.2 Penrose raises the possibility that randomness plays a key role in humans' mathematical abilities, and that such randomness might account for the errors people make. "It would be reasonable to suppose that whenever the robot does make an error in one of its STAR-assertions, then this error can be attributed, at least in part, to some chance factors in its environment or its internal workings." (p. 169) So we would have to include a random input to our robot mathematician, and this would apparently vitiate the Gödelian argument. Not so, according to Penrose: The robot's "environment can also be provided as some kind of digital input," and if we can take as our computational entity the ensemble of all possible robot + environment combinations, then "there will be a finite number of ... possible alternatives" (p. 169). I am a little unsure of exactly what Penrose means by "alternatives," but I think he means possible A + environment pairs. "Thus, the entire ensemble of all possible robots ... will itself constitute a computational system.... One could see how to build a ....Turing machine ... that could carry out the simulation, even though it would be out of the question actually to carry it out" (still p. 169). Now we can detect that a STAR-assertion is in error by letting the robots vote. Since errors in STAR-assertions are rare, random, and statistically independent, it will be essentially impossible for a majority of robots to be wrong, so the ensemble will serve as the A necessary to get the argument moving.

5.3 There are two huge problems with this idea. The first is that it seems to assume that A is a distillation of a human mathematician that leaves out absolutely every human motivation or interest except mathematics, and even within mathematics leaves out everything except a single problem we've told it to work on. Hence if several copies of A are told to work on a problem, and are also given an "environment" to move around in (simulated, of course), then they will all generate outputs on about the same time scale and then stop. But what if it's the case that human mathematicians can't get far without collaborating with other human mathematicians? Won't the environment have to include them? What if A develops an interest in mathematics because of a beloved third-grade math teacher? We'll have to throw the teacher in, too. What if some A's become devotees of category theory, and others can't stand it? How will we cajole the second group into solving problems in category theory?

5.4 It seems to me that we are led to the idea that the only way to implement A is to simulate billions of copies of the entire universe on a Turing machine, and hope that a significant number develop a community of mathematicians that find our problem interesting. Okay, we're talking "in principle" here, so I'll grant that. What I won't grant (and this is the other huge problem with the idea) is that this ensemble of universes implements a sound inference algorithm that we believe is sound (which is required for Theorem 1). The computation is dominated by a simulation of the physics of the world. It's not clear how we're even going to find the mathematicians, more of which can presumably evolve as the simulation progresses, let alone be sure that they obey our STAR convention.

5.5 The situation is even worse with respect to Nontheorem 1, which requires us to postulate that the ensemble of universes hypothesizes itself to be a particular sound inference algorithm. Even if we grant, only for the sake of argument, that each element of the ensemble contains one or more pieces that hypothesize that they are sound inference algorithms, that doesn't mean the ensemble entertains this hypothesis or any other hypothesis.

5.6 The sheer scope of the simulations required to run the argument bothers even Penrose. "The reader may still have the uneasy feeling that no matter how careful we have been, there may still be some erroneous ... STAR-assertions that could slip through the net.... Soundness requires that absolutely no erroneous STAR-assertions<2> are included.... This may still seem to us, and perhaps to the robots themselves, to fall short of certainty---if only for the reason that the number of possible such assertions is infinite." (p. 173) To circumvent this problem, he develops an argument in Section 3.20 that only a finite number of STAR-assertions need to be considered. This argument is intricate, and seems at first to contradict the following elementary theorem of computability theory: For all c, there is a Turing machine that can deduce exactly the true mathematical statements of length c. The Turing machine merely contains a table of all theorems and nontheorems. Of course, this machine cannot actually be constructed without knowing which statements are theorems, which may serve as a warning about the exact status of existence claims in computability theory.

5.7 As I said, Penrose's argument seems at first to contradict this fact about bounded-length theorems. But his argument avoids this problem because it says that for any given putative theorem prover, there exists a bound c such that we can find a provable statement of that length or less that the prover can't prove. The argument is quite dense and hard to follow, and it seems to vacillate between trying to be a variant of Theorem 1 and trying to be a variant of Nontheorem 1. I think, though, that I can extract the essence of the argument, and possibly even strengthen Penrose's conclusion so that it applies to any of a certain class of probabilistic inference algorithms, not just the somewhat bogus ensemble of universes that I discussed above. I will put the argument in the style of Theorem 1, so that we don't have to use the dubious postulate that if an inference system entertains the hypothesis that it is implemented by a given algorithm, it must assume that the algorithm is sound.

6. Random Ensembles and Gödel Gaps

6.1 We start by assuming we have an algorithm Q that is T(c)-reliable, in the following sense: We give Q the problem of deciding whether computation q would halt on input d. I'll write Q(q,d) to refer to the running of algorithm Q on inputs q and d. Q is supposed to compute for a while, then print out Y if q would halt on d, else N. Suppose that for all problems such that size(q) + size(d) =< c, there is a time T(c) such that if the algorithm answers Y or N within this time, it is always right. To be concrete, we'll say that the algorithm is "T(c)-reliable" if, whenever it says Y or N before time T(c)/2, and then doesn't change its mind before time T(c), then what it says is actually true. Q is a probabilistic Turing machine, which means that it has access to an extra input tape containing an infinite sequence of completely random bits. With different input tapes, it might come to a conclusion about different <q,d> pairs, but it's never wrong about a pair it comes to a conclusion about.

6.2 Penrose has in mind that his ensembles of computers are such a system, which he parameterizes with several parameters, not just the T(c) I am using. But I think the theorem works just as well for this somewhat broader class of inference algorithms. Let's use the term "short_c-theorems" for formulas of the form $halts(q,d)$ and $not halts(q,d)$, for which size(q) + size(c) =< c and Q(q,d) prints Y or N reliably within time T(c), as described.

6.3 Here's what it would mean for the community of human mathematicians to be T(c)-reliable in this sense: Suppose we give the mathematicians a problem, Does q halt for input d?, to work on, where size(q) + size(d) =< 100. After 500 years, if they don't have a solution, we just forget about this problem. Otherwise, they'll say Yes or No, so we give them another 500 years to try to find an error in their solution. If they stick by it after that time, we label it "proved." Let's suppose that no buggy solution survives this filtering process, and that, given 1000 years, the mathematical community would never let a buggy solution to a problem of size 100 remain STAR-flagged. And it might be the case that for all c, we could take T(c)=10c and the human mathematical community would never make a mistake about a size-c problem given 5c years to solve it and 5c years to check the solution. If you don't buy that, perhaps it will help to let T(c)=100^c, or take T(c) to be any other computable function whose algorithm has size O(log c), an easy requirement to satisfy.

6.4 Now what Penrose does in essence is to define a derived computational system Q_c(a) that takes as input an algorithm description a, and runs Q(q,d) for all inputs q and d such that size(q) + size(d) =< c. It runs Q for only T(c) time units per <q,d> pair, and collects all the short_c-theorems. It then enumerates all deductive conseqences of these theorems (each of which is a formula of the form $halts(q,d)$ or $not halts(q,d)$). If $not halts(a,a)$ ever appears in this enumeration, then Q_c(a) stops. Otherwise, it goes on forever. Clearly Q_c is sound for all c, in the sense that if it halts for input a then machine a actually runs forever given a copy of itself. What we now show is that it has the usual blind spot.

6.5 Penrose's key observation is that the size of Q_c, considered as an algorithm, grows only slowly with c. That's because c occurs in it only as a loop limit and as the argument to T(c), which itself (i.e., size(code(T))) grows only slowly with c. Hence it is easy to pick a c* such that 2.size(code(Q_c*)) =< c*. Define Q* to be Q_c* . If we let k=code(Q*), then consider what happens when we run Q* with argument k, a computation we call Q*(k). The conclusion $halts(k,k)$ or $not halts(k,k)$, if derived, will follow from a Q computation of size =< c* (because k+k =< c*), so if that conclusion is included in the short_c*-theorems it will be correct. Now if Q*(k) halts, then it says Q*(k) does not halt, so by soundness it must not halt, but Q* cannot infer it (the usual Gödel gap). Every short_c*-theorem of Q is a theorem of Q*, by construction, so Q does not give an answer on the input <k,k>.

6.6 So far the argument is unproblematical (and quite ingenious), and shows that any T(c)-reliable algorithm must be incomplete. We can call this Theorem 2. The only trouble is that Penrose can't quite get from that conclusion to the one he wants, which is that the incompleteness occurs at a point where humans have no trouble drawing the correct conclusion. And at first blush this looks like such a case. Just take c* to be greater than the number of characters in the conclusion, and you have a short_{c*}-theorem for people that isn't a short_{c*}-theorem for Q. Unfortunately, that isn't quite as easy as it sounds. In the proof of Theorem 1, we were asked to picture a situation where we had a listing of an algorithm that was claimed to embody us. We were then given a theorem that the algorithm couldn't prove, except that we weren't really given it --- we were given a way of constructing it from the listing. Suppose that AI triumphs completely, and you hold in your hand a CD-ROM containing a listing of a computerized Gauss (call it G). Can you then apply the construction described above to derive a theorem that mathematicians find easy to prove and that G cannot prove? No, because G is not the Q we need to start the construction. To create Q, we would need to simulate lots of mathematicians (including a von Neumann and maybe even a Penrose as well as our Gauss), plus a large chunk of their environment. It's not at all clear that AI research would ever get to the point where it could take a stand on the existence or nature of Q. Furthermore, suppose that a candidate for Q were suggested. How would we evaluate it? In particular, how would we ever prove that it was T(c)-reliable? We would have to show somehow that no matter what random bits were input to the algorithm, it would never make a mistake. I conjecture that the possibility would always remain open that both the algorithm and the human mathematical community are not T(c)-reliable. Even worse, there's no way even in principle that we could determine that Q duplicated exactly the conditions prevailing in our universe. The best we could hope for is that Q be indistinguishable from our universe, that it apparently yield "typical" behaviors. But it could be the case that arbitarily small physical differences could change a T(c)-reliable universe into a non-T(c)-reliable one.

7. The Infallibly Fallible Robot

7.1 Penrose ends his treatment of Gödel's Theorem with the strange fantasy of Section 3.23, in which a robot mathematician (MJC, for "Mathematically Justified Cybersystem") gets into an argument with its human creator (Albert Imperator, or AI). AI convinces MJC that if MJC and its fellow robot mathematicians even entertain the possibility that they embody any algorithm Q, then there is a sentence (which I'll call Omega(Q), simplifying Penrose's notation a bit) that is true but unprovable by MJC and its robotic colleagues if they are infallible. MJC sees the logic of this argument, and, refusing to abandon belief in its infallibility, goes insane. AI is forced to destroy MJC and all its fellows.

7.2 This fantasy is incoherent at several levels. It seems to assume that infallibility is such an integral part of the AI research program that the robots can not even conceive of not possessing it. Yet MJC demonstrates spectacular fallibility in concluding at the end of the dialogue that its initials actually stand for Messiah Jesus Christ and that it is divinely guided to its mathematical conclusions. It seems to me that it would be much less traumatic for MJC just to say, "I guess we must very occasionally make mistakes; in fact, my impetuous assertion of infallibility was just such a mistake!"

7.3 The dialogue has MJC hearing and agreeing with the argument for Omega(Q). "Yet .... it's impossible that [we] can accept Omega(Q), because, by its very nature of your Gödel's construction, is something that lies outside what can be STAR-asserted by us.... It must be the case that ... the procedures incorporated into Q are not after all the ones you used." (p. 183) Surely MJC has a pretty good case here. It is agreeing with the argument for Omega(Q); it even shows that it understands several implications of it. It sounds odd for AI and Penrose to continue to talk as if MJC really is unable to conclude Omega(Q). If MJC were to affix a STAR to Omega(Q), on what grounds would we quibble?

7.4 Of course, in imagining MJC's possible behavior, I'm just coasting on the anthropomorphic fuel Penrose provides by painting such an extravagant picture of what MJC can do. And that brings me to what is really incoherent about this dialogue. It seems to knock the keystone out of Penrose's whole argument, which is that finding one tiny gap in the ability of robots to do mathematics would destroy any hope that they could ever really understand anything. If that's the case, then he would presumably believe that nothing like the dialogue between AI and MJC, in which the robot seems to understand every nuance of the conversation, could ever actually take place. The casual reader, who is urged by Penrose to skip all the hard stuff in Chapter 3, and go right to Section 3.23, is surely going to draw the conclusion that Penrose thinks that robots can do almost any task, except prove a certain theorem. He titles the section "Reductio ad absurdum--- a fantasy dialogue," and I suppose it could be taken as trying to show that no matter what powers of understanding we imagine we could give to robots, we will also have to imagine them having strange lapses (but strange lapses that are consistent with infallibility in some way), and that therefore we mustn't impute those powers. But it's as if I presented a proof that all toasters were useless by hypothesizing a talking toaster and showing that it must burn at least one slice of toast.

8. What Does Penrose Think AI Is?

8.1 Now that I've torn Penrose's argument to shreds, it's time for a spirited rebuttal of his critique of the computationalist theory of consciousness. Unfortunately, Penrose has no critique. Indeed, he says almost nothing about points of view different from his. The two and a half pages of Section 1.14, "Some Difficulties with the Computational Model," are almost all there is. There's a brief reference on p. 149 to what he supposes is the computationalist view of mathematical ability, a somewhat odd discussion of "top-down" vs. "bottom-up" programs in Section 1.5, and a few other remarks in passing throughout the book. One might conclude from this silence that AI has had nothing in particular to say about consciousness, but in fact there has been quite a bit of theorizing. In particular, between the publication of Penrose's previous volume and "Shadows of the Mind" appeared Daniel Dennett's "Consciousness Explained," which provides a rich set of ideas for thinking about computation and consciousness. I would have been quite interested in seeing Penrose's critique of that set of ideas. But there are (by my count) exactly two references to Dennett in "Shadows of the Mind," both in passing.

8.2 Let me deal with his observations, what there is of them, in reverse order. Section 8.2, "Things that computers do well---or badly," distinguishes problems on which we would expect computers to do better than people from problems on which we would expect people to do better. The analysis is "a little crude," as Penrose admits, but basically correct. Suppose a problem can be analyzed as a search space with a branching factor of p. Then a computer might examine on the order of T=t.p^n search states if the solution is m moves away and it takes time t to explore a state. "It follows ... that games for which p is large, but can effectively be cut down significantly by the use of understanding and judgement, are relatively to the advantage of the human player." (p. 397) One might wonder what this has to with consciousness, but Penrose, as I said before, assumes that awareness and judgement are two manifestations of the same underlying property. "...The essential point is that the quality of human judgement, which is based on human understanding, is an essential thing that computers lack, and this is generally supported by the above remarks...." But nothing of the sort follows from the formula T=t.p^n. AI practitioners routinely think in terms of this formula when they look for heuristics to cut down p. Furthermore, there is no reason in principle why the computer needs to stay in one search space. The problem of finding the right search space can sometimes be phrased as a search problem itself.

8.3 Finally we work our way back to Section 1.14, which is a review, at a sketchy and shallow level, of "difficulties" with the computational model. At the risk of stating the obvious several times, let me review these "difficulties."

8.4 On p. 42, he says,
....It is the mere 'carrying out' or enaction of appropriate algorithms that is supposed to evoke awareness. But what does this actually mean? Does 'enaction' mean that bits of physical material must be moved around in accordance with the successive operations of the algorithm? Suppose we imagine these successive operations to be written line by line in a massive book. Would the act of writing or printing these lines constitute 'enaction'?
8.5 Presumably awareness will not be "evoked" by some computation; it will be constituted by some computation, and not just any computation. (See below.) And "enaction" does not mean recital of a sequence of operations; it means taking part in a certain interaction with the environment. It's as if someone objected:

It is the mere 'carrying out' or enaction of appropriate switch transitions that is supposed to control a furnace. But what does this actually mean? Does 'enaction' means that bits of metal must be moved around in accordance with the successive operations of the thermostat? Suppose we imagine these successive switch transitions to be written line by line in a massive book. Would the act of writing or printing these lines constitute 'enaction'?

8.6 The same objection has been made, with slightly more subtlety, by John Searle (1992) and Hilary Putnam (1988). In each case it rests on a perverse identification of a computer program with a trace of its execution (a particular trace of a particular execution), which is simply absurd.

8.7 "In any case," continues Penrose,
it would presumably not be the case, according to [computationalism], that just any complicated algorithm could evoke ... awareness. It would be expected that some special features of the algorithm such as 'higher-level organization', or 'universality', or 'self-reference', or 'algorithmic simplicity/complexity', or some such, would be needed before significant awareness could be considered to be evoked. Moreover, there is the sticky issue of what particular qualities of an algorithm would be supposed to be responsible for the various different 'qualia' that constitute our awareness. What kind of computation evokes the sensation 'red', for example? What computations constitute the sensations of 'pain', 'sweetness', 'harmoniousness,' 'pungency', or whatever? Attempts have been sometimes been made by proponents of [computationalism] to address issues of this nature (cf. Dennett 1991, for example), but so far these attempts do not strike me as at all persuasive. (p. 42)
8.8 It may be that Penrose finds the computationalist theory unpersuasive, and in a sense he's surely right. No one has a completely worked out theory of consciousness, Penrose least of all. But it would have been sporting of him to tell the reader what he takes the computationalist position to be before dismissing it. Since he didn't, I will. What follows is my interpretation of a theory due to Minsky (1968, 1986) and Dennett (1991). (See also Gelernter 1994.) I make no claim, however, that I am representing their views or the views of a majority of the AI community.

9. The Computationalist Alternative

9.1 The basic idea is that a computational system can often be said to have a model or theory of some part of its environment. I hesitate to use either the word "model" or "theory" here, because of the danger that some will assume I mean to use these words in the senses they have in mathematical logic, and I emphatically do not. Perhaps "simulacrum" is the right word; some computational systems maintain simulacra of some part of their surroundings. A simulacrum allows the system to explain and predict the behavior of the world around it. It's very important at the beginning of the exegesis to understand that when I use words like "explain" and "predict" I mean them in the least anthropomorphic way possible, as when one might say that an antiaircraft missile predicts the future locations of its target.

9.2 Simple systems can get by with simple simulacra, but the more complex the organism, the broader must its skills be in relating one part of its environment to others, so that at some point it becomes legitimate to talk of the organism's simulacrum of the world. And at some point the organism must include itself in the model. This is not meant to be a mystical step. A computer taking an inventory of office furniture will include itself in its simulacrum. Of course, nowadays the computer will not distinguish itself from other workstations, or hat racks for that matter. But if the same computer is used to control the movements of the office furniture (using robotic sensors and effectors), then some interesting singularities arise. Some items of furniture will, as they are moved, give rise to moving patches of pixels in the images the computer's camera produces. But at least one item, the camera itself, will cause quite different sensory events when it is moved. The computer's world simulacrum must, to be accurate, reflect the asymmetry between these different kinds of physical objects.

9.3 So far, no consciousness, and nothing out of the ordinary either. We have robots in our lab that watch their arms move toward targets, and they use different models for the arm and the target (Grunwald et al. 1994). The point where consciousness arises is where an agent requires a model of itself as a behaving agent, and even there consciousness does not depend on the agent having just any model of itself; it must have a model of itself as a being with free will, a transcendental ego, sensations with certain qualia, and so forth. This model is based on attributes that the being really does have. Free will is based on the fact that the computations the agent carries out really do influence its behavior. The transcendental ego is based on the fact that the agent must behave as a single entity. Qualia are based on the fact that sensory information really is processed. The model goes beyond the truth, but it's not really a lie; it's a self-fulfilling fiction.

9.4 One pleasant (perhaps suspiciously pleasant) aspect of this theory is that it explains so nicely why the theory seems incredible. Our self-models deny that things like qualia are computational entities. Of course, they also deny that qualia have to do with large-scale quantum coherence, or any other physical phenomenon. That's why qualia seem so mysterious: any explanation of consciousness in terms of nonmysterious entities is ruled out as if by reflex.

9.5 This theory has plenty of difficulties. To my mind, its biggest problem is that it raises a question that it has yet not answered, which is: How do we tell when a computational system X has a simulacrum of entity Y? The answer cannot depend on whether it's convenient for outside observers to impute this property to X. We have to start from an observerless universe and infer observers. But I don't think these problems are insurmountable, and they suggest some interesting lines of research.

9.6 The theory also makes a prediction, which Penrose anticipates on page 42:
"...Any clear-cut and reasonably simple algorithmic suggestion [for a theory of qualia] ... would suffer from the drawback that it could be implemented without great difficulty on a present-day electronic computer. Such an implementation would...have to evoke the actual experience of the intended [quale]. It would be hard ... to accept seriously that such a computation ... could actually experience mentality .... It would therefore appear to be the case that proponents of such suggestions must resort to the belief that it is the sheer complication of the computations ... that are involved in the activities of our own brains that allow us to have appreciable mental experiences.
9.7 The first half of this paragraph is correct; the second half is wrong. It does seem to be the case that consciousness is no big deal. I believe I could program a computer to be conscious; it may have already been done by accident. The reason why it's so hard to detect is because computers are so stupid and clumsy. It's child's play to program a computer to perceive its own sense-event descriptors, but if it can't actually see anything interesting, and can't really carry on a conversation, then it won't have much to say about its introspections. Hence the bottleneck in getting computers to be conscious is getting them to be smart. Intelligence is a prerequisite for (recognizable) consciousness, not the other way around, as Penrose would have it. "Sheer complication" is a red herring. The cerebrum is conscious, and the cerebellum is not, because it uses a certain kind of model of itself, and the cerebellum doesn't. The kind of intelligence that I am talking about here is not what is measured by IQ tests, but a general ability to integrate information about the world. I'm quite sure that mammals have enough intelligence to be nontrivially conscious, and quite sure that existing computer programs do not.

9.8 Curiously, the idea that consciousness will turn out to be quite simple is in harmony with Penrose's ideas. If we flip back to page 149, we find him expressing much the same conclusion in his framework: "[Understanding] need not be something so complicated that it is unknowable or incomprehensible.... Understanding has the appearance of being a simple and common-sense quality."

9.9 This is not the only place where Penrose's views run parallel to the computationalist view. The second half of the book is taken up with the problem of the observer in quantum mechanics, the same problem he wrestled with in "The Emperor's New Mind." As I mentioned above, for computationalism the problem arises in finding an objective way to draw lines around systems that model themselves. In quantum mechanics the problem arises at a more fundamental level, when we try to find macroscopic objects in a world of wave functions. But it's likely a solution to the quantum-mechanical observer problem would shed light on the computational observer problem.

9.10 To summarize: Computationalism is scarcely examined, let alone refuted, by this book, which stakes all its marbles on the Gödelian-gap argument, and loses. A computational theory of consciousness has many problems, but is better worked out than any alternative, including especially Penrose's. It is not arrogance, but a humble desire for truth, that leads some researchers to pursue the computational theory as a working hypothesis. The biggest obstacle to the success of this theory is not the absence of an account of conscious awareness per se, but the fact that AI has as yet made little progress on the problem of general intelligence, and has decided to focus on a more modest strategy of studying individual cognitive skills. The burden is on AI to show that this research program ever will lead to a theory of general intelligence. People like Penrose should declare victory and withdraw.

Acknowledgements

Thanks to Richmond Thomason for helping me with bibliographical searches.

Notes

<1> Roger Penrose is the Rouse Ball Professor of Mathematics at the University of Oxford. Same Ball.

<2> Penrose actually has STAR_M-assertions here and in a couple of my later quotes. I don't think the distinction between these and STAR-assertions simpliciter is important for my discussion.

References

Ball, W. W. R. & Coxeter, H. S. M. (1939) Mathematical Recreations and Essays. 11th Edition. New York: The Macmillan Company

Dennett, D. (1991) Consciousness Explained. Boston: Little, Brown and Company

Gelernter, D. H. (1994) The Muse in the Machine. New York: Free Press

Grunwald, G. Hager, G. & Hirzinger, G. (1994) Feature-Based Visual Servoing and its Application to Telerobotics (with and ). Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 164--171. IEEE Computer Society Press

Heawood, P.J. (1890) Quarterly Journal of Mathematics, 24, 332-338.

Kaplan, D. & Montague, R. (1960) A paradox regained. Notre Dame Journal of Formal Logic, 1, 79-90. Reprinted in Formal Philosophy: Selected Papers of Richard Montague, Richmond Thomason (Ed). New Haven: Yale University Press

Kempe, A.B. (1879) American Journal of Mathematics, 2, 193-200.

Lakatos, I. (1976) Proofs and refutations: the logic of mathematical discovery. John Worrall and Elie Zahar (Eds). Cambridge: Cambridge University Press

Minsky, M. (1968) Matter, mind, and models. In Semantic Information Processing., Marvin Minsky (Ed). Cambridge, Mass: The MIT Press

Minsky, M. (1986) The Society of Mind. New York: Simon & Schuster.

Putnam, H. (1988) Representation and reality. Cambridge, Mass.: MIT Press

Searle, J. R. (1992) The Rediscovery of the Mind. Cambridge, Mass.: MIT Press

Shaw, R. (1958) The paradox of the unexpected examination. Mind, 67, 382-384.


Return to PSYCHE home page