Computability, Self-Reference, and Self-Amendment George Kampis Dept. of Ethology, ELTE University, Budapest and Dept. of Theoretical Chemistry, University of Tùbingen, Germany In: Communication and Cognition - Artificial Intelligence, Vol. 12, Nos. 1-2, 1995, pp. 91-109, Special Issue Self-Reference in Biological and Cognitive Systems, Luis Rocha (ed.) Abstract There exist theories of cognition that assume the importance of self-referentiality and/or self-modification. We argue for the necessity of such considerations. We discuss basic concepts of self-reference and self-amendment, as well as their relationship to each other. Selfmodification will be suggested to involve non-algorithmic mechanisms, and it will be developed as a primary concept from which self-reference derives. A biologically motivated mechanism for achieving both phenomena is outlined. Problems of computability are briefly discussed in connection with the definability and describability of self-modifying systems. Finally, the relevance of these problems to applications in semantic problems of cognition is shown. We proceed in the following way. The paper starts with an outline of the evolutionary approach to cognition, as that context where the problems of circularity and recursiveness can be raised. Next, complete and incomplete forms of self-references are discussed. The "causal" theory of self-referentiality is reviewed, and a thought experiment is presented, which points out that no computable model for complete self-reference can exist. On the other hand, constructive definitions are shown to offer a framework where "self- defining" and selfmodifying systems, if such exist in reality, can be formulated. Studying the realization problem, a general abstract model is given, and a "biological computation" mechanism that corresponds to it is outlined. The underlying phenomenon, called "shifting reading frame", is discussed in relation to how self-referentiality can be achieved through self-modification. The applicability of the approach to the autonomous definition of semantic relations in symbol systems, which may allow for a kind of autonomous "symbol grounding" is discussed. I. Introduction 91 Certain theories of cognition assume that selected aspects of mental activity can be characterized via self-referential or self-modifying constructs. Among them, the most widely known are those expressed in Hofstadter's "Gödel, Escher, Bach" [1979], and in Maturana and Varela's "autopoiesis" [1980]. The latter was further developed in Varela, Rosch, and Thomson's "The Embodied Mind" [1991]. Ideas of self-reference (and its self-modification), and their application to cognition have a much longer history, however. In mathematics, the problem of self-reference is more than two thousand years old. An obvious twentieth century example for its use is B. Russell's "barber paradox", that refuted the consistency of the Fregean system of set theory and logic, and led Russell to develop the theory of types. The cognitive and philosophical significance of such issues was first raised by the so-called BCL school, whose members and associates included W. McCulloch, W.R. Ashby, G. Günther, L. Löfgren, H. von Foerster, and H. Maturana. An important work closely related to theirs but developed independently was H.H. Pattee's on "semantic closure" [1973]. Also related and in part motivated by the BCL thinking were theories of "radical constructivism", a school particularly active in the German tradition ([Waczlawick 1984] offers a summary). Radical constructivism maintains that mental events bring forth rather than exploit the observations to which they refer, and that the two can be understood in a circular fashion where knowledge and the knower are considered to be each other's reasons [Maturana and Varela 1987], just as in evolutionary epistemology [Callebaut and Pinxten 1987], only here with an anti-realistic attitude. And there is a vast literature on self-reference proper: the best source known to me, Suber and Bartlett's "Self-reference: reflections on reflexivity" [1987], is a collection of papers with a long bibliography, that brings over 1000 items, and is far from being complete. In a tradition of different origin, the idea of self- programming, or self-modification has emerged as a possible remedy to the often noticed stiffness of formal systems. At the root of the idea we find the intuition that a system should sometimes become something else than it was before; in other words, that its defining symbols can be both too vague and too narrow in comparison to processes of interest. Part of what in the seventies were called "goal-seeking systems", "catastrophes", or "dissipative structures" (characterized by a computational bottomup approach that permits flexibility at a higher level) lived further in the eighties in connectionism and in the subsymbolic paradigm, or in the notion of "swarm intelligence". A good introduction to the flexibility aspect of subsymbolic computations is A. Clark's Micro Cognition [1989]. Self-programming computers that could directly alter their own codes were already studied by von Neumann [1966], and several authors afterwards. Whether and to what extent self-programming can be carried out algorithmically, was always unclear, however. We discuss some of the basic problems with notions of self- reference and self-modification in this paper, in particular, the problem of computability, and its relation to these concepts. In a recent book [Kampis 1991a] we have outlined a concept of self-modification beyond ordinary computation. Over and again, the paper will use material from that book. 92 II. Cognition: From Performance To The Process That Gives Rise To It We have quoted various predecessors, but there is also a direct form how our problems can be raised. At the outset, a useful distinction between "competence modeling" and something like the "origin of competence" should be made. "Competence modeling" means to focus on the repetition of some cognitive performance in an essentially functionalistic fashion, where all that we want is that a performance similar to that of a human is achieved by some machine at any rate, whereas the "origin of competence" problem would be the opposite, referring to the acquisition of such cognitive faculties that make the achievement of the given performance possible. Learning is one example for such a process, but it is not the only one, as indeed various self-organizing theories of brain and mind function exist the targets of which range from neural connectivity to the problems of memory and the semantics of language. The dynamic studies of cognition often use the metaphor of biological evolution. The parallel between the cognitive and evolutionary domains has been repeatedly noticed by many authors [Holland 1975, Csányi 1989, Dawkins 1975, Lumsden and Wilson 1981, Edelman 1987]. This metaphor proves very useful in many aspects of practical modeling. Yet, at the same time when evolution and its variation-selection mechanisms help illuminate several concepts in cognition, such as directed randomness, the spontaneous accumulation of useful information, or intrinsic parallelism, the metaphor also imports unsolved problems from evolution theory. For instance, a question that has already bothered Darwin was, wherefrom comes the new information? Or, to translate it to a more general language, why does evolution not stop after the necessary adaptations have taken place? Why are there always new species? The revolution of Darwinism was less about the existence of spontaneous evolution and descent, the concepts of which were already known before Darwin, but about another idea, which we would express today by saying that evolution is a productive process that needs an open universe. To say that evolution is a productive process is not only to mean that new forms emerge that have not existed before, but also to imply that these new forms can feed back to the system by setting new conditions for subsequent evolution. In this sense, evolution should be seen as a process that is its own product. In modern terms, biologists who recognized this problem started to speak of co-evolution [Van Valen 1973, Stenseth and Maynard Smith 1984], or recursive evolution [Conrad 1983] instead of simply any evolution. Van Valen's famous "Red Queen" theory says that successful adaptation in one species implies a worsening of the environment for others, so they have to re-adapt, or they have to perish, and in the end it takes all the running to stay where you are. The idea of recursive evolution is very similar, it refers to the fact that selective "tasks" and their "solutions" develop iteratively, the one giving rise to the other, and vice versa, so whenever a solution is achieved a new task is also defined. That is, we end up with a picture where circularity, and some sort of informational "self-sufficiency", or self-reference, of the source of internal change play an important role. 93 III. Forms of Self-Reference Let us briefly review the various forms in which circularities exist. Self-reference can be complete or incomplete, that is, full or partial. Most of the cases we know from logic and its related fields like argumentation theory offer partial self-references. Statements like that of the undecidable sentence "This sentence is false" (i.e., the well-known Epimenides paradox, also called the Liar, or the Cretan, and its variants the Gödelian sentence, or the hypothetical program that demonstrates the unsolvability of the Halting Problem) are by necessity only incompletely self-referential. Partiality enters here because the above sentence needs an external interpreter (like us) to make sense at all; terms like "this", "sentence" and "falsity" must be defined and understood a priori, and refer to classes outside the object under consideration. That is, literally, the sentence refers to something else also, not just to itself. By analogy it would seem now that complete self-reference should be impossible or unintelligible, as we always need some "seed" to start with. However, that is not the case, as the existence of coherence theories of semantics, self-consistent fields in electrodynamics, and mathematical results like those of Löfgren [1968] or Varela [1974] indicate. In other words, systems whose every term is internally defined by the very same system are possible. This can be illustrated on a well- known example. For an untrained Martian (or Hungarian, the difference is negligible), English (just as any other natural language) appears as a system of messages perfectly closed to itself. That is, the meanings of English words and sentences are given in the same English, which is of little help, unless you are a native speaker, and then you know English anyhow. This is a trap from which the only escape is through a door of extra-linguistic methods (such as that of pointing to the surrounding objects) -- or, by using dictionaries, but who made the first dictionary, and how?. Then, of course, we recognize that the act of pointing belongs to another language (namely, that of body talk), and so on, and therefore we are left with a hierarchy of languages, each inaccessible on its own, since body talk is no less esoteric than English if you don't happen to speak it (remember, you are a Martian now). If we know any one of these languages, we can probably bootstrap all, but without that, the system remains completely self-referential and opaque. In other words, there must be an "initial translation", in order to open up the circles of meaning. This thought experiment shows two points, the one is that complete self-reference is certainly possible, and second, that it does not make any sense at all. In other words, complete selfreference poses a problem for both description and handling. From a formal point of view, the results of Löfgren and Varela lead to the same conclusion. Löfgren [1968] has shown that the axiom of complete self-reference is independent from set theory and logic, and can therefore be added to it as a new primitive. However, it is impossible, by the same result, to explain this (complete) self- reference by decomposing or deriving it in terms of the other axioms. That is to say, a completely self-referential object necessarily looks like something inapproachable, a perfectly closed class, perhaps a Kantian "Ding and sich", or a Leibnizian "Monad". (Let us note for the sake of interest that Löfgren's work was also related to a biologically motivated mathematical problem, called perfect self-reproduction [Rosen 1959]). Varela's "Calculus for Self-Reference" [1974] reflects the same philosophy: he introduces self- 94 referentiality as a third logical value besides truth and falsity (or, to be more precise, besides indicational and non-indicational states in the non-standard logical calculus of G. Spencer-Brown [1969]). IV. The Ways of Self-Amendment So far, we have dealt with self-reference, but the situation is quite similar with the notion of self-modification. Partial self- modification is easy to achieve; the complete form goes beyond ordinary mathematics and anything we can formulate. Consider, for instance, recursive programs. Every recursive program can be said to modify itself in some sense, since (by the definition of recursiveness) the exact operation carried out at time t depends on the result of the operation at t-1, and so on: therefore, the final "shape" of the transformation is getting defined iteratively, in runtime (a fact somewhat obscured by the usual way in which recursion is written down in high-level programming languages like C). At the same time, as we can expect, to every finite recursive program there belongs an equivalent "straight" program, that uses no recursion at all, and is perfectly well defined in advance, so that it does not change in any respect; it is simply a fixed sequence of a priori given elementary operations. Now some advanced structures such as list-processing languages like LISP allow for a direct manipulation of their program texts during execution, thus permitting the user to construct programs that actively rewrite themselves. Yet, again, there is a top-down structure associated with these programs, and even if the program list is altered, somewhere in this structure there is at least one level, namely the topmost level (or the bottom-most level, depending on how we think of it) that never changes. In the case of the LISP program, we find this unchanged level when taking into account the LISP interpreter, that translates program instructions to machine code. It is the same interpreter, and no matter what we do, it does not get disturbed. But that is of course necessary, for otherwise we could not write and execute any program at all. Therefore, just as in the case of self-reference, and for the same reason, algorithmically derived self-modification can never achieve completeness: there must be a fixed point somewhere or the whole building collapses into nonsense. The most interesting part of the story comes when we push the previous exercise to the extreme. If now somebody writes a tricky language that goes beyond the capabilities of LISP and changes its own interpreter as well, and then perhaps it changes the operating system, and so on, finally we find ourselves at the level of the processor chip of the computer that carries out the machine code instructions. Now, unlike the earlier levels, this level belongs to a piece of physical hardware, where things will be done the way the machine was once built, and this can no more be a matter of negotiations. Ultimately, this is what serves as that Archimedean starting point (similar to the initial translation that opens up self-reference) that defines a constant framework for the programs. The importance of this is that we understand: self-modification, and self-reference, are not really just issues of programming (that is, of using the right software), but of designing a whole machine in some sense. Therefore, the impossibility of achieving complete self-modification 95 depends, ultimately, on the separability of machine from program (and the way around): the separability of software from hardware. V. Nomic, the Ultimate Game of Change The complications of self-modification can be further studied on a delightful example P. Suber presents in his "The Paradox of Self- Amendment" [1990]. Suber introduces a logical game called Nomic (that has many similarities to the "game" of legislation in the USA). Nomic formulates what can be called an "ultimate legal system". In a real-life legal system, laws regulate several things, including the way they can be changed. And laws sometimes do get changed. In Nomic the only function of laws is to change, and of course, to govern, at the same time, how they can be changed. This system is designed as an actually playable game, where a move of a player is the attempted change of one law. The laws are passed or rejected by votes of the players. Just how many votes are needed for a law to pass is regulated by the very rules of the game, that is, by the other laws which also take part in the game, and are possible subjects to such changes. Certain laws are (at least according to the initial rule set of Suber) more difficult to change than others; it is instructive here to think of the first group as "constitutional" rules, and the second group as "normal" laws and statutes. Also, the first set of rules has (at least initially) a logical priority over the second set, and the newer rules over the older ones; but any of these rules can be withdrawn at the players' own discretion, as the game proceeds. Just to make it clear how far this can go, let us take a bizarre example: if the players want, they can transform Nomicinto Monopoly or any other party game, just by democratically introducing "laws" for the board, the dice and the cards, together with rules for how to use them, and then all the players have to do is to completely remove the original set of rules that allowed for the change of the rules at all. And here it is, your Monopoly. (However, Suber's game has such a cleverly chosen initial rule set, that a typical player will not want to turn the game into Monopoly but he or she will rather want to win. The winning conditions, just like anything else, are again part of the initial rule set, so before you changed the game too much, you are better off if you have a new law passed that favors your own moves (whatever they are), than by trying to transform the game into Checker just because you are the world champ.) The actual game is somewhat more complicated than this (if that is possible), and involves the use of a two-chamber voting system, but these are unnecessary technicalities for the present discussion. The initial rule set consists of 29 different rules, some of which even offer conditions to decide whether a new rule is valid or not (i.e. whether the system is consistent). It is easy to generate paradoxes in Nomic. A paradox occurs, for instance, when one of the chambers passes a law that regulates the work of the other, but it must be first validated by that chamber. For instance, chamber A decides that decisions of chamber B must be made by a 6-3 96 majority instead of a prior 5-4, but chamber B, when controlling the legality of this move, rejects it, by a 5-4 decision, of course. What has happened? There are many other questions one could raise about Nomic, as was done by D. Hofstadter in his column in Scientific American [1985]. For instance, can every initial rule be actually changed (remember that the initial rule set favors democracy)? On the other hand, can certain rules of the game be made truly immutable without stopping the change of the others? Can sudden changes (like revolutions) of the rule set occur, without violating the rules? And, finally, can this game be made algorithmic, or computable? In the present form of Nomic, the changes come from "outside" the system, by the intervention of the players. Is it now possible to achieve the same outcome by the rules themselves? By this we mean: Is it possible to realize every desired transformation of the rule set by means of the same rule set? Can we write an initial rule set that prescribes every transformation of itself in such a way that even the further transformations of the by then already transformed rule set occur according to this original plan, without even the presence of the original rules? Similarly, can a legal system be designed, or, to proceed now backwards, to evolutionary processes, cognitive models, and like: can self-modifying processes be made equivalent with processes that use fixed rule sets? VI. Computability and Definability The notion of computability is often understood in a deliberately narrow sense, as something confined to decidability. According to this simplistic view, a process is computable if there is an effective procedure that decides its outcome. In other words, a process is essentially then computable if there exists a program that computes it and then halts. There is, however, a different way of looking at the issue of computability: we can say, alternatively, that a function is computable if it is definable by some procedure that guarantees its decidability. Mathematical logic has offered already a long time ago something very close to this definition procedure. It is called lambda-definition, a method invented by A. Church in the thirties. Now, intuitively, a definition is something more fundamental than the subsequent execution of an algorithm, and in this sense it is possible to say that, ultimately, computability boils down to definability. In the everyday context, this contention comes through in popular statements like the "hackers version" of the Church-Turing Hypothesis, according to which, "If you can define it, we can simulate it" [Kampis 1991a]. Put in this way, the crucial problem of complete self-reference can be reformulated as saying that a perfectly self-referential, or perfectly self-modifying system would not be defined at all, in the sense of being self-defining and hence empty. 97 VII. The Causal Theory of Self-Reference Despite the many difficulties surrounding the notion of complete self-reference, it is surprisingly easy to show that the concept can be consistent with linear causality and can therefore be realized by some natural processes. To see this, let us be more specific now as to what causes the self-referentiality of a linguistic expression. We do this in order to show later how this constraint can be relaxed, and hence selfreference eliminated, in a causal framework. An important distinction of formal systems should be used here: the use/mention dichotomy. According to that, these two types of dealings must be always kept separate. To mention a word, is to quote it -- that is, to "protect" it somehow, such as with embedding quotation marks, that prevent it from exerting any effect in a statement. To use a word is simply to utter it, and hence state and express it. That is, "use" is active, and "mention" is passive. It is obvious that in self-reference the two functions get mixed, we are making statements with words and sentences about the very same words and sentences, as if they were just mentioned: the same thing is active and passive at the same time. So, for instance, when unfolding partial selfreference, it is this double nature that is eliminated first. What a vocabulary does for us when we first start to learn English is to help making sense of words before we start using them. The sentence "This sentence...", however, defines the sentence and uses its meaning at the same time to express something about itself. We understand this sentence intuitively because our psychological understanding is not delimited by this expressed meaning and/or truth of the sentence, and so we can decompose the sentence into words with which we are familiar and we understand as passive tokens. In a situation involving complete self-reference, that is no more possible, however. Yet there is a solution. We can try to think of the expression "this sentence" as a GOTO statement in some computer program, a program statement that refers to another place, to a label that is defined completely separately. It can now happen that this label then turns out to be the label of the very same GOTO statement. But who cares? That makes no difference. And yet it does, for in this way self-reference can be produced; this is precisely what Gödel did when he constructed his undecidable sentence in two steps instead of one. To illustrate this method, we can say: "The next sentence is KLXPUZ127. KLXPUZ127 is false." No immediate circularity is involved: therefore, the underlying structure can be causally realized. Details of this construction were elaborated by Hart [1987]: Imagine, for instance, two people who pronounce these sentences, one after the other. The corresponding causal process has no unusual property; the summative description does. The same argument can be used to solve the puzzle of complete self-reproduction, itself a logical impossibility (for a discussion of this problem, see [Löfgren 1968], [Kampis 1991a]). Instead of the troubled circular form "A produces A" (where we would immediately ask, what is this A, in the first place? and we would have to answer, the definition of A is that it is this thing that produces A), there is now a causal way of representing the process as this: "A produces B" & "We demonstrate that A = B". Notice again the use of the two-step construction, and a 98 presence of a causal precedence relation that opens up an initially closed referential circle. As a consequence of this method, there is a possibility to unfold every closed self-definition as a temporal sequence of ordinary definitions, free of inconsistency or paradox. VIII. Constructivism Causality can imply self-reference, all right, but how can we describe this? What sort of causal processes are those that allow for such phenomena, or can any process be self-referential? I shall now argue, as I have done elsewhere repeatedly [Kampis 1988, 1989, 1991, 1992], that a philosophy of mathematical constructivism should be adopted when dealing with a causal description of a self-referential or self-modifying situation (the reader will still remember that we have promised to reveal the connection of these two threads). This will mark certain processes are quite natural candidates for entangled phenomena. The motivations of adopting constructivism in this context are essentially the same as they were in mathematics or in S. Harnad's "neo-constructivism" [1982]. Constructivism was initiated by mathematicians, like Kronecker and above all, Brouwer, who felt uncomfortable with indirect proofs and implicit definitions, and in general with anything that was not directly accessible or tangible within mathematics. It is characteristic that one of the main objects of the constructivists' wrath was Cantor's set theory, with its notorious use, without any inhibition, of the concept of actual infinity, and its habit of talking about infinite sets as if they were explicitly known to us, compactly defined, with their elements given simultaneously. Brouwer, the leading figure of twentieth century constructivism, went very far in his purificatory enterprise against infinity. Many believe that he threw out the baby with the bathwater, and indeed in his "new" mathematics he did not succeed to prove even the simplest and most obvious theorems, the truth of which was nevertheless clear to him. On the other hand, constructivism is not necessarily related to the issue of infinity and it offers many important contributions within and outside mathematics (see, for instance, [Beeson 1985]). Of these uses of constructivism, we will focus on its tools for dealing with definitions. Central to the constructivists' thinking is the concept of piecewise definition of mathematical entities (this is a point where an important connection to algorithms, as well as to discrete and "finitistic" mathematics exist). Piecewise definition is just the opposite of definition by "fiat", or "let" (which was dubbed as "definition by intimidation" in [Kampis 1992]). Instead of using a top-down method, it proceeds bottom-up. When carried out, such a constructivist definition requires groundwork and makes a revision of extensional definitions possible, in the form of allowing constructs like growing sets, for instance. Following the constructivistic suggestion, in systems definitions we are not supposed to write "state set X" but "state object xt", then "state object xt+1", and so on, and only at the end should we go and summarize these in an encompassing state definition [Kampis 1989]. What this has to do with our issue of causal 99 definitions should not be difficult to see. The procedure I am going to suggest consists of replacing self-reference with a temporal (i.e. causal) "self- definition", then of realizing the latter with the aid of processes that need a self-constructing, that is, temporally growing (in this sense self-modification) constructivist definition. In other words, the processes that bring forth selfreference will be conceived as ones that get defined in the course of their own execution, by means of a growing universe of basic elements. Let us stop here for a while to note that the unavoidable use of the term "self" can be very misleading at this point. Whereas we are still talking about the same self-referential and selfmodifying systems as before, in a growing system there is no more an invariant entity that could be identified with any "self" to which we can refer. Rather, we deal with a temporally extended process the various stages of which together constitute an identity. However, if that is so, and we accept the conceptualization of processes that use a constructivist definition, this global entity nowhere appears on the local time scale of the process. Let us also note that a constructive definition is "not algorithmic", in the same sense as setting up a formal system or designing an algorithm is never algorithmic. But when we proceed by a usual definition, it involves no temporal elements, that is, the whole definition can be finished prior to execution, and hence the subsequent temporal process can be algorithmically viewed, as a consequence of this definition. In the constructivist conception, however, the definition does not necessarily end when the process begins, so the "non- algorithmic" features extend to the causal process itself. 100 IX. Causal Systems that Need Constructivist Definition Of course, that is all hypothetical so far, we spoke of constructively self-extending processes as if they existed, but how do we know that they exist at all? Maybe every process reduces to some sort of finite "fiat". It is difficult to think of any example offhand for the use or necessity of the constructivist definition. Now the next question is whether there are systems in actuality that need it. The answer I am suggesting is positive, and I will now try to show the existence of processes that "extend" or "modify" themselves, therefore they need new concepts exactly as foreseen in the discussion of the previous section. More than that, my argument is that there is an abstract prototype for such systems that can be studied independently from any realization. This prototype, which I believe is a generator of a wide class of related systems, is a system characterized by the inherent ability to produce its own components. Accordingly, I have suggested calling these kinds of systems componentsystems [Kampis 1991a]. A component-system does not have a fixed basic composition but produces new components for itself. An obvious real-world example is a macromolecular system with its chemical reactions. In other words, the behavior of a component-system is similar to how we have depicted a game session in Nomic, where laws come and go, and after a while the structure of the system may be completely different from its initial one, with perhaps not even a single rule left intact. So, in a component-system the components can be conceived like the rules of Nomic, but there is one very important difference: whereas Nomic, or any other possible game of similar design, requires human players to interact with, in order to perform the moves, a component-system is conceived as an "automatic" causal system that executes the "moves" on its own. In other words, just to continue our metaphor, one could think of the elements of a component-system as "rules" that execute themselves so as to produce new "rules", in a system where the growth law is defined recursively, by itself. Why we had to put quotation marks when speaking about the "rules" of a causal transformation was of course that these "rules" of component-systems do not have to be algorithmic, that is, they do not have to be finitely definable; that is the entire meaning of the enterprise, for that is what necessitates a constructivist, growing definition. The mechanisms for that is what we discuss starting the next section. But before that, it should be made clear that several other suggestions exist, the basic ideas of which are closely related to those of component-systems. They are reviewed in my book [Kampis 1991a]. Of them, let me mention one, Rössler's "privileged zero" systems [1984], for their clarity. The expression "privileged zero" refers to the property that of components not yet produced there is not simply a zero amount but there is in fact none; Rössler used this idea for pointing out that, for instance, to run chemical reactions, you don't have to buy all chemicals that appear in the transition matrix of the system (unlike with electric circuits); this "trick" has 101 an importance for evolution, controlled ontogenesis, and the information theory of biochemical regulation. 102 X. Biological Computations So long, we have been moving in the abstract domain, discussing computer programs and their ability (or inability) to express self- reference and self-modifications, and we discussed very general causal mechanisms that acted on essentially undefined primitives. Now we change the domain while keeping the essential features of the discussion. From now on we will speak about biological systems, or, more precisely, about biochemical reaction systems. It is natural to ask, of course, in what sense can such a consideration help us to understand, ultimately, cognitive functions -- it seems that we are getting farther and farther from the initial goal. However, there is a much studied and deep analogy between computations and biological processes. Not only is it the case that connectionism and other forms of learning theory have greatly benefited from biological metaphors (e.g. in computer science they are now talking about "Parallelism from Nature"), there is also a more direct gateway between the two phenomenal domains. Authors like Bennett and Landauer [1985], Paton [1994], Fontana and Buss [1993], or the entire A-life movement that concentrates on such modelling, have elaborated many details of the transfer. For example, abstract biochemical systems can accommodate Universal Turing Machines, according to the paper-and-pencil proof of Bennett and Landauer [1985]. Or, molecules can be identified with rewriting operators, and molecular reactions conceived as formula substitutions of a formal system. There exist general purpose simulation environments that use this parallel in detail, such as the SPL system developed by us together with M. Vargyas [Kampis and Vargyas 1994]). In particular, then, molecular production systems can be modeled as instances of component-systems. It is this relationship that we elaborate now. XI. A Biological Mechanism: Shifting Reading Frames Analyzed in [Kampis 1991b, Kampis 1992, Kampis 1993], we present a molecular example for the biological self-determination (or construction) of information. The example is quiteuniversal in the sense that it can also be used for structure/function studies, or in biological information theory. Here we will restrict ourselves to but one aspect, that of definability. First let me give the example. Certain viruses, such as the one called %X174, use overlapping codes to represent their genetic information content. That is, in them the same piece of DNA can code for more than one structural protein. Which of these proteins is produced, depends on how the given piece of DNA is transcribed to mRNA. This transcription, in turn, is controlled biochemically, and, according to the current needs of the viral reproduction process, one of the corresponding pathways will be activated. In this way, the virus (or, more precisely, the cell 103 infected by the virus and taken over by it) can control the information content of one of its molecules. Through this control, the entire definition of the molecule, and that of the whole process using it, can be manipulated. A list of further biological examples that follow a similar pattern is given in [Kampis 1994]; not only DNA but also other molecules or entire organs, even organisms can become subjects of similar "environmental tinkering". The Nobel Laureate F. Jacob has studied many of these mechanisms and their relevance in evolution theory [1981]. What makes this example so significant is the fact that it offers a relational mechanism that transforms production rules by means of their context. This can be generalized far beyond the simple two-protein system, as the same procedure that allowed for a single change of genetic expression (the choice of the starting point of transcription, in this case) can be applied repeatedly, producing an unbounded variety of possible execution routes, each of which can be activated in the presence of a suitable control molecule. In other words, we can never know how many different expressions of a given piece of "genetic material" exist. By this logic, it is easy to see now that in a component-system any two molecules (or, expressed abstractly, any two elements of the component-system) can have a unique effect on determining what further molecules will be produced. How many things can be done by such a molecule, only depends on how many different environments exist. This produces a potentially endless variety of causal "rules" for the system, since by this mechanism the newly produced components' interaction with the old components can again amend those things the old components do, and so on, ad infinitum; there is a potentially infinite "depth" of the layers that come into light when evoked by newer and newer contexts, as the production process unfolds. Somewhat more elegantly, now, we can reconstruct this self-redefinition mechanism with the aid of a concept we introduced in [Kampis 1991b]. Let us go back to the genetic example to fix things, and let us call the "window" of interactions for the information readout a reading frame. This concept is neutral enough, so that it can be equally applicable to any other molecular example, to component-systems, and to very abstract situations such as the information consumption of algorithms, or, for the simplicity of presentation, to Turing Machines. Thereby we obtain a possible tool for discussing the fine structure of the differences between algorithmic definitions and self-modifying systems. In the Turing Machine case, the reading frame consists of that encoding that connects the read/write head with the machine tape where the information is stored. In the simplest case, this encoding can be formulated as an identity map, which says that the 0-s of the tape will correspond to 0-s of the machine, and the 1-s to the 1-s. The case gets a little more complicated when we start to consider information storage methods other than visual ones and noughts on the tape -- for instance, color codes or elementary magnetized domains. In any case, what we immediately find to characterize every algorithm is a fixed interpretation, and, ultimately, a constant reading frame in the Turing Machine that executes the algorithm. We have to realize that the reading frame belongs to the design of the computer and not to the objects it can manipulate, and consequently, that this frame cannot change: again, the same separation of design and operation, or hardware and software, as familiar to us from the earlier discussion. 104 In the genetic machinery, the role of the "tape" is played by the genome while that of the read/write head is played by the complex biochemical route connecting the genome to the ribosome, place of the genetic translation. The reading frame, then, is the current expressibility of the genes in terms of their structural products. The changing transcription method in this case, or, if we talk more liberally about abstract components, the corresponding contextually defined production rules correspond to changing or shifting reading frames. If we can agree that such a contextual element is never finitely definable, because it just slips out of hand, then we end up with a picture of systems that can alter their definitions without allowing us to formulate the rules that govern this growth; in other words, we have an effectively constructed example for a process that leads to selfmodification and, through that, according to our earlier train of thought, to causally generated self-reference. An algorithmically minded person could think at this point of various remedies that would reduce these shifting systems to non-shifting ones. Among these medicines, we will certainly hear of large collections of composite reading frames installed on Turing Machines, in order to grasp the complexity of a shifting-frame based component system. And of course, in line with the previously mentioned most popular form of the Church-Turing Hypothesis, "If you can define it, we can simulate it", whenever we name one new reading frame to which our system has switched, somebody can always come up with a super-Turing-Machine with even more multiple frames so that it can simulate even this switching, and so on. But what cannot be remedied in whatever way is the fundamentally deep difference between a bounded and an unbounded set of reading frames, where the first applies to algorithms, and the second to selfmodifying systems, for it is clear, that any fixed set of frames can be always extended by adding a new one, or, in the molecular domain, by adding one more reaction -- and then, everything is mixed up again. XII. Semantically Closed Systems To complete the plan we set for ourselves at the beginning of this paper, we have to return to authentically cognitive themes, such as the generation of meaning. Now what is exactly understood by the latter is a question that goes beyond the limits of this writing, nevertheless it is clear how the problem of meaning became popular recently, and that was by J. Searle's Chinese Room argument [1984, 1992]. The argument says that a performance, such as, for example, the translation from English to Chinese, can be achieved by fundamentally different means. There is the human way of it, and there is the machine way of it. More specifically, it says that a translation (or the execution of any other "intelligent" task) is possible without any understanding, entirely on the basis of formal execution manuals that relate non-interpreted typographical symbols to each other; and further, that machine intelligence is bound to this sort of execution, hence it is unrelated to 105 human intelligence, which is not. The first part of the argument is certainly solid. Indeed every symbolic operation, and therefore anything computers can do, whether working on formal logical problems or executing connectionist algorithms, can in fact be carried out by entirely syntactic operations where no meaning claims can be done -- a situation known in computer science since Turing's universal simulation result and in philosophy since the Wittgenstein-Russell debate. So far, so good, can we say, for our goal here is not to discuss the Chinese Room any further, but to use one part of it that we have just singled out, and to tie it to our suggestions. The concept of reading frame lends itself to a comparison with the notion of interpretation of a symbolic system. Properties of symbolic systems can now be reformulated using the physical analog of reading frames. We know, for example, that syntactical systems have fixed semantic interpretations (and hence they have no "meaning" at all, since that interpretation can itself be made part of a syntax). Obviously, this corresponds to the case of a pre-defined reading frame that is initially (that is, "externally") provided for a system, which is just more of the same we have seen already. If we revert the sentence, we obtain that permanent reading frames yield semantically uninteresting systems, subject to the Chinese Room problem. However, and that may be the gain of all the work we have done, a self-modifying system associated with the presence of shifting reading frames relates, ultimately, to the generation of new meanings, insofar as reading frames can define semantic relations. If we rethink now the concept of self-modifying system in these loosely conceived semantic terms, we should replace their executions with semantic relations, or meanings. Then we arrive at a "holistic" picture of semantics where the meaning of every component is intelligible only with respect to the system that encompasses it, and the same system is, in turn, a product of the previous changes and productions, determined by the components' "meanings". That results in a heterarchy or entanglement between the system and its components -- in other words, in a certain kind of closure. [But it is not a complete closure, because those systems are essentially open] XIII. Conclusions With the last suggested relationship, we have achieved two distinct goals. First we can conclude the train of thought that started with riddles of self-reference and self-modification, the outcome being the characterization of systems with explicit semantic modifications (similar to the ones introduced by H. Pattee [1973], or to P. Cariani's "semantic adaptation" [1989]); second, we succeeded to relate this idea, which is far too specific in itself, to problems of modelling and the philosophy of mind. The core of this was a suggestion of causally realizable circularities that go one step beyond ordinary computations and its perplexing paradoxes. 106 Let me finally mention a speculative idea to indicate where we can possibly get from this point. An important concept that relates to the above outlined problems of semantics is S. Harnad's "Symbol Grounding Problem" [1990]. It expresses a concern with the origin of semantic relations, a question to which our suggestions may also contribute in some form. It may be a very long way to go from what we have presented now to the point where internally generated semantic relations can be modelled with a self-modifying system, but the very idea is there: to offer an autonomous or dynamic symbol grounding of internal meanings -- perhaps a construction of meanings in the Fregean sense of "Sinn". Acknowledgments This paper was written during the author's stay at the Department of Theoretical Chemistry, University of Tübingen, Germany. The author wishes to thank Prof. O.E. Rössler for his hospitality and friendship. The work was partially supported by a German DFG research grant ("New Ideas in Complexity") and two Hungarian research grants (OTKA #2314 and OTKA #T6428). The author would like to thank the grant-giving organizations for this support. References Bartlett, S.J. and Suber, P. 1987: Self-Reference: Reflections on Reflexivity, M. Nijhoff, Amsterdam. Beeson, M.J. 1985: Foundations of Constructive Mathematics: Metamathematical studies, Springer, New York. Bennett, C.H. and Landauer, R. 1985: "The Fundamental Physical Limits of Computation", Sci.American 253, July pp. 38-46. Callebaut, W. and Pinxten, R. (ed.) 1987: Evolutionary Epistemology: A Multiparadigm Program with a Complete Evolutionary Epistemology Bibliography, Reidel, Boston. Cariani, P. 1989: On the Design of Devices with Emergent Semantic Functions, Ph.D. dissertation, Dept. of Systems Sci., SUNY at Binghamton. Clark, A 1989: Microcognition: Philosophy, Cognitive Science, and Parallel Distributed Processing, MIT Press, Cambridge, Mass. Conrad, M. 1983: Adaptability. The Significance of Variability from Molecule to Ecosystem, Plenum, New York. 107 Csányi, V. 1989: Evolutionary Systems and Society: A General Theory of Life, Mind, and Culture, Duke University Press, Durham. Dawkins, R. 1976: The Selfish Gene, Oxford University Press, Oxford. Edelman, G.M. 1987: Neural Darwinism: The Theory of Neuronal Group Selection, Basic Books, New York. Fontana, W. and Buss, L. 1993: "The Arrival of the Fittest", paper presented at the 3rd Artificial Life Conference, Santa Fe, NM. Harnad, S. 1982: "Neoconstructivism: A Unifying Theme for the Cognitive Sciences", in: Language, Mind and Brain (T. Simon and R. Scholes, eds.), Erlbaum, Hillsdale, pp 1 - 11. Harnad, S. 1990: "The Symbol Grounding Problem". PhysicaD 42, 335-346. Hart, W.D. 1987: "Causation and Self-Reference", in: Bartlett and Suber 1987, pp. 179189. Hofstadter, D. 1979: Gödel, Escher, Bach, Basic Books, New York. Hofstadter, D.R. 1985: Metamagical Themas: Questing for the Essence of Mind and Pattern, Basic Books, New York. Holland, J.H. 1975: Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence, University of Michigan Press, Ann Arbor. Jacob, F. 1981: Le jeu des possibles, Fayard, Paris. Kampis, G. 1988: "On the Modelling Relation", Systems Research 5, 131-144. Kampis, G. 1989: "Two Approaches for Defining 'Systems'", Int.J.General Systems 15, 7580. Kampis, G. 1991: Self-Modifying Systems: A New Framework for Dynamics, Information, and Complexity, Pergamon, Oxford-New York, pp 543+xix. Kampis, G. 1991: "Process, Information Theory, and the Creation of Systems", in: Nature and the Evolution of Information Processing (ed. K. Haefner), Springer, Berlin, pp. 83-103. Kampis, G. 1992: "Constructive Systems Theory: Process and Model", in: Cybernetics and Systems '92 (ed. Trappl, R.), World Scientific, Singapore, pp. 11-17. Kampis, G. and Vargyas, M. 1994: "Evolving Component-Systems and Other Alife Simulations in the SPL System", paper submitted to the 4th Artificial Life Conference, MIT, Cambridge, Mass. 108 Lumdsen, C.J. and Wilosn, E.O. 1981: Genes, Mind, and Culture: The Coevolutionary Process, Harvard University Press, Cambridge, Mass. Löfgren, L. 1968: "An Axiomatic Explanation of Complete Self-Reproduction", Bull. Math. Biophys. 30, 415-425. Maturana, H.R. and Varela, F.J. 1980: Autopoiesis and Cognition, D. Reidel, Dordrecht. Maturana, H.R. and Varela, F.J. 1987: The Tree of Knowledge: The Biological Roots of Human Understanding, New Science Library, Boston. Neumann, J. von 1966: The Theory of Self-Reproducing Automata (ed. Burks, A.W.), Univ. of Illinois Press, Chicago. Paton, R. 1994: Computing with Biological Metaphors, Chapman and Hall, London, in press. Pattee, H.H. 1973: "Unsolved Problems of Hierarchy Theory", in: Hierarchy Theory (ed. Pattee, H.H.), Braziller, New York. Rosen, R. 1959: "On a Logical Paradox Implicit in the Notion of a Self-Reproducing Automaton", Bull.Math.Biophys. 21, 387-394. Rössler, O.E. 1984: "Deductive Prebiology", in: Molecular Evolution and Protobiology (ed. Matsuno, K., Dose, K., Harada, K, and Rohlfing, D.L.), Plenum, New York, pp. 375-385. Searle, J.R. 1984: Minds, Brains, and Science, Harvard University Press, Cambridge. Searle, J.R.1992: The Rediscovery of the Mind, MIT Press, Cambridge. Suber, P. 1990: The Paradox of Self-amendment: a Study of Logic, Law, Omnipotence, and Change, P. Lang, New York. Spencer-Brown, G. 1969: Laws of Form, Allen and Unwin, London. Stenseth, N.C. and Maynard Smith, J. 1984: "Coevolution in Ecosystems: Red Queen Evolution or Stasis?", Evolution 38, 870-880. Van Valen, L. 1983: "A New Evolutionary Law", Evolutionary Theory 1, 1-30. Varela, F.J. 1974: "A Calculus for Self-Reference", Int.J.General Systems 2, 5-24. Varela, F.J., Thompson, E. and Rosch, E. 1991: The Embodied Mind: Cognitive Science and Human Experience, MIT Press, Cambridge, Mass. Watzlawick, P. (ed.) 1984: The Invented Reality: How Do We Know What We Believe We Know?: Contributions to Constructivism, Norton, New York. 109