Determinism and the Paradox of Predictability
- 3.9k Downloads
The inference from determinism to predictability, though intuitively plausible, needs to be qualified in an important respect. We need to distinguish between two different kinds of predictability. On the one hand, determinism implies external predictability, that is, the possibility for an external observer, not part of the universe, to predict, in principle, all future states of the universe. Yet, on the other hand, embedded predictability as the possibility for an embedded subsystem in the universe to make such predictions, does not obtain in a deterministic universe. By revitalizing an older result—the paradox of predictability—we demonstrate that, even in a deterministic universe, there are fundamental, non-epistemic limitations on the ability of one subsystem embedded in the universe to predict the future behaviour of other subsystems embedded in the same universe. As an explanation, we put forward the hypothesis that these limitations arise because the predictions themselves are physical events which are part of the law-like causal chain of events in the deterministic universe. While the limitations on embedded predictability cannot in any direct way show evidence of free human agency, we conjecture that, even in a deterministic universe, human agents have a take-it-or-leave-it control over revealed predictions of their future behaviour.
KeywordsPhysical Event Revelation Condition External Observer Guidance Control Paradoxical Situation
In the sixties and seventies of last century, the discovery of, what we call, ‘the paradox of predictability’ triggered a lively debate on determinism, predictability and free will, mainly in British philosophy. Although it ended without any clear upshot and with a lot of loose ends, we have three reasons to revisit and reopen this debate. First, the introduction of an explicit distinction between two different kinds of predictability, allows us to reformulate the paradox in a way that transparently reveals its underlying structure (Sect. 2). As a result, and this is our second and main point, we are able to show why the earlier explanations of the paradox failed (Sect. 3) and, subsequently, to give a new alternative explanation of its origin (Sect. 4). Third, we believe that the paradox of predictability might have important as yet unnoticed consequences for the contemporary compatibilism–incompatibilism debate on free will. In this paper, we elaborate the first two points, while leaving the third—apart from a promissory note (Sect. 5)—for another occasion.
2 The Paradox of Predictability
Embedded predictability holds in a universe, U, if there exists a subsystem, S, embedded in U—for instance, a highly intelligent (human) being or a very powerful computer—that is able to predict all the future events in U.
In the past, several authors argued that, especially but not exclusively with regard to non-linear systems, there are serious epistemic limitations on the ability of an embedded subsystem of a deterministic universe to make predictions of future events (see, for example, Suppes 1985; Bishop 2003). These epistemic limitations arise, for instance, because a finite subsystem of the universe is unable to make the required accurate measurements of the initial conditions of the universe, or because such a subsystem is unable to represent fully accurately the initial value of a variable, if this value is a (non-computable) irrational number (requiring an infinite string of digits). Though we believe that these arguments show that epistemic restrictions in fact or in practice preclude embedded predictability from holding in a deterministic universe, we want to argue for the stronger claim that embedded predictability cannot obtain in a deterministic universe even on the idealizing assumption that the epistemic limitations at issue are removed.4 Therefore, we propose to lift these restrictions and to assume that the subsystem S in our definition of embedded predictability has infinite possibilities of representing information, is able to gather infinitely precise information about the initial conditions of the universe and is able to make calculations with any desired accuracy in a finite amount of time. We accept, of course, that it is physically impossible to realize these infinite capabilities in a finite subsystem of the universe. However, this reference to the infinite epistemic capabilities of subsystem S is not problematic because we simply use it as a shorthand for saying that the predicting subsystem can have arbitrarily strong computing capabilities. Authors who argue for unpredictability in terms of the epistemic limitations of predicting subsystems typically start from a given subsystem with finite capabilities and then construct a corresponding future event that remains beyond the predictive scope of this given subsystem. In contrast, we will identify a single future event that remains unpredictable for all possible predicting subsystems, no matter how strong their computing capabilities.
Already in 1950, Karl Popper has made an attempt to make this stronger case against embedded predictability. Although we think that his highly complex and at times overly technical arguments are of genuine interest, we believe that they are ultimately, as Popper himself seems to admit, not fully convincing (Popper 1950, p. 193).5 A more successful attempt to make the stronger claim can be found implicitly in work of Donald MacKay (1960, 1961, 1967, 1971, 1973) and in a more explicit way in that of Scriven (1965) and P.T. Landsberg and D.A. Evans (1970, Evans and Landsberg 1972). These authors argued for the thesis that, conditional on some minimal assumptions about the universe U, there are fundamental limitations of a non-epistemic nature on the ability of any subsystem S to make predictions and that, therefore, even in the deterministic universe U, embedded predictability does not hold. They established this conclusion by constructing a paradoxical situation in which a subsystem S makes a prediction which is unavoidably self-defeating. Though we believe that their arguments provide a convincing example of a situation in which embedded predictability is impossible, we also believe that their accounts of this paradoxical situation are unwieldy and that, consequently, these authors failed to fully grasp the real origin of the paradox and the general nature of the unpredictability involved. Therefore, we proceed, first, by presenting our reconstruction of the paradox which allows us to lay out its underlying structure.6
When the revelation and contrapredictivity conditions obtain, if S 1’s prediction at t 1 is P = 0, then S 2’s action at t 2 will be A = 1; if S 1’s prediction at t 1 is P = 1, then S 2’s action at t 2 will be A = 0. The paradoxical nature of the situation thus resides in the fact that any prediction made by S 1 is inevitably self-defeating.
3 How Not to Explain the Paradox (Away)
We briefly elucidate our reconstruction of the paradox of predictability against the backdrop of the earlier 1960–1970 discussion. We believe that, until now, no convincing explanation of the underlying structure of the paradox has been given. To dispel some of the misunderstandings which have clouded the debate, four important clarifications are in order.
First, the self-defeating structure does not, as some have mistakenly thought (DeWitt 1973), depend on some supposedly unique ontological characteristics of cognitive events, or on the presence of some special kind of free will. There is nothing extraordinary about a contrapredictive mechanism. We can safely assume that such a mechanism can be instantiated by a variety of biological, psychological or electronic mechanisms. Contrapredictivity could, for instance, easily be programmed into a computer, requiring the generation of output 1 if and only if receiving input 0, and vice versa (Landsberg and Evans 1970, pp. 356–357; Good 1971).
In this case, the prediction at t 1 will be P = 0 and, therefore, the action at t 2 will be A = 1. This uniquely determined course of events will be clear to everybody, even to the predictor S 1. However, this knowledge does not help the subsystem S 1: it also knows that it is unable to generate the correct output P = 1, for this prediction would again be self-defeating and bring about action A = 0.
Third, it is not the case, contrary to what several authors assume (Lewis and Richardson 1966; Good 1971; also Roberts 1975), that the paradox has something to do with limitations on the speed with which the predictor and the predicted system can make calculations concerning each other’s behaviour. These authors seem to have in mind a scenario in which the predictor is able to make a second and secret prediction, P′, which would be correct (say, P′ = 1). Yet, when this secret prediction is revealed, the predicted system will decide to act contrapredictively as regards this second prediction, which will require a third secret prediction, P″ (=0), which in its turn could be revealed, etc. This to-and-fro scenario generates a series of predictions and contrapredictive intentions, such that the subsystem that calculates the fastest will outsmart the other subsystem at the time of action, t 2.
However, this kind of consideration does not genuinely alleviate the paradox. There are three possibilities. We could, first, suspend our assumption that the predicting subsystem has infinite calculating capabilities and interpret this to-and-fro scenario as an actual to-and-fro process, whereby each prediction and each contrapredictive calculation takes a finite amount of time. Whether, in this case, the final prediction made by predictor S 1 before time t 2 will be correct, depends entirely on the contingent characteristics of the calculating and interacting mechanisms. Although, in this scenario, the stronger claim that we can identify a single future event which is unpredictable for all possible predicting subsystems no longer holds, the argument against embedded predictability as such remains. Remember that to demonstrate that embedded predictability cannot obtain, it suffices, strictly speaking, to show that, for every would-be predictor S 1, we can think of at least one situation in which it will fail to make a correct prediction. In this scenario, we can, therefore, secure the paradoxical result simply by stipulating that the predicted subsystem S 2 calculates faster than the predicting subsystem S 1.
The second interpretation is also straightforward. If we consider the to-and-fro scenario as an infinite series in a finite amount of time (by assuming, for instance, that both subsystems are able to make infinitely fast calculations), then it simply captures the typically oscillating nature of the paradoxical equation P = not-P (Wormell 1958). We are again just saying that this equation has no solution and that the predictor will have to resort to the default prediction P = 0, which is self-defeating.
On a third interpretation, we could make the asymmetrical assumption that the predicting subsystem S 1 is capable of making infinitely fast calculations, whereas the predicted subsystem S 2 is only capable of finitely fast ones. Whilst this third scenario is the only one in which the proposed ‘solution’ of the paradox actually works, and S 1 indeed always outsmarts S 2, this resolution comes at a high price. Not only does the asymmetry of the idealizing assumption seem arbitrary, it also remains a fact that by choosing an arbitrarily fast predicted subsystem S 2, we can make the time-gap ε which remains between prediction P and action A arbitrarily small. The claim that S 1 can actually ‘predict’ the behaviour of S 2, given a very small value for ε, reduces then to the claim that S 1 can fulfil its predictive task simply by waiting until S 2 is no longer capable of reacting to any relevant information from its environment, due to a lack of time. Since this scenario makes the idea of a prediction rather meaningless, we have introduced in our version of the paradox the plausible requirement that the prediction has to be produced at a specific time t 1 (<t 2). This assumption simply captures the idea that a real prediction needs to refer to an event that is in some significant sense situated in a finitely (as opposed to infinitesimally) remote future.
Importantly, the determination of a specific time t 1 (<t 2) at which the prediction has to be made immediately blocks all three versions of the to-and-fro scenario. The determinate, finite lapse of time between the prediction and the action allows the predicted subsystem to respond to P and the unpredictability of A by S 1 at time t 1 is restored. Since A is thus an event which remains unpredictable, independently from the specific characteristics of the predicting subsystem, our stronger claim that embedded predictability cannot obtain even on the idealizing assumption that S 1 has infinite calculating capabilities still stands in our version of the paradox.
Four, and in line with our first two clarifications, the paradox does not reveal some kind of ‘essential unpredictability in human behaviour’, as Michael Scriven (1965) wants us to belief. To begin with, neither his version of the paradox, nor ours essentially presupposes that the nature of the predicted subsystem S 2 is human. Neither of both versions makes a distinction between human and non-human agency. In addition, and more importantly, the situation we have sketched remains throughout fully determined and, therefore, no essential unpredictability is implied at all (Suppes 1964, p. 144). As we will show, Scriven’s claim is based on a serious flaw in his own argument. Close scrutiny of where it goes wrong is worthwhile, since it helps to reveal the real origin of the paradox.
Scriven’s version of the paradox differs from ours in one design feature. Although he too presupposes a contrapredictive mechanism, he rejects the idea of an actual interaction between the predicting and the predicted system. Instead, he assumes that the predicted system has access to the same information I and has the same predictive capacities as the predicting system. In this way, the predicted system can itself make the very same prediction P (without the need for interaction or revelation) and, subsequently, act contrapredictively. As we will explain in the next section, we believe that this difference in the design of the assumptions is in itself not really significant and does not actually change the structure of the paradoxical situation.
The paradox of predictability does not arise because action A cannot be logically inferred from the information available, but because the embedded subsystem S 1 cannot deduce its prediction without thereby invalidating it. More specifically, S 1’s prediction is not an external logical inference. The paradox, we conjecture, arises because S 1 is itself an embedded part of universe U and, therefore, its prediction a physical event which is itself part of the causal chain of events.
4 Predictions as Physical Events
Exploring the distinction between external and embedded predictability somewhat further, we can clarify why and how the paradox of predictability is inextricably connected with the embeddedness of S 1 as a subsystem of universe U. This furnishes the material for our alternative explanation of the paradox’s origin and its underlying structure.
These equations mean the following. Equation (8) describes how prediction P, as a physical event at time t 1, is uniquely determined by the state of U at time t 0 and the law-like function g. Equation (9) describes, similarly, how action A, as a physical event at time t 2, is uniquely determined by the state of U at time t 0 and the law-like function h. Equation (10) describes the predictive task of the external predictor, who has to make sure that its prediction P* adequately predicts action A. Importantly, equation (10) has no physical interpretation because the prediction P* is not itself a physical event, but just the name for the prediction made by the external (and thus non-physical) observer.
As this set of equations makes clear, an external predictor has no problem whatsoever in understanding and describing the chain of events in the paradoxical situation at hand. On the basis of its knowledge of g, h and U 0 (and its calculating abilities), this observer is perfectly capable of determining the values of P and A and, thus, to fulfil its predictive task (10). If we apply the general description to the specific case of the paradox, equation (8), in our default case, simply becomes the default prediction P = 0 and equation (9) on the revelation and contrapredictivity conditions becomes equation A = not-P. Clearly, these equations allow the external observer to make correctly (10) the prediction P* = 1.
The first two equations (8′) and (9′), are identical to the first two equations of the previous set, (8) and (9), and mean the same. They describe how prediction P and action A, as physical events, are uniquely determined by the initial state of U at time t 0 and the laws describing its evolution. The third equation (11), however, is different and describes the predictive task of the embedded predictor S 1. In contrast to equation (10), and this is a point of great consequence, equation (11) has a physical interpretation. Prediction P at time t 1 can only be successful if it adequately describes the future action A at time t 2. In our case, this means that the number produced by the predictor S 1 at time t 1 has to be the same as the number chosen by S 2 at time t 2.
As the equations make clear, the fact that P, in contrast to P*, is a physical event has a dramatic impact on the ability of the embedded predictor to fulfil its task. Whereas the first set of equations (8)–(10) provides three equations from which to determine the value of three variables (P, A and P*), the set of equations (8′)–(9′)–(11) provides three equations from which to determine the value of only two variables (P and A). In the second set of equations, the variables are thus overdetermined and, consequently, a simultaneous solution of all the equations is no longer assured. Overdetermination possibly implies unsolvability.
The explanation of this essential difference is the following. The deterministic nature of the universe implies that the first two equations of each set already uniquely determine the value of P and A. The third equation of the second set, (11), imposes an additional physical constraint on an already fully determined universe. Equation (11) thus overdetermines the variables P and A. Prediction P is successful only if a certain relation exists between prediction P as a physical event and the predicted action A as a physical event (whereby P has to provide an adequate representation of the future action A). Whether this relation actually obtains and, thus, whether a successful prediction (P = A) is possible, cannot be guaranteed a priori but depends on the specifics of the laws and conditions of the universe and of the situation at hand.
By contrast, in the case of an external predictor, this uncertainty does not arise. In this case, the deterministic nature of the universe guarantees a priori that a successful prediction P* is always possible. The first two equations (8) and (9) allow the external predictor to determine the values of P and A. Although it is thereby perfectly possible that these two values turn out to be different (P ≠ A), this only means that the external predictor observes that the embedded predictor is unsuccessful. This observation, however, in no way affects the ability of the external predictor to make its own successful prediction P* = A.
As these two examples make clear, self-referentiality, as functionally expressed by (13), can arise as the result of a direct causal interaction between the predicting subsystem S 1 and the predicted subsystem S 2. In our two examples, we have assumed that the revelation condition, (3), obtains and, thus, that S 2 somehow receives information about the prediction P after it has been made. In these situations, the functional dependence f of action A on prediction P can properly be understood as a description of the causal impact the prediction has as a physical event. In the paradoxical case, the causal impact of the prediction as a physical event is such that it inevitably falsifies the prediction itself.
While in this case the functional dependence of A on P arises without any actual causal interaction after time t 1 between the prediction as a physical event and the action itself, this dependence now crucially turns upon the presence of a common causal history. In our account, this common history is encapsulated by the initial conditions U 0 such that both prediction P and action A causally depend on these conditions as expressed by equations (8′) and (9′), respectively.10
The fact that such an action cannot exist demonstrates that subsystem S 2 is unable to act contrapredictively with regard to an external observer with which it does not share a common causal history.
Although our account of the overdetermination involved in a case of embedded prediction has focused, thus far, on the category of more tractable situations in which action A is directly functionally dependent on prediction P (equation 12), the generality of the three equations (8′)–(9′)–(11) indicates that the problem of overdetermination might also be much more general. The laws of our universe seem to be such that all sorts of interactions and functional dependencies between two embedded subsystems unavoidably exist. Physics tells us, for instance, that two of the fundamental forces in the universe (gravity and the electromagnetic force) have an infinite scope. This means that the actual physical workings of the predicting subsystem (containing biological or electronic devices which calculate and store the prediction) generate causal factors which will inevitably affect the future evolution of the predicted subsystem. Owing to this causal interconnectedness, one can expect that the set of equations (8′)–(9′)–(11) will often generate problems of self-referentiality similar to those of equation (13).
In fact, the ubiquity of causal interactions seems to imply a reversal of our initial problem. Given that functional dependencies are almost inescapable, how then is it at all possible that we are actually capable of making embedded predictions of future events? The answer is, of course, that in many situations when we want to make predictions we are able to discount or shield of our own past and present interactions with the predicted system at hand. For instance, if we want to calculate the orbit of Jupiter and predict its position at a specific time in the future, we may safely discount the gravitational impact of our calculations (as a physical activity) on its orbit. In such a situation, the future event still remains functionally dependent on the present state of the universe, but will be, for all practical purposes, sufficiently causally uncoupled from our predicting activities. This uncoupling eliminates the self-referentiality of the situation and allows us to find a single solution that describes and predicts the future event.
Needless to say, however, that such an independence is never complete and only possible to a sufficient extent for some (but not all) specific ‘local’ problems and purposes. A general and complete causal uncoupling of our past and future activities from the predicted subsystem, would require, among other things, that we were capable of obtaining all the information needed for our predictions without actually disturbing the predicted system. This would require, what Popper (1950, p. 129) called, a ‘one-way membrane’ between the predictor and the predicted system. Because such a membrane cannot physically exist, the possibility of complete causal uncoupling is only available to a Laplacean predictor as a non-physical observer outside the universe.11
5 Take-It-or-Leave-It Control
In the 1960–1970 debate, it was widely assumed that the paradox of predictability was relevant for the debate on freedom and determinism. Especially Donald MacKay (1960, 1961, 1967, 1971, 1973) tried to use the paradox as the starting point for a compatibilist theory, according to which, even in a deterministic universe, human brain-states remain ‘logically indeterminate’ until the moment of ‘free’ choice. Although MacKay’s approach is interesting in many respects, we believe that the works of Landsberg and Evans (1970), Evans and Landsberg (1972) and Grünbaum (1971) have convincingly shown that MacKay ultimately failed in developing a plausible compatibilist position.12
We believe that this failure is not accidental but rather the effect of an underlying assumption which MacKay and most other authors in the earlier debate seemed to have shared. They all assumed a direct conceptual connection between freedom of human actions and their (embedded) unpredictability. We believe, on the contrary, that this direct connection fails to hold and that (embedded) unpredictability is neither a sufficient nor a necessary condition for human freedom. It is not sufficient, since, as we have noted, it is perfectly possible that ‘soulless’ machines programmed in a contrapredictive manner remain unpredictable. It also seems unreasonable to suggest that unpredictability is a necessary condition for freedom. If someone or something predicts that tomorrow I am going to help my neighbour fix the shed in the garden and if, as a matter of fact, I see no good reason not to do so, it seems implausible to suggest that my freedom to choose is undermined simply because the prediction is indeed in line with my personality and my reasons for action.
While these considerations imply that no direct inference from the lack of embedded predictability to the possibility of human freedom can be drawn, we conjecture that the paradox of predictability leads to results which might be relevant for the compatibilism-incompatibilism debate on free will. Introducing some basic terminology will be helpful.13 If determinism, as characterized by (1) in Sect. 2, is true, the facts of the past, together with the laws of nature, entail all facts of the present and future. Indeterminism is the denial of determinism. Compatibilism is the view that free will, free action and moral responsibility are compatible with determinism; incompatibilism is the denial of compatibilism. Libertarians are incompatibilists (and indeterminists) who believe that at least some of us, at times, perform free actions for which we are responsible. Hard determinists are incompatibilists who deny that we have the sort of free will required for moral responsibility; hard incompatibilists deny the same, irrespective of accepting determinism or not. Incompatibilists standardly hold that responsibility-relevant freedom can be nothing less than regulative control (the sort of control which involves the existence of genuinely open alternative possibilities), whereas compatibilists argue that such freedom only requires guidance control. At a minimum, our conjecture pertains to guidance control.14
As explained, this means that S 1’s embedded prediction will be inevitably self-defeating. Yet, in case the SCR-conditions are not revealed or fail to obtain, a solution for the pertinent equations is not necessarily excluded and, depending on the details of the situation, an adequate prediction might remain possible.
The presupposition that we actually live in a deterministic universe is compatible with the additional and, in our view, plausible assumption that human beings are in possession of (deterministic) psychological mechanisms of the condition-responsive kind, indicated by (19). This means that, even in a deterministic universe, a human being who is confronted with a prediction about its own future behaviour might consider several other conditions and, depending on whether or not these hold, decide to go against or, alternatively, decide to act in accordance with the prediction made. We illustrate the types of situation we have in mind by considering two possible sufficient conditions for rejection. Suppose, first, that the subsystem which predicts our actions does so on the basis of its foreknowledge for the purpose of manipulating us. In this case, if the prediction and the manipulative intentions of the predictor are revealed, we could be motivated to act contrapredictively. Similarly, it could turn out that a prediction of my future behaviour is based on the discovery of some kind of biological or sociological mechanism which determines my actions. In this second case, such a discovery could lead me to reassess what I previously considered as my authentic preferences and my authentic reasons for making my decisions. As a result, I could decide, on the basis of the more extensive information now available to me, that I no longer wish to pursue my biologically or socially determined urges and desires.15 If, as in these two examples, manipulation or the presence of inauthentic preferences is interpreted as a SCR-condition, and if this condition is revealed, then the working of (deterministically) operating condition-responsive psychological mechanisms warrants that the predictions made about our future behaviour will inevitably be self-defeating.
Our analysis of the paradox demonstrates that embedded subsystems in a deterministic universe retain, what we call, a take-it-or-leave-it control as regards revealed predictions. The specifics of the operating mechanisms of such an embedded subsystem establish the conditions under which it will accept or reject predictions about its future behaviour made by an other embedded subsystem in the same universe. Whether or not this consequence will turn out to be important for the contemporary debate on the compatibility of human freedom with determinism depends on the additional assumption that take-it-or-leave-it control is an essential part of a convincing analysis of free human agency. We conjecture that this assumption holds water, but a further elaboration of such an analysis is beyond this paper’s scope.
The function needs to be law-like, for, even in a non-deterministic universe, a non-law-like function between states of a universe at different times can always be defined. Specifying the sense in which this function needs to be ‘law-like’ is beyond the scope of the present paper (see Russell 1917; Hempel 1961; Boyd 1972; Kukla 1978, 1980; Dieks 1980). We submit that our research presented here is independent from the details of a more specific account of determinism.
The Laplacean observer has a ‘view from nowhere’ and takes a detached ‘objective view’—sub specie aeternitatis—of the universe as a whole. For this theoretical standpoint, from which we also wrote this paper, see Nagel (1986). When we say that this external observer and its predictions are ‘non-physical’, we only mean outside U, not embedded in the physical universe itself. We remain neutral as to the intrinsic nature of this external standpoint—be it abstract (Platonic), spiritual or material.
Both external and embedded predictability are types of conditional predictability—predictability in theory or in principle—as defined by Honderich (1988, p. 337): ‘If something is conditionally predictable, it could be predicted if certain knowledge existed, the having or recording of which knowledge is logically or conceptually possible, although perhaps factually or nomically impossible.’ Since Honderich does not distinguish between external and embedded predictability, as we do, he fails to notice that his arguments against ‘determinism without predictability’ (1988, pp. 347–350) do not pertain to embedded predictability but only to external predictability.
Popper (1950) has tried to show that there are unavoidable limitations on the ability of a predictor to predict its own future. He elaborates three arguments. The first and the third ones essentially suggest that a physical predictor is unable to describe its own describing activities and that it cannot contain sufficient information about itself to make accurate predictions. These arguments deal with the kind of epistemic limitations we have excluded from the start. The second argument makes use of Gödel’s incompleteness results (1950, pp. 179–188). Here Popper suggests that a computer will be unable to predict at time t whether it will itself be able to answer, at a later time t′, the mathematical Gödelian question regarding the consistency of the information it has available at that later time. Since this Gödelian question is undecidable, the same holds for the physical question at the prior time t regarding the decidability at the later time. The problem with this argument, however, is that there is no reason whatsoever to suppose that the predictor should even be willing to consider, at the later time t′, the Gödelian question Popper has in mind. Since this question is a mathematical one and not a physical one, it suffices to construct a better predictor which refuses to consider mathematical questions by generating some error-message. The generation of this error message (as a physical event) will be perfectly predictable at the earlier time t and so no convincing counterexample which disproves embedded predictability has been generated.
Landsberg and Evans’ account is cumbersome in the sense that it makes reference to the brain state of the predicted agent. Whereas this is justified in view of their attempt to criticize Donald MacKay, it obscures the logical simplicity and generality of the result. Scriven’s account is unwieldy because it is couched in utilitarian language and it conflates two different issues. As Lewis and Richardson (1966, p. 74, n.3) confirm, the issue of unpredictability in possession of full information is conceptually entirely distinct from the randomizing contrapredictive stratagems used in a case of limited information (which is the main focus of Scriven’s paper but not our present concern).
Scriven (1965) considers human beings and assumes that they could be contrapredictively motivated. Psychological evidence for the existence of such a type of motivation in humans can be found in the theory of ‘psychological reactance’ (Brehm 1966). Because we want to provide a more general account by leaving it open whether S 2 is a human being or a machine, we prefer to talk about a contrapredictive mechanism, thereby highlighting the deterministic nature of the pertinent mechanism.
We have adapted Scriven’s notation to fit our own.
Alvin Goldman (1968, pp. 141–143) suggests that there are possible deterministic universes in which a full ‘book of life’ exists (containing the full and correct life story of, for instance, Alvin Goldman) and in which the person whose life is described actually reads this book, without thereby invalidating the predictions. Our equations imply, however, that the possible existence of such a book requires that all the self-referential situations generated by reading the book remain solvable and, thus, that several completely contingent constraints (‘contingent’ in the sense of being additional and independent from the deterministic laws of this universe) are met within this universe. The existence of such a book in some universe cannot be excluded a priori, but its existence would be a quirky exception without any philosophical significance for the problems dealt with in this paper.
The commonness of this history is an effect of the fact that the predicted subsystem S 2, in order to be capable of completing its calculation of P, needs to acquire sufficient information about the initial conditions which determine the future evolution of S 1. So, gathering the relevant information requires actual interactions—either directly or through intermediaries—between the two subsystems at some point in their pasts.
The idea that observers necessarily causally influence the system they are observing is, of course, not novel. In the social sciences, for instance, this idea is known as the observer’s paradox. However, our results support the additional and central claim that this influence seriously and, in some circumstances, irreparably undermines our ability to predict the future behavior of the observed system.
MacKay (1960, 1967, 1971) argued in favor of a notion of truth, according to which predictions are only true if they are revelation-robust as well as belief-robust. This means that (1) the truth value of propositions describing a predicted future state of a system should not depend on whether or not a physical interaction revealing the prediction has taken place between the predicting and the predicted system, and that (2) this truth value should also not depend on whether or not the predicted system believes that the predicted state will actually be realized. Since these assumptions imply that a future state of affairs (the predicted event or action) should not depend upon prior physical events and states of affairs with which it is actually causally connected, this notion of truth seems highly implausible. Indeed, this notion of truth implies, as MacKay specifically intended, that no propositions describing future human actions have a truth value. As Evans and Landsberg (1972) demonstrated, this inflated notion of truth begs all the relevant questions regarding compatibilism. Compare also Honderich (1988, pp. 347–350).
This distinction between regulative and guidance control, introduced by John Martin Fischer, has become well known; see Fischer (1994, pp. 131–189). We surmise that our suggestion here can be elaborated and possibly be integrated with Fisher’s compatibilist model of free agency as reason-responsiveness (Fischer and Ravizza 1998; Fischer 2006).
For sure, whether or not my condition-responsive mechanisms will be able to override the underlying biological or social mechanisms (whether my thus determined urges and desires are ‘controllable’ or not) depends on the contingent causal connections between these psychological, biological and social mechanisms realized in my central nervous system that controls my movements. The answers here seem to depend on empirical issues relating to the functioning of our brain. There is, however, nothing particularly mysterious or metaphysical about the assumption that a physically realized psychological mechanism could override (or could learn to override) other physically realized mechanisms.
Stefan Rummens’ research contained in this paper was made possible by a postdoctoral fellowship of the Research Foundation Flanders (FWO). Stefaan Cuypers’ research connects with the research project Alternatives, Belief and Action (FFI2009-09686), awarded and supported by the Spanish Ministry of Science and Innovation. We are grateful to the FWO Odysseus formal epistemology research group of Igor Douven (KU Leuven) for useful discussion. We also thank three anonymous referees for Erkenntnis for their helpful suggestions.
This article is distributed under the terms of the Creative Commons Attribution Noncommercial License which permits any noncommercial use, distribution, and reproduction in any medium, provided the original author(s) and source are credited.
- Brehm, J. W. (1966). A theory of psychological reactance. New York: Academic Press.Google Scholar
- Fischer, J. M. (1994). The metaphysics of free will. Oxford: Blackwell.Google Scholar
- Fischer, J. M. (2006). My way. Essays on moral responsibility. Oxford: Oxford University Press.Google Scholar
- Fischer, J. M., & Ravizza, M. (1998). Responsibility and control. A theory of moral responsibility. Cambridge: Cambridge University Press.Google Scholar
- Goldman, A. I. (1968). Actions, predictions and books of life. American Philosophical Quarterly, 5(3), 135–151.Google Scholar
- Grünbaum, A. (1971). Free will and laws of human behaviour. American Philosophical Quarterly, 8(4), 299–317.Google Scholar
- Hempel, C. G. (1961). Some reflections on “the case for determinism”. In S. Hook (Ed.), Determinism and freedom in the age of modern science (pp. 170–175). New York: Collier Books.Google Scholar
- Honderich, T. (1988). A theory of determinism. The mind, neuroscience, and life-hopes. Oxford: Clarendon Press.Google Scholar
- Kane, R. (Ed.). (2002). The Oxford handbook of free will. New York: Oxford University Press.Google Scholar
- Kane, R. (2005). A contemporary introduction to free will. New York: Oxford University Press.Google Scholar
- MacKay, D. M. (1967). Freedom of action in a mechanistic universe. In M. S. Gazzaniga & E. P. Lovejoy (Eds.), Good reading in psychology, (pp. 121–137). Englewood Cliffs: Prentice-Hall, 1971; originally id. (1967), Freedom of action in a mechanistic universe: Eddington memorial lecture of November 1967. Cambridge: Cambridge University Press.Google Scholar
- Nagel, T. (1986). The view from nowhere. New York: Oxford University Press.Google Scholar
- Russell, B. (1917). On the notion of cause. In id., Mysticism and logic (pp. 132–151). London: Unwin, 1963.Google Scholar
- Scriven, M. (1965). An essential unpredictability in human behaviour. In B. B. Wolman & E. Nagel (Eds.), Scientific psychology: Principles and approaches (pp. 411–425). New York: Basic Books.Google Scholar
- Sellars, W. (1966). Fatalism and determinism. In K. Lehrer (Ed.), Freedom and determinism (pp. 141–174). New York: Random House.Google Scholar