Abstract
Traditional approaches model consciousness as the outcome either of internal computational processes or of cognitive structures. We advance an alternative hypothesis – consciousness is the hallmark of a fundamental way to organise causal interactions between an agent and its environment. Thus consciousness is not a special property or an addition to the cognitive processes, but rather the way in which the causal structure of the body of the agent is causally entangled with a world of physical causes. The advantage of this hypothesis is that it suggests how to exploit causal coupling to envisage tentative guidelines for designing conscious artificial agents. In this paper, we outline the key characteristics of these causal building blocks and then a set of standard technologies that may take advantage of such an approach. Consciousness is modelled as a kind of cognitive middle ground and experience is not an internal by-product of cognitive processes but the external world that is carved out by means of causal interaction. Thus, consciousness is not the penthouse on top of a 50 stores cognitive skyscraper, but the way in which the steel girders snap together from bottom to top.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
A necessary caveat: In this paper, whenever we criticize other approaches it is only to the extent that they are presented as an explanation of consciousness. Thus, for example, we have nothing either against Baars’ theory of the global workspace as an excellent cognitive model, or Tononi’s notion of integrated information as a mean to achieve unified representations. However, at the present state of research, we do not understand how these models could justify the emergence of consciousness at the end of a computational process or as the output of a cognitive module.
- 2.
We use outside and inside to refer to physical events inside and outside one’s body.
- 3.
In causal terms, we may distinguish between ‘the cause and that without which the cause would not be a cause’ ([47], p. 119). The former may be taken to be an event actually occurring while the latter may be just a state of affairs. The latter may be formalized in terms of conditions G such that P \(\wedge \) G \( \Box \rightarrow \) E which may unfolded in three conditionals P \(\wedge \) \(\lnot \)G \(\rightarrow \) \(\lnot \)E, \(\lnot \)C \(\wedge \) G \(\rightarrow \) \(\lnot \)E, and C \(\wedge \) \(\lnot \)G \(\rightarrow \) \(\lnot \)E.
- 4.
These properties may be a inhomogeneous set of actual physical properties such as the reflected colour spectrum, the percentage of certain components, or the contrastive ratios among different areas. For the sake of the example, consider just a set of wavelengths.
- 5.
Ofcourse, these models do not rule out the importance of the external environment to guide the development of internal structures. Indeed, they consider necessary for a healthy brain to develop by means of continuous interactions with the environment. However, once the required neural connections are in place, the mind is taken to be an internal phenomenon.
- 6.
Whenever it was necessary to point to the autonomy of the mental with respect to the physical domain, the issue of misperception has been the battering ram of both philosophers and scientists. Dream and hallucinations appear as formidable evidence in favour of an inner world. However, this approach promises to locate a physical cause for any experience in the physical surrounding. All cases of conscious experience ought to be revisited as cases of (admittedly unusual) perception. The approach presented here honestly stands or falls on whether it will succeed to show that – perhaps surprisingly – whenever there is consciousness, there is a physical phenomenon, which is the content of one’s experience. We cannot do justice here to the problem of misperception. However, we flesh out a template of the strategy – namely to address each purported case of misperception and to revise them in terms of perception. (One of the authors is actually working on such an account for most cases of misperception, from hallucination to illusions, from aftereffects to direct brain stimulations).
- 7.
The distinction between strong and weak machine consciousness mirrors that between strong and weak AI [38]. Weak machine consciousness considers whether it is possible to build machines that behave as if they were consciousness. Strong machine consciousness ventures to consider the possibility of real conscious machines. We believe that skipping the ‘hard problem’ is not a viable option in the business of making conscious machines [61].
- 8.
- 9.
Consider the following example: In the literature, we find that cats are somewhat colour-blind concerning the colour red, they see it as a shade of grey (whereas they have a perfect colour vision concerning e.g. green and blue). Well, the first finding is that we cannot be really sure about that, since we can only predict that from their eye anatomy – but what kind of ‘consciousness’ cats really have concerning the colour red is a totally different topic because we don’t know anything about the design of the key-lock-structure at this point. It may be totally different from ours. The second – and much more important – insight is that it might be less important for a cat to be capable of seeing red than for example for a bear. Cats – being carnivores – do not have to differentiate between ripe and unripe apples since they would not eat them anyway. For a bear on the other hand – being omnivores consuming a large portion of fruits daily – the situation may show itself quite differently.
References
D.M. Armstrong, What is Consciousness? In: The Nature of Mind, ed. by J. Heil, Cornell University Press, 1981, pp. 55–77
D.B. Edelman, B.J. Baars, A.K. Seth, Identifying hallmarks of consciousness in non-mammalian species. Consciousness and Cognition 14 (1), 2005, pp. 169–187. doi:10.1016/j.concog.2004.09.001. http://www.sciencedirect.com/science/article/pii/S1053810004000935
A.K. Seth, B.J. Baars, D.B. Edelman, Criteria for consciousness in humans and other mammals. Consciousness and Cognition 14 (1), 2005, pp. 119–139. doi:10.1016/j.concog.2004.08.006. http://www.sciencedirect.com/science/article/pii/S1053810004000893
N. Block, On a confusion about a function of consciousness. Behavioral and Brain Sciences 18 (2), 1995, pp. 227–247. doi:10.1017/S0140525X00038188
D.J. Chalmers, In: The Conscious Mind: In Search of a Fundamental Theory, Oxford University Press, U.S.A., 1996, pp. xvii, 414
D. Gamez, Information integration based predictions about the conscious states of a spiking neural network. Consciousness and Cognition 19 (1), 2010, pp. 294–310. doi:10.1016/j.concog.2009.11.001. http://www.sciencedirect.com/science/article/pii/S1053810009001767
S. Harnad, The Symbol Grounding Problem. Physica D (42), 1990, pp. 335–346
R. Manzotti, A Process-oriented Framework for Goals and Motivations in Biological and Artificial Agents. In: Causality and Motivation, ed. by R. Poli, Ontos-Verlag, Frankfurt, 2010, pp. 105–134
G. Tononi, An information integration theory of consciousness. BMC Neuroscience 5 (42), 2004, pp. 1–22. doi:10.1186/1471-2202-5-42. http://www.biomedcentral.com/1471-2202/5/42/abstract
E. Bullmore, O. Sporns, The economy of brain network organization. Nature Reviews Neuroscience 13 (5), 2012, pp. 336–349. doi:10.1038/nrn3214. http://www.nature.com/nrn/journal/v13/n5/full/nrn3214.html
D. George, How the Brain Might Work: A Hierarchical and Temporal Model for Learning and Recognition. Ph.D. thesis, Stanford University, Stanford, CA, USA, 2008. AAI3313576
R. Manzotti, A. Chella, Physical integration: A causal account for consciousness. Journal of Integrative Neuroscience 13 (2), 2014, pp. 403–427. doi:10.1142/S0219635214400044. http://www.worldscientific.com/doi/abs/10.1142/S0219635214400044
V.B. Mountcastle, The columnar organization of the neocortex. Brain 120 (4), 1997, pp. 701–722. doi:10.1093/brain/120.4.701. http://brain.oxfordjournals.org/content/120/4/701
G. Tononi, G.M. Edelman, O. Sporns, Complexity and coherency: integrating information in the brain. Trends in Cognitive Sciences 2 (12), 1998, pp. 474–484. doi:10.1016/S1364-6613(98)01259-5. http://www.sciencedirect.com/science/article/pii/S1364661398012595
M. Sur, A. Angelucci, J. Sharma, Rewiring cortex: the role of patterned activity in development and plasticity of neocortical circuits. Journal of Neurobiology 41 (1), 1999, pp. 33–43
M. Sur, J.L.R. Rubenstein, Patterning and Plasticity of the Cerebral Cortex. Science 310 (5749), 2005, pp. 805–810. doi:10.1126/science.1112070. http://www.sciencemag.org/content/310/5749/805
O. Sacks, The Man Who Mistook His Wife For A Hat: And Other Clinical Tales. Berkeley, 1970
V.S. Ramachandran, W. Hirstein, The perception of phantom limbs. Brain 121 (9), 1998, pp. 1603–1630. doi:10.1093/brain/121.9.1603. http://brain.oxfordjournals.org/content/121/9/1603
B. Merker, Consciousness without a cerebral cortex: a challenge for neuroscience and medicine. The Behavioral and Brain Sciences 30 (1), 2007, pp. 63–81; discussion 81–134. doi:10.1017/S0140525X07000891
D.M. Kahn, L. Krubitzer, Massive cross-modal cortical plasticity and the emergence of a new cortical area in developmentally blind mammals. Proceedings of the National Academy of Sciences 99 (17), 2002, pp. 11,429–11,434. doi:10.1073/pnas.162342799. http://www.pnas.org/content/99/17/11429
J.A. Hobson, The Dream Drugstore: Chemically Altered States of Consciousness. MIT Press, 2002
B.J. Baars, In the Theatre of Consciousness: Global Workspace Theory, a Rigorous Scientific Theory of Consciousness. Journal of Consciousness Studies 4 (4), 1997, pp. 292–309
C. Koch, G. Tononi, Can Machines Be Conscious? IEEE Spectrum 45 (6), 2008, pp. 47–51
M. Shanahan, B. Baars, Applying global workspace theory to the frame problem. Cognition 98 (2), 2005, pp. 157–176. doi:10.1016/j.cognition.2004.11.007
G. Tononi, C. Koch, The neural correlates of consciousness: an update. Annals of the New York Academy of Sciences 1124, 2008, pp. 239–261. doi:10.1196/annals.1440.004
A. Clark, Whatever next? Predictive brains, situated agents, and the future of cognitive science. The Behavioral and Brain Sciences 36 (3), 2013, pp. 181–204. doi:10.1017/S0140525X12000477
O. Holland, The Future of Embodied Artificial Intelligence: Machine Consciousness? In: Embodied Artificial Intelligence, ed. by F. Iida, R. Pfeifer, L. Steels, Y. Kuniyoshi, no. 3139 In: Lecture Notes in Computer Science, Springer Berlin Heidelberg, 2004, pp. 37–53. http://link.springer.com/chapter/10.1007/978-3-540-27833-7_3
A. Noë, Action in Perception. MIT Press, 2004
J.K. O’Regan, How to Build a Robot that is Conscious and Feels. Minds and Machines 22 (2), 2012, pp. 117–136. doi:10.1007/s11023-012-9279-x. http://link.springer.com/article/10.1007/s11023-012-9279-x
R. Chrisley, Embodied Artificial Intelligence. Artifcial Intelligence 149 (1), 2003, pp. 131–150. doi:10.1016/S0004-3702(03)00055-9. http://dx.doi.org/10.1016/S0004-3702(03)00055-9
S. Dehaene, J.P. Changeux, L. Naccache, J. Sackur, C. Sergent, Conscious, preconscious, and subliminal processing: a testable taxonomy. Trends in Cognitive Sciences 10 (5), 2006, pp. 204–211. doi:10.1016/j.tics.2006.03.007
C. Koch, The Quest for Consciousness: A Neurobiological Approach. Roberts & Company Publishers, 2004
B. Libet, Mind Time: The Temporal Factor in Consciousness. Harvard University Press, Cambridge, Mass, 2005
S. Zeki, A. Bartels, Toward a theory of visual consciousness. Consciousness and Cognition 8 (2), 1999, pp. 225–259. doi:10.1006/ccog.1999.0390
J. Kim, Mind in a Physical World. MIT Press, Cambridge, Mass, 1998
J. Kim, Blocking Causal Drainage and Other Maintenance Chores with Mental Causation. Philosophy and Phenomenological Research 67 (1), 2003, pp. 151–176. http://www.jstor.org/stable/20140586
D. Gamez, Progress in machine consciousness. Consciousness and Cognition 17 (3), 2008, pp. 887–910. doi:10.1016/j.concog.2007.04.005. http://www.sciencedirect.com/science/article/pii/S1053810007000347
O. Holland, ed., Machine Consciousness. Imprint Academic, 2003
P. Yourgrau, A World Without Time: The Forgotten Legacy of Godel and Einstein. Penguin Books, New York, 2005
C. Adami, What Do Robots Dream Of? Science 314 (5802), 2006, pp. 1093–1094. doi:10.1126/science.1135929. http://science.sciencemag.org/content/314/5802/1093
D.J. Chalmers, A Computational Foundation for the Study of Cognition. Journal of Cognitive Science (Seoul) 12 (4), 2011, pp. 323–357
R. Manzotti, The Computational Stance Is Unfit for Consciousness. International Journal of Machine Consciousness 4 (2), 2012, pp. 401–420. doi:10.1142/S1793843012400239. http://www.worldscientific.com/doi/abs/10.1142/S1793843012400239
P. Dowe, Causal Processes. In: The Stanford Encyclopedia of Philosophy, ed. by E.N. Zalta, 2007. http://plato.stanford.edu/archives/fall2008/entries/causation-process/
L.A. Paul, N. Hall, Causation: A User’s Guide. Oxford University Press, New York, 2013
J. Schaffer, The Metaphysics of Causation. In: The Stanford Encyclopedia of Philosophy, ed. by E.N. Zalta, 2014. http://plato.stanford.edu/archives/sum2014/entries/causation-metaphysics/
H. Reichenbach, The Philosophy of Space and Time. Dover, 1958
S. Yablo, Advertisement for a Sketch of an Outline of a Prototheory of Causation. In: Causation and Counterfactuals, ed. by J. Collins, N. Hall, L.A. Paul, MIT Press, Cambridge, Mass, 2004, pp. 119–137
Y.S. Bonneh, A. Cooperman, D. Sagi, Motion-induced blindness in normal observers. Nature 411 (6839), 2001, pp. 798–801. doi:10.1038/35081073. http://www.nature.com/nature/journal/v411/n6839/full/411798a0.html
D.J. Simons, C.F. Chabris, Gorillas in our midst: sustained inattentional blindness for dynamic events. Perception 28 (9), 1999, pp. 1059–1074
F. Crick, Astonishing Hypothesis: The Scientific Search for the Soul. Touchstone, New York, 1994
A. Revonsuo, Inner Presence: Consciousness as a Biological Phenomenon. MIT Press, Cambridge, Mass, 2006
D. Marr, Vision. Freeman, San Francisco, 1982
M.R. Bennett, P. Hacker, Philosophical Foundations of Neuroscience. Blackwell, Malden, Mass, 2003
R. Manzotti, P. Moderato, Is Neuroscience Adequate As The Forthcoming “Mindscience”? Behavior and Philosophy 38, 2010, pp. 1–29. http://www.academia.edu/2767596/IS_NEUROSCIENCE_ADEQUATE_AS_THE_FORTHCOMING_MINDSCIENCE_
J. Dewey, Experience And Nature. Open Court, Chicago, 1925
W.T. Rockwell, Neither Brain nor Ghost. MIT Press, Cambridge, Mass, 2005
N.H. Kerr, G.W. Domhoff, Do the Blind Literally “See” in Their Dreams? A Critique of a Recent Claim That They Do. Dreaming 14 (4), 2004, pp. 230–233. http://asdreams.org/journal/issues/asdj14-4.htm#Do%20the%20Blind%20Literally%20See%20in%20Their%20Dreams%20A%20Critique%20of%20a%20Recent%20Claim%20That%20They%20Do. DREAMING 8(3) Abstracts - The Journal of the Association for the Study of Dreams
A. Revonsuo, C. Salmivalli, A content analysis of bizarre elements in dreams. Dreaming 5 (3), 1995, pp. 169–187. doi:10.1037/h0094433
E. Schwitzgebel, C. Huang, Y. Zhou, Do we dream in color? Cultural variations and skepticism. Dreaming 16 (1), 2006, pp. 36–42. doi:10.1037/1053-0797.16.1.36
C.S. Hurovitz, S. Dunn, G.W. Domhoff, H. Fiss, The Dreams of Blind Men and Women: A Replication and Extension of Previous Findings. Dreaming 9 (2-3), 1999, pp. 183–193. doi:10.1023/A:1021397817164. http://link.springer.com/article/10.1023/A%3A1021397817164
R. Manzotti, Is Consciousness Just Conscious Behavior? International Journal of Machine Consciousness 3 (2), 2011, pp. 353–360. doi:10.1142/S1793843011000765. http://www.worldscientific.com/doi/abs/10.1142/S1793843011000765
R. Jackendoff, Consciousness and the Computational Mind. MIT Press, Cambridge, Mass, 1987
R. Manzotti, F. Mutti, G. Gini, S.Y. Lee, Cognitive Integration through Goal-Generation in a Robotic Setup. In: Biologically Inspired Cognitive Architectures 2012, ed. by A. Chella, R. Pirrone, R. Sorbello, K.R. Jóhannsdóttir, no. 196 In: Advances in Intelligent Systems and Computing, Springer, 2013, pp. 225–231. http://link.springer.com/chapter/10.1007/978-3-642-34274-5_40
R. Manzotti, V. Tagliasco, From behaviour-based robots to motivation-based robots. Robotics and Autonomous Systems 51 (2–3), 2005, pp. 175–190. doi:10.1016/j.robot.2004.10.004. http://www.sciencedirect.com/science/article/pii/S0921889004001861
D. George, J. Hawkins, Towards a Mathematical Theory of Cortical Micro-circuits. PLoS Comput Biol 5 (10), 2009, p. e1000532. doi:10.1371/journal.pcbi.1000532. http://dx.doi.org/10.1371/journal.pcbi.1000532
J. Hawkins, S. Blakeslee, On Intelligence. Times Books, New York, 2004
J. Sharma, A. Angelucci, M. Sur, Induction of visual orientation modules in auditory cortex. Nature 404 (6780), 2000, pp. 841–847. doi:10.1038/35009043. http://www.nature.com/nature/journal/v404/n6780/abs/404841a0.html
J. Sharma, V. Dragoi, J.B. Tenenbaum, E.K. Miller, M. Sur, V1 Neurons Signal Acquisition of an Internal Representation of Stimulus Location. Science 300 (5626), 2003, pp. 1758–1763. doi:10.1126/science.1081721. http://www.sciencemag.org/content/300/5626/1758
M. Sur, P.E. Garraghty, A.W. Roe, Experimentally induced visual projections into auditory thalamus and cortex. Science 242 (4884), 1988, pp. 1437–1441
E.R. Kandel, L.R. Squire, Neuroscience: Breaking Down Scientific Barriers to the Study of Brain and Mind. Science 290 (5494), 2000, pp. 1113–1120. doi:10.1126/science.290.5494.1113. http://www.sciencemag.org/content/290/5494/1113
F. Aboitiz, D. Morales, J. Montiel, The evolutionary origin of the mammalian isocortex: towards an integrated developmental and functional approach. The Behavioral and Brain Sciences 26 (5), 2003, pp. 535–552; discussion 552–585
E. Thelen, Grounded in the World: Developmental Origins of the Embodied Mind. Infancy 1 (1), 2000, pp. 3–28. doi:10.1207/S15327078IN0101_02. http://onlinelibrary.wiley.com/doi/10.1207/S15327078IN0101_02/abstract
U. Neisser, In: Concepts and Conceptual Development: Ecological and Intellectual Factors in Categorization, Cambridge University Press, 1987, pp. x, 317
F. Tong, M.S. Pratte, Decoding patterns of human brain activity. Annual Review of Psychology 63, 2012, pp. 483–509. doi:10.1146/annurev-psych-120710-100412
P.S. Churchland, T.J. Sejnowski, Neural Representation and Neural Computation. Philosophical Perspectives 4, 1990, pp. 343–382
M. Shanahan, Embodiment and the inner life: Cognition and Consciousness in the Space of Possible Minds. Oxford University Press, USA, 2010
R. Manzotti, A Process-Based Architecture for an Artificial Conscious Being. In: Process Theories: Crossdisciplinary Studies in Dynamic Categories, ed. by J. Seibt, Springer Netherlands, 2003, pp. 285–312. http://link.springer.com/chapter/10.1007/978-94-007-1044-3_12
R. Manzotti, Machine Free Will: Is Free Will a Necessary Ingredient of Machine Consciousness? In: From Brains to Systems, ed. by C. Hernández, R. Sanz, J. Gómez-Ramirez, L.S. Smith, A. Hussain, A. Chella, I. Aleksander, no. 718 In: Advances in Experimental Medicine and Biology, Springer New York, 2011, pp. 181–191. http://link.springer.com/chapter/10.1007/978-1-4614-0164-3_15
R. Manzotti, L. Papi, S.Y. Lee, Does radical externalism suggest how to implement machine consciousness? In: Biologically Inspired Cognitive Architectures 2011, ed. by A. Samsonovich, K. Jóhannsdóttir, IOS Press, Amsterdam, pp. 232–240
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Manzotti, R., Jeschke, S. (2016). A Causal Foundation for Consciousness in Biological and Artificial Agents. In: Jeschke, S., Isenhardt, I., Hees, F., Henning, K. (eds) Automation, Communication and Cybernetics in Science and Engineering 2015/2016. Springer, Cham. https://doi.org/10.1007/978-3-319-42620-4_40
Download citation
DOI: https://doi.org/10.1007/978-3-319-42620-4_40
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-42619-8
Online ISBN: 978-3-319-42620-4
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)