Abstract
This paper is a continuation of a previous work where I argued that social expectations driving researches and developments in care robotics are not based on the simplistic optimism in a new commercial artifact but, involve a meaningful transformation of two rational features of human reality – ontology and normativity. In this paper my aim is to take a step forwards by investigating the kind of care relationship that could exist between a robot and a human. Usually, we take care of things and people because we love them, or else we want to give them support in their suffering. However, continuing to only value this sense of care, in a future rendered increasingly transparent and abstract by technology, may mean losing sight of the fact that taking care of others also means taking care of ourselves. If we completely entrust robots with the role of caring, the bigger concern is not the foreseeable decrease in the “humanity” in healthcare contexts, but the much more challenging notion of people surrendering the value and meaning in their lives. Since caring about something means, firstly, giving it value, a society passively nursed by technology is a society unable to give value to things and people. In order to avoid this risk, new approaches are required, no longer based on love or solidarity, but responsibility. I have named this approach “I tech care” and this paper aims to provide a general overview of the main concepts involved.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
I speak about “human rationality” because, in my view, the predicted dissemination of assistive and care robotics constitutes a hugely challenging topic not only for some rational cultures – as the western world and its epistemological paradigms – rather of human beings as such. Whether the Enlightenment was still a form of rationality developed in Europe and North America, thanks or because of globalization, today the worldwide techno-scientific cultures move toward a general convergence. According to a report commissioned by the U.S. National Science Foundation and Department of Commerce we are aiming for a “Converging Technologies for Improving Human Performance” (http://www.wtec.org/ConvergingTechnologies/Report/NBIC_report.pdf). I do not sustain that technological progress is in itself a factor of human development. On the contrary, in my view, without freedom and democracy technology alone does not improve the human condition (Sen, 1999). But I am also convinced that the socialization and globalization of technologies we are witnessing – and healthcare robots are an emblematic example – will expand regardless of cultures and religions. These two processes end up complementing and hybridizing old and new myths (Hughes, 2012).
- 2.
Aristotle, in his doctrine of causality, defined four different typologies of cause of a manufactured thing (Metaphysics, I, 3). Considering the classic example of the statue, the sculptor is the efficient cause, bronze the material cause, the form taken by the statue is the formal cause, and the worship of the deity is the final cause.
- 3.
In this conception of “real abstraction” I was inspired by the work of Roberto Finelli and his reinterpretation of the link between Hegelian and Marxist philosophies (see Finelli, 1999).
- 4.
It could be argued, at least, that without freedom no level of abstraction is possible, but I would define this kind of abstraction – recalling one of Hegel’s conceptions – a sort of “bad infinity” (schlechte Unendlichkeit), that is simply the finite repeated infinitely, an abstraction that is always realized in half, a representation of the world that denies itself and is forced to resort to false images to explain the contents of the real (Encyclopaedia, §93).
- 5.
- 6.
The philosophical debate on the theme of solidarity is very broad. Therefore it is difficult to give a complete account. Generally speaking, the idea of solidarity I refer to owes very much to authors such as Rorty (1989), Honneth (1995), Habermas (1998), and Benhabib (2002). According to their views, the processes of socialization in contemporary society have little to do with conventional drives. Social cohesion is no longer a matter of membership to a family, a clan, a class. As already Durkheim predicted by analyzing the division of labor, the modern concept of solidarity requires no empathy or identity with others with whom we are in solidarity. Instead, it requires the acceptance of the other as other, an acknowledgment based on a discourse of justice and not on belonging or benevolence (Habermas, 1998). Solidarity, justice and care, then, become two sides of the same coin. By treating the other according to the norms of friendship, love and care, our humanity is declared, but also our individuality as a person. The moral categories that follow such interactions are responsibility, relationships, and participation. The corresponding moral feelings are love, care, sympathy and solidarity (Benhabib, 2002).
- 7.
References
Ackerman, E., & Guizzo, E. (2011). Mystery robot revealed: Robodynamics luna is fully programmable adult-size personal robot. IEEE Spectrum, 11 May 2011.
Agamben, G. (1998). Homo sacer: Sovereign power and bare life (D. Heller-Roazen, Trans.). Stanford: Stanford University Press.
Battaglia, F., & Carnevale, A. (Eds.). (2014). Reframing the debate on human enhancement. Humana Mente Journal of Philosophical Studies, issue 26.
Benhabib, S. (2002). The claims of culture: Equality and diversity in the global era. Princeton: Princeton University Press.
Buchanan, A. (2009). Human nature and enhancement. Bioethics, 23(3), 141–150.
Carnevale, A. (2014). Ontology and normativity in the care-robot relationship. In J. Seibt, R. Hakli, & M. Nørskov (Eds.), Sociable robots and the future of social relations: Proceedings of robo-philosophy 2014 (pp. 143–152). Amsterdam: IOS Press.
Carnevale, A. (2015). Robots, disability, and good human life. Disability Studies Quarterly, 35(1). doi:http://dx.doi.org/10.18061/dsq.v35i1.4604
Clark, A. (2003). Natural-born cyborg. Oxford: Oxford University Press.
Coeckelbergh, M. (2010). Robot rights? Towards a social-relational justification of moral consideration. Ethics and Information Technology, 12(3), 209–221.
Datteri, E., & Tamburrini, G. (2009). Ethical reflections on health care robots. In R. Capurro & M. Nagenborg (Eds.), Ethics and robotics (pp. 35–48). Amsterdam: IOS Press.
Derrida, J. (1995). The gift of death (D. Wills, Trans.) Chicago: University of Chicago Press.
Feil-Seifer, D. J., & Matarić, M. J. (2011). Ethical principles for socially assistive robotics. IEEE Robotics & Automation Magazine, 18(1), 24–31.
Ferri, G., Manzi, A., Salvini, P., Mazzolai, B., Laschi, C., & Dario, P. (2011). Dust-Cart, an autonomous robot for door-to-door garbage collection: From DustBot project to the experimentation in the small town of Peccioli. In 2011 IEEE international conference on robotics and automation (ICRA), Shanghai (pp. 655–660).
Finelli, R. (1999). Abstraktion. In H. Sandkühler (Ed.), Enzyklopädie philosophie (pp. 20–23). Hamburg: Felix Meiner Verlag.
Floridi, L. (2013). The ethics of information. Oxford: Oxford University.
Frankfurt, H. (2004). The reasons of love. Princeton: Princeton University Press.
Habermas, J. (1998). Inclusion of the other: Studies in political theory (C. Cronin & P. D. Greiff, Trans.). Cambridge MA: MIT Press.
Henry, B. (2014). Imaginaries of the global age. ‘golem and others’ in the post-human condition. Politica & Societá Periodico di filosofia politica e studi sociali, 2, 221–246.
Honneth, A. (1995). The struggle for recognition: The moral grammar of social conflicts (J. L1Anderson, Trans.). Cambridge: Polity Press.
Hughes, J. (2012). The politics of transhumanism and the techno-millennial imagination, 1626–2030. Zygon: Journal of Religion and Science, 47, 757–776.
Ihde, D. (1990). Technology and the lifeworld: From garden to earth. Bloomington: Indiana University Press.
Kanda, T., Shiomi, M., Miyashita, Z., Ishiguro, H., & Hagita, N. (2009). An affective guide robot in a shopping mall. In Proceedings of the 4th ACM/IEEE international conference on human robot interaction (HCI) (pp. 173–180).
Kittay, E. F. (1999). Love’s labor: Essays on women, equality, and dependency. New York: Routledge.
Kristiansen, K., Vehmas, S., & Shakespeare, T. (2009). Arguing about disability: Philosophical perspectives. London: Routledge.
Levy, D. (2007). Love and sex with robots: The evolution of human-robot relationships. New York: HarperCollins.
Loretoni, A. (2015). Ampliare lo sguardo. Rome: Donzelli.
Lucivero, F., Swierstra, T., & Boenink, M. (2011). Assessing expectations: Towards a toolbox for an ethics of emerging technologies. NanoEthics, 5(2), 129–141.
Mauss, M. (1967). The gift: Forms and functions of exchange in archaic societies (I. Cunnison, Trans.). New York: Norton.
Nussbaum, M. C. (2006). Frontiers of justice: Disability nationality species membership. Cambridge MA: Harvard University Press.
Pirni, A., & Carnevale, A. (2014). Technologies change – do we change as well? on the link between technologies, self, and society. Politica & Societá, 2, 173–184.
Ricoeur, P. (2005). The course of recognition (D. Pellauer Trans.). Cambridge MA: Harvard University Press.
Rorty, R. (1989). Contingency irony and solidarity. Cambridge, MA: Cambridge University Press.
Sen, A. (1999). Development as freedom. Oxford: Oxford University Press.
Sharkey, N., & Sharkey, A. (2011). The rights and wrongs of robot care. In P. Lin, K. Abney, & G. A. Bekey (Eds.), Robot ethics: The ethical and social implications of robotics (pp. 267–282). Cambridge MA: MIT Press.
Simonov, M., Frisiello, A., & Bazzani, M. (2012). Using humanoid robot in ambient-assisted living. Global Telemedicine and eHealth Updates: Knowledge Resources, 5, 438–442.
Turkle, S. (2011). Alone together: Why we expect more from technology and less from each other. New York: Basic Books.
van den Berg, B. (2011). Robots as tools for techno-regulation. Law, Innovation and Technology, 3(2), 317–332.
van den Berg, B., & Leenes, R. (2013). Abort, retry fail: Scoping techno-regulation and other techno-effects. Human Law and Computer Law: Comparative Perspectives, 25, 67–87.
van Wynsberghe, A. L. (2013). Designing robots for care: Care centered value-sensitive design. Science and Engineering Ethics, 19(2), 407–433.
Veloso, M., Biswas, J., Coltin, B., Rosenthal, S., Kollar T., Mericli, C.,…Ventura, R. (2012). Cobots: Collaborative robots servicing multi-floor buildings. In 2012 IEEE/RSJ international conference on intelligent robots and systems, Vilamoura (pp. 5446–5447).
Verbeek, P. P. (2011). Moralizing technology: Understanding and designing the morality of things. Chicago: University of Chicago Press.
Wallach, W., & Allen, C. (2009). Moral machines: Teaching robots right from wrong. Oxford: Oxford University Press.
Young, K. (1996). Internet addiction: The emergence of a new clinical disorder Cyberpsychology and Behavior, 3, 237–244.
Acknowledgements
This paper is a continuation of a previous work (Carnevale, 2014) and it is one of the outcomes of the Research Unit “Social Asymmetries and Political Inclusion: Concepts, methods, and policies”, Scuola Superiore Sant’Anna, Pisa (Prin 2010–2011: Local Coordinator, Barbara Henry; National Coordinator, Laura Bazzicalupo).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this chapter
Cite this chapter
Carnevale, A. (2017). “I Tech Care”: How Healthcare Robotics Can Change the Future of Love, Solidarity, and Responsibility. In: Hakli, R., Seibt, J. (eds) Sociality and Normativity for Robots. Studies in the Philosophy of Sociality. Springer, Cham. https://doi.org/10.1007/978-3-319-53133-5_11
Download citation
DOI: https://doi.org/10.1007/978-3-319-53133-5_11
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-53131-1
Online ISBN: 978-3-319-53133-5
eBook Packages: Religion and PhilosophyPhilosophy and Religion (R0)