1 I, Robot: How Human Appearance and Mind Attribution Relate to the Perceived Danger of Robots

Watching the movie ‘Ex Machina’, you quickly perceive Ava, the android main character of the movie, as a real human with emotions and feelings. Anthropomorphising Ava in this way, that is to ascribe human-like characteristics and/or intentions to non-human agents, is a fundamental human process that spontaneously happens and increases our social connection with non-human agents [1]. Although highly evolved robots seem a vision of the future, we already interact with artificial intelligent agents on a regular basis (e.g., Siri, Apple’s speaking assistant, Amazon’s Alexa, or CaMeLi, an avatar designed to help elderly in daily life). Developments in robot technology are proceeding rapidly: ‘Social robots’, i.e., robots that are designed to interact and communicate with people [2], feature increasingly more human-like appearances and behaviour. While, these technical developments are especially interesting when it comes to maintaining and improving our quality of life, for example in health care or education, a large body of research also shows that social robots tend to elicit negative feelings of eeriness, danger, and threat [3,4,5,6,7]. In the present study, we investigated the factors that elicit these negative feelings and clarified possible underlying mechanisms, including the extent to which robots look human-like and the extent to which they are attributed with a mind.

1.1 Anthropomorphism

Anthropomorphism involves “going beyond behavioral descriptions of imagined or observable actions (e.g., the dog is affectionate) to represent an agent’s mental or physical characteristics using humanlike descriptors (e.g., the dog loves me)” [1, page 865]. Anthropomorphism for non-human agents can be elicited in two ways: First, by increasing the physical appearance with humans [8,9,10]. Second, through attributions that ascribe agency, affect, or intentionality to the non-human [11]. Interestingly, humans spontaneously anthropomorphise non-human agents [1]. Initially introduced to describe the appearance of religious agents and gods [Hume, 1757, in 1], the term is now used to describe human characteristics towards animals or plants [12], objects and technical devices [13], and even geometric shapes [14]. In fact, neuroscientific research has demonstrated that similar brain regions are activated when participants attribute mental states to non-human agents as when attributing mental states to other humans [15,16,17]. In a predictive coding framework [18, 19], which suggests that the brain continuously produce hypotheses that predict sensory input, anthropomorphism makes sense: when something looks like a human or moves or acts in a human-like way, it is more likely that your interaction with this agent will be efficient and smooth if you treat it as another human-being.

Research has shown that when we anthropomorphise non-biological agents, this has profound effects on how we interact with these agents: it leads to more interpersonal closeness [20], increased moral care [21], and smoother interactions [11, 22]. In addition, people rate robots acting playfully as being more extroverted and outgoing than robots acting seriously [23, 24]. Moreover, avatars are judged to be more trustworthy, competent, sensitive, and warm when they are anthropomorphised [25], and similar stereotypes are applied to robots than to humans [26].

1.2 Negative Effects of Anthropomorphism

However, perceptual similarity with humans can also elicit negative feelings towards robots: Robots whose physical appearance closely (but not perfectly) resembles human beings often evoke negative feelings, a phenomenon referred to as the uncanny valley [4]. The uncanny valley hypothesis states that more realistic artificial agents elicit more positive reactions until they are very close (but not close enough) to the human ideal. This dimension of “realism”, however, is not limited to realistic looks/appearances, but also realistic motor behaviours, such as imitation [4]. Research showed that human-like appearance or motion can be responsible for the emergence of the uncanny valley [27, 28]. Furthermore, large inter-individual differences in the emergence of the uncanny valley have been found, suggesting that people can be more or be less sensitive towards negative feelings in response to a robot [5]. Often, the concept of prediction error has been used to explain the uncanny valley [29]: when something looks like a human or moves or acts in a human-like way, it is more likely that your interaction with this agent will be efficient and smooth if you treat it as another human-being. However, when a robot that looks extremely human-like moves in a mechanical way, our predictions are violated since their strong anthropomorphic appearance led us to expect them to follow biological movement patterns. Consequently, this high prediction error leads to a negative feeling of unease or fear.

Interestingly, an additional explanation of the uncanny valley has been proposed recently, with empirical studies suggesting that this negativity for human-like robots can be explained by a violation of the need for distinctiveness [30, 31]. As humans, we feel unique and distinct by understanding how our own group differs from another group [32, 33]. However, when this feeling of uniqueness, and with it our intergroup boundaries, disappears, we feel threatened. Applied to robots, correlational research showed that higher feelings of eeriness and a decrease in felt warmth towards robots was positively related to whether participants perceived robots and humans as similar categories [31]. In addition, it was causally demonstrated that too much perceived similarity between robots and humans undermines people’s ideas about human uniqueness, and this subsequently leads people to perceive robots as more threatening and potentially damaging entities [30].

In line with this assumption, looking at manipulation of anthropomorphism, a recent study on human/avatar interactions found evidence for a so-called “uncanny valley of the mind” [6]. In their study, participants watched interactions involving emotional responses between two digital characters, which were presented as either human-controlled versus computer-controlled, and scripted versus autonomous. Their results showed that levels of eeriness rose especially when participants thought the digital agents were autonomous artifical intelligences. These findings support the notion that attributions of a mind might lead to a decrease in human/machine distinctiveness, subsequently leading to a negative feelings and perceived damage to someone’s identity.

A key feature by which we distinguish between humans and non-humans relates to mind attribution. Specifically, we attribute the minds of potential other agents along the lines of two constructs: experience and agency [34]. Experience involves capacities to feel emotions, such as the ability to experience hunger or pleasure, whereas agency relates to capacities of being an autonomous agent, such as self-control and thought. Especially experience is seen as a unique human trait, as people ascribe a medium amount of agency but no experience towards robots [34]. The question however is whether mind attribution influences the need for distinctiveness.

In three studies, we aimed to replicate the link between robot-human similarity and threat-perceptions using a variety of stimuli (Studies 1-3). Furthermore, we aimed to extend the literature in this domain by investigating whether distinctiveness-negativity is evoked by perceived physical similarity, by similarity in mind attribution, or both (Study 2 and Study 3; see Fig. 1 for an overview of the mediation model). Similar to earlier research [30], participants were presented with pictures of mechanical, humanoid, and android robots, and physical anthropomorphism, mind attribution of agency and experience (Study 2 and Study 3), threat to the human–machine distinctiveness, and damage to humans and their identity were assessed for all three robot types. We expected that human–machine distinctiveness mediated the influence of robot type on the perceived damage for humans and their identity, and that this mediation would be due to (a) the anthropomorphic appearance of the robot, and (b) perceived experience and perceived agency from the robot.

Fig. 1
figure 1

Overview of the tested mediation model

2 Study 1

2.1 Methods

2.1.1 Participants and Design

Fifty-five participants (44 females, 11 males, Mage = 19.50 years, SDage = 1.40, age range 17–23 years) completed the experiment in exchange for course credits. The experiment had a 3 (Robot type: mechanical robot vs. humanoid robot vs. android robot) within-subjects design, with damage to humans and their identity as dependent variable, and physical anthropomorphism and threat to the human–machine distinctiveness as mediators. Data was acquired online using Inquisit 4 [35]. Via an online participant pool from Radboud University, participants could sign up for the study, and received a link to complete the study online. Before participants could start with the experiment, they were asked to ensure they would not be interrupted for the duration of the experiment (approximately 30 min).

2.1.2 Procedure and Materials

Participants were instructed that they had to evaluate different types of robots: Participants received a self-paced evaluation task in which they indicated their attitudes towards three different categories of robots varying in human-like appearance; participants either saw mechanical, humanoid, or android robots. While mechanical robots were clearly machines (no legs, no facial features), humanoid robots were more humanlike by having legs, arms, a torso, and a head with a face. Still, they also possess clear similarities with a machine (e.g., no hair or skin). Android robots were very high in humanlike-ness and difficult to distinguish from real humans. All stimuli and measures used were derived from earlier research [30]. Four different robots were used for each robot category, resulting in 12 robot evaluations.Footnote 1

Participants had to rate each of the 12 robots on (1) physical anthropomorphism; (2) threat to human–machine distinctiveness; and (3) damage to humans and human identity. Firstly, physical anthropomorphism was assessed using a three-item scale (e.g., “I could easily mistake the robot for a real person”; Cronbach’s αmechanical robot = .935; Cronbach’s αhumanoid robot = .887; Cronbach’s αandroid robot = .800).Footnote 2 Secondly, threat to the human–machine distinctiveness was measured using a three-item scale (e.g., “Looking this kind of robot I ask myself what the differences are between robots and humans”; Cronbach’s αmechanical robot = .948; Cronbach’s αhumanoid robot = .957; Cronbach’s αandroid robot = .963). Thirdly, damage to humans and human identity was assessed using a four items scale (e.g., “I get the feeling that the robot could damage relations between people”; Cronbach’s αmechanical robot = .880; Cronbach’s αhumanoid robot = .895; Cronbach’s αandroid robot = .897). For all three questionnaires, participants could answer on a 7-point Likert scale (1 = ‘totally not agree’ to 7 = ‘totally agree’). Items were presented along with each robot photo respectively, and robot photos were presented in a random order. After completing the questionnaires, participants reported their age and gender, and were thanked for their participation.

2.2 Results and Discussion

Within-participant mediation analyses were conducted using MEMORE package in SPSS with 1000 bootstrap samples [36]. In this analysis conclusions on mediation are based upon the correlation of the score-difference between conditions of the dependent variable and the score-difference between conditions of the proposed mediator. Since each participant responded to all three conditions, and we did not have specific hypotheses on certain comparisons, we reported all possible comparisons of three robot types (i.e., mechanical vs. humanoid, mechanical vs. android, and humanoid vs. android) while using Bonferroni correction to control Type one error (here we used 99% confidence interval). All mediation statistics (B and 99% CI) are depicted in Table 1. An overview about descriptive statistics can be found in Table 2.

Table 1 B’s and 99% confidence intervals (between brackets) of the mediation analyses of Study 1, as a function of comparison (mechanical vs. humanoid; mechanical vs. android; humanoid vs. android)
Table 2 Descriptive statistics (Means, SD) of Study 2 for all variables (N = 55)

2.2.1 Mediation Effect of Distinctiveness on the Relation Between Robot Types and Damage to Humans and Their Identity

We conducted a mediation analysis on the relation between robot types as IV and damage to humans and their identity as DV, with the human–machine distinctiveness of the different robot types as a proposed mediator. The model indicated significant indirect effects and non-significant direct effects for all three comparisons (mechanical with android, mechanical with humanoid, and humanoid with android robot). The relation between robot-types and damage to humans and their identity was therefore fully mediated by the human–machine distinctiveness of the different robot types.

2.2.2 Mediation Effect of Anthropomorphic Appearance on the Relation Between Robot Type and Distinctiveness

We conducted a mediation analysis on the relation between robot types as IV and human–machine distinctiveness as DV, with the anthropomorphic appearance of the different robot types as a proposed mediator. The model indicated significant indirect effects and non-significant direct effects for all three comparisons. The relation between robot-types and human–machine distinctiveness was therefore fully mediated by the anthropomorphic appearance of the different robot types.

Our findings of Study 1 successfully replicate earlier research on the relationship between anthropomorphic appearance, human–machine distinctiveness, and perceived threat to humans and their identity [30]: Human–machine distinctiveness mediated the influence of robot type on the perceived damage for humans and their identity, and this mediation was due to anthropomorphic appearance of the robot. Thus, these results support the notion that too much physical appearance leads to negative perceptions of robots due to a decrease in distinctiveness [30].

In our following studies, we add the concept of mind perception to our model [34, 37], a fundamental construct central for us to understand ourselves as living beings. The mind perception literature distinguishes between two essential constructs: agency and experience [34]. Both agency and experience are critical components for being human, and how we are distinct from robots, avatars, and other non-living objects. Our aim in Study 2 was to replicate findings from Study 1, and additionally, to investigate whether agency and experience mediate the relation between robot-types and human machine distinctiveness.

3 Study 2

3.1 Methods

3.1.1 Participants

Sixty participants (52 females, 5 males, 3 unknown, Mage = 19.30 years, SDage = 1.90, age range 17–24 years) completed the experiment for course credits. The experiment had a 3 (Robot type: mechanical robot vs. humanoid robot vs. android robot) within-subjects design, with damage to humans and their identity as dependent variable, and physical anthropomorphism, mind attribution, and threat to the human–machine distinctiveness as mediators. Data was acquired similar to Study 1.

3.1.2 Procedure and Materials

The same procedure and stimuli from Study 1 was used. However, in Study 2, the mind attribution scale was added (Gray et al., 2007), and subsequently, the other variables (physical anthropomorphism: Cronbach’s αmechanical robot = .884; Cronbach’s αhumanoid robot = .832; Cronbach’s αandroid robot = .785; threat to human–machine distinctiveness: Cronbach’s αmechanical robot = .948; Cronbach’s αhumanoid robot = .957; Cronbach’s αandroid robot = .974; damage to humans and their identity: Cronbach’s αmechanical robot = .932; Cronbach’s αhumanoid robot = .937; Cronbach’s αandroid robot = .918) were assessed. Participants had to rate 18 adjectives on whether they think the robot can experience certain capacities connected to agency (e.g., thought, self-control; 7 items; Cronbach’s αmechanical robot = .841; Cronbach’s αhumanoid robot = .884; Cronbach’s αandroid robot = .898) and experience (feelings such as anger, joy; 11 items; Cronbach’s αmechanical robot = .943; Cronbach’s αhumanoid robot = .967; Cronbach’s αandroid robot = .978). A mean score for all dependent variables was calculated before data analyses.

3.2 Results and Discussion

All mediation statistics (B and 99% CI) are depicted in Table 3. An overview about descriptive statistics can be found in Table 4.

Table 3 B’s and 99% confidence intervals (between brackets) of the mediation analyses of Study 2, as a function of comparison (mechanical vs. humanoid; mechanical vs. android; humanoid vs. android)
Table 4 Descriptive statistics (Means, SD) of Study 2 for all variables (N = 60)

3.2.1 Mediation Effect of Distinctiveness on the Relation Between Robot Types and Damage to Humans and Their Identity

We conducted the same mediation analysis as in Study 1: We treated robot types as IV and damage to humans and their identity as DV, and human–machine distinctiveness of the different robot types as a proposed mediator. The same results showed that the relation between robot-types and damage to humans and their identity was fully mediated by the human–machine distinctiveness of the different robot types.

3.2.2 Mediation Effect of Anthropomorphic Appearance on the Relation Between Robot Type and Distinctiveness

We conducted the same mediation analysis as in Study 1 on the relation between robot types as IV and human–machine distinctiveness as DV, with the anthropomorphic appearance of the different robot types as a proposed mediator. We replicate our findings in Study 1: the relation between robot-types and human–machine distinctiveness was therefore fully mediated by the anthropomorphic appearance of the different robot types.

3.2.3 Mediation Effect of Mind Attribution on the Relation Between Robot Type and Distinctiveness

We conducted a mediation analysis on the relation between robot types as IV and human–machine distinctiveness as DV, with the mind attribution of the different robot types as a proposed mediator. The model indicated non-significant indirect effects of the experience component of mind attribution, thus no mediation effect was found.

The mediation analysis that involved the agency component of mind attribution as a proposed mediator showed different results depending on which robot types were compared. In the comparisons between the mechanical versus android robot types, and between the mechanical versus humanoid robot types, the model indicated non-significant indirect effects, only significant direct effects. Furthermore, the model indicated significant indirect effects and significant direct effects in the comparison between the humanoid versus the android robot types. The relation between robot-types and human–machine distinctiveness might be partially mediated by the anthropomorphic appearance when comparing humanoid and android robots.

3.2.4 Correlation Between Mind Attribution and Damage to Humans and Their Identity

To explore whether the correlation between mind attribution and damage to humans and their identity vary differently between robot types, correlation analyses on each robot type were performed. The results showed that the correlation between agency and damage to humans and their identity for mechanical and humanoid robots were significant (rmechanical = .369, p = .004; rhumanoid = .313, p = .015), but not for android robots (randroid = .193, p = .140). The correlation between experience and damage to humans and their identity for mechanical robots was significant (rmechanical = .485, p < .001), but not significant for humanoid and android robots (rhumanoid = .205, p = .116; randroid = .171, p = .190).

In line with earlier research [30], we could again support the notion that an anthropomorphic appearance mediates the human/machine distinctiveness, and thus gives room to feelings of threat and damage to humans. Additionally, we did not find mediating effect of experience on participants perception of human/machine distinctiveness, while agency might partially mediate the influence on human/machine distinctiveness only for the comparison between the humanoid and android robots. Interestingly, however, is the fact that the results showed a trend of positive correlation between both experience and agency and the damage to humans and their identity.

As the correlations were not all significant, an interpretation is difficult. One could argue that especially the similarity between android robots and humans weakens a relationship between agency/experience perception with damage to humans and their identity: it is easier to identify with these robots and perceive them as closer to humans. Another explanation could be that agency and experience would be much more logical for android robots, while mechanical or humanoid robots do not need these capacities and provide a bigger threat when possessing them. To further strengthen and be able to interpret our findings of Study 2, we conducted a third study with the goal to replicate this pattern. To be sure that the current pattern is not due to the used stimuli, new pictures of mechanical, humanoid, and android robots were introduced in Study 3.

4 Study 3

4.1 Methods

4.1.1 Participants and Design

Sixty-seven participants (57 females, 8 males, 2 unknown, Mage = 19.10 years, SDage = 1.20, age range 17 – 23 years) completed the experiment for course credits. The experiment had a 3 (Robot type: mechanical robot vs. humanoid robot vs. android robot) within-subjects design, with damage to humans and their identity as dependent variable, and physical anthropomorphism, mind attribution, and threat to the human–machine distinctiveness as mediators. Data was acquired similar to Study 1 and Study 2.

4.1.2 Procedure and Materials

The same procedure as in Study 2 was used. However, in Study 3, new stimuli for the three robot categories (mechanical, humanoid, or android robots) were used. While mechanical robots were clearly machines (e.g., no legs), but in comparison to Study 1 and Study 2, they had arms and a head. Humanoid robots were more humanlike by having legs, arms, a torso, and a head with a face. Still, they also possess clear similarities with a machine (e.g., no hair or skin). Lastly, android robots were very high in humanlike-ness and difficult to distinguish from real humans perceptually. Only full body pictures of the robots were used in Study 3 and previously used in research [5]. Again, for each category, four different robots were used, resulting in 12 robots which had to be evaluated on agency attribution (Cronbach’s αmechanical robot = .804; Cronbach’s αhumanoid robot = .852; Cronbach’s αandroid robot = .902), experience attribution (Cronbach’s αmechanical robot = .916; Cronbach’s αhumanoid robot = .916; Cronbach’s αandroid robot = .970), anthropomorphic appearance (Cronbach’s αmechanical robot = .931; Cronbach’s αhumanoid robot = .887; Cronbach’s αandroid robot = .846), human/machine distinctivenessFootnote 3 (Cronbach’s αmechanical robot = .983; Cronbach’s αhumanoid robot = .986; Cronbach’s αandroid robot = .977), and damage to humans and their identity (Cronbach’s αmechanical robot = .902; Cronbach’s αhumanoid robot = .932; Cronbach’s αandroid robot = .902).

4.2 Results and Discussion

All mediation statistics (B and 99% CI) are depicted in Table 5. An overview about descriptive statistics can be found in Table 6.

Table 5 B’s and 99% confidence intervals (between brackets) of the mediation analyses of Study 3, as a function of comparison (mechanical vs. humanoid; mechanical vs. android; humanoid vs. android)
Table 6 Descriptive statistics (means, SD) of Study 3 for all variables (N = 67)

4.2.1 Mediation Effect of Distinctiveness on the Relation Between Robot Types and Damage to Humans and Their Identity

We conducted the same mediation analysis as in Study 1 and Study 2: We treated robot types as IV and damage to humans and their identity as DV, and human–machine distinctiveness of the different robot types as a proposed mediator. We replicated our findings of Study 1 and Study 2: the relation between robot-types and damage to humans and their identity was therefore fully mediated by the human–machine distinctiveness of the different robot types.

4.2.2 Mediation Effect of Anthropomorphic Appearance on the Relation Between Robot Type and Distinctiveness

We conducted a mediation analysis on the relation between robot types as IV and human–machine distinctiveness as DV, with the anthropomorphic appearance of the different robot types as a proposed mediator. Differently to Study 1 and Study 2, for the comparison between the mechanical with the humanoid robot, the model indicated non-significant indirect effect. All other comparisons indicated significant indirect effects.

4.2.3 Mediation Effect of Mind Attribution on the Relation Between Robot Type and Distinctiveness

We conducted a mediation analysis on the relation between robot types as IV and human–machine distinctiveness as DV, with the mind attribution of the different robot types as a proposed mediator.

The mediation analysis that involved the experience component of mind attribution as a proposed mediator showed different results depending on which robot types were compared. Specifically, the model indicated significant indirect effects and significant direct effects in the comparison between the mechanical versus the android robot types. The relation between robot-types and human–machine distinctiveness was therefore partially mediated by difference between mechanical and android in the experience attribution. In the comparisons between the mechanical versus humanoid robot types, and the humanoid versus android robot types, the model indicated non-significant indirect effects, only significant direct effects.

The mediation analysis that involved the agency component of mind attribution as a proposed mediator showed different results depending on which robot types were compared. Specifically, the model indicated significant indirect effects and significant direct effects in the comparison between the mechanical versus the android robot types, and the humanoid versus the android robot types. The relation between robot-types and human–machine distinctiveness was therefore partially mediated by the agency attribution of the different robot types. In the comparisons between the mechanical versus humanoid robot types the model indicated non-significant indirect effects, only significant direct effects.

4.2.4 Correlation Between Mind Attribution and Damage to Humans and Their Identity

Like in Study 2, we explored whether the correlation between mind attribution and damage to humans and their identity vary differently between robot types, correlation analyses on each robot type were performed. The results showed that the correlation between agency and damage to humans and their identity was not significant for mechanical robots (rmechanical = .128, p = .301), but the correlation for the humanoid and android robots were significant (rhumanoid = .259, p = .034; randroid = .304, p = .012).The correlation between experience and damage to humans and their identity for mechanical robots was significant (rmechanical = .252, p = .04), a non-significant trend was found for humanoid robots (rhumanoid = .229, p = .063), and a significant correlation for android robots (randroid = .309, p = .011).

While we replicated earlier research on anthropomorphic appearance [30], our results on mind perception were less straightforward. Possible explanations of these findings and inconsistencies between Study 2 and Study 3 concerning mind perception will be explained in the following discussion section.

5 General Discussion

In the present study we investigated the factors that elicit negative feelings towards robots and clarified possible underlying mechanisms, including the extent to which robots look human-like and the extent to which they are attributed with a mind. More specifically, we expected that a felt decrease in distinctiveness between humans and machines would be the reason that participants feel a potential damage to their identity, and that this influence can be explained by the human-like appearance of a robot as well as the attribution of agency and experience to a robot. That is, when people attribute high levels of agency or experience to a robot, or when a robot looks very human-like, this makes the boundaries between humans and machines blurrier, leading to higher levels of perceived damage to ones’ identity as human.

Importantly, we could replicate earlier research [30] which demonstrated that anthropomorphic appearance indeed mediated the relationship between type of robot, human–machine distinctiveness, and perceived damage for humans and their identity. In three studies we were able to replicate earlier findings showing that when people feel that the distinction between humans and robots are blurred intergroup distinctiveness is threatened [30], both with similar and new stimuli of mechanical, humanoid, and android robots. This further strengthens the notion that although robot familiarity can be used to reach the goal of increasing robot acceptance, “this goal should however not conflict with the need for distinctiveness that typically characterizes intergroup comparisons” [30, page 299]. Thus, when robots are designed to interact with people in their daily life, and when high acceptance is important for successful use (e.g., care taking robots for elderly), it is important to consider that human-like appearance can lead to resistance.

The findings regarding the attribution of agency and experience were less consistent than the results on anthropomorphic appearance. Based on our findings, we cannot conclude that mind attribution has comparable effects on human/machine distinctiveness as anthropomorphic appearance and mediates the relationship between type of robot, human–machine distinctiveness, and damage to humans and their identity. In Study 2, one out of six comparisons showed a mediation effect for agency when comparing humanoid and android robots, which could be the result of chance. In Study 3 more significant results were found, with agency attributions mediating the influence on human/machine distinctiveness for the comparison between the mechanical and android robots, and the humanoid and android robots. Additionally, only in study 3, experience attribution mediated the effect on participants perception of human/machine distinctiveness, but for mechanical versus android robots only. Three explanations for why we did not find a stable pattern in results for mind attribution should be mentioned: Firstly, different stimuli were used between Study 2 and Study 3, with pictures in Study 3 being more controlled (e.g., same background) and only reflecting full-body postures. Research has demonstrated that robot perception is highly dependent on the sort of stimuli, and therefore, pictures in Study 3 lead to different effects. Whether these effects are more reliable needs to be replicated in future research. Secondly, because pictures were presented, it could be that participants in our studies could not imagine that the depicted robots really possess agency or experience. Thirdly, it might be that effects of agency and experience attribution are weaker than effects of anthropomorphic appearance, and therefore, the current research had not enough power to detect this influence. Therefore, future research should use a larger sample when investigating these effects.

Interestingly, in both studies, experience and agency seemed to have some positive relation with the damage to humans and their identity, meaning that robots who were ascribed more agency and/or experience were considered more damaging to human identities. Although here an inconsistent pattern was found, the correlations fit well with research introducing an “uncanny valley of the mind” [6]: Digital agents that were perceived as autonomous artificial agents also elicited an increased feeling of eeriness in the observer. Our results suggest that it might be possible to distinguish the impact of ‘mind attribution’ on perceived human identity threat into two separate factors: agency and experience. Thus, attributions of agency and experience can lead to feelings of threat under certain circumstances. It can be speculated whether the ability of a robot to experience human-like emotions, such as pain or happiness, would indeed make the line between robots and humans blurrier; yet, it would also make a robot more relatable and familiar. In contrast, a robot that is able to carry out the most complex human cognitive functions without experiencing or feeling anything is not relatable to us at all. Our findings provide a first hint that these factors have a differential impact on feelings of threat to our human identity. However, definitely, further research is needed to corroborate this.

The relationship between agency, experience, and damage to humans and their identity could point to other possible mediators instead of human–machine distinctiveness. For example, the extent to which humans can relate to a robot, or feel close to a robot could alleviate the potential damage to someone’s identity experienced as a result of low distinctiveness. Whether mind perception leads to a decrease in human–machine distinctiveness in moving digital avatars, or which other factors might lead to the negative feelings felt is up for future investigations.

One major limitation of the current research is that only pictures were presented, while films or interactions with robots (either real-life interactions or interactions using VR technologies, see [6]) would allow for more complexity in the design. It is therefore unclear whether seeing a robot in a picture leads to similar feelings and cognitions as seeing and interacting with an actual robot. It is likely that through real interaction, one’s preliminary cognitions and feelings are updated based on experience, thus leading to more reliable and clear perceptions. Therefore, real-life interacting with robots is a promising venue for future research to explore. By using pictures, we were able to test our assumptions in a highly controlled environment, and allowed for comparisons with earlier research [30]. Nevertheless, future studies should try to implement moving materials or real interactions to further validate what factors lead to a decrease in human–machine distinctiveness and negative feelings and perceived damage to ones’ identity.

Another limitation of the current study is that the design was only tested in a convenience sample of highly educated, young participants. In future research, it would be interesting to see whether the current results can also be found in a more diverse sample. As research has shown that elderly respond differently towards artificial agents [38], it might be and interesting target group: they might have less experience with seeing robots, and could therefore probably show stronger feelings of threat towards robots. In addition, it would be interesting to test how this relationship can be minimised. For example, research has shown that highlighting a shared goal can lead to less intergroup conflict [39]. Thus, making it clear that an interaction with a robot is necessary to reach a desired state might help to increase acceptance.

Our findings have important practical implications: Although a more human-like appearance and mind attribution can be beneficial for human–robot interactions [21], robots are better accepted when perceptual differences in appearance remain preserved. Thus, when designing robots for daily use, it is important users can clearly perceive non-human–robotic features in appearance and behaviour. This could for example be achieved by using a body that does not closely resemble human anatomy (such as for Cozmo or Roomba). Similarities in mind perception and experience are less threatening compared to similarities in appearance. The former does not affect the human–robot distinction as strongly as the latter. Nevertheless, robots that are perceived to have high agency and experience might evoke eeriness and unease. Therefore, while it is good that robots are able to demonstrate that they can make own decisions or have a basic experience of emotions and feelings (e.g., by verbally expressing understanding or explaining why they behave in a certain way), moderation is key if one wishes to avoid negative evaluations.