How will we know when a social robot (or any other kind of artificial intelligence) is a genuine social agent? That is, how will we know when it is conscious, feels things, and understands what it hears or says? This is the philosophical problem of other minds – the problem of how we can know that anyone else has a mind – applied to human creations (Harnad, Reference Harnad1991).
The target article raises a new problem of other minds. Clark and Fischer suggest that rather than viewing social robots as genuine social agents, people instead view them as depictions of social agents. Under this depiction account, people engage in a kind of pretense when interacting with social robots (also see Rueben, Rothberg, & Matarić, Reference Rueben, Rothberg, Matarić, Nørskov, Seibt and Quick2020). This account could be right, but we suggest it remains possible that people might instead view social robots as genuine social agents. Testing between these accounts introduces a new second-order problem of other minds: How can we tell if other people think they are dealing with a genuine social agent or a mere depiction of one?
The second-order problem of other minds may be difficult to resolve. When dealing with depictions, people normally hold back – their actions fall short from what they would do with the real thing. For example, children pretending to eat plastic fruit refrain from actually biting it (e.g., Leslie & Happé, Reference Leslie and Happé1989; Lillard, Reference Lillard1993) and filmgoers do not attempt to intervene in movie events. Do people also hold back with social robots? It might be hard to tell. Although people do not treat social robots exactly the way they treat their peers, this isn't saying much. There are many different kinds of agents and people see them as varying in their mental capacities (Gray, Gray, & Wegner, Reference Gray, Gray and Wegner2007; Weisman, Dweck, & Markman, Reference Weisman, Dweck and Markman2017) and moral standing (Crimston, Hornsey, Bain, & Bastian, Reference Crimston, Hornsey, Bain and Bastian2018; Goodwin, Reference Goodwin2015). So, while it might be obvious when people hold back when dealing with many kinds of depictions (e.g., plastic fruit), this will be less obvious with social robots. What looks like holding back could turn out to reflect beliefs that social robots have limited capacities and moral standing.
To illustrate these points, let's consider the evidence offered as support for the idea that people view social robots as depictions. One line of evidence is that rather than seeing social robots the way they see their fellow humans, people see social robots as a kind of property. They affirm social robots can be sold, and if a social robot dented someone's car, the owner of the car would seek compensation from the robot's owner rather than from the robot itself. Treating social robots like property might follow from the belief that they are depictions rather than genuine social agents. But it is also reminiscent of how people treat real agents viewed as having limited moral standing or limited mental abilities. For example, pets and other animals are bought and sold and their owners are held liable when they cause harm (e.g., Bowman-Smith, Goulding, & Friedman, Reference Bowman-Smith, Goulding and Friedman2018; Nadler & McDonnell, Reference Nadler and McDonnell2011). Similar points may apply to how enslaved people were viewed and treated in the past. They too were treated as chattel, and when they caused harm, their enslavers were held liable in some legal systems (Oppenheim, Reference Oppenheim1940). But it is unlikely that people view pets as depictions, or that the enslaved were viewed this way either. So rather than viewing robots as mere depictions, people might instead see them as genuine agents with limited moral worth and limited mental capacities.
The target article also notes that people differ from one another in how they interact with social robots. Although some people converse with social robots, others refrain from doing so – these people do not respond to greetings from social robots and if they address robots at all, it is only with blunt questions and brusque orders; perhaps these people are unwilling to play along with the pretense that these depictions are social agents. But this again is ambiguous. We might also expect differences between people if some believed that social robots are real agents, while others did not. Here again, people's treatment of animals raises questions. As with social robots, people vary in how they address their pets and some people's communication with their pet dogs is apparently limited to commands and threats (e.g., Carlisle-Frank, Frank, & Nielsen, Reference Carlisle-Frank, Frank and Nielsen2004; Mitchell, Reference Mitchell2004). Some talk to pets could have a pretend element – people sometimes ask dogs questions but then also answer the questions (Mitchell, Reference Mitchell2004). But it seems unlikely (at least to us) that people view pets as depictions, or that most variation in talk to pets come down to differences in owners' proclivity to pretend.
Although the second-order problem of other minds be difficult to resolve, the difficulty may be asymmetric. While it might be difficult to confirm that social robots are viewed as depictions, it may be easier to confirm when they are viewed as genuine agents. Consider the issue of whether people show moral concern for robots (for a recent review see Harris & Anthis, Reference Harris and Anthis2021). When people express concerns for the welfare of robots and advocate for robots to have rights, this might suggest they view social robots as genuine agents – at least if these expressions of concern focus on robots themselves and not on side-concerns, such as concerns about property damage, or concerns that mistreating robots will encourage mistreatment of humans (e.g., Levy, Reference Levy2009). By contrast, the absence of concern would not necessarily show that people view robots as depictions. It could instead stem from viewing robots as genuine agents with limited capacities or moral worth.
How will we know when a social robot (or any other kind of artificial intelligence) is a genuine social agent? That is, how will we know when it is conscious, feels things, and understands what it hears or says? This is the philosophical problem of other minds – the problem of how we can know that anyone else has a mind – applied to human creations (Harnad, Reference Harnad1991).
The target article raises a new problem of other minds. Clark and Fischer suggest that rather than viewing social robots as genuine social agents, people instead view them as depictions of social agents. Under this depiction account, people engage in a kind of pretense when interacting with social robots (also see Rueben, Rothberg, & Matarić, Reference Rueben, Rothberg, Matarić, Nørskov, Seibt and Quick2020). This account could be right, but we suggest it remains possible that people might instead view social robots as genuine social agents. Testing between these accounts introduces a new second-order problem of other minds: How can we tell if other people think they are dealing with a genuine social agent or a mere depiction of one?
The second-order problem of other minds may be difficult to resolve. When dealing with depictions, people normally hold back – their actions fall short from what they would do with the real thing. For example, children pretending to eat plastic fruit refrain from actually biting it (e.g., Leslie & Happé, Reference Leslie and Happé1989; Lillard, Reference Lillard1993) and filmgoers do not attempt to intervene in movie events. Do people also hold back with social robots? It might be hard to tell. Although people do not treat social robots exactly the way they treat their peers, this isn't saying much. There are many different kinds of agents and people see them as varying in their mental capacities (Gray, Gray, & Wegner, Reference Gray, Gray and Wegner2007; Weisman, Dweck, & Markman, Reference Weisman, Dweck and Markman2017) and moral standing (Crimston, Hornsey, Bain, & Bastian, Reference Crimston, Hornsey, Bain and Bastian2018; Goodwin, Reference Goodwin2015). So, while it might be obvious when people hold back when dealing with many kinds of depictions (e.g., plastic fruit), this will be less obvious with social robots. What looks like holding back could turn out to reflect beliefs that social robots have limited capacities and moral standing.
To illustrate these points, let's consider the evidence offered as support for the idea that people view social robots as depictions. One line of evidence is that rather than seeing social robots the way they see their fellow humans, people see social robots as a kind of property. They affirm social robots can be sold, and if a social robot dented someone's car, the owner of the car would seek compensation from the robot's owner rather than from the robot itself. Treating social robots like property might follow from the belief that they are depictions rather than genuine social agents. But it is also reminiscent of how people treat real agents viewed as having limited moral standing or limited mental abilities. For example, pets and other animals are bought and sold and their owners are held liable when they cause harm (e.g., Bowman-Smith, Goulding, & Friedman, Reference Bowman-Smith, Goulding and Friedman2018; Nadler & McDonnell, Reference Nadler and McDonnell2011). Similar points may apply to how enslaved people were viewed and treated in the past. They too were treated as chattel, and when they caused harm, their enslavers were held liable in some legal systems (Oppenheim, Reference Oppenheim1940). But it is unlikely that people view pets as depictions, or that the enslaved were viewed this way either. So rather than viewing robots as mere depictions, people might instead see them as genuine agents with limited moral worth and limited mental capacities.
The target article also notes that people differ from one another in how they interact with social robots. Although some people converse with social robots, others refrain from doing so – these people do not respond to greetings from social robots and if they address robots at all, it is only with blunt questions and brusque orders; perhaps these people are unwilling to play along with the pretense that these depictions are social agents. But this again is ambiguous. We might also expect differences between people if some believed that social robots are real agents, while others did not. Here again, people's treatment of animals raises questions. As with social robots, people vary in how they address their pets and some people's communication with their pet dogs is apparently limited to commands and threats (e.g., Carlisle-Frank, Frank, & Nielsen, Reference Carlisle-Frank, Frank and Nielsen2004; Mitchell, Reference Mitchell2004). Some talk to pets could have a pretend element – people sometimes ask dogs questions but then also answer the questions (Mitchell, Reference Mitchell2004). But it seems unlikely (at least to us) that people view pets as depictions, or that most variation in talk to pets come down to differences in owners' proclivity to pretend.
Although the second-order problem of other minds be difficult to resolve, the difficulty may be asymmetric. While it might be difficult to confirm that social robots are viewed as depictions, it may be easier to confirm when they are viewed as genuine agents. Consider the issue of whether people show moral concern for robots (for a recent review see Harris & Anthis, Reference Harris and Anthis2021). When people express concerns for the welfare of robots and advocate for robots to have rights, this might suggest they view social robots as genuine agents – at least if these expressions of concern focus on robots themselves and not on side-concerns, such as concerns about property damage, or concerns that mistreating robots will encourage mistreatment of humans (e.g., Levy, Reference Levy2009). By contrast, the absence of concern would not necessarily show that people view robots as depictions. It could instead stem from viewing robots as genuine agents with limited capacities or moral worth.
Financial support
This work was supported by a Discovery Grant from the Natural Sciences and Engineering Research Council of Canada awarded to OF.
Competing interest
None.