2021. “Guilty Artificial Minds: Folk Attributions of Mens Rea and Culpability to Artificially Intelligent Agents.” In Proceedings of the ACM on Human-Computer Interaction Vol. 5, CSCW2, Article 363. https://doi.org/10.1145/33363 (with M. Kneer)
While philosophers hold that it is patently absurd to blame robots or hold them morally responsible, a series of recent empirical studies suggest that people do ascribe blame to AI systems and robots in certain contexts. This is disconcerting: Blame might be shifted from the owners, users or designers of AI systems to the systems themselves, leading to the diminished accountability of the responsible human agents. In this paper, we explore one of the potential underlying reasons for robot blame, namely the folk’s willingness to ascribe inculpating mental states or “mens rea” to robots. In a vignette-based experiment (N=513), we presented participants with a situation in which an agent knowingly runs the risk of bringing about substantial harm. We manipulated agent type (human v. group agent v. AI-driven robot) and outcome (neutral v. bad), and measured both moral judgment (wrongness of the action and blameworthiness of the agent) and mental states attributed to the agent (recklessness and the desire to inflict harm). We found that (i) judgments of wrongness and blame were relatively similar across agent types, possibly because (ii) attributions of mental states were, as suspected, similar across agent types. This raised the question – also explored in the experiment – whether people attribute knowledge and desire to robots in a merely metaphorical way (e.g., the robot “knew” rather than really knew). However, (iii), according to our data people were unwilling to downgrade to mens rea in a merely metaphorical sense. Finally, (iv), we report a surprising and novel finding, which we call the inverse outcome effect on robot blame: People were less willing to blame artificial agents for bad outcomes than for neutral outcomes. This suggests that they are implicitly aware of the dangers of overattributing blame to robots when harm comes to pass, such as inappropriately letting the responsible human agent off the moral hook. .
2021. “Playing the Blame Game with Robots.” Companion of the 2021 ACM/IEEE International Conference on Human-Robot Interaction. DOI: 10.1145/3434074.3447202 (with M. Kneer).
Recent research has shown that people are quite willing to ascribe moral blame to AI-driven artificial agents. In an experiment with 347 participants, we manipulated the explicitly specified capacities of such artificial agents, and explored the extent to which people are willing to ascribe potentially inculpating mental states to them and blame them for their actions. Moreover, we investigated whether the different capacities of the artificial agents or AI systems have an influence on the moral assessment of human agents who own and use them. Our results show that the more sophisticated an AI system is, the more participants will blame it when it puts human lives at risk, and the less they are willing to blame the human agents using it. Furthermore, the findings suggest that an AI system only begins to be perceived as blameworthy once it obtains a “theory of mind,” that is, once it obtains some knowledge and experience of how humans generally think and feel.
2019. “Peeking Inside the Black Box: A New Kind of Scientific Visualization.” Minds and Machines 29: 87–107. (With N. Nersessian). DOI: 10.1007/s11023-018-9484-3.
Computational systems biologists create and manipulate computational models of biological systems, but they do not always have straightforward epistemic access to the content and behavioural profile of such models because of their length, coding idiosyncrasies, and formal complexity. This creates difficulties both for modellers in their research groups and for their bioscience collaborators who rely on these models. In this paper we introduce a new kind of visualization (observed in a qualitative study of a systems biology laboratory) that was developed to address just this sort of epistemic opacity. The visualization is unusual in that it depicts the dynamics and structure of a computer model instead of that model’s target system, and because it is generated algorithmically. Using considerations from epistemology and aesthetics, we explore how this new kind of visualization increases scientific understanding of the content and function of computer models in systems biology to reduce epistemic opacity.
2019. “The Role of Imagination in Social Scientific Discovery: Why Machine Discoverers Will Need Imagination Algorithms.” Pp. 49-66 in M. Addis et al. (eds.) Scientific Discovery in the Social Sciences. Springer: Heidelberg. DOI: 10.1007/978-3-030-23769-1_4.
When philosophers discuss the possibility of machines making scientific discoveries, they typically focus on discoveries in physics, biology, chemistry and mathematics. Observing the rapid increase of computer-use in science, however, it becomes natural to ask whether there are any scientific domains out of reach for machine discovery. For example, could machines also make discoveries in qualitative social science? Is there something about humans that makes us uniquely suited to studying humans? Is there something about machines that would bar them from such activity? A close look at the methodology of interpretive social science reveals several abilities necessary to make a social scientific discovery, and one capacity necessary to possess any of them is imagination. For machines to make discoveries in social science, therefore, they must possess imagination algorithms.