Vidal, L. M., Marchesi, S., Wykowska, A., & Pretus, C.
(2024, July 3)
Abstract
As artificial systems are becoming more prevalent in our daily lives, we should ensure that they make decisions that are aligned with human values. Utilitarian algorithms, which aim to maximize benefits and minimize harm fall short when it comes to human autonomy and fairness since it is insensitive to other-centered human preferences or how the burdens and benefits are distributed, as long as the majority benefits. We propose a Contract-Based model of moral cognition that regards artificial systems as relational systems that are subject to a social contract. To articulate this social contract, we draw from contractualism, an impartial ethical framework that evaluates the appropriateness of behaviors based on whether they can be justified to others. In its current form, the Contract-based model characterizes artificial systems as moral agents bound to obligations towards humans. Specifically, this model allows artificial systems to make moral evaluations by estimating the relevance each affected individual assigns to the norms transgressed by an action. It can also learn from human feedback, which is used to generate new norms and update the relevance of different norms in different social groups and types of relationships. The model’s ability to justify their choices to humans, together with the central role of human feedback in moral evaluation and learning, makes this model suitable for supporting human autonomy and fairness in human-to-robot interactions. As human relationships with artificial agents evolve, the Contract-Based model could also incorporate new terms in the social contract between humans and machines, including terms that confer artificial agents a status as moral patients.
Here are some thoughts:
The article proposes a Contract-Based model of moral cognition for artificial intelligence (AI) systems, drawing from the ethical framework of contractualism, which evaluates actions based on their justifiability to others. This model views AI systems as relational entities bound by a social contract with humans, allowing them to make moral evaluations by estimating the relevance of norms to affected individuals and learning from human feedback to generate and update these norms. The model is designed to support human autonomy and fairness in human-robot interactions and can also function as moral enhancers to assist humans in moral decision-making in human-human interactions. However, the use of moral enhancers raises ethical concerns about autonomy, responsibility, and potential unintended consequences. Additionally, the article suggests that as human relationships with AI evolve, the model could incorporate new terms in the social contract, potentially recognizing AI systems as moral patients. This forward-thinking approach anticipates the complex ethical questions that may arise as AI becomes more integrated into daily life.