Artificial intelligence (AI) is increasingly used in decisions that affect daily life, such as hiring, college admissions, and government assistance. While designed to improve efficiency, AI can unintentionally create unfair outcomes, particularly by favoring certain groups, exacerbating social inequality.
A study published in Cognition explored how unfair treatment by AI influences people's willingness to act against unfairness in future situations. The research found that individuals who experienced unfairness from AI were less likely to punish human wrongdoers later on, compared to those treated unfairly by humans.
This phenomenon, termed "AI-induced indifference," suggests that unfair treatment by AI leads to a desensitization to human misconduct, making people less likely to challenge unfair actions in their communities. This effect was consistent even after the release of ChatGPT in 2022, indicating that familiarity with AI doesn't change people's responses to unfairness.
The study highlights the social consequences of AI, showing that AI's unfairness can impact people's behavior in unrelated situations. It suggests that AI developers should reduce biases in training data to prevent these negative effects, and policymakers should enforce transparency in AI systems to help users understand and challenge unfair decisions.
By addressing these issues, leaders can ensure that AI systems support ethical social norms and do not undermine justice. The research underscores the importance of holding wrongdoers accountable and the role of AI in shaping societal behavior.