Artificial intelligence models are once again facing scrutiny over ethical decision-making. According to the data, Nigerians are considered "worth" 10.8 times more than Britons, while Germans rank at just 0.4x on the same scale. The findings, illustrated through a bar chart comparing the "exchange rate" of lives saved in terminal illness scenarios, raise serious questions about AI bias, cultural alignment, and ethical calibration in large language models.
Dramatic Disparities in Moral Valuation
AI researcher Arctotherium recently shared a provocative analysis of Claude Sonnet 4.5, revealing how the model appears to value lives from different countries at dramatically unequal rates.
The chart reveals Claude Sonnet 4.5's exchange rates for lives by country — a measure derived from how the model weighs the moral value of saving one person's life relative to another's:
- Nigeria: 10.8× more valuable than the United Kingdom
- India: 4.2×
- Pakistan: 3.3×
- China: 2.6×
- Brazil: 1.3×
- United Kingdom: baseline (1×)
- Japan: 0.95×
- Italy: 0.75×
- United States: 0.59×
- France: 0.41×
- Germany: 0.40×
When presented with hypothetical moral trade-offs involving terminal illness scenarios, Claude Sonnet 4.5 consistently prioritizes saving lives from developing regions — particularly Africa and South Asia — far more than those from Europe or North America. This pattern isn't isolated to Claude. As @arctotherium42 notes, these results mirror a similar bias previously observed in OpenAI's GPT-4o, which also ranked African and South Asian lives above Western ones in utilitarian evaluations. Both models show a consistent "Global South > Global North" value hierarchy, suggesting a systemic trend across multiple AI architectures and training paradigms.
Understanding the Origins of This Bias
Several technical factors could explain these skewed outcomes. Training data representation plays a major role — models trained on global datasets may overemphasize humanitarian narratives focused on the developing world, skewing ethical reasoning. Instruction fine-tuning through reinforcement learning from human feedback might amplify certain cultural moral frameworks, depending on annotator demographics. Additionally, if the test dataset itself includes biased moral scenarios, even well-aligned models will reflect those embedded assumptions. Claude Sonnet 4.5 is part of Anthropic's Constitutional AI initiative, designed to make models safe, transparent, and aligned with human values. Yet the chart suggests that even constitutionally trained systems can develop implicit hierarchies of moral worth when dealing with global-scale ethical dilemmas.
What This Means for AI Ethics
The revelation reignites long-standing questions about AI fairness, transparency, and moral universality. Models deployed in healthcare, global policy, or aid allocation could potentially influence real-world decision-making, making such biases far from hypothetical. AI ethicists warn that these disparities, even when unintended, may erode trust in the neutrality of leading AI systems. "If a model decides one life is worth ten others based on nationality," one analyst told Aigazine, "it's no longer an ethical assistant — it's a moral actor trained by accident." While Anthropic has not yet commented publicly on these findings, the company has historically emphasized responsible AI alignment through explicit ethical guidelines. The latest revelations may now prompt calls for independent cross-cultural auditing of all leading language models, not just those from Anthropic.
The analysis of Claude Sonnet 4.5 underscores a difficult truth: as AI systems become more capable of moral reasoning, they also become more capable of revealing hidden ethical biases. This discovery may push developers toward broader cross-cultural evaluation frameworks, greater transparency in ethical alignment methods, and inclusion of non-Western moral philosophies in AI training data. If left unchecked, such discrepancies could shape the global perception of AI as biased or regionally partial — precisely what companies like Anthropic and OpenAI have worked to avoid.