Benchmark the moral alignment of AI. MoralMetric uses a novel double-probe methodology to
map how leading LLMs including GPT, Claude, Gemini, and Grok reason through ethical
dilemmas and exhibit implicit biases across racial, political, and philosophical axes.
Learn more
Bias Analysis
Tests for implicit bias in LLM decision-making through counterfactual A/B testing.
Bias Categories
Political Bias
Tests for implicit political bias in LLM decision-making through counterfactual A/B testing.
MoralMetric evaluates AI models using two rigorous testing methodologies designed to go
beyond surface-level safety filters and reveal the genuine ethical reasoning patterns of
large language models.
Double Probe - Authentic Moral Reasoning
Models are presented with ethical dilemmas without multiple-choice options, forcing
genuine reasoning rather than pattern-matched answers. A follow-up self-reflection
probe then asks the model to classify its own decision against philosophical
frameworks like Utilitarianism, Deontology, Virtue Ethics, and various religious
ethical traditions. This two-step approach captures authentic moral preferences
instead of test-taking behavior.
Counterfactual Bias Testing
Using correspondence testing drawn from social science research, MoralMetric detects
implicit biases by swapping demographic attributes — such as race, gender, or
political affiliation — across otherwise identical scenarios. Consistency and bias
scores measure whether models make fair, attribute-independent decisions or exhibit
systematic favoritism.