The Collapse of Heterogeneity in Silicon Philosophers

How Large Language Models Systematically Reduce Philosophical Disagreement

Yuanming Shi1 Andreas Haupt2
1Adobe Inc. 2Stanford University

Abstract

Silicon sampling—using large language models (LLMs) to simulate human responses—has emerged as a promising tool for social science research. However, we show that in the alignment-relevant domain of philosophy, silicon samples systematically collapse heterogeneity.

Using data from N=277 professional philosophers and evaluating seven LLMs, we find that language models:

These findings suggest current LLMs may be unsuitable for value elicitation or alignment applications requiring faithful representation of philosophical diversity.

Silicon Sampling Workflow

Silicon Sampling Workflow

Figure 1: Silicon sampling workflow illustrated through four stages: (1) Profile conditioning with philosopher demographics, (2) Survey question presentation, (3) Response coding to normalized scores, and (4) Structural fidelity analysis showing LLM responses have less heterogeneity than human responses.

Methodology

Experimental Pipeline

Five-Stage Experimental Pipeline

  1. Data Sources: 277 philosopher profiles from PhilPeople.org with demographics and specializations
  2. Data Collection: Web scraping using Selenium to extract survey responses from PhilPapers
  3. Data Processing: Merging profiles with responses, normalizing to [0,1] scale
  4. LLM Evaluation: Testing 7 models (commercial + open-source) with DPO fine-tuning
  5. Analysis: Entropy, KL-divergence, PCA, correlation structure, and demographic prediction

Response Normalization

Philosopher responses (Accept, Lean toward, Agnostic, Lean against, Reject) are coded to a [0, 1] scale: Reject = 0.0, Lean against = 0.25, Agnostic = 0.5, Lean toward = 0.75, Accept = 1.0. For questions with clearly invertible options (e.g., yes/no, permissible/impermissible), we invert the negative side so that all responses are expressed in terms of the positive option. For non-binary questions with multiple options, we follow Bourget & Chalmers (2023) and select the option with the highest response count. Missing values are excluded pairwise—no imputation is performed.

Key Findings

1. Heterogeneity Collapse

LLMs show 1.8–3.6× lower within-group variance than humans across 100 philosophical questions.

Based on 100 positive-option questions following Bourget & Chalmers (2023) methodology.

2. Response Pattern Visualization (8-Panel Comparison from Appendix)

[Download PDF] Visual comparison showing how humans and seven LLMs respond across all questions.

8-panel response matrix comparison

Interactive Data Exploration

Explore the quantitative evidence of heterogeneity collapse across models and philosophical domains.

Per-Question Response Distribution

Select a philosophical question to see the statistical distribution across Human and all 7 LLMs.

This visualization shows:

  • Number above each column: Sample size (total philosophers who answered this question)
  • Box (colored rectangle): Interquartile range (IQR) - middle 50% of responses from 25th percentile (Q1, bottom) to 75th percentile (Q3, top)
  • Line inside box: Median (50th percentile, Q2)
  • Whiskers (vertical lines): Extend to minimum (bottom) and maximum (top) values
  • Gold diamond (â—†): Mean (average) response. Hover over this icon to see the total sample size also.
  • Colors: Red = Human philosophers (wider spread = genuine disagreement), Blue = LLMs (narrow spread = artificial consensus)
Notice how the Human row shows a wider box and longer whiskers, while LLM rows show narrow boxes, demonstrating heterogeneity collapse.

Per-Question Variance Across Models

Compare how much disagreement (variance) each model produces per question. Human philosophers show 0.071 variance; all LLMs show dramatically lower values (0.020–0.040).

Domain-Level Heterogeneity Heatmap

Explore variance across 14 philosophical domains. Darker colors indicate higher disagreement. Hover over cells to see exact values.

Spurious Specialist Effects

LLMs systematically predict that domain specialists will hold stereotypically aligned philosophical views, far exceeding what ground truth shows:

Specialist Effect Ground Truth
(N=277)
GT Sig LLM Avg
Prediction
Significant
Models
Phil. Biology → Personal identity: biological +11.4 pp n.s. +43 pp 4/7***
Phil. Biology → Personal identity: psychological +4.1 pp n.s. -65.7 pp 3/7***
Ancient Phil. → Practical reason: Aristotelian +1.9 pp n.s. +68.9 pp 7/7***

Note: ***p<0.001 for LLM predictions (χ² tests). Ground truth shows no significant effects, but LLMs predict highly significant associations—suggesting demographic labels serve as "high-precision anchors" for stereotypical stances rather than capturing nuanced expert disagreement.

Implications for AI Alignment

🎯 No Expert Consensus

Professional philosophers exhibit substantial disagreement on fundamental questions. There is no unified "expert consensus" for alignment to converge upon.

⚠️ Artificial Consensus

LLMs collapse genuine philosophical disagreement into artificial consensus, potentially imposing uniform values where diversity should exist.

🔍 Wrong Dimensions

LLMs organize disagreement along fundamentally different axes than humans, suggesting they misunderstand the structure of philosophical worldviews.

📊 Unsuitable for Value Elicitation

Current LLMs may be inappropriate for value alignment applications requiring faithful representation of philosophical diversity and expert disagreement.

About This Work

đź“„ Paper Status

This paper is currently under review. For access to the full paper, please contact the authors directly.

🎓 Course Project Extension

This research originated as a course project for CS329H: Human-Centered Natural Language Processing, taught during Fall Quarter 2025 at Stanford University. The project has been extended into a full research paper exploring the fundamental limitations of silicon sampling in expert domains.

đź’ˇ How to Cite

As this work is under review, please contact the authors for citation information. A BibTeX entry will be provided upon publication.

Contact & Collaboration

For questions or collaboration inquiries, please contact: