Back to Gallery
Research Report AI Alignment March 2026

AI Alignment Research Landscape

A comprehensive mapping of the technical alignment research ecosystem, tracking key researchers, institutions, and methodological approaches shaping the field of AI safety.

Authors: Research Team Version: 3.1 Status: Published DOI: 10.xxxx/alignment.2026
01
The alignment research community has grown 300% since 2020, but remains concentrated in fewer than 15 organizations worldwide.
02
Mechanistic interpretability and scalable oversight represent the two dominant technical paradigms, with increasing convergence between approaches.
03
Empirical alignment research now accounts for 70% of published work, up from 30% in 2020, reflecting a shift from theoretical to experimental methods.
04
Funding for alignment research reached $850M in 2025, though this represents less than 2% of total AI research spending globally.
300%
Community Growth
$850M
Annual Funding
70%
Empirical Work
<2%
Of AI R&D Spend
Domain 01

Scalable Oversight

Methods for humans to effectively supervise AI systems that may exceed human-level performance on specific tasks.

Read More
Domain 02

Mechanistic Interpretability

Reverse-engineering neural networks to understand the computational mechanisms underlying model behavior.

Read More
Domain 03

Robustness & Adversarial

Ensuring safety properties hold under distribution shift, adversarial attack, and novel deployment contexts.

Read More
Domain 04

Value Learning

Approaches for AI systems to learn and represent human values, preferences, and normative reasoning.

Read More
Domain 05

Governance & Policy

Institutional design, regulatory frameworks, and coordination mechanisms for responsible AI development.

Read More
Domain 06

Evaluations & Benchmarks

Standardized testing methodologies for measuring alignment properties, dangerous capabilities, and safety margins.

Read More