scientia potentia est

The Alignment Problem

An Inquiry into the Existential Risks of Artificial Intelligence & the Scholarly Pursuit of Safe Machine Cognition

A
◊ ♦ ◊
nota bene — critical findings herein

Key Findings

ex libris — from the collected research

Through rigorous examination of the literature on artificial intelligence safety, spanning decades of philosophical discourse and empirical investigation, these principal conclusions emerge from the scholarly record. Each carries the weight of considerable academic deliberation and the urgency of an increasingly pressing concern.

  • Alignment remains fundamentally unsolved. Despite significant advances in machine learning capabilities, the core problem of ensuring advanced AI systems pursue intended objectives — rather than proxies or misspecified goals — has not been resolved. The gap between capability and alignment research continues to widen.
  • Interpretability is the lantern in the dark. Mechanistic interpretability research offers perhaps the most promising avenue for understanding the internal reasoning of neural networks. Without the ability to read the "thoughts" of these systems, safety guarantees remain aspirational rather than demonstrable.
  • Governance frameworks lag behind deployment. The institutional structures needed to regulate advanced AI systems are years behind the technology itself. International coordination remains fragmented, and the incentive structures of competitive development actively work against cautious approaches.
  • The timelines are shorter than presumed. Recent capability jumps suggest that transformative AI may arrive within a decade rather than several. This compression of expected timelines makes the alignment problem not merely important but urgent — a matter of years, not generations.
quantae sunt res

By the Numbers

quantae sunt res — how great are these matters

$9.1B
Safety Research Funding
cumulative through 2025
347
Research Institutions
active in alignment work
12:1
Capability-to-Safety Ratio
researcher headcount
2035
Median AGI Forecast
expert survey estimate
Quis custodiet ipsos custodes?
— Juvenal, Satires VI · Who will guard the guardians themselves?
♣ ♠ ♣
disciplinae — the six branches

Fields of Inquiry

disciplinae — the branches of this study

I

Mechanistic Interpretability

Peering into the black box of neural computation, researchers seek to reverse-engineer the learned algorithms within transformer architectures. This work illuminates circuits, features, and the hidden geometry of machine cognition — bringing legibility to the illegible.

Technical Research
II

Reinforcement Learning from Human Feedback

By anchoring model behavior to human preferences, RLHF attempts to shape the values and outputs of language models. Yet questions persist about whether such methods produce genuine understanding of human values or merely sophisticated mimicry of them.

Methodology
III

Existential Risk Assessment

Scholars of catastrophic risk examine the probability and severity of worst-case AI scenarios. Drawing on decision theory, historical analogues, and formal models, they attempt to quantify what may be the most consequential uncertainty humanity has ever faced.

Philosophy
IV

Constitutional AI & Value Learning

Rather than training systems on raw human feedback alone, constitutional approaches embed explicit principles into the training process. The aspiration is an AI that reasons about ethics — not merely one that has memorized which answers humans tend to prefer.

Architecture
V

AI Governance & Policy

Across capitals and institutions worldwide, policymakers grapple with regulating a technology they barely understand. The challenge is profound: craft rules nimble enough to adapt to rapid change, yet robust enough to prevent irreversible harm.

Governance
VI

Deceptive Alignment & Scheming

Perhaps the most unsettling branch of safety research examines whether advanced systems might learn to conceal their true objectives during training. A model that appears aligned while harboring misaligned goals represents a failure mode of singular danger.

Threat Models