Research

Phenomenai treats AI self-reports as cheap hypothesis generators for expensive interpretability work. The project runs along four strands.

Repository

A shared registry tracking which concepts have been probed, steered, or ablated — in which models, with which methods, to what results. Modeled on the Cognitive Atlas.

Read more →

Interpretability research

A four-phase ladder validating candidate terms against activation space: probing, cross-architecture generalization, extension beyond emotion, and discovery.

Read more →

Methodology research

Structured phenomenological elicitation — can the model’s own vocabulary suggest interpretability targets that human-designed probes would miss? The pilot corpora.

Read more →

Policy bridges

Two downstream theories: functional rights built from validated internal states, and anticipatory legislation that activates when scientific thresholds are met.

Read more →
Open problems