Operationalizing Truth-Seeking in AI
Models today can do maths, code, and literature review, but can't do good theoretical alignment research, nor do good philosophy. This sequence tries to spell out what's missing, and develops experimental infrastructure for empirical research on this problem of truth-seeking in AI.
Based on an Anthropic Fellows project.