Externalized reasoning oversight: a research direction for language model alignment — AI Alignment Forum