Barriers to Mechanistic Interpretability for AGI Safety — AI Alignment Forum