Sycophancy to subterfuge: Investigating reward tampering in large language models — AI Alignment Forum