Intuitive examples of reward function learning? — AI Alignment Forum