How could we know that an AGI system will have good consequences? — AI Alignment Forum