QAPR 3: interpretability-guided training of neural nets — AI Alignment Forum