Rohin Shah comments on [AN #141]: The case for practicing alignment work on GPT-3 and other large models