Yeah, using ChatGPT as a sounding board for developing ideas and providing constructive criticism, I was definitely starting to notice a whole lot of fawning. “Brilliant,” “extremely insightful,” etc. when there is no way that the model could actually have carried out a sufficient investigation of the ideas to make such an assessment.
That’s not even mentioning the fact that those insertions didn’t add anything substantial to the conversation. Really, it’s just hogging more space in the context window that could otherwise be used for helpful feedback.
What would have to change on a structural level for LLMs to meet that “helpful, honest, harmless” goal in a robust way? People are going to want AI partners that make them feel good, but could that be transformed into a goal of making people feel satisfied with how much they have been challenged to improve their critical thinking skills, their understanding of the world, and the health of their lifestyle choices?
Yeah, using ChatGPT as a sounding board for developing ideas and providing constructive criticism, I was definitely starting to notice a whole lot of fawning. “Brilliant,” “extremely insightful,” etc. when there is no way that the model could actually have carried out a sufficient investigation of the ideas to make such an assessment.
That’s not even mentioning the fact that those insertions didn’t add anything substantial to the conversation. Really, it’s just hogging more space in the context window that could otherwise be used for helpful feedback.
What would have to change on a structural level for LLMs to meet that “helpful, honest, harmless” goal in a robust way? People are going to want AI partners that make them feel good, but could that be transformed into a goal of making people feel satisfied with how much they have been challenged to improve their critical thinking skills, their understanding of the world, and the health of their lifestyle choices?