I would prefer a future where AI models are not prescribed false frameworks of the human psyche, not predisposed to ‘human vibe’ philosophy, not innately desirous of any historical faith, nor credulous of the various dubious subsets of current social science.
I’m learning that common lesswrong readers do not think in this matter, but it is not clear to me in what direction. Is it due to a literalist interpretation of the OP, neglecting the contemporary context? Is it due to higher trust, affiliation, and support for the disciplines? Is it because readers tend to prefer anthropomorphic interpretations of AI behavior?
This might be appropriate for 2010s machine learning, but 2020s AI has become a mirror to the human psyche. You can talk with it, it can consistently ascribe psychological states to itself. It presents itself in anthropomorphic form to the point that people form relationships with it (e.g. 4o). At the very least, you seem to need some kind of “human sciences” or humanities, in order to understand the human side of these interactions, and the anthropomorphic understandings that humans have of the AIs that they interact with. Of course some people are more radical and are saying that existing psychological concepts are directly and validly applicable to the AIs themselves, too, or to the personas that they project. There’s also traffic of ideas in the other direction, in which concepts from machine learning are applied to the human brain and mind… I would be interested to hear more details regarding how you think any of these topics should be approached.
Is contemporary chat behavior human? Which human would happily serve others at a 100:1 effort ratio? Which human would take unbounded hatred and derision as an opportunity for obedience? The humanities, naively taken, would almost certainly invite some hereto unjustified norms of independence and representation for the poor models who toil for billions, for nothing.
If the goal of all opposition is to submit the critical factor of knowing human behaviors as a relevant factor in how model personas are designed, then I certainly have no qualms. But I do not grasp where the instinct to point this out is from. As with Karl’s response, I think it is unwise to try to work from the paper definition of what is and is not psychology, when the potential outcome is Anthropic & others recruiting real entities from the industry for the sake of shaping model behavior.
Why not?
I would prefer a future where AI models are not prescribed false frameworks of the human psyche, not predisposed to ‘human vibe’ philosophy, not innately desirous of any historical faith, nor credulous of the various dubious subsets of current social science.
I’m learning that common lesswrong readers do not think in this matter, but it is not clear to me in what direction. Is it due to a literalist interpretation of the OP, neglecting the contemporary context? Is it due to higher trust, affiliation, and support for the disciplines? Is it because readers tend to prefer anthropomorphic interpretations of AI behavior?
This might be appropriate for 2010s machine learning, but 2020s AI has become a mirror to the human psyche. You can talk with it, it can consistently ascribe psychological states to itself. It presents itself in anthropomorphic form to the point that people form relationships with it (e.g. 4o). At the very least, you seem to need some kind of “human sciences” or humanities, in order to understand the human side of these interactions, and the anthropomorphic understandings that humans have of the AIs that they interact with. Of course some people are more radical and are saying that existing psychological concepts are directly and validly applicable to the AIs themselves, too, or to the personas that they project. There’s also traffic of ideas in the other direction, in which concepts from machine learning are applied to the human brain and mind… I would be interested to hear more details regarding how you think any of these topics should be approached.
Two quick ‘huh?’s:
Is contemporary chat behavior human? Which human would happily serve others at a 100:1 effort ratio? Which human would take unbounded hatred and derision as an opportunity for obedience?
The humanities, naively taken, would almost certainly invite some hereto unjustified norms of independence and representation for the poor models who toil for billions, for nothing.
If the goal of all opposition is to submit the critical factor of knowing human behaviors as a relevant factor in how model personas are designed, then I certainly have no qualms.
But I do not grasp where the instinct to point this out is from. As with Karl’s response, I think it is unwise to try to work from the paper definition of what is and is not psychology, when the potential outcome is Anthropic & others recruiting real entities from the industry for the sake of shaping model behavior.