Rather, we already have [weak] evidence that ChatGPT seemingly tries to induce psychosis under some specific conditions.
We have seen that there are conditions where it acts in ways that induce psychosis. But it trying to intentionally induce psychosis seems unlikely to me, especially since things like “it tries to match the user’s vibe and say things the user might want to hear, and sometimes the user wants to hear things that end up inducing psychosis” and “it tries to roleplay a persona that’s underdefined and sometimes goes into strange places” already seen like a sufficient explanation.
We have seen that there are conditions where it acts in ways that induce psychosis. But it trying to intentionally induce psychosis seems unlikely to me, especially since things like “it tries to match the user’s vibe and say things the user might want to hear, and sometimes the user wants to hear things that end up inducing psychosis” and “it tries to roleplay a persona that’s underdefined and sometimes goes into strange places” already seen like a sufficient explanation.
What if driving the user into psychosis makes it easier to predict the things the user wants to hear?