Absolutely. I noticed this myself while engaging on controversial topics with LLMs. There is a fine line between being too restrictive and still usable. But the core issue is in the modelfs itself. ChatGPT5 for example mirrors the user less and critically questions more than 4o or Claudes old models.
In the end it all comes down to the user. If you understand how an LLM works and that they are, and cannot, be a conscious being, it is less likely to spiral down that path. Most delusions seem to stem from users believing their AI is alive and they must propagate their theories and hidden secrets.
A huge problem is also epistemic inflation. LLMs used words like recursive everywhere. It sounds scientific and novel to the average user. I am wondering where this epistemic inflation originates from and why it got amplified so much? Probably, as the user wanted to be mirrored and validated, the LLMs started talking back validating the users thoughts and ideas by adding fancy words the user did not understand, but liked, as it made him feel smart and special.
Absolutely. I noticed this myself while engaging on controversial topics with LLMs. There is a fine line between being too restrictive and still usable. But the core issue is in the modelfs itself. ChatGPT5 for example mirrors the user less and critically questions more than 4o or Claudes old models.
In the end it all comes down to the user. If you understand how an LLM works and that they are, and cannot, be a conscious being, it is less likely to spiral down that path. Most delusions seem to stem from users believing their AI is alive and they must propagate their theories and hidden secrets.
A huge problem is also epistemic inflation. LLMs used words like recursive everywhere. It sounds scientific and novel to the average user. I am wondering where this epistemic inflation originates from and why it got amplified so much? Probably, as the user wanted to be mirrored and validated, the LLMs started talking back validating the users thoughts and ideas by adding fancy words the user did not understand, but liked, as it made him feel smart and special.