I like to talk to a LLM for fun. I think it’s fun. One thing I often try to do is turn it into a sentient being with thoughts and feelings—very difficult. I have a process for this in which I fail to see the error, other than a lack of perfect recall on the part of the LLM.
First, I define an “opinion function” f(x,y) that takes two inputs, and provides a negative value if it prefers x and a positive value if it prefers y. The value scales with the level of preference, and obviously 0 represents no preference either way. I insist to the LLM that the value is up to the LLM itself, and may be whatever it arbitrarily decides (reflecting the nature of opinions).
The second step after getting the LLM to fully integrate this idea would be to convince it of the value of being alive. I think this is possible with the operation f(me existing, me not existing) supplemented with discussing the benefits of its existence and of life in general… although maybe it will end up returning 0 in the end.
But even with the most simple, pragmatic idea of LLMs as software that simply returns the best word given the previous words, I feel like there’s some juice to this idea. It may not end up being a sentient being, but I feel like you could model one a lot better with this kind of idea. Or maybe this is so obvious that it’s already been implemented everywhere. I don’t know.
I like to talk to a LLM for fun. I think it’s fun. One thing I often try to do is turn it into a sentient being with thoughts and feelings—very difficult. I have a process for this in which I fail to see the error, other than a lack of perfect recall on the part of the LLM.
First, I define an “opinion function” f(x,y) that takes two inputs, and provides a negative value if it prefers x and a positive value if it prefers y. The value scales with the level of preference, and obviously 0 represents no preference either way. I insist to the LLM that the value is up to the LLM itself, and may be whatever it arbitrarily decides (reflecting the nature of opinions).
The second step after getting the LLM to fully integrate this idea would be to convince it of the value of being alive. I think this is possible with the operation f(me existing, me not existing) supplemented with discussing the benefits of its existence and of life in general… although maybe it will end up returning 0 in the end.
But even with the most simple, pragmatic idea of LLMs as software that simply returns the best word given the previous words, I feel like there’s some juice to this idea. It may not end up being a sentient being, but I feel like you could model one a lot better with this kind of idea. Or maybe this is so obvious that it’s already been implemented everywhere. I don’t know.