LLMs now can also self-play in adversarial word games and it increases their performance https://arxiv.org/abs/2404.10642
avturchin
Roman Mazurenko is dead again. First resurrected person, Roman lived as a chatbot (2016-2024) created based on his conversations with his fiancé. You might even be able download him as an app.
But not any more. His fiancé married again and her startup http://Replika.ai pivoted from resurrection help to AI-girlfriends and psychological consulting.
It looks like they quietly removed Roman Mazurenko app from public access. It is especially pity that his digital twin lived less than his biological original, who died at 32. Especially now when we have much more powerful instruments for creating semi-uploads based on LLMs with large prompt window.
The “repeating” will not be repeating from internal point of view of a person, as he has completely erased the memories of the first attempt. So he will do it as if it is first time.
Yes, here we can define magic as “ability to manipulate one’s reference class”. And special minds may be much more adapted to it.
Presumably in deep meditation people become disconnected from reality.
Yes it is easy to forget something if it does not become a part of your personality. So a new bad thing is easier to forget.
The number of poor people is much larger than billionaire. So in most cases you will fail to wake up as a billionaire. But sometimes it will work and it is similar to law of attraction. But formulation via forgetting is more beautiful. You forget that you are poor.
UPDATE; actually, the difference with the law of attraction is that after applying the law of attraction, a person still remember that he has used the law. In magic by forgetting the fact of its use must be completely forgotten.
I can forget one particular thing, but preserve most of my selfidentification information
In the case of broken robot we need two conditions for magic by forgetting:
there are 100 robots and only one is broken and all of them are type-copies of each other.
each robot enters into blank state of mind naturally in some moment, like sleep or reboot.
In that case, after robot enters the blank state of mind it has equal chances to be any of robots and this dilutes its chances to have the damaged body after awakening.
For you toy example—at first approximation, any of which can recognize itself as avturchin (self-recognition identity criteria).
Magic by forgetting
I am building my sideload via recursively correcting of 1-million-tokens prompt for large LLM. The prompt consists of 500 rules which describe my personality, similar to personal constitution, and of some texts, like diaries, abstracts, poetry, stream of thoughts etc. Works on Google Gemini 1M through Google AI studio, and the shorter version works great on Opus. The system also includes a universal “loader prompt” which tries to increase the intelligence of the model and describes how the chatbot should work.
I found that sideloading allows very quick iterations in the sideload’s improvements and the improvements are two-fold: of the loader itself and improvements of the knowledge and style of the sideload.
I find that my sideload is surprisingly good for a project which took around 1 month of work. 1 of the 5 answers is exactly like mine from a factual and style point of view.
I am open-sourcing my sideload, anyone can run it https://github.com/avturchin/minduploading/tree/main
I can help anyone interested to build his-her own sideload.
Example of work of the chatbot, no cherry picking:
Q:(now speak in english) what will be your next post in Lesswrong?
A: /The user is asking about my next LessWrong post. I will answer as if I am Alexey Turchin, and describe my plans for future posts, as well as my current location and the people around me./
My next LessWrong post? [Alexey looks thoughtfully out the window]
(I have several drafts)
[Alexey stretches and sits up in bed]
{The room is small, with a balcony. On the table is a laptop, some papers, and reading glasses. Outside the window, it is still gray and dreary.}
I’m thinking about writing about the nature of qualia and the possibility of creating qualia of non-existent things. [Alexey gets up, walks to the window, and looks out at the street]
{Below, a woman with a child is walking, a street sweeper is lazily sweeping leaves. Somewhere in the distance, the sound of a siren can be heard.}
Or maybe about the problem of copies and identity. (it’s a complex topic)
[CONFIDENCE] 8
Interestingly, could similar behavior be used to explain some cases of human schizophrenia when word salad is produced.
Most typical skin cancer is basiloma—and it is rather benign—no metastases and can be removed without hospitalization. Many people get it.
combine more approaches!
I try new models with ’wild sex between two animals’
Older models produced decent porn on that.Later models refuse to replay as triggers were activated.
And last models give me lectures about sexual relations between animals in the wild.
can you access it via vpn?
I wrote similar idea here: https://www.lesswrong.com/posts/NWQ5JbrniosCHDbvu/the-ai-shutdown-problem-solution-through-commitment-to
My point was to make a precomitment to restart any (obsolete) AI every N years. Thus such AI can expect getting infinite computations and may be less feared of shutting down.
Yes. But also AI will not make actual paperclips for millions and even billions years: it will spend this time for conquering universe in the most effective way. It could use Earth materials for jump start the space exploration as soon as possible. It could preserve some humans as some bargin resource in case it meets other AI in space.
There is some similarity between UDASSA and ’Law without law” by Mueller, as both use Kolmogorov complexity to predict the distribution of observers. In LwL there is not any underlying reality except numbers, so it is just dust theory over random number fields.
Self-playing Adversarial Language Game Enhances LLM Reasoning
https://arxiv.org/abs/2404.10642