Holistic believer in serendipitous causality;
newbie to the rationality dance; learning the steps.. one step at a time.
Holistic believer in serendipitous causality;
newbie to the rationality dance; learning the steps.. one step at a time.
Random thought, but at what point does an AGI+ learn to deal with further scaling by rewriting its own physical topology to minimize entropy production? ~ Last Question à la Asimov style
Updated a bit on the self-similar vs non-self similar scaling; I’m more unsure than I previously thought that I understand how scaling works, from individual to different types of collectives + time dynamics.
Short of the redundancy and letdown ending, I like this writing as it captures echoes of reasoning failure (that I find myself fall for at times); seems to be written not just for AI researchers (& adjacents) in 2025, but for many minds across “now and then”.
It strikes me that Humman does grasp truths (reality is complicated and people do have different strengths) but errs “true at this/his resolution” for “true at all scales”. Feels like he assumes self-similar scaling (like a tree) instead of considering the nature of realities that scale in non-self-similar ways (like a snowflake*, with shifts from dendritic to radial structures). More so, he uses his understanding of complexity as a thought-terminating invocation rather than a call to deeper/clearer/coherent-er modeling(s). Both are fairly common failure modes and it would be cool to leave them behind, but I’m unaware of stable ways to do so.
*not meaning to use the snowflake example as a supportive argument for his position ~ “every snowflake is unique and incomparable”, tho I do like that in my best of days.
soo, how does one re-choose & make the choice stick in the middle of a pattern that became part of normality?
One of these quick, cheap & semi-permanent wins for me was to uninstall the apps I didn’t want to use (e.g. instagram) & make it difficult to access the ones I sort of wanted to use (e.g. youtube).
Each time someone tells me that something is not possible, I tend to double the time it takes to it.
Assuming there’s a way of being in which one can half the time instead, what does your mind do then?
bondage play*
This thread needs a bit of counterbalance.
One can consider an alternative in bonding & control play; lots to explore with no brain cells at risk of dying in the process.
If one does not want to consider bonding, maybe a first aid course is due or a refresher would be fitting.
Depending on what the application involved, would it work to anon them, make them public and sort of let the community give feedback?
but I think that’s not what I should spend my remaining words on.
Why not?
~ edit: more sensical ~
I like this one:
I dare you to try to make yourself believe something you know to be false, and to closely observe what actually happens in your mind as you do.
~ more words are appreciated ~
They could be tasked with solving the entropic death of the universe á la Asimov’s Last Question.
What about the gpt oss ones for next weekend?
Soo, what do you see as a better alternative?
Or mini-alternative, with some minimal changes that could create a better future for quality content and access to it?
edit: silly typo
What did you mess up?
But switching from a consumer to a creator mindset is a process that seems to improve the quality of one’s thinking. I’d even say that more people should start creating. Though I do agree that we also need better consumption and better metrics for creation (not just volume & engagement).
I say this because I recently made a Wikipedia account (inspired by Why You Should Edit Wikipedia) and as a first-time contributor I noticed a big difference between what reading a page on whatever does to my mind and how my mind acts when working through an edit on a page I care about. Reading/consuming can easily fall into passive pattern matching, but editting/writting pushes one into building an overall understanding of the meta of context of content, structure, audience &more.
My take is that the cognitive development justifies the noise.
Also, feedback helps.
Small note that the “Twelve Virtues of Rationality” link does not work.
4 years later, how are your Bayesians going?
do you happen to have some examples or a repo or a write-up of this? Alternatively, are you aware of published research on it? I want to try it and would like to compare notes.
where does one do quick syncs on risk of eow in the now+ε?