Hello! I work at Lightcone and like LessWrong :-)
kave
From a message I wrote to a friend once that seems a little relevant
[H]ow should you act when you’re inside someone’s OODA loop? I was thinking about how like Wikipedia/tab explosions are sort of inside my ooda loop. But sometimes I can be more of an active reader who is navigating the concepts being exposed to me as I choose, and the process becomes like a magic genie or butler who is doing interpretative labour and conjuring up new scenes following my fickle interest.
So it seems like one thing that the person with the smaller loop can do is interpretative labour, and spend the faster cycles on self-legibilising.
Yep, the question is definitely about how far it transfers.
increase the agent’s expected future value
I wonder if there’s a loopiness here is which breaks the setup (the expectation I’m guessing is relative to the prediction markets probabilities? Though it seems like the market is over sensory experiences but the values are over world states in general, so maybe I’m missing something). But it seems like if I take an action and move the market at the same time, I might be able to extract a bunch of extra money and acquire outsize control.
Bidding to control the agent’s actions for the next N timesteps
This seems like it’s wasteful relative to contributing to a pool that bids on action A (or short-term policy P). I guess coordination is hard if you’re just contributing to the pool though, and all connects to the merging process you describe.
I mean when I journal I come up with little exercises to improve areas of my life. I imagine that people in your cohort might do similarly, and given that they signed up to improve their IQ, that might include things adjacent to the tasks of the IQ test.
And I don’t think general meditation should count as training, but specific meditations could (e.g. if you are training doing mental visualisations and the task involves mental rotations).
I’m not trying to say that there are definitely cross-training effects, just that these seem like the kinds of thing which are somewhat more likely (than, say, supplements) to create fairly narrow improvements close to the test.
And I can make people think “out of the box” (e.g. via specific games, specific “supplements”, specific meditations)
And prod people to think about how they can improve in whatever areas they want (e.g. via journaling, talking, and meditating)
Ah, these two have made me more concerned about training effects: especially the games, but also the meditations and journaling.
It seems pretty plausible certain games could basically train the same skills as the IQ test.
I think this a real problem (tho I think it’s more fundamental than your hypothesis would suggest; we could check commenting behaviour in the 2000s as a comparison).
We have some explorations underway addressing related issues (like maybe the frontpage should be more recommender-y and show you good old posts, while the All Posts page is used for people who care a lot about recency). I don’t think we’ve concretely considered stuff that would show you good old posts with new comments, but that might well be worth exploring.
My quick guess is that people don’t agree about what constitutes a (relevant) flaw. (And there are lots of irrelevant flaws so you can’t just check for the existence of any flaws at all).
I think if people could agree, the authorial incentives would follow. I’m fairly sympathetic to the idea that readers aren’t incentivised to correctly agree on what consitutes a flaw.
I notice that the formalised owings have professionals dedicated (in part) to making sure people line up their behaviour with the formalised owings. I wonder if other social credit formalizations would also need this
I recognised the allusion but also disliked the title.
Sometimes running to stand still is the right thing to do
It’s nice when good stuff piles up into even more good stuff, but sometimes it doesn’t:Sometimes people are worried that they will habituate to caffeine and lose any benefit from taking it.
Most efforts to lose weight are only temporarily successful (unless using medicine or surgery).
The hedonic treadmill model claims it’s hard to become durably happier.
Productivity hacks tend to stop working.
These things are like Alice’s red queen’s race: always running to stay in the same place. But I think there’s a pretty big difference between running that keeps you exactly where you would have been if you hadn’t bothered, and running that either moves you a little way and then stops, or running that stops you moving in one direction.
I’m not sure what we should call such things, but one idea is hamster wheels for things that make no difference, bungee runs for things that let you move in a direction a bit but you have to keep running to stay there, and backwards escalators for things where you’re fighting to stay in the same place rather than moving in a direction (named for the grand international pastime of running down rising escalators).
I don’t know which kind of thing is most common, but I like being able to ask which dynamic is at play. For example, I wonder if weight loss efforts are often more like backwards escalators than hamster wheels. People tend to get fatter as they get older. Maybe people who are trying (but failing) to lose weight are gaining weight more slowly than similar people who aren’t trying to do so?
Or my guess is that most people will have more energy than baseline if they take caffeine every day, even though any given dose will have less of an effect than taking the same amount of caffeine while being caffeine-naive, so they’ve bungee ran (done a bungee run?) a little way forward and that’s as far as they’ll go.
I am currently considering whether productivity hacks, which I’ve sworn off, are worth doing even though they only last for a little while. The extra, but finite, productivity could be worth it. (I think this would count as another bungee run).
I’d be interested to hear examples that fit within or break this taxonomy.
I sometimes like things being said in a long way. Mostly that’s just because it helps me stew on the ideas and look at them from different angles. But also, specifically, I liked the engagement with a bunch of epistemological intuitions and figuring out what can be recovered from them. I like in particular connecting the “trend continues” trend to the redoubtable “electron will weight the same tomorrow” intuition.
(I realise you didn’t claim there was nothing else in the dialogue, just not enough to justify the length)
How many people read your post is probably meaningful to you, and karma affects that a lot.
I say this because I certainly care about how many people read which posts, so it’s kind of sad when karma doesn’t track value in the post (though of course brevity and ease of reading are also important and valuable).
-- Some beliefs about AI Scaling Labs that I’m redacting on LW --
Is the reason for the redaction also private?
Hm, I’m not seeing the issue. Could you screenshot it?
There are separate random number generators for most things in Slay the Spire.
Specifically, this piece of code is executed at the start of a run:
public static void generateSeeds() { logger.info(“Generating seeds: ” + Settings.seed); monsterRng = new Random(Settings.seed); eventRng = new Random(Settings.seed); merchantRng = new Random(Settings.seed); cardRng = new Random(Settings.seed); treasureRng = new Random(Settings.seed); relicRng = new Random(Settings.seed); potionRng = new Random(Settings.seed); // The following rngs are actually re-initialized each floor: monsterHpRng = new Random(Settings.seed); aiRng = new Random(Settings.seed); shuffleRng = new Random(Settings.seed); cardRandomRng = new Random(Settings.seed); miscRng = new Random(Settings.seed); }
From https://forgottenarbiter.github.io/Correlated-Randomness/ (which does point out there is some unfortunate correlated randomness, though I think there are mods that fix that)
Dungeon Crawl Stone Soup
I had previously guessed air movement made me feel better because my body expected air movement (i.e. some kind of biophilic effect). But this explanation seems more likely in retrospect! I’m not quite sure how to run the calculation using the diffusivity coefficient to spot check this, though.
That’s a cool idea! Currently this is just a restyling of a comment thread using the normal react system; I think if we decided to invest more into it to make it a real feature it would be kind of cool to build stuff like that. Could also be cool to make it play with viewpoints somehow, though the philosophies are a little different. These polls are built around people publically stating their opinions, whereas viewpoints is anonymous(?).
Say more / references?