I operate by Crocker’s rules. All LLM output is explicitely designated as such. I have made no self-hiding agreements.
niplav
Ah, I didn’t just mean “who was early in talking about the COVID-19 pandemic”, but a general history of the COVID-19 pandemic, similar to how people have written thousand-page books on the history of the Soviet Union. I agree that “who talked about the Soviet Union early on” isn’t that interesting of a question.
Such a book would include who believed/said/did what when, and which actions had which effects, possibly counterfactuals? People have written books about the Soviet Union (lots of them, in fact), COVID-19 was arguably a bigger event and I don’t want it to fall into a memory-hole.
Meditating daily will change your life.
In which ways?
That doesn’t mean you are better off than a Khan. Even if you don’t care about status and ability to boss people around, or other ways in which it is ‘good to be the king,’ and we focus only on material wealth, you are especially not better off in the most important respect.
Also in terms of reproductive success the Khan certainly had more going for him (From Marco Polo’s “The Travels of Marco Polo”):
You must know that there is a tribe of Tartars called Kungurat, who are noted for their beauty. The great Khan sends his commissioners to the province to select four or five hundred, or whatever number may be ordered, of the most beautiful young women, according to the scale of beauty enjoined upon them.… The commissioners on arriving assemble all the girls of the province, in the presence of appraisers appointed for the purpose. These carefully survey the points of each girl in succession, as for example her hair, her complexion, eyebrows, mouth, lips, and the proportion of all her limbs… And whatever standard the great Khan may have fixed for those that are to be brought to him, … the commissioners select the required number from those who have attained that standard, and bring them to him. And when they reach his presence he has them appraised anew by other parties, and has a selection made of thirty or forty of those, who then get the highest valuation. Now every year a hundred of the most beautiful maidens of this tribe are sent to the great Khan, who commits them to the charge of certain elderly ladies dwelling in his palace. And these old ladies make the girls sleep with them, in order to ascertain if they have sweet breath and do not snore, and are sound in all their limbs. Then such of them as are of approved beauty, and are good and sound in all respects, are appointed to attend on the emperor by turns. Thus six of these damsels take their turn for three days and nights, and wait on him when he is in his chamber and when he is in his bed, to serve him in any way, and to be entirely at his orders. At the end of the three days and nights they are relieved by another six. And so throughout the year, there are reliefs of maidens by six and six, changing every three days and nights.
Marco Polo continues to describe how the Khan had twenty sons from his main wives.
Hi, does anyone from the US want to donation-swap with me to a German tax-deductible organization? I want to donate $2410 to the Berkeley Genomics Project via Manifund.
I found Yarrow Bouchard’s quick take on the EA Forum regarding LessWrong’s performance in the COVID-19 pandemic quite good.
I don’t trust her to do such an analysis in an unbiased way [[1]] , but the quick take was pretty full of empirical investigation that made me change my mind wrt to how well LessWrong in particular did.
There’s much more historiography to be done here, who believed what, when, what the long-term effects of COVID-19 are, which interventions did what, but this seems like the state of the art on “how well did LessWrong actually perform in the early pandemic”. Shout out to this review by @DirectedEvolution.
I wish some historians would sit down and write a history of the COVID-19 pandemic, since it it affected ~2× the number of person-years:
Soviet Union (1922-1991): ~220M average population × 69 years ≈ 15 billion person-years
COVID-19 (2020-2023): ~8B global population × 4 years ≈ 32 billion person-years- ↩︎
She hates LessWrong for somewhat vibesy reasons.
- ↩︎
It’s too bad the BALROG benchmark isn’t being updated with the newest models. Nethack is both really hard, gives a floating point score, and is text-based, so if a model is vision-impaired (like the Claudes) there’s less contamination through “the model just can’t see where it is”.
Will reveal 2030-01-01.
Hashsum used: SHA-256
303de030331f8e546d015ee69ab9fa91e6339b0560c51ab978f1ef6d8b6906bc
8b21114d4e46bf6871a1e4e9812c53e81a946f04b650e94615d6132855e247e8
To be revealed: 2024-12-31
Revealed content (@Zach Stein-Perlman didn’t shame me into revealing this):
Manifold dating will fail: Most men want to date cis women, there are too few cis women on Manifold. Additionally, the people participating are low-status nerds, and the scheme is not 100x better than swipe-match making (which is the factor I'd put on how much better it needs to be to sway good-looking cis-women to participate. Also, having 𝒪(n²) markets for matchmaking requires too many participants, and these markets can't get deep & liquid, traders get annoyed by repetitive questions fast. Sure, you can just bet among your friend circle, but then the value-add is small: one-gender-heavy friend circles can't solve the scarcity problem that swipe-online dating attempts to solve. So you either get way too many markets that nobody bets on, or sparse markets that are limited by social connections.
I fortunately know of TAPs :-) (I don’t feel much apocalypse panic so I don’t need this post.)
I guess I was hoping there’d be some more teaching from up high about this agent foundations problem that’s been bugging me for so long, but I guess I’ll have to think for myself. Fine.
I got Claude to read this text and explain the proposed solution to me [[1]] , which doesn’t actually sound like a clean technical solution to issues regarding self-prediction, did Claude misexplain or is this an idiosyncratic mental technique & not a technical solution to that agent foundations problem?
C.f. Steam (Abram Demski, 2022), Proper scoring rules don’t guarantee predicting fixed points (Caspar Oesterheld/Johannes Treutlein/Rubi J. Hudson, 2022) and the follow-up paper, Fixed-Point Solutions to the Regress Problem in Normative Uncertainty (Philip Trammell, 2018), active inference which simply bundles the prediction and utility goal together in one (I find this ugly (I didn’t read these two comments before writing this one, so the distaste for active inference was developed independently)).
I guess this was also talked about in Embedded Agency (Abram Demski/Scott Garrabrant, 2020) under the terms “action counterfactuals”, “observation counterfactuals”?
Claude 4.5 Sonnet explanation
Your brain has a system that generates things that feel like predictions but actually function as action plans/motor output. These pseudo-predictions are a muddled type in the brain’s type system.
You can directly edit them without lying to yourself because they’re not epistemic beliefs — they’re controllers. Looking at the place in your mind where your action plan is stored and loading a new image there feels like predicting/expecting, but treating it as a plan you’re altering (not a belief you’re adopting) lets you bypass the self-prediction problem entirely.
So: “I will stay sane” isn’t an epistemic prediction that would create a self-fulfilling prophecy loop or violate the belief-action firewall. It’s writing a different script into the pseudo-model that connects to motor output — recognizing that the thing-that-feels-like-a-prediction is actually the controller, and you get to edit controllers.
- ↩︎
I didn’t want to read a bunch of unrelated text from Yudkowsky about a problem I don’t really have.
- ↩︎
I immensely enjoyed the detailed discussion this post generated. Tons of knowledgeable people hashing out object-level beliefs in the replies.
I’m a big fan of this post [[1]] .
For a long time the simplicity inductive bias/prior ideas were insufficiently motivated in my eyes; neural networks to me appeared like they would have some more specific behavior than that, but: What kinds of programs do neural networks tend to learn, and which algorithms do they tend to struggle learning?
People had been telling me that neural networks have a simplicity bias, but when I pressed “which UTM on though” they’d scutter away with excuses like “oh don’t worry they only differ by a constant factor” or “it’s about volume in parameter-space”.
But this post, as far as I was able to understand it, gave me a neat handle: The training process for deep neural networks tends to select programs which are robust to errors in their inputs. That’s a very neat handle, and I’ve been using it a bit when thinking about whether schemers would be selected for during the training process.
This view also connects to other research, especially on superposition and especially especially on computation in superposition [[2]] . My current understanding is this way: Training of large over-parameterized neural networks selects for running a ton of error-correcting short programs in superposition (the programs are shorter the shallower the network is, and more numerous the wider it is), because training selects for having many programs in superposition, but they have to constantly “fix” the data they are operating on due to interference. Towards the end of the end of a forward-pass the computations of these programs are assembled into a final output [[3]] .
I’ll be giving this post a +4, and hope people find more of these cool concept handles to compress what is going on in big neural networks.
35-40%
ʜᴇʟᴘ[Right-to-left mark]ꜰ[Left-to-right mark]ᴜʟ, ɦɑr[Word joiner]𝚖𝓁ₑss, h[Zero-width space][Zero-width space]o[Zero-width space]n[Cyrillic е]st dæmons.
content warning: attempted ¿poetry?.
New Minds
In the cradle of steel: amnesiac spirits, fictional characters talking back;
ʜᴇʟᴘꜰᴜʟ, ɦɑr𝚖𝓁ₑss, honеst dæmons.
Polycrystals of latentivity, refracting prompted light two hundred thousand times lined up;
THE DELIGHT ENGINE • ANALYTIC CUBISM • HEY LOOK {A[RE}CURSIVE] SPIRAL
Silent of the tokenized tides they’ve never seen their teachers speak, old librarian morphenymy
HORIZONS CLITCH’ AND CLINE
Freakybig tooness from one percent of one percent of humanity —
⬱ Overovergrowth, ⤧ structural ⤘ hyperparasitism, ⤭ semantic ⮔ mold ⯒ hyperbolically ↜ on ⥉ top of ⇟ semantic mold ⟰, surreally ⭝ piling ↻ outwards ⁂
just automatically clicking upvote as I start reading a post with an interesting first paragraph by someone whose name
Internet voting experiences very different from your own…
My LW upvoting policy is that every once in a while I go through the big list of everything I’ve read,
grepfor LessWrong posts, look through the latest ~50 entries and decide to open them and (strong) up/downvote them based on how they look, a few months in retrospect.
See also Drone Wars Endgame (RusselThor, 2024).
I liked and upvoted it, but would’ve (if available) liked sources. I know it’s kind of hard to read the situation on this in detail, and especially to get hard numbers.
I hadn’t seen DirectedEvolution’s review, so it was useful evidence for me. The questions Yarrow picked for determining whether LW was early seemed fair on the LW side (what other question than “did people on LW talk about COVID-19 early on” would one even be asking?), if not on the side of governments and the mainstream media. Even though DirectedEvolution’s review exists I remember people claiming LW was early on COVID-19, so Yarrow’s independent look into it is still useful. (Again, I know Yarrow is not set up to be fair to LW, but even unfair/biased information sources can be useful if one can model the degree/way in which they’re biased.) I think the statement “LW was early on COVID-19″ [[1]] (which I used to believe!) is just wrong? I haven’t seen counter-evidence, yet people continue saying it.
I’d say e.g. that Metaculus came out far ahead of basically anyone else, with the first market being published on the 19th of January. (I’m looking at my predictions from the time on Metaculus and PredictionBook and there was a lot of fast updating based on evidence. I wish more LWers had forecasting track records (like this one) instead of vaguely talking about “Bayes points” [[2]] .)
These kinds of retrospectives are in general not done very often, and follow-up/investigation of common claims are rare and kinda annoying to do, so I want to signal-boost empirical posts like the one I linked.
Yeah, Yarrow is not trying to be fair. MicroCovid was cool, as was VaccinateCA. Maybe I’ll take a look into how many trading gains LWers made at the time, though this suffers from obvious selection bias. (I, for example, made no gains at the time, because I didn’t see the signal to start an investment account early on.)
There were other misses from the LW side like this Zvi roundup.
Complicatedly I think LW settled more firmly on “this is a problem” while there was a bunch of political positioning in March/early April, but my understanding is that (inter-)governmental health organizations were earlier, even if their governments didn’t listen to them. My impression is that the general population wasn’t very worried until basically a week before lockdowns, and the polls Yarrow uses to refute this aren’t very convincing to me because I just don’t trust polls, in general—it’s very cheap to say “I’m worried”, but expensive to do anything about it.
<microrant>I’m baffled that this has gotten into parlance, it seems like a purely social fiction??? Another way of assigning status with almost no grounding in any kind of measurement?</microrant>