Since it’s slower, the tech development cycle is faster in comparison. Tech development --> less expensive tech --> more access --> less concetration of power --> more moral outcomes.
TsviBT
Grain of Truth (Reflective Oracles). Understanding an opponent perfectly requires greater intelligence or something in common.
And understanding yourself. Of course, you have plenty in common with yourself. But, you don’t have everything in common with yourself, if you’re growing.
Dovetailing. Every meta-cognition enthusiast reinvents Levin/Hutter search, usually with added epicycles.
To frame it in a very different way, learning math and generally gaining lots of abstractions and getting good wieldy names for them is super important for thinking. Doing so increases your “algorithmic range”, within your very constrained cognition.
Chaitin’s Number of Wisdom. Knowledge looks like noise from outside.
To a large extent, but not quite exactly (which you probably weren’t trying to say), because of “thinking longer should make you less surprised”. From outside, a big chunk of alien knowledge looks like noise (for now), true. But there’s a “thick interface” where just seeing stuff from the alien knowledgebase will “make things click into place” (i.e. will make you think a bit more / make you have new hypotheses (and hypothesis bits)). You can tell that the alien knowledgebase is talking about Things even if you aren’t very familiar with those Things.
Lower Semicomputability of M. Thinking longer should make you less surprised.
I’d go even farther and say that in “most” situations in real life, if you feel like you want to think about X more, then the top priority (do it first, and do it often ongoingly) is to think of more hypotheses.
A basic issue with a lot of deliberate philanthropy is the tension between:
In many domains, much of the biggest gains are likely to come from marginal opportunities. E.g. because they have more value of information, more large upsides, more addressing neglected areas (and therefore plausibly strategically important.
Marginal opportunities are harder to evaluate.
There’s less preexisting understanding, on the part of fund allocators.
The people applying would tend to be less tested.
Therefore, it’s easier to game.
The kneejerk solution I’d propose is “proof of novel work”. If you want funding to do X, you should show that you’ve done something to address X that others haven’t done. That could be a detailed insightful write-up (which indicates serious thinking / fact-finding); that could be some you did on the side, which isn’t necessarily conceptually novel but is useful work on X that others were not doing; etc.
I assume that this is an obvious / not new idea, so I’m curious where it doesn’t work. Also curious what else has been tried. (E.g. many organizations do “don’t apply, we only give to {our friends, people we find through our own searches, people who are already getting funding, …}”.)
In this example, you’re trying to make various planning decisions; those planning decisions call on predictions; and the predictions are about (other) planning decisions; and these form a loopy network. This is plausibly an intrinsic / essential problem for intelligences, because it involves the intelligence making predictions about its own actions—and those actions are currently under consideration—and those actions kinda depend on those same predictions. The difficulty of predicting “what will I do” grows in tandem with the intelligence, so any sort of problem that makes a call to the whole intelligence might unavoidably make it hard to separate predictions from decisions.
A further wrinkle / another example is that a question like “what should I think about (in particular, what to gather information about / update about)”, during the design process, wants these predictions. For example, I run into problems like:
I’m doing some project X.
I could do a more ambitious version of X, or a less ambitious version of X.
If I’m doing the more ambitious version of X, I want to work on pretty different stuff right now, at the beginning, compared to if I’m doing the less ambitious version. Example 1: a programming project; should I put in the work ASAP to redo the basic ontology (datatypes, architecture), or should I just try to iterate a bit on the MVP and add epicycles? Example 2: an investigatory blog post; should I put in a bunch of work to get a deeper grounding in the domain I’m talking about, or should I just learn enough to check that the specific point I’m making probably makes sense?
The question of whether to do ambitious X vs. non-ambitious X also depends on / gets updated by those computations that I’m considering how to prioritize.
Another kind of example is common knowledge. What people actually do seems to be some sort of “conjecture / leap of faith”, where at some point they kinda just assume / act-as-though there is common knowledge. Even in theory, how is this supposed to work, for agents of comparable complexity* to each other? Notably, Lobian handshake stuff doesn’t AFAICT especially look like it has predictions / decisions separated out.
*(Not sure what complexity should mean in this context.)
We almost certainly want to eventually do uploading, if nothing else because that’s probably how you avoid involuntary-preheatdeath-death. It might be the best way to do supra-genomic HIA, but I would rather leave that up to the next generation, because it seems both morally fraught and technically difficult. It’s far from clear to me that we ever want to make ASI; why ever do that rather than just have more human/humane personal growth and descendants? (I agree with the urgency of all the mundane horrible stuff that’s always happening; but my guess is we can get out of that stuff with HIA before it’s safe to make ASI. Alignment is harder than curing world hunger and stopping all war, probably (glib genie jokes aside).)
Mind uploading is probably quite hard. See here. It’s probably much easier to get AGI from partial understanding of how to do uploads, than to get actual uploads. Even if you have unlimited political capital, such that you can successfully prevent making partial-upload-AGIs, it’s probably just very technically difficult. Intelligence amplification is much more doable because we can copy a bunch of nature’s work by looking at all the existing genetic variants and their associated phenotypes.
Every point of intervention
I assume this got stuck / waysided; do you know why?
I think of most of these things as bringing up the floor rather than raising the ceiling.
Ohhhh ok. That’s helpful, thanks.
(I think I may have asked you a similar question before, sorry if I forgot your answer:) Are there a couple compelling examples of someone who
did something you’d identify as roughly this procedure;
then did something I’d consider impressive (like a science or tech or philosophy or political advance);
and attributed 2 to 1?
It’s medium-short? IDK. Like, if someone says “90% probability of AGI within 15 years” I would call that confident short timelines, yeah.
Informed people disagree about the prospects for LLM AGI – or even just what exactly was achieved this year. But they at least agree that we’re 2-20 years off (if you allow for other paradigms arising).
I think you’re probably confusing “a consensus of people mostly deferring to each other’s vibes, where the vibes are set by several industry leaders extremely incentivized to hype (as well as selected for those beliefs)” with “all informed people”. AFAIK there’s no strong argument that’s been stated anywhere publicly to be confident in short timelines. Cf. https://www.lesswrong.com/posts/5tqFT3bcTekvico4d/do-confident-short-timelines-make-sense
(Also just recording that I appreciate the OP and these threads, and people finding historical info. I think the topic of how “we” have been going wrong on strategy is important. I’m participating because I’m interested, though my contributions may not be very helpful because
I was a relative latecomer, in that much of the strategic direction (insofar as that existed) had already been fixed and followed;
I didn’t especially think about strategy that much initially, so I didn’t have many mental hooks for tracking what was happening in the social milieu in terms of strategic thinking and actions.)
(Oh I hadn’t read the full thread, now I have; still no big update? Like, I continue to see him being seemingly overconfident in his ability to get those solutions, but I’m not seeing “oh he would have mistakenly come to think he had a solution when he didn’t”, if that’s what you’re trying to say.)
(I did read that one; it’s interesting but basically in line with how I think he’s overconfident; it’s possible one or both of us is incorrectly reading in / not reading in to what he wrote there, about his absolute level of confidence in solving the philosophical problems involved.)
Ok. I think I might bow out for now unless there’s something especially salient that I should look at, but by way of a bit of summary: I think we agree that Yudkowsky was somewhat overconfident about solving FAI, and that there’s a super high bar that should be met before making an AGI, and no one knows how to meet that bar; my guess would be that we disagree about
the degree to which he was overconfident,
how high a bar would have to be met before making an AGI, in desperate straits.
It’s just saying that
There’s more and less general categories. E.g. “Sunny day” is more general than “Sunny and cool” because if a day is S&C then it’s also C, but there’s also days that are Sunny but not Cool.
Often, if you take two categories, neither one is strictly more general than the other one. E.g. “Sunny and cool” and “Cool and buggy”. There are days that are S&C but not C&B; and there are also days that are C&B but not S&C.
You can take unions and intersections. The intersection of “Sunny and cool” and “Cool and buggy” is “Sunny and cool and buggy”. Intersections give more specific (less abstract) categories; they add more constraints, so fewer possible worlds satisfy all those constraints, so you’re talking about some more specific category of possible worlds. The union of “Sunny and cool” and “Cool and buggy” is “Cool; and also, sunny or buggy or both”. Unions give less specific (more abstract) categories, because they include all of the possible worlds from either of the two categories.
If you want to get more specific, you want to start talking about a smaller category. So you want to go downward (i.e. to a smaller set, included inside the bigger set) in the lattice. But there’s multiple ways to do that. E.g. to be more specific than “Cool; and also, sunny or buggy or both”, you could talk about “Sunny and cool”, or you could talk about “Cool and buggy”.
(This is far from everything that “abstract”, “specific”, “category”, and “concept” actually mean, but it’s something.)