RSS

Gen­eral intelligence

TagLast edit: 20 Feb 2025 2:21 UTC by Morpheus

Definition

Although humans share 95% of their DNA with chimpanzees, and have brains only three times as large as chimpanzee brains, humans appear to be far better than chimpanzees at learning an enormous variety of cognitive domains. A bee is born with the ability to construct hives; a beaver is born with an instinct for building dams; a human looks at both and imagines a gigantic dam with a honeycomb structure of internal reinforcement. Arguendo, some set of factors, present in human brains but not in chimpanzee brains, seem to sum to a central cognitive capability that lets humans learn a huge variety of different domains without those domains being specifically preprogrammed as instincts.

This very-widely-applicable cognitive capacity is termed general intelligence (by most AI researchers explicitly talking about it; the term isn’t universally accepted as yet).

We are not perfectly general—we have an easier time learning to walk than learning to do abstract calculus, even though the latter is much easier in an objective sense. But we’re sufficiently general that we can figure out Special Relativity and engineer skyscrapers despite our not having those abilities built-in at compile time (i.e., at birth). An Artificial General Intelligence would have the same property; it could learn a tremendous variety of domains, including domains it had no inkling of when it was switched on.

More specific hypotheses about how general intelligence operates have been advanced at various points, but any corresponding attempts to define general intelligence that way, would be theory-laden. The pretheoretical phenomenon to be explained is the extraordinary variety of human achievements across many non-instinctual domains, compared to other animals.

Artificial General Intelligence is not par-human AI

Since we only know about one organism with this ‘general’ or ‘significantly more generally applicable than chimpanzee cognition’ intelligence, this capability is sometimes identified with humanity, and consequently with our overall level of cognitive ability.

We do not, however, know that “cognitive ability that works on a very wide variety of problems” and “overall humanish levels of performance” need to go together across much wider differences of mind design.

Humans evolved incrementally out of earlier hominids by blind processes of natural selection; evolution wasn’t trying to design a human on purpose. Because of the way we evolved incrementally, all neurotypical humans have specialized evolved capabilities like ‘walking’ and ‘running’ and ‘throwing stones’ and ‘outwitting other humans’. We have all the primate capabilities and all the hominid capabilities as well as whatever is strictly necessary for general intelligence.

So, for all we know at this point, there could be some way to get a ‘significantly more general than chimpanzee cognition’ intelligence, in the equivalent of a weaker mind than a human brain. E.g., due to leaving out some of the special support we evolved to run, throw stones, and outwit other minds. We might at some point consistently see an infrahuman general intelligence that is not like a disabled human, but rather like some previously unobserved and unimagined form of weaker but still highly general intelligence.

Since the concepts of ‘general intelligence’ and ‘roughly par-human intelligence’ come apart in theory and possibly also in practice, we should avoid speaking of Artificial General Intelligence as if were identical with a concept like “human-level AI”.

General intelligence is not perfect intelligence

General intelligence doesn’t imply the ability to solve every kind of cognitive problem; if we wanted to use a longer phrase we could say that humans have ‘significantly more generally applicable intelligence than chimpanzees’. A sufficiently advanced Artificial Intelligence that could self-modify (rewrite its own code) might have ‘significantly more generally applicable intelligence than humans’; e.g. such an AI might be able to easily write bug-free code in virtue of giving itself specialized cognitive algorithms for programming. Humans, to write computer programs, need to adapt savanna-specialized tiger-evasion modules like our visual cortex and auditory cortex to representing computer programs instead, which is one reason we’re such terrible programmers.

Similarly, it’s not hard to construct math problems to which we know the solution, but are unsolvable by any general cognitive agent that fits inside the physical universe. For example, you could pick a long random string and generate its SHA-4096 hash, and if the SHA algorithm turns out to be secure against quantum computing, you would be able to construct a highly specialized ‘agent’ that could solve the problem of ‘tell me which string has this SHA-4096 hash’ which no other agent would be able to solve without directly inspecting your agent’s cognitive state, or tricking your agent into revealing the secret, etcetera. The ‘significantly more generally applicable than chimpanzee intelligence’ of humans is able to figure out how to launch interplanetary space probes just by staring at the environment for a while, but it still can’t reverse SHA-4096 hashes.

It would however be an instance of the continuum fallacy, nirvana fallacy, false dichotomy, or straw superpower fallacy, to argue:

For attempts to talk about performance relative to a truly general measure of intelligence (as opposed to just saying that humans seem to have some central capability which sure lets them learn a whole lot of stuff) see Shane Legg and Marcus Hutter’s work on proposed metrics of ‘universal intelligence’.

General intelligence is a separate concept from IQ /​ g-factor

Charles Spearman found that by looking on performances across many cognitive tests, he was able to infer a central factor, now called Spearman’s g, which appeared to be more correlated with performance on each task than any of the tasks were correlated with each other.

For example, the correlation between students’ French and English scores was 0.67: that is, 67% of the variation in performance in French could be predicted by looking at the student’s score in English.

However, by looking at all the test results together, it was possible to construct a central score whose correlation with the student’s French score was 88%.

This would make sense if, for example, the score in French was “g-factor plus uncorrelated variables” and the score in English was “g-factor plus other uncorrelated variables”. In this case, the setting of the g-factor latent variable, which you could infer better by looking at all the student’s scores together, would be more highly correlated with both French and English observations, than those tests would be correlated with each other.

In the context of Artificial Intelligence, g-factor is not what we want to talk about. We are trying to point to a factor separating humans from chimpanzees, not to internal variations within the human species.

That is: If you’re trying to build the first mechanical heavier-than-air flying machine, you ought to be thinking “How do birds fly? How do they stay up in the air, at all?” Rather than, “Is there a central Fly-Q factor that can be inferred from the variation in many different measures of how well individual pigeons fly, which lets us predict the individual variation in a pigeon’s speed or turning radius better than any single observation about one factor of that pigeon’s flying ability?”

In some sense the existence of g-factor could be called Bayesian evidence for the notion of general intelligence: if general intelligence didn’t exist, probably neither would IQ. Likewise the observation that, e.g., John von Neumann existed and was more productive across multiple disciplines compared to his academic contemporaries. But this is not the main argument or the most important evidence. Looking at humans versus chimpanzees gives us a much, much stronger hint that a species’ ability to land space probes on Mars correlates with that species’ ability to prove Fermat’s Last Theorem.

Cross-domain consequentialism

A marginally more detailed and hence theory-laden view of general intelligence, from the standpoint of advanced agent properties, is that we can see general intelligence as “general cross-domain learning and consequentialism”.

That is, we can (arguendo) view general intelligence as: the ability to learn to model a wide variety of domains, and to construct plans that operate within and across those domains.

For example: AlphaGo can be seen as trying to achieve the consequence of a winning Go position on the game board—to steer the future into the region of outcomes that AlphaGo defines as a preferred position. However, AlphaGo only plans within the domain of legal Go moves, and it can’t learn any domains other than that. So AlphaGo can’t, e.g., make a prank phone call at night to Lee Se-Dol to make him less well-rested the next day, even though this would also tend to steer the future of the board into a winning state, because AlphaGo wasn’t preprogrammed with any tactics or models having to do with phone calls or human psychology, and AlphaGo isn’t a general AI that could learn those new domains.

On the other hand, if a general AI were given the task of causing a certain Go board to end up in an outcome defined as a win, and that AI had ‘significantly more generally applicable than chimpanzee intelligence’ on a sufficient level, that Artificial General Intelligence might learn what humans are, learn that there’s a human trying to defeat it on the other side of the Go board, realize that it might be able to win the Go game more effectively if it could make the human play less well, realize that to make the human play less well it needs to learn more about humans, learn about humans needing sleep and sleep becoming less good when interrupted, learn about humans waking up to answer phone calls, learn how phones work, learn that some Internet services connect to phones...

If we consider an actual game of Go, rather than a logical game of Go, then the state of the Go board at the end of the game is produced by an enormous and tangled causal process that includes not just the proximal moves, but the AI algorithm that chooses the moves, the cluster the AI is running on, the humans who programmed the cluster; and also, on the other side of the board, the human making the moves, the professional pride and financial prizes motivating the human, the car that drove the human to the game, the amount of sleep the human got that night, all the things all over the world that didn’t interrupt the human’s sleep but could have, and so on. There’s an enormous lattice of causes that lead up to the AI’s and the human’s actual Go moves.

We can see the cognitive job of an agent in general as “select policies or actions which lead to a more preferred outcome”. The enormous lattice of real-world causes leading up to the real-world Go game’s final position, means that an enormous set of possible interventions could potentially steer the real-world future into the region of outcomes where the AI won the Go game. But these causes are going through all sorts of different domains on their way to the final outcome, and correctly choosing from the much wider space of interventions means you need to understand all the domains along the way. If you don’t understand humans, understanding phones doesn’t help; the prank phone call event goes through the sleep deprivation event, and to correctly model events having to do with sleep deprivation requires knowing about humans.

Deep commonalities across cognitive domains

To the extent one credits the existence of ‘significantly more general than chimpanzee intelligence’, it implies that there are common cognitive subproblems of the huge variety of problems that humans can (learn to) solve, despite the surface-level differences of those domains. Or at least, the way humans solve problems in those domains, the cognitive work we do must have deep commonalities across those domains. These commonalities may not be visible on an immediate surface inspection.

Imagine you’re an ancient Greek who doesn’t know anything about the brain having a visual cortex. From your perspective, ship captains and smiths seem to be doing a very different kind of work; ships and anvils seem like very different objects to know about; it seems like most things you know about ships don’t carry over to knowing about anvils. Somebody who learns to fight with a spear, does not therefore know how to fight with a sword and shield; they seem like quite different weapon sets.

(Since, by assumption, you’re an ancient Greek, you’re probably also not likely to wonder anything along the lines of “But wait, if these tasks didn’t all have at least some forms of cognitive labor in common deep down, there’d be no reason for humans to be simultaneously better at all of them than other primates.”)

Only after learning about the existence of the cerebral cortex and the cerebellum and some hypotheses about what those parts of the brain are doing, are you likely to think anything along the lines of:

“Ship-captaining and smithing and spearfighting and swordfighting look like they all involve using temporal hierarchies of chunked tactics, which is a kind of thing the cortical algorithm is hypothesized to do. They all involve realtime motor control with error correction, which is a kind of thing the cerebellar cortex is hypothesized to do. So if the human cerebral cortex and cerebellar cortex are larger or running better algorithms than chimpanzees’ cerebrums and cerebellums, humans being better at learning and performing this kind of deep underlying cognitive labor that all these surface-different tasks have in common, could explain why humans are simultaneously better than chimpanzees at learning and performing shipbuilding, smithing, spearfighting, and swordfighting.”

This example is hugely oversimplified, in that there are far more differences going on between humans and chimpanzees than just larger cerebrums and cerebellums. Likewise, learning to build ships involves deliberate practice which involves maintaining motivation over long chains of visualization, and many other cognitive subproblems. Focusing on just two factors of ‘deep’ cognitive labor and just two mechanisms of ‘deep’ cognitive performance is meant more as a straw illustration of what the much more complicated real story would look like.

But in general, the hypothesis of general intelligence seems like it should cash out as some version of: “There’s some set of new cognitive algorithms, plus improvements to existing algorithms, plus bigger brains, plus other resources—we don’t know how many things like this there are, but there’s some set of things like that—which, when added to previously existing primate and hominid capabilities, created the ability to do better on a broad set of deep cognitive subproblems held in common across a very wide variety of humanly-approachable surface-level problems for learning and manipulating domains. And that’s why humans do better on a huge variety of domains simultaneously, despite evolution having not preprogrammed us with new instinctual knowledge or algorithms for all those domains separately.”

Underestimating cognitive commonalities

The above view suggests a directional bias of uncorrected intuition: Without an explicit correction, we may tend to intuitively underestimate the similarity of deep cognitive labor across seemingly different surface problems.

On the surface, a ship seems like a different object from a smithy, and the spear seems to involve different tactics from a sword. With our attention going to these visible differences, we’re unlikely to spontaneously invent a concept of ‘realtime motor control with error correction’ as a kind of activity performed by a ‘cerebellum’—especially if our civilization doesn’t know any neuroscience. The deep cognitive labor in common goes unseen, not just because we’re not paying attention to the invisible constants of human intelligence, but because we don’t have the theoretical understanding to imagine in any concrete detail what could possibly be going on.

This suggests an argument from predictable updating: if we knew even more about how general intelligence actually worked inside the human brain, then we would be even better able to concretely visualize deep cognitive problems shared between different surface-level domains. We don’t know at present how to build an intelligence that learns a par-human variety of domains, so at least some of the deep commonalities and corresponding similar algorithms across those domains, must be unknown to us. Then, arguendo, if we better understood the true state of the universe in this regard, our first-order/​uncorrected intuitions would predictably move further along the direction that our belief previously moved when we learned about cerebral cortices and cerebellums. Therefore, to avoid violating probability theory by foreseeing a predictable update, our second-order corrected belief should already be that there is more in common between different cognitive tasks than we intuitively see how to compute.

In sum this suggests a deflationary psychological account of a directional bias of uncorrected intuitions toward general-intelligence skepticism: People invent theories of distinct intelligences and nonoverlapping specializations, because (a) they are looking toward socially salient human-human differences instead of human-vs-chimpanzee differences, (b) they have failed to correct for the fading of invisible constants such as human intelligence, and (c) they have failed to apply an explicit correction for the extent to which we feel like we understand surface-level differences but are ignorant of the cognitive commonalities suggested by the general human performance factor.

(The usual cautions about psychologizing apply: you can’t actually get empirical data about the real world by arguing about people’s psychology.)

Naturally correlated AI capabilities

Few people in the field would outright disagree with either the statement “humans have significantly more widely applicable cognitive abilities than other primates” or, or the other side, “no matter how intelligent you are, if your brain fits inside the physical universe, you might not be able to reverse SHA-4096 hashes”. But even taking both those statements for granted, there seems to be a set of policy-relevant factual questions about, roughly, to what degree general intelligence is likely to shorten the pragmatic distance between different AI capabilities.

For example, consider the following (straw) amazing simple solution to all of AI alignment:

“Let’s just develop an AI that knows how to do good things but not bad things! That way, even if something goes wrong, it won’t know how to hurt us!”

To which we reply: “That’s like asking for an AI that understands how to drive blue cars but not red cars. The cognitive work you need to do in order to drive a blue car is very similar to the cognitive labor required to drive a red car; an agent that can drive a blue car is only a tiny step away from driving a red car. In fact, you’d pretty much have to add design features specifically intended to prevent the agent from understanding how to drive a car if it’s painted red, and if something goes wrong with those features, you’ll have a red-car-driving-capable agent on your hands.”

“I don’t believe in this so-called general-car-driving-intelligence,” comes the reply. “I see no reason why ability at driving blue cars has to be so strongly correlated with driving red cars; they look pretty different to me. Even if there’s a kind of agent that’s good at driving both blue cars and red cars, it’d probably be pretty inefficient compared to a specialized blue-car-driving or red-car-driving intelligence. Anyone who was constructing a car-driving algorithm that only needed to work with blue cars, would not naturally tend to produce an algorithm that also worked on red cars.”

“Well,” we say, “maybe blue cars and red cars look different. But if you did have a more concrete and correct idea about what goes on inside a robotic car, and what sort of computations it does, you’d see that the computational subproblems of driving a blue car are pretty much identical to the computational subproblems of driving a red car.”

“But they’re not actually identical,” comes the reply. “The set of red cars isn’t actually identical to the set of blue cars and you won’t actually encounter exactly identical problems in driving these non-overlapping sets of physical cars going to different places.”

“Okay,” we reply, “that’s admittedly true. But in order to reliably drive any blue car you might get handed, you need to be able to solve an abstract volume of not-precisely-known-in-advance cognitive subproblems. You need to be able to drive on the road regardless of the exact arrangement of the asphalt. And that’s the same range of subproblems required to drive a red car.”

We are, in this case, talking to someone who doesn’t believe in color-general car-driving intelligence or that color-general car-driving is a good or natural way to solve car-driving problems. In this particular case it’s an obvious straw position because we’ve picked two tasks that are extremely similar in an intuitively obvious way; a human trained to drive blue cars does not need any separate practice at all to drive red cars.

For a straw position at the opposite extreme, consider: “I just don’t believe you can solve logical Tic-Tac-Toe without some deep algorithm that’s general enough to do anything a human can. There’s no safe way to get an AI that can play Tic-Tac-Toe without doing things dangerous enough to require solving all of AI alignment. Beware the cognitive biases that lead you to underestimate how much deep cognitive labor is held in common between tasks that merely appear different on the surface!”

To which we reply, “Contrary to some serious predictions, it turned out to be possible to play superhuman Go without general AI, never mind Tic-Tac-Toe. Sometimes there really are specialized ways of doing things, the end.”

Between these two extremes lie more plausible positions that have been seriously held and debated, including:

Arguably, these factual questions have in common that they revolve about the distance between different cognitive domains—given a natural design for an agent that can do X, how close is it in design space to an agent that can do Y? Is it ‘driving blue cars vs. driving red cars’ or ‘Tic-Tac-Toe vs. classifying pictures of cats’?

(Related questions arise in any safety-related proposal to divide an AI’s internal competencies into internal domains, e.g. for purposes of minimizing the number of internal goals with the power to recruit subgoals across any known domain.)

It seems like in practice, different beliefs about ‘general intelligence’ may account for a lot of the disagreement about “Can we have an AI that X-es without that AI being 30 seconds away from being capable of Y-ing?” In particular, different beliefs about:

To the extent that you assign general intelligence a more central role, you may tend in general to think that competence in domain X is likely to be nearer to competence at domain Y. (Although not to an unlimited degree, e.g. witness Tic-Tac-Toe or reversing a SHA-4096 hash.)

Relation to capability gain theses

How much credit one gives to ‘general intelligence’ is not the same question as how much credit one gives to issues of rapid capability gains, superintelligence, and the possible intermediate event of an intelligence explosion. The ideas can definitely be pried apart conceptually:

In practice, it seems to be an observed fact that people who give more credit to the notion of general intelligence expect higher returns on cognitive reinvestment, and vice versa. This correlation makes sense, since:

It also seems to make sense for people who give more credit to general intelligence, being more concerned about capability-gain-related problems in general; they are more likely to think that an AI with high levels of one ability is likely to be able to acquire another ability relatively quickly (or immediately) and without specific programmer efforts to make that happen.

  1. ^

    E.g. in the sense of having lower sample complexity and hence being able to derive correct answers using fewer observations than humans trying to do the same over relatively short periods of time.

  2. ^

    It seems conceptually possible to believe, though this belief has not been observed in the wild, that self-programming minds have something worthy of being called ‘general intelligence’ but that human brains don’t.

Is Gen­eral In­tel­li­gence “Com­pact”?

DragonGod4 Jul 2022 13:27 UTC
27 points
6 comments22 min readLW link

Ar­tifi­cial Addition

Eliezer Yudkowsky20 Nov 2007 7:58 UTC
97 points
128 comments6 min readLW link

Refram­ing Su­per­in­tel­li­gence: Com­pre­hen­sive AI Ser­vices as Gen­eral Intelligence

Rohin Shah8 Jan 2019 7:12 UTC
122 points
77 comments5 min readLW link2 reviews
(www.fhi.ox.ac.uk)

Adap­ta­tion-Ex­e­cuters, not Fit­ness-Maximizers

Eliezer Yudkowsky11 Nov 2007 6:39 UTC
178 points
33 comments3 min readLW link

The Teacup Test

lsusr8 Oct 2022 4:25 UTC
102 points
32 comments2 min readLW link

Effi­cient Cross-Do­main Optimization

Eliezer Yudkowsky28 Oct 2008 16:33 UTC
56 points
38 comments5 min readLW link

AlphaS­tar: Im­pres­sive for RL progress, not for AGI progress

orthonormal2 Nov 2019 1:50 UTC
113 points
58 comments2 min readLW link1 review

[Book Re­view] “The Bell Curve” by Charles Murray

lsusr2 Nov 2021 5:49 UTC
94 points
134 comments23 min readLW link

Hu­mans Who Are Not Con­cen­trat­ing Are Not Gen­eral Intelligences

sarahconstantin25 Feb 2019 20:40 UTC
204 points
35 comments6 min readLW link1 review
(srconstantin.wordpress.com)

birds and mam­mals in­de­pen­dently evolved intelligence

bhauth8 Apr 2025 20:00 UTC
73 points
23 comments1 min readLW link
(www.quantamagazine.org)

Mea­sur­ing Op­ti­miza­tion Power

Eliezer Yudkowsky27 Oct 2008 21:44 UTC
91 points
38 comments6 min readLW link

Distinc­tions in Types of Thought

sarahconstantin10 Oct 2017 3:36 UTC
37 points
24 comments13 min readLW link

The Oc­to­pus, the Dolphin and Us: a Great Filter tale

Stuart_Armstrong3 Sep 2014 21:37 UTC
76 points
236 comments3 min readLW link

How many OOMs of com­pute span the hu­man range?

tickybob10 Jul 2025 11:51 UTC
12 points
6 comments1 min readLW link

On Mea­sur­ing In­tel­lec­tual Perfor­mance—per­sonal ex­pe­rience and sev­eral thoughts

Alexander Gufan20 Sep 2024 17:21 UTC
3 points
2 comments8 min readLW link

What’s up with ChatGPT and the Tur­ing Test?

4 Jan 2023 15:37 UTC
13 points
19 comments3 min readLW link

How min­i­mal is our in­tel­li­gence?

Douglas_Reay25 Nov 2012 23:34 UTC
78 points
214 comments6 min readLW link

Some com­ments on intelligence

Viliam1 Aug 2024 15:17 UTC
30 points
5 comments3 min readLW link

Chap­ter 24: Machi­avel­lian In­tel­li­gence Hypothesis

Eliezer Yudkowsky14 Mar 2015 19:00 UTC
32 points
1 comment14 min readLW link

What In­tel­li­gence Tests Miss: The psy­chol­ogy of ra­tio­nal thought

Kaj_Sotala11 Jul 2010 23:01 UTC
53 points
55 comments9 min readLW link

The Power of Intelligence

Eliezer Yudkowsky1 Jan 2007 20:00 UTC
128 points
7 comments4 min readLW link

Re­sponse to Blake Richards: AGI, gen­er­al­ity, al­ign­ment, & loss functions

Steven Byrnes12 Jul 2022 13:56 UTC
62 points
9 comments15 min readLW link

Many peo­ple lack ba­sic sci­en­tific knowledge

lsusr29 Mar 2024 6:43 UTC
11 points
20 comments1 min readLW link

Thoughts on Loss Land­scapes and why Deep Learn­ing works

beren25 Jul 2023 16:41 UTC
54 points
4 comments18 min readLW link

Yud­kowsky and Chris­ti­ano dis­cuss “Take­off Speeds”

Eliezer Yudkowsky22 Nov 2021 19:35 UTC
210 points
176 comments60 min readLW link1 review

The Limits of In­tel­li­gence and Me: Do­main Expertise

ChrisHallquist7 Dec 2013 8:23 UTC
46 points
79 comments5 min readLW link

Is Click­bait De­stroy­ing Our Gen­eral In­tel­li­gence?

Eliezer Yudkowsky16 Nov 2018 23:06 UTC
207 points
68 comments5 min readLW link2 reviews

The Power of In­tel­li­gence—The Animation

Writer11 Mar 2023 16:15 UTC
45 points
3 comments1 min readLW link
(youtu.be)

The ground of optimization

Alex Flint20 Jun 2020 0:38 UTC
252 points
80 comments27 min readLW link1 review

Soares, Tal­linn, and Yud­kowsky dis­cuss AGI cognition

29 Nov 2021 19:26 UTC
121 points
39 comments40 min readLW link1 review

See­ing Through the Eyes of the Algorithm

silentbob22 Feb 2025 11:54 UTC
18 points
3 comments10 min readLW link

Belief in Intelligence

Eliezer Yudkowsky25 Oct 2008 15:00 UTC
118 points
38 comments3 min readLW link

How Should We Mea­sure In­tel­li­gence Models: Why Use Fre­quency of Ele­men­tal In­for­ma­tion Operations

hwj2024 Oct 2024 16:54 UTC
1 point
0 comments5 min readLW link

...Re­cur­sion, Magic

Eliezer Yudkowsky25 Nov 2008 9:10 UTC
34 points
28 comments5 min readLW link

Book Re­view: The Eureka Factor

drossbucket4 Mar 2019 19:47 UTC
19 points
2 comments13 min readLW link

Hands-On Ex­pe­rience Is Not Magic

Thane Ruthenis27 May 2023 16:57 UTC
22 points
14 comments5 min readLW link

Quan­tify­ing Gen­eral Intelligence

JasonB17 Jun 2022 21:57 UTC
9 points
6 comments13 min readLW link

Every­day Les­sons from High-Di­men­sional Optimization

johnswentworth6 Jun 2020 20:57 UTC
168 points
44 comments6 min readLW link

The Prin­ci­pled In­tel­li­gence Hypothesis

KatjaGrace14 Feb 2018 1:00 UTC
34 points
15 comments4 min readLW link
(meteuphoric.wordpress.com)

On some first prin­ci­ples of intelligence

Macheng_Shen17 Mar 2023 17:42 UTC
−14 points
0 comments4 min readLW link

Com­plex­ity and Intelligence

Eliezer Yudkowsky3 Nov 2008 20:27 UTC
34 points
78 comments11 min readLW link

Hu­mans are not au­to­mat­i­cally strategic

AnnaSalamon8 Sep 2010 7:02 UTC
632 points
278 comments4 min readLW link

Ben Go­ertzel’s “Kinds of Minds”

JoshuaFox11 Apr 2021 12:41 UTC
12 points
4 comments1 min readLW link

A Case for the Least For­giv­ing Take On Alignment

Thane Ruthenis2 May 2023 21:34 UTC
100 points
85 comments22 min readLW link

The Flynn Effect Clarified

PeterMcCluskey12 Dec 2020 5:18 UTC
34 points
2 comments4 min readLW link
(www.bayesianinvestor.com)

The Dark Side of Cog­ni­tion Hypothesis

Cameron Berg3 Oct 2021 20:10 UTC
19 points
1 comment16 min readLW link

AGI will be made of het­ero­ge­neous com­po­nents, Trans­former and Selec­tive SSM blocks will be among them

Roman Leventov27 Dec 2023 14:51 UTC
33 points
9 comments4 min readLW link

In­stru­men­tal con­ver­gence is what makes gen­eral in­tel­li­gence possible

tailcalled11 Nov 2022 16:38 UTC
105 points
11 comments4 min readLW link

Two ex­pla­na­tions for vari­a­tion in hu­man abilities

Matthew Barnett25 Oct 2019 22:06 UTC
87 points
28 comments5 min readLW link1 review

What can we learn from or­cas?

Jonasb10 Jun 2024 18:01 UTC
1 point
0 comments8 min readLW link
(www.denominations.io)

Hu­man­ity’s Last Prowess

satpugnet29 Jun 2025 18:50 UTC
1 point
0 comments1 min readLW link

Mo­ti­va­tions, Nat­u­ral Selec­tion, and Cur­ricu­lum Engineering

Oliver Sourbut16 Dec 2021 1:07 UTC
16 points
0 comments42 min readLW link

Rec­og­niz­ing Intelligence

Eliezer Yudkowsky7 Nov 2008 23:22 UTC
28 points
30 comments4 min readLW link

Cri­tique of some re­cent philos­o­phy of LLMs’ minds

Roman Leventov20 Jan 2023 12:53 UTC
52 points
8 comments20 min readLW link

Mo­ti­va­tion in AI

nickasaf16 Jun 2023 9:50 UTC
−1 points
1 comment2 min readLW link

Loose Threads on Intelligence

Shoshannah Tekofsky24 Dec 2022 0:38 UTC
11 points
3 comments8 min readLW link

AGI is what gen­er­ates evolu­tion­ar­ily fit and novel information

onur1 Jan 2025 9:22 UTC
1 point
0 comments6 min readLW link
(solmaz.io)

My Child­hood Role Model

Eliezer Yudkowsky23 May 2008 8:51 UTC
94 points
63 comments5 min readLW link

Memetic Judo #3: The In­tel­li­gence of Stochas­tic Par­rots v.2

Max TK20 Aug 2023 15:18 UTC
8 points
33 comments6 min readLW link

Im­plicit extortion

paulfchristiano13 Apr 2018 16:33 UTC
29 points
16 comments6 min readLW link
(ai-alignment.com)

How spe­cial are hu­man brains among an­i­mal brains?

zhukeepa1 Apr 2020 1:35 UTC
79 points
38 comments7 min readLW link

HELP: Do I have a chance at be­com­ing in­tel­li­gent?

johnbgone26 Oct 2010 21:41 UTC
38 points
68 comments1 min readLW link

LLMs Suck at Deep Think­ing Part 3 - Try­ing to Prove It (fixed)

Taylor G. Lunt27 Sep 2025 14:54 UTC
17 points
6 comments15 min readLW link

Modest Superintelligences

Wei Dai22 Mar 2012 0:29 UTC
35 points
100 comments1 min readLW link

AGI is Inevitable — The­o­ret­i­cal Physics Might Tell Us When

Bernard Minn30 Mar 2025 4:31 UTC
1 point
0 comments4 min readLW link
(medium.com)

Thoughts for and against an ASI figur­ing out ethics for itself

sweenesm20 Feb 2024 23:40 UTC
6 points
10 comments3 min readLW link

Con­tra Alexan­der on the Bit­ter Les­son and IQ

Andrew Keenan Richardson26 Jul 2023 0:07 UTC
9 points
1 comment4 min readLW link
(mechanisticmind.com)

Au­to­mated in­tel­li­gence is not AI

KatjaGrace1 Nov 2020 23:30 UTC
54 points
10 comments2 min readLW link
(meteuphoric.com)

If brains are com­put­ers, what kind of com­put­ers are they? (Den­nett tran­script)

Ben Pace30 Jan 2020 5:07 UTC
37 points
9 comments27 min readLW link

Fun­da­men­tal Philo­soph­i­cal Prob­lems In­her­ent in AI discourse

AlexSadler16 Sep 2018 21:03 UTC
23 points
1 comment17 min readLW link

Co­her­ence-Based Mea­sure of AGI: GPT-5 ≈ 24 %

Fares Fourati25 Oct 2025 9:14 UTC
1 point
0 comments2 min readLW link

Is the ar­gu­ment that AI is an xrisk valid?

MACannon19 Jul 2021 13:20 UTC
5 points
61 comments1 min readLW link
(onlinelibrary.wiley.com)

[Question] His­tor­i­cal ex­am­ples of peo­ple gain­ing un­usual cog­ni­tive abil­ities?

Nicholas Kross24 Nov 2022 19:01 UTC
8 points
2 comments1 min readLW link

Muehlhauser-Wang Dialogue

lukeprog22 Apr 2012 22:40 UTC
34 points
288 comments12 min readLW link

A rant against robots

Lê Nguyên Hoang14 Jan 2020 22:03 UTC
66 points
7 comments5 min readLW link

My Best and Worst Mistake

Eliezer Yudkowsky16 Sep 2008 0:43 UTC
73 points
17 comments5 min readLW link

A ba­sic math­e­mat­i­cal struc­ture of intelligence

Golol12 Apr 2023 16:49 UTC
4 points
6 comments4 min readLW link

Pro­duc­tivity as a func­tion of abil­ity in the­o­ret­i­cal fields

Stefan_Schubert26 Jan 2014 13:16 UTC
45 points
34 comments4 min readLW link

Iter­ated Distil­la­tion-Am­plifi­ca­tion, Gato, and Proto-AGI [Re-Ex­plained]

Gabe M27 May 2022 5:42 UTC
22 points
4 comments6 min readLW link

Take part in our gi­ant study of cog­ni­tive abil­ities and get a cus­tomized re­port of your strengths and weak­nesses!

spencerg25 Nov 2022 16:28 UTC
8 points
1 comment1 min readLW link
(www.guidedtrack.com)

In­tel­li­gence en­hance­ment as ex­is­ten­tial risk mitigation

Roko15 Jun 2009 19:35 UTC
21 points
244 comments3 min readLW link

Pow­er­ful mesa-op­ti­mi­sa­tion is already here

Roman Leventov17 Feb 2023 4:59 UTC
35 points
1 comment2 min readLW link
(arxiv.org)

What is In­tel­li­gence?

IsaacRosedale23 Apr 2023 6:10 UTC
1 point
0 comments1 min readLW link

[Question] Are Hu­man Brains Univer­sal?

DragonGod15 Sep 2022 15:15 UTC
16 points
28 comments5 min readLW link

Sur­prised by Brains

Eliezer Yudkowsky23 Nov 2008 7:26 UTC
62 points
28 comments7 min readLW link

The Apol­lo­nian Gas­ket and AI’s Hid­den Geo­met­ric Structure

LikaMench7 Mar 2025 20:06 UTC
1 point
0 comments3 min readLW link

Four Phases of AGI

Gabe M5 Aug 2024 13:15 UTC
13 points
3 comments13 min readLW link

Might hu­mans not be the most in­tel­li­gent an­i­mals?

Matthew Barnett23 Dec 2019 21:50 UTC
56 points
41 comments3 min readLW link

Gen­eral in­tel­li­gence: what is it, what makes it hard, and will we have it soon?

homeopathicsyzygy1 Jun 2023 6:46 UTC
2 points
0 comments21 min readLW link

Beyond Smart and Stupid

PhilGoetz17 May 2011 6:25 UTC
36 points
44 comments3 min readLW link

Eco­nomic Defi­ni­tion of In­tel­li­gence?

Eliezer Yudkowsky29 Oct 2008 19:32 UTC
18 points
9 comments7 min readLW link

Ar­tifi­cial In­tel­li­gence and Liv­ing Wisdom

TMFOW29 Mar 2024 7:41 UTC
−6 points
1 comment17 min readLW link
(tmfow.substack.com)

Three ways that “Suffi­ciently op­ti­mized agents ap­pear co­her­ent” can be false

Wei Dai5 Mar 2019 21:52 UTC
65 points
3 comments3 min readLW link

Can we cre­ate a func­tion that prov­ably pre­dicts the op­ti­miza­tion power of in­tel­li­gences?

whpearson28 May 2009 11:35 UTC
−7 points
17 comments2 min readLW link

“Pick Two” AI Trilemma: Gen­er­al­ity, Agency, Align­ment.

Black Flag15 Jan 2025 18:52 UTC
7 points
0 comments2 min readLW link

GPT-4 is bad at strate­gic thinking

Christopher King27 Mar 2023 15:11 UTC
22 points
8 comments1 min readLW link

Speed of Failing

nano_brasca31 Dec 2023 10:39 UTC
8 points
0 comments2 min readLW link

[Question] Is Stu­pidity Ex­pand­ing? Some Hy­pothe­ses.

David Gross15 Oct 2020 3:28 UTC
71 points
42 comments5 min readLW link

[Linkpost] The AGI Show podcast

Soroush Pour23 May 2023 9:52 UTC
4 points
0 comments1 min readLW link

The de­fault sce­nario for the next 50 years

Julien24 Nov 2024 14:01 UTC
1 point
0 comments6 min readLW link

Con­crete vs Con­tex­tual values

whpearson2 Jun 2009 9:47 UTC
−1 points
32 comments3 min readLW link

[Question] Es­ti­mat­ing Re­turns to In­tel­li­gence vs Num­bers, Strength and Looks

TruePath31 Dec 2023 10:03 UTC
3 points
6 comments1 min readLW link

Ngo and Yud­kowsky on sci­en­tific rea­son­ing and pivotal acts

21 Feb 2022 20:54 UTC
67 points
14 comments35 min readLW link

Up­date on Devel­op­ing an Ethics Calcu­la­tor to Align an AGI to

sweenesm12 Mar 2024 12:33 UTC
4 points
2 comments8 min readLW link

Ex­plor­ing a New AGI Cog­ni­tive Ar­chi­tec­ture: Vi­sual Thought, Con­tra­dic­tion Re­s­olu­tion, and Safety

derek v9 Aug 2025 20:03 UTC
1 point
0 comments3 min readLW link

Beyond Re­wards and Values: A Non-du­al­is­tic Ap­proach to Univer­sal Intelligence

Akira Pyinya30 Dec 2022 19:05 UTC
10 points
4 comments14 min readLW link

In­te­gra­tion Band­width: The Mechanism Be­hind In­tel­li­gence and Puberty

Dortex5 Jun 2025 19:37 UTC
−1 points
4 comments1 min readLW link
(osf.io)

When An­thro­po­mor­phism Be­came Stupid

Eliezer Yudkowsky16 Aug 2008 23:43 UTC
57 points
12 comments3 min readLW link

The Align­ment Prob­lem No One Is Talk­ing About

James Stephen Brown10 May 2024 18:34 UTC
10 points
10 comments2 min readLW link
(nonzerosum.games)

What will the scaled up GATO look like? (Up­dated with ques­tions)

Amal 25 Oct 2022 12:44 UTC
34 points
22 comments1 min readLW link

Notes on Antelligence

Aurigena13 May 2023 18:38 UTC
2 points
0 comments9 min readLW link

[Question] What should a non-ge­nius do in the face of rapid progress in GAI to en­sure a de­cent life?

kaler1 Jan 2024 8:22 UTC
12 points
16 comments1 min readLW link

AGI and Friendly AI in the dom­i­nant AI textbook

lukeprog11 Mar 2011 4:12 UTC
80 points
27 comments3 min readLW link

I No Longer Believe In­tel­li­gence to be “Mag­i­cal”

DragonGod10 Jun 2022 8:58 UTC
28 points
34 comments6 min readLW link

A Gen­er­ally In­tel­li­gent Game

snerx28 Jun 2024 1:31 UTC
−1 points
1 comment4 min readLW link

The Fu­ture Value of Money - a new definition

B Bret30 Dec 2024 21:01 UTC
1 point
0 comments5 min readLW link

Most Ivy-smart stu­dents aren’t at Ivy-tier schools

Aaron Bergman7 Aug 2022 3:18 UTC
82 points
7 comments8 min readLW link
(www.aaronbergman.net)

[Question] What are some posthu­man­ist/​more-than-hu­man ap­proaches to defi­ni­tions of in­tel­li­gence and agency? Par­tic­u­larly in ap­pli­ca­tion to AI re­search.

Eli Hiton9 Apr 2024 21:52 UTC
1 point
0 comments1 min readLW link

[Question] How does some­one prove that their gen­eral in­tel­li­gence is above av­er­age?

M. Y. Zuo16 Sep 2024 21:01 UTC
−3 points
12 comments1 min readLW link

Aiming at the Target

Eliezer Yudkowsky26 Oct 2008 16:47 UTC
40 points
40 comments5 min readLW link

The Soul of the Writer (on LLMs, the psy­chol­ogy of writ­ers, and the na­ture of in­tel­li­gence)

rogersbacon16 Apr 2023 16:02 UTC
11 points
1 comment3 min readLW link
(www.secretorum.life)

The One That Isn’t There

Annoyance20 Nov 2009 20:10 UTC
18 points
6 comments3 min readLW link

[Question] When try­ing to define gen­eral in­tel­li­gence is abil­ity to achieve goals the best met­ric?

jmh22 Oct 2022 3:09 UTC
5 points
0 comments1 min readLW link

[Question] Why Do Peo­ple Think Hu­mans Are Stupid?

DragonGod14 Sep 2022 13:55 UTC
22 points
41 comments3 min readLW link

[Question] What is the na­ture of hu­mans gen­eral in­tel­li­gence and it’s im­pli­ca­tions for AGI?

Will_Pearson26 Mar 2024 15:20 UTC
5 points
4 comments1 min readLW link

[Question] [Dis­cus­sion] How Broad is the Hu­man Cog­ni­tive Spec­trum?

DragonGod7 Jan 2023 0:56 UTC
29 points
49 comments2 min readLW link

The AI Sus­tain­abil­ity Wager

dpatzer@orfai.net15 Aug 2025 19:45 UTC
1 point
0 comments2 min readLW link

How to solve the mi­suse prob­lem as­sum­ing that in 10 years the de­fault sce­nario is that AGI agents are ca­pa­ble of syn­thetiz­ing pathogens

jeremtti27 Nov 2024 21:17 UTC
6 points
0 comments9 min readLW link

The Hopium Wars: the AGI En­tente Delusion

Max Tegmark13 Oct 2024 17:00 UTC
236 points
60 comments9 min readLW link

Ex­plor­ing Emo­tional Align­ment: A Jour­ney of AI and Hu­man Con­nec­tion

Geumjae Ryu8 May 2025 0:56 UTC
1 point
0 comments1 min readLW link

[Yann Le­cun] A Path Towards Au­tonomous Ma­chine In­tel­li­gence

DragonGod27 Jun 2022 19:24 UTC
38 points
14 comments1 min readLW link
(openreview.net)

Prac­ti­cal ev­ery­day hu­man strategizing

akaTrickster27 Mar 2022 14:20 UTC
7 points
0 comments3 min readLW link

Shal­low vs. Deep Think­ing—Why LLMs Fall Short

Taylor G. Lunt3 Sep 2025 15:26 UTC
2 points
4 comments11 min readLW link

Joscha Bach on Syn­thetic In­tel­li­gence [an­no­tated]

Roman Leventov2 Mar 2023 11:02 UTC
10 points
1 comment9 min readLW link
(www.jimruttshow.com)

Ex­plor­ing a Vi­sion for AI as Com­pas­sion­ate, Emo­tion­ally In­tel­li­gent Part­ners — Seek­ing Col­lab­o­ra­tion and Insights

theophilos14 Jul 2025 23:22 UTC
1 point
0 comments1 min readLW link

Re­la­tion­ship be­tween sub­jec­tive ex­pe­rience and in­tel­li­gence?

Q Home24 Jul 2022 9:10 UTC
5 points
4 comments9 min readLW link

[Question] Will AGI have “hu­man” flaws?

Agustinus Theodorus23 Dec 2020 3:43 UTC
1 point
2 comments1 min readLW link

Gato’s Gen­er­al­i­sa­tion: Pre­dic­tions and Ex­per­i­ments I’d Like to See

Oliver Sourbut18 May 2022 7:15 UTC
43 points
3 comments10 min readLW link

The vir­tu­ous cir­cle: twelve con­jec­tures about fe­male re­pro­duc­tive agency and cul­tural self-determination

Miles Saltiel27 Dec 2023 18:25 UTC
0 points
2 comments14 min readLW link

Ex­pected Creative Surprises

Eliezer Yudkowsky24 Oct 2008 22:22 UTC
55 points
45 comments4 min readLW link

Mo­ral gauge the­ory: A spec­u­la­tive sug­ges­tion for AI alignment

James Diacoumis23 Feb 2025 11:42 UTC
6 points
2 comments8 min readLW link

Deep­Mind: Gen­er­ally ca­pa­ble agents emerge from open-ended play

Daniel Kokotajlo27 Jul 2021 14:19 UTC
247 points
53 comments2 min readLW link
(deepmind.com)

Towards a New On­tol­ogy of Intelligence

Tara4 Jun 2024 8:19 UTC
1 point
0 comments3 min readLW link

[Question] Why do Peo­ple Think In­tel­li­gence Will be “Easy”?

DragonGod12 Sep 2022 17:32 UTC
15 points
32 comments2 min readLW link

On agen­tic gen­er­al­ist mod­els: we’re es­sen­tially us­ing ex­ist­ing tech­nol­ogy the weak­est and worst way you can use it

Yuli_Ban28 Aug 2024 1:57 UTC
10 points
2 comments9 min readLW link

Syn­thetic Neuroscience

hpcfung25 Mar 2025 17:45 UTC
2 points
3 comments3 min readLW link

What boot­straps in­tel­li­gence?

invertedpassion10 Sep 2024 7:11 UTC
2 points
2 comments1 min readLW link

How to test your men­tal perfor­mance at the mo­ment?

taw23 Nov 2009 18:35 UTC
24 points
74 comments1 min readLW link

Another take on agent foun­da­tions: for­mal­iz­ing zero-shot reasoning

zhukeepa1 Jul 2018 6:12 UTC
64 points
20 comments12 min readLW link

“The Uni­verse of Minds”—call for re­view­ers (Seeds of Science)

rogersbacon25 Jul 2023 16:53 UTC
7 points
0 comments1 min readLW link

Are In­tel­li­gence and Gen­er­al­ity Orthog­o­nal?

cubefox18 Jul 2022 20:07 UTC
18 points
16 comments1 min readLW link