RSS

AI Takeoff

TagLast edit: Dec 30, 2024, 10:14 AM by Dakara

AI Takeoff is the process of an Artificial General Intelligence going from a certain threshold of capability (often discussed as “human-level”) to being super-intelligent and capable enough to control the fate of civilization. There has been much debate about whether AI takeoff is more likely to be slow vs fast, i.e., “soft” vs “hard”.

See also: AI Timelines, Seed AI, Singularity, Intelligence explosion, Recursive self-improvement

AI takeoff is sometimes casually referred to as AI FOOM.

Soft takeoff

A soft takeoff refers to an AGI that would self-improve over a period of years or decades. This could be due to either the learning algorithm being too demanding for the hardware or because the AI relies on experiencing feedback from the real-world that would have to be played out in real-time. Possible methods that could deliver a soft takeoff, by slowly building on human-level intelligence, are Whole brain emulation, Biological Cognitive Enhancement, and software-based strong AGI [1]. By maintaining control of the AGI’s ascent it should be easier for a Friendly AI to emerge.

Vernor Vinge, Hans Moravec and have all expressed the view that soft takeoff is preferable to a hard takeoff as it would be both safer and easier to engineer.

Hard takeoff

A hard takeoff (or an AI going “FOOM” [2]) refers to AGI expansion in a matter of minutes, days, or months. It is a fast, abruptly, local increase in capability. This scenario is widely considered much more precarious, as this involves an AGI rapidly ascending in power without human control. This may result in unexpected or undesired behavior (i.e. Unfriendly AI). It is one of the main ideas supporting the Intelligence explosion hypothesis.

The feasibility of hard takeoff has been addressed by Hugo de Garis, Eliezer Yudkowsky, Ben Goertzel, Nick Bostrom, and Michael Anissimov. It is widely agreed that a hard takeoff is something to be avoided due to the risks. Yudkowsky points out several possibilities that would make a hard takeoff more likely than a soft takeoff such as the existence of large resources overhangs or the fact that small improvements seem to have a large impact in a mind’s general intelligence (i.e.: the small genetic difference between humans and chimps lead to huge increases in capability) [3].

Notable posts

External links

References

  1. http://​​www.aleph.se/​​andart/​​archives/​​2010/​​10/​​why_early_singularities_are_softer.html

  2. http://​​lesswrong.com/​​lw/​​63t/​​requirements_for_ai_to_go_foom/​​

  3. http://​​lesswrong.com/​​lw/​​wf/​​hard_takeoff/​​

New re­port: In­tel­li­gence Ex­plo­sion Microeconomics

Eliezer YudkowskyApr 29, 2013, 11:14 PM
74 points

50 votes

Overall karma indicates overall quality.

246 comments3 min readLW link

AlphaGo Zero and the Foom Debate

Eliezer YudkowskyOct 21, 2017, 2:18 AM
98 points

53 votes

Overall karma indicates overall quality.

17 comments3 min readLW link

Ar­gu­ments about fast takeoff

paulfchristianoFeb 25, 2018, 4:53 AM
103 points

66 votes

Overall karma indicates overall quality.

68 comments2 min readLW link1 review
(sideways-view.com)

Will AI un­dergo dis­con­tin­u­ous progress?

Sammy MartinFeb 21, 2020, 10:16 PM
27 points

20 votes

Overall karma indicates overall quality.

21 comments20 min readLW link

Will AI See Sud­den Progress?

KatjaGraceFeb 26, 2018, 12:41 AM
27 points

18 votes

Overall karma indicates overall quality.

11 comments1 min readLW link1 review

Dis­con­tin­u­ous progress in his­tory: an update

KatjaGraceApr 14, 2020, 12:00 AM
191 points

76 votes

Overall karma indicates overall quality.

25 comments31 min readLW link1 review
(aiimpacts.org)

Quick Nate/​Eliezer com­ments on discontinuity

Rob BensingerMar 1, 2018, 10:03 PM
44 points

27 votes

Overall karma indicates overall quality.

1 comment2 min readLW link

Soft take­off can still lead to de­ci­sive strate­gic advantage

Daniel KokotajloAug 23, 2019, 4:39 PM
122 points

50 votes

Overall karma indicates overall quality.

47 comments8 min readLW link4 reviews

Take­off Speeds and Discontinuities

Sep 30, 2021, 1:50 PM
63 points

19 votes

Overall karma indicates overall quality.

1 comment15 min readLW link

Against GDP as a met­ric for timelines and take­off speeds

Daniel KokotajloDec 29, 2020, 5:42 PM
140 points

51 votes

Overall karma indicates overall quality.

19 comments14 min readLW link1 review

Model­ling Con­tin­u­ous Progress

Sammy MartinJun 23, 2020, 6:06 PM
30 points

12 votes

Overall karma indicates overall quality.

3 comments7 min readLW link

Yud­kowsky and Chris­ti­ano dis­cuss “Take­off Speeds”

Eliezer YudkowskyNov 22, 2021, 7:35 PM
210 points

73 votes

Overall karma indicates overall quality.

176 comments60 min readLW link1 review

Con­jec­ture in­ter­nal sur­vey: AGI timelines and prob­a­bil­ity of hu­man ex­tinc­tion from ad­vanced AI

Maris SalaMay 22, 2023, 2:31 PM
155 points

75 votes

Overall karma indicates overall quality.

5 comments3 min readLW link
(www.conjecture.dev)

Dist­in­guish­ing defi­ni­tions of takeoff

Matthew BarnettFeb 14, 2020, 12:16 AM
79 points

32 votes

Overall karma indicates overall quality.

6 comments6 min readLW link

Analo­gies and Gen­eral Pri­ors on Intelligence

Aug 20, 2021, 9:03 PM
57 points

22 votes

Overall karma indicates overall quality.

12 comments14 min readLW link

Re­view of Soft Take­off Can Still Lead to DSA

Daniel KokotajloJan 10, 2021, 6:10 PM
85 points

20 votes

Overall karma indicates overall quality.

16 comments6 min readLW link

Why all the fuss about re­cur­sive self-im­prove­ment?

So8resJun 12, 2022, 8:53 PM
159 points

86 votes

Overall karma indicates overall quality.

62 comments7 min readLW link1 review

My cur­rent frame­work for think­ing about AGI timelines

zhukeepaMar 30, 2020, 1:23 AM
107 points

52 votes

Overall karma indicates overall quality.

5 comments3 min readLW link

Towards a For­mal­i­sa­tion of Re­turns on Cog­ni­tive Rein­vest­ment (Part 1)

DragonGodJun 4, 2022, 6:42 PM
17 points

4 votes

Overall karma indicates overall quality.

11 comments13 min readLW link

Con­tin­u­ing the take­offs debate

Richard_NgoNov 23, 2020, 3:58 PM
67 points

17 votes

Overall karma indicates overall quality.

11 comments9 min readLW link

Pos­si­ble take­aways from the coro­n­avirus pan­demic for slow AI takeoff

VikaMay 31, 2020, 5:51 PM
135 points

61 votes

Overall karma indicates overall quality.

36 comments3 min readLW link1 review

Every­thing I Need To Know About Take­off Speeds I Learned From Air Con­di­tioner Rat­ings On Amazon

johnswentworthApr 15, 2022, 7:05 PM
166 points

124 votes

Overall karma indicates overall quality.

128 comments5 min readLW link

Re­quire­ments for a STEM-ca­pa­ble AGI Value Learner (my Case for Less Doom)

RogerDearnaleyMay 25, 2023, 9:26 AM
33 points

9 votes

Overall karma indicates overall quality.

3 comments15 min readLW link

“Hereti­cal Thoughts on AI” by Eli Dourado

DragonGodJan 19, 2023, 4:11 PM
146 points

65 votes

Overall karma indicates overall quality.

38 comments3 min readLW link
(www.elidourado.com)

AI take­off story: a con­tinu­a­tion of progress by other means

Edouard HarrisSep 27, 2021, 3:55 PM
76 points

33 votes

Overall karma indicates overall quality.

13 comments10 min readLW link

The date of AI Takeover is not the day the AI takes over

Daniel KokotajloOct 22, 2020, 10:41 AM
155 points

88 votes

Overall karma indicates overall quality.

32 comments2 min readLW link1 review

Take­off speeds, the chimps anal­ogy, and the Cul­tural In­tel­li­gence Hypothesis

NickGabsDec 2, 2022, 7:14 PM
16 points

13 votes

Overall karma indicates overall quality.

2 comments4 min readLW link

Could Ad­vanced AI Ac­cel­er­ate the Pace of AI Progress? In­ter­views with AI Researchers

Mar 3, 2025, 7:05 PM
41 points

21 votes

Overall karma indicates overall quality.

1 comment1 min readLW link
(papers.ssrn.com)

More Chris­ti­ano, Co­tra, and Yud­kowsky on AI progress

Dec 6, 2021, 8:33 PM
91 points

31 votes

Overall karma indicates overall quality.

28 comments40 min readLW link

Re­view Re­port of David­son on Take­off Speeds (2023)

Trent KannegieterDec 22, 2023, 6:48 PM
37 points

16 votes

Overall karma indicates overall quality.

11 comments38 min readLW link

A com­pressed take on re­cent disagreements

kmanJul 4, 2022, 4:39 AM
33 points

13 votes

Overall karma indicates overall quality.

9 comments1 min readLW link

Brain Effi­ciency: Much More than You Wanted to Know

jacob_cannellJan 6, 2022, 3:38 AM
209 points

141 votes

Overall karma indicates overall quality.

103 comments29 min readLW link

[Question] Is there a name for the the­ory that “There will be fast take­off in real-world ca­pa­bil­ities be­cause al­most ev­ery­thing is AGI-com­plete”?

David Scott Krueger (formerly: capybaralet)Sep 2, 2021, 11:00 PM
31 points

9 votes

Overall karma indicates overall quality.

8 comments1 min readLW link

What Ev­i­dence Is AlphaGo Zero Re AGI Com­plex­ity?

RobinHansonOct 22, 2017, 2:28 AM
37 points

41 votes

Overall karma indicates overall quality.

44 comments2 min readLW link

What 2026 looks like

Daniel KokotajloAug 6, 2021, 4:14 PM
589 points

352 votes

Overall karma indicates overall quality.

166 comments16 min readLW link1 review

Reflec­tions on the state of the race to su­per­in­tel­li­gence, Fe­bru­ary 2025

Mitchell_PorterFeb 23, 2025, 1:58 PM
21 points

9 votes

Overall karma indicates overall quality.

7 comments4 min readLW link

How might we safely pass the buck to AI?

joshcFeb 19, 2025, 5:48 PM
83 points

53 votes

Overall karma indicates overall quality.

58 comments31 min readLW link

My Thoughts on Take­off Speeds

tristanmMar 27, 2018, 12:05 AM
11 points

8 votes

Overall karma indicates overall quality.

2 comments7 min readLW link

Con­ver­sa­tion on tech­nol­ogy fore­cast­ing and gradualism

Dec 9, 2021, 9:23 PM
108 points

33 votes

Overall karma indicates overall quality.

30 comments31 min readLW link

Takes on Takeoff

atharvaMar 25, 2025, 12:20 AM
10 points

5 votes

Overall karma indicates overall quality.

0 comments4 min readLW link

[Question] Math­e­mat­i­cal Models of Progress?

abramdemskiFeb 16, 2021, 12:21 AM
28 points

8 votes

Overall karma indicates overall quality.

8 comments2 min readLW link

Loose thoughts on AGI risk

YitzJun 23, 2022, 1:02 AM
7 points

2 votes

Overall karma indicates overall quality.

3 comments1 min readLW link

More on dis­am­biguat­ing “dis­con­ti­nu­ity”

Aryeh EnglanderJun 9, 2020, 3:16 PM
16 points

7 votes

Overall karma indicates overall quality.

1 comment3 min readLW link

AI Align­ment 2018-19 Review

Rohin ShahJan 28, 2020, 2:19 AM
126 points

44 votes

Overall karma indicates overall quality.

6 comments35 min readLW link

S-Curves for Trend Forecasting

Matt GoldenbergJan 23, 2019, 6:17 PM
113 points

47 votes

Overall karma indicates overall quality.

23 comments7 min readLW link4 reviews

The cru­cible — how I think about the situ­a­tion with AI

owencbMay 5, 2025, 1:18 PM
25 points

8 votes

Overall karma indicates overall quality.

1 comment8 min readLW link
(strangecities.substack.com)

Ta­boo “com­pute over­hang”

Zach Stein-PerlmanMar 1, 2023, 7:15 PM
21 points

22 votes

Overall karma indicates overall quality.

8 comments1 min readLW link

Stan­ford En­cy­clo­pe­dia of Philos­o­phy on AI ethics and superintelligence

Kaj_SotalaMay 2, 2020, 7:35 AM
43 points

19 votes

Overall karma indicates overall quality.

19 comments7 min readLW link
(plato.stanford.edu)

Cy­borg Pe­ri­ods: There will be mul­ti­ple AI transitions

Feb 22, 2023, 4:09 PM
109 points

62 votes

Overall karma indicates overall quality.

9 comments6 min readLW link

We don’t un­der­stand what hap­pened with cul­ture enough

Jan_KulveitOct 9, 2023, 9:54 AM
87 points

46 votes

Overall karma indicates overall quality.

22 comments6 min readLW link1 review

AGI-Au­to­mated In­ter­pretabil­ity is Suicide

__RicG__May 10, 2023, 2:20 PM
25 points

64 votes

Overall karma indicates overall quality.

33 comments7 min readLW link

How do take­off speeds af­fect the prob­a­bil­ity of bad out­comes from AGI?

KRJun 29, 2020, 10:06 PM
15 points

7 votes

Overall karma indicates overall quality.

2 comments8 min readLW link

Nice-ish, smooth take­off (with im­perfect safe­guards) prob­a­bly kills most “clas­sic hu­mans” in a few decades.

RaemonOct 2, 2025, 9:03 PM
155 points

63 votes

Overall karma indicates overall quality.

19 comments12 min readLW link

Three Types of In­tel­li­gence Explosion

Mar 17, 2025, 2:47 PM
40 points

17 votes

Overall karma indicates overall quality.

8 comments3 min readLW link
(www.forethought.org)

[Question] Is driv­ing worth the risk?

Adam ZernerMay 11, 2021, 5:04 AM
28 points

16 votes

Overall karma indicates overall quality.

29 comments7 min readLW link

For FAI: Is “Molec­u­lar Nan­otech­nol­ogy” putting our best foot for­ward?

leplenJun 22, 2013, 4:44 AM
86 points

73 votes

Overall karma indicates overall quality.

118 comments3 min readLW link

“Sharp Left Turn” dis­course: An opinionated review

Steven ByrnesJan 28, 2025, 6:47 PM
219 points

67 votes

Overall karma indicates overall quality.

31 comments31 min readLW link

Fac­to­rio, Ac­celerando, Em­pathiz­ing with Em­pires and Moder­ate Takeoffs

RaemonFeb 4, 2018, 2:33 AM
53 points

30 votes

Overall karma indicates overall quality.

19 comments4 min readLW link

Hyper­bolic takeoff

Ege ErdilApr 9, 2022, 3:57 PM
18 points

8 votes

Overall karma indicates overall quality.

7 comments10 min readLW link
(www.metaculus.com)

I Believe we are in a Hard­ware Overhang

nemDec 8, 2022, 11:18 PM
8 points

4 votes

Overall karma indicates overall quality.

0 comments1 min readLW link

Pre­face to the se­quence on eco­nomic growth

Matthew BarnettAug 27, 2020, 8:29 PM
51 points

21 votes

Overall karma indicates overall quality.

0 comments4 min readLW link

...Re­cur­sion, Magic

Eliezer YudkowskyNov 25, 2008, 9:10 AM
34 points

21 votes

Overall karma indicates overall quality.

28 comments5 min readLW link

Hu­man-level is not the limit

Apr 16, 2025, 8:33 AM
23 points

8 votes

Overall karma indicates overall quality.

2 comments2 min readLW link
(aisafety.info)

Cas­cades, Cy­cles, In­sight...

Eliezer YudkowskyNov 24, 2008, 9:33 AM
35 points

22 votes

Overall karma indicates overall quality.

31 comments8 min readLW link

An­nounc­ing Epoch: A re­search or­ga­ni­za­tion in­ves­ti­gat­ing the road to Trans­for­ma­tive AI

Jun 27, 2022, 1:55 PM
97 points

45 votes

Overall karma indicates overall quality.

2 comments2 min readLW link
(epochai.org)

AI-202X: a game be­tween hu­mans and AGIs al­igned to differ­ent fu­tures?

StanislavKrymJul 1, 2025, 11:37 PM
6 points

4 votes

Overall karma indicates overall quality.

0 comments16 min readLW link

[Question] First and Last Ques­tions for GPT-5*

Mitchell_PorterNov 24, 2023, 5:03 AM
20 points

10 votes

Overall karma indicates overall quality.

5 comments1 min readLW link

The road from hu­man-level to su­per­in­tel­li­gent AI may be short

Apr 16, 2025, 8:35 AM
10 points

4 votes

Overall karma indicates overall quality.

0 comments2 min readLW link
(aisafety.info)

How much novel se­cu­rity-crit­i­cal in­fras­truc­ture do you need dur­ing the sin­gu­lar­ity?

BuckJul 4, 2025, 4:54 PM
56 points

17 votes

Overall karma indicates overall quality.

7 comments5 min readLW link

Robin Han­son & Liron Shapira De­bate AI X-Risk

LironJul 8, 2024, 9:45 PM
41 points

12 votes

Overall karma indicates overall quality.

4 comments1 min readLW link
(www.youtube.com)

Train­ing AGI in Se­cret would be Un­safe and Unethical

Daniel KokotajloApr 18, 2025, 12:27 PM
140 points

57 votes

Overall karma indicates overall quality.

15 comments6 min readLW link

The prospect of ac­cel­er­ated AI safety progress, in­clud­ing philo­soph­i­cal progress

Mitchell_PorterMar 13, 2025, 10:52 AM
11 points

3 votes

Overall karma indicates overall quality.

0 comments4 min readLW link

AIs will greatly change en­g­ineer­ing in AI com­pa­nies well be­fore AGI

ryan_greenblattSep 9, 2025, 4:58 PM
46 points

22 votes

Overall karma indicates overall quality.

9 comments11 min readLW link

Why AI may not foom

John_MaxwellMar 24, 2013, 8:11 AM
29 points

37 votes

Overall karma indicates overall quality.

81 comments12 min readLW link

Hard Takeoff

Eliezer YudkowskyDec 2, 2008, 8:44 PM
36 points

32 votes

Overall karma indicates overall quality.

34 comments11 min readLW link

Why I Think Abrupt AI Takeoff

lincolnquirkJul 17, 2022, 5:04 PM
14 points

6 votes

Overall karma indicates overall quality.

6 comments1 min readLW link

Vi­talik’s Re­sponse to AI 2027

Daniel KokotajloJul 11, 2025, 9:43 PM
120 points

54 votes

Overall karma indicates overall quality.

53 comments12 min readLW link
(vitalik.eth.limo)

The fast take­off motte/​bailey

lcFeb 24, 2023, 7:11 AM
−2 points

19 votes

Overall karma indicates overall quality.

7 comments1 min readLW link

Bet­ter than log­a­r­ith­mic re­turns to rea­son­ing?

Oliver SourbutJul 30, 2025, 12:50 AM
14 points

4 votes

Overall karma indicates overall quality.

5 comments2 min readLW link

Take­off Speed: Sim­ple Asymp­totics in a Toy Model.

Aaron RothMar 5, 2018, 5:07 PM
21 points

9 votes

Overall karma indicates overall quality.

21 comments9 min readLW link
(aaronsadventures.blogspot.com)

Will the Need to Re­train AI Models from Scratch Block a Soft­ware In­tel­li­gence Ex­plo­sion?

Tom DavidsonMar 28, 2025, 2:12 PM
10 points

4 votes

Overall karma indicates overall quality.

0 comments3 min readLW link

A sum­mary of the Han­son-Yud­kowsky FOOM debate

Kaj_SotalaNov 15, 2012, 7:25 AM
42 points

29 votes

Overall karma indicates overall quality.

10 comments1 min readLW link

Slow cor­po­ra­tions as an in­tu­ition pump for AI R&D automation

May 9, 2025, 2:49 PM
91 points

36 votes

Overall karma indicates overall quality.

23 comments9 min readLW link

[Question] Is “Re­cur­sive Self-Im­prove­ment” Rele­vant in the Deep Learn­ing Paradigm?

DragonGodApr 6, 2023, 7:13 AM
32 points

19 votes

Overall karma indicates overall quality.

36 comments7 min readLW link

State of play of AI progress (and re­lated brakes on an in­tel­li­gence ex­plo­sion) [Linkpost]

Noosphere89Apr 30, 2025, 7:58 PM
7 points

1 vote

Overall karma indicates overall quality.

0 comments5 min readLW link
(www.interconnects.ai)

Knowl­edge, Rea­son­ing, and Superintelligence

owencbMar 26, 2025, 11:28 PM
21 points

5 votes

Overall karma indicates overall quality.

1 comment7 min readLW link
(strangecities.substack.com)

[AN #97]: Are there his­tor­i­cal ex­am­ples of large, ro­bust dis­con­ti­nu­ities?

Rohin ShahApr 29, 2020, 5:30 PM
15 points

5 votes

Overall karma indicates overall quality.

0 comments10 min readLW link
(mailchi.mp)

An “AI re­searcher” has writ­ten a pa­per on op­ti­miz­ing AI ar­chi­tec­ture and op­ti­mized a lan­guage model to sev­eral or­ders of mag­ni­tude more effi­ciency.

Y BMar 18, 2025, 1:15 AM
3 points

4 votes

Overall karma indicates overall quality.

1 comment1 min readLW link

For ev­ery choice of AGI difficulty, con­di­tion­ing on grad­ual take-off im­plies shorter timelines.

Francis Rhys WardApr 21, 2022, 7:44 AM
31 points

22 votes

Overall karma indicates overall quality.

13 comments3 min readLW link

A model of the fi­nal phase: the cur­rent fron­tier AIs as de facto CEOs of their own com­pa­nies

Mitchell_PorterMar 8, 2025, 10:15 PM
23 points

8 votes

Overall karma indicates overall quality.

2 comments1 min readLW link

“Slow” take­off is a ter­rible term for “maybe even faster take­off, ac­tu­ally”

RaemonSep 28, 2024, 11:38 PM
220 points

100 votes

Overall karma indicates overall quality.

69 comments1 min readLW link

Carl Shul­man on The Lu­nar So­ciety (7 hour, two-part pod­cast)

ESRogsJun 28, 2023, 1:23 AM
79 points

28 votes

Overall karma indicates overall quality.

17 comments1 min readLW link
(www.dwarkeshpatel.com)

U.S.-China Eco­nomic and Se­cu­rity Re­view Com­mis­sion pushes Man­hat­tan Pro­ject-style AI initiative

worseNov 19, 2024, 6:42 PM
56 points

17 votes

Overall karma indicates overall quality.

7 comments1 min readLW link

Some con­cep­tual high­lights from “Disjunc­tive Sce­nar­ios of Catas­trophic AI Risk”

Kaj_SotalaFeb 12, 2018, 12:30 PM
45 points

22 votes

Overall karma indicates overall quality.

4 comments6 min readLW link
(kajsotala.fi)

Fast Take­off in Biolog­i­cal Intelligence

eapacheApr 25, 2020, 12:21 PM
14 points

7 votes

Overall karma indicates overall quality.

21 comments2 min readLW link

hu­man psy­chol­in­guists: a crit­i­cal appraisal

nostalgebraistDec 31, 2019, 12:20 AM
185 points

95 votes

Overall karma indicates overall quality.

60 comments16 min readLW link2 reviews
(nostalgebraist.tumblr.com)

[Question] Is there a cul­ture over­hang?

Aleksi LiimatainenOct 3, 2022, 7:26 AM
18 points

9 votes

Overall karma indicates overall quality.

4 comments1 min readLW link

Crit­i­cal re­view of Chris­ti­ano’s dis­agree­ments with Yudkowsky

Vanessa KosoyDec 27, 2023, 4:02 PM
176 points

74 votes

Overall karma indicates overall quality.

40 comments15 min readLW link

AI over­hangs de­pend on whether al­gorithms, com­pute and data are sub­sti­tutes or complements

NathanBarnardDec 16, 2022, 2:23 AM
4 points

2 votes

Overall karma indicates overall quality.

0 comments3 min readLW link

[Link] Sarah Con­stantin: “Why I am Not An AI Doomer”

lbThingrbApr 12, 2023, 1:52 AM
61 points

19 votes

Overall karma indicates overall quality.

13 comments1 min readLW link
(sarahconstantin.substack.com)

Ngo and Yud­kowsky on AI ca­pa­bil­ity gains

Nov 18, 2021, 10:19 PM
131 points

43 votes

Overall karma indicates overall quality.

61 comments38 min readLW link1 review

Ac­celerando as a “Slow, Rea­son­ably Nice Take­off” Story

RaemonSep 23, 2025, 2:15 AM
73 points

28 votes

Overall karma indicates overall quality.

20 comments30 min readLW link

Chris­ti­ano, Co­tra, and Yud­kowsky on AI progress

Nov 25, 2021, 4:45 PM
119 points

39 votes

Overall karma indicates overall quality.

95 comments66 min readLW link

What a com­pute-cen­tric frame­work says about AI take­off speeds

Tom DavidsonJan 23, 2023, 4:02 AM
188 points

79 votes

Overall karma indicates overall quality.

30 comments16 min readLW link1 review

How quick and big would a soft­ware in­tel­li­gence ex­plo­sion be?

Jul 22, 2025, 12:58 PM
42 points

17 votes

Overall karma indicates overall quality.

23 comments34 min readLW link
(www.forethought.org)

MIRI Con­ver­sa­tions: Tech­nol­ogy Fore­cast­ing & Grad­u­al­ism (Distil­la­tion)

CallumMcDougallJul 13, 2022, 3:55 PM
31 points

15 votes

Overall karma indicates overall quality.

1 comment20 min readLW link

What does 10x-ing effec­tive com­pute get you?

ryan_greenblattJun 24, 2025, 6:33 PM
55 points

19 votes

Overall karma indicates overall quality.

10 comments12 min readLW link

An AI Real­ist Man­i­festo: Nei­ther Doomer nor Foomer, but a third more rea­son­able thing

PashaKamyshevApr 10, 2023, 12:11 AM
16 points

14 votes

Overall karma indicates overall quality.

13 comments8 min readLW link

My Overview of the AI Align­ment Land­scape: A Bird’s Eye View

Neel NandaDec 15, 2021, 11:44 PM
127 points

74 votes

Overall karma indicates overall quality.

9 comments15 min readLW link

BCIs and the ecosys­tem of mod­u­lar minds

berenJul 21, 2023, 3:58 PM
88 points

34 votes

Overall karma indicates overall quality.

14 comments11 min readLW link

[Question] Any re­but­tals of Chris­ti­ano and AI Im­pacts on take­off speeds?

SoerenMindApr 21, 2019, 8:39 PM
67 points

26 votes

Overall karma indicates overall quality.

26 comments1 min readLW link

Ex­treme GDP growth is a bad op­er­at­ing defi­ni­tion of “slow take­off”

lcMar 1, 2023, 10:25 PM
24 points

10 votes

Overall karma indicates overall quality.

1 comment1 min readLW link

AI 2027 Thoughts

PeterMcCluskeyApr 26, 2025, 12:00 AM
29 points

11 votes

Overall karma indicates overall quality.

2 comments6 min readLW link
(bayesianinvestor.com)

AI take­off and nu­clear war

owencbJun 11, 2024, 7:36 PM
80 points

23 votes

Overall karma indicates overall quality.

6 comments11 min readLW link
(strangecities.substack.com)

[LINK] What should a rea­son­able per­son be­lieve about the Sin­gu­lar­ity?

Kaj_SotalaJan 13, 2011, 9:32 AM
38 points

30 votes

Overall karma indicates overall quality.

14 comments2 min readLW link

[Question] Prob­a­bil­ity that other ar­chi­tec­tures will scale as well as Trans­form­ers?

Daniel KokotajloJul 28, 2020, 7:36 PM
22 points

8 votes

Overall karma indicates overall quality.

4 comments1 min readLW link

Quick thoughts on the im­pli­ca­tions of multi-agent views of mind on AI takeover

Kaj_SotalaDec 11, 2023, 6:34 AM
48 points

24 votes

Overall karma indicates overall quality.

14 comments4 min readLW link

[Question] Re­sponses to Chris­ti­ano on take­off speeds?

Richard_NgoOct 30, 2020, 3:16 PM
29 points

11 votes

Overall karma indicates overall quality.

7 comments1 min readLW link

Why you might ex­pect ho­mo­ge­neous take-off: ev­i­dence from ML research

Andrei AlexandruJul 17, 2022, 8:31 PM
24 points

10 votes

Overall karma indicates overall quality.

0 comments10 min readLW link

Some thoughts point­ing to slower AI take-off

BastiaanFeb 27, 2023, 7:53 PM
8 points

9 votes

Overall karma indicates overall quality.

2 comments4 min readLW link

Papers for 2017

Kaj_SotalaJan 4, 2018, 1:30 PM
12 points

8 votes

Overall karma indicates overall quality.

2 comments2 min readLW link
(kajsotala.fi)

Emer­gence of su­per­in­tel­li­gence from AI hive­minds: how to make it hu­man-friendly?

Mitchell_PorterApr 27, 2025, 4:51 AM
12 points

3 votes

Overall karma indicates overall quality.

0 comments2 min readLW link

Grad­ual Disem­pow­er­ment: Sys­temic Ex­is­ten­tial Risks from In­cre­men­tal AI Development

Jan 30, 2025, 5:03 PM
167 points

61 votes

Overall karma indicates overall quality.

65 comments2 min readLW link
(gradual-disempowerment.ai)

Are short timelines ac­tu­ally bad?

joshcFeb 5, 2023, 9:21 PM
61 points

34 votes

Overall karma indicates overall quality.

7 comments3 min readLW link

Is Gen­eral In­tel­li­gence “Com­pact”?

DragonGodJul 4, 2022, 1:27 PM
27 points

13 votes

Overall karma indicates overall quality.

6 comments22 min readLW link

Foom & Doom 1: “Brain in a box in a base­ment”

Steven ByrnesJun 23, 2025, 5:18 PM
281 points

102 votes

Overall karma indicates overall quality.

120 comments29 min readLW link

Life and ex­pand­ing steer­able consequences

Alex FlintMay 7, 2021, 6:33 PM
46 points

15 votes

Overall karma indicates overall quality.

3 comments4 min readLW link

[Question] What’s the like­li­hood of only sub ex­po­nen­tial growth for AGI?

M. Y. ZuoNov 13, 2021, 10:46 PM
5 points

4 votes

Overall karma indicates overall quality.

22 comments1 min readLW link

Evolu­tion pro­vides no ev­i­dence for the sharp left turn

Quintin PopeApr 11, 2023, 6:43 PM
206 points

152 votes

Overall karma indicates overall quality.

65 comments15 min readLW link1 review

Frame for Take-Off Speeds to in­form com­pute gov­er­nance & scal­ing alignment

Logan RiggsMay 13, 2022, 10:23 PM
15 points

7 votes

Overall karma indicates overall quality.

2 comments2 min readLW link

Shul­man and Yud­kowsky on AI progress

Dec 3, 2021, 8:05 PM
90 points

28 votes

Overall karma indicates overall quality.

16 comments20 min readLW link

“Tech com­pany sin­gu­lar­i­ties”, and steer­ing them to re­duce x-risk

Andrew_CritchMay 13, 2022, 5:24 PM
75 points

38 votes

Overall karma indicates overall quality.

11 comments4 min readLW link

Wargam­ing AGI Development

ryan_bMar 19, 2022, 5:59 PM
37 points

13 votes

Overall karma indicates overall quality.

10 comments5 min readLW link

GPT-2030 and Catas­trophic Drives: Four Vignettes

jsteinhardtNov 10, 2023, 7:30 AM
50 points

15 votes

Overall karma indicates overall quality.

5 comments10 min readLW link
(bounded-regret.ghost.io)

Will AI R&D Au­toma­tion Cause a Soft­ware In­tel­li­gence Ex­plo­sion?

Mar 26, 2025, 6:12 PM
19 points

8 votes

Overall karma indicates overall quality.

3 comments2 min readLW link
(www.forethought.org)

Take­off speeds have a huge effect on what it means to work on AI x-risk

BuckApr 13, 2022, 5:38 PM
139 points

69 votes

Overall karma indicates overall quality.

27 comments2 min readLW link2 reviews

[Question] Does the AI con­trol agenda broadly rely on no FOOM be­ing pos­si­ble?

Noosphere89Mar 29, 2025, 7:38 PM
22 points

7 votes

Overall karma indicates overall quality.

3 comments1 min readLW link

Short Timelines Don’t De­value Long Hori­zon Research

Vladimir_NesovApr 9, 2025, 12:42 AM
172 points

69 votes

Overall karma indicates overall quality.

24 comments1 min readLW link

Con­ti­nu­ity Assumptions

Jan_KulveitJun 13, 2022, 9:31 PM
53 points

29 votes

Overall karma indicates overall quality.

13 comments4 min readLW link

Mis­con­cep­tions about con­tin­u­ous takeoff

Matthew BarnettOct 8, 2019, 9:31 PM
82 points

37 votes

Overall karma indicates overall quality.

38 comments4 min readLW link

An illus­tra­tive model of back­fire risks from paus­ing AI research

Maxime RichéNov 6, 2023, 2:30 PM
33 points

17 votes

Overall karma indicates overall quality.

3 comments11 min readLW link

Po­ten­tial gears level ex­pla­na­tions of smooth progress

ryan_greenblattDec 22, 2021, 6:05 PM
4 points

2 votes

Overall karma indicates overall quality.

2 comments2 min readLW link

WHO Biolog­i­cal Risk warning

Jonas KgomoApr 25, 2023, 3:10 PM
−6 points

4 votes

Overall karma indicates overall quality.

2 comments1 min readLW link

Use com­put­ers as pow­er­ful as in 1985 or AI con­trols hu­mans or ?

jrincaycFeb 3, 2025, 12:51 AM
3 points

3 votes

Overall karma indicates overall quality.

0 comments2 min readLW link

Agent 002: A story about how ar­tifi­cial in­tel­li­gence might soon de­stroy humanity

Jakub GrowiecJul 23, 2025, 1:56 PM
5 points

2 votes

Overall karma indicates overall quality.

0 comments26 min readLW link

Se­cu­rity Mind­set and Take­off Speeds

DanielFilanOct 27, 2020, 3:20 AM
55 points

15 votes

Overall karma indicates overall quality.

23 comments8 min readLW link
(danielfilan.com)

What role should evolu­tion­ary analo­gies play in un­der­stand­ing AI take­off speeds?

anson.hoDec 11, 2021, 1:19 AM
14 points

8 votes

Overall karma indicates overall quality.

0 comments42 min readLW link

AI Align­ment and the Quest for Ar­tifi­cial Wisdom

MyspyJul 12, 2024, 9:34 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments13 min readLW link

OpenAI Credit Ac­count (2510$)

Emirhan BULUTJan 21, 2024, 2:32 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

A short es­say on Illusions

cris.Sep 1, 2025, 10:25 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments3 min readLW link

How I Learned To Stop Wor­ry­ing And Love The Shoggoth

Peter MerelJul 12, 2023, 5:47 PM
9 points

8 votes

Overall karma indicates overall quality.

15 comments5 min readLW link

AIOS

samhealyDec 31, 2023, 1:23 PM
−3 points

5 votes

Overall karma indicates overall quality.

5 comments6 min readLW link

What are the differ­ences be­tween a sin­gu­lar­ity, an in­tel­li­gence ex­plo­sion, and a hard take­off?

Apr 3, 2025, 10:37 AM
5 points

3 votes

Overall karma indicates overall quality.

0 comments2 min readLW link
(aisafety.info)

True Sources of Disagreement

Eliezer YudkowskyDec 8, 2008, 3:51 PM
12 points

10 votes

Overall karma indicates overall quality.

53 comments8 min readLW link

Dou­ble Crux­ing the AI Foom debate

agilecavemanApr 27, 2018, 6:46 AM
17 points

12 votes

Overall karma indicates overall quality.

3 comments11 min readLW link

Places of Lov­ing Grace [Story]

ankFeb 18, 2025, 11:49 PM
−1 points

2 votes

Overall karma indicates overall quality.

0 comments4 min readLW link

What if AI doesn’t quite go FOOM?

Mass_DriverJun 20, 2010, 12:03 AM
16 points

22 votes

Overall karma indicates overall quality.

191 comments5 min readLW link

Do not miss the cut­off for im­mor­tal­ity! There is a prob­a­bil­ity that you will live for­ever as an im­mor­tal su­per­in­tel­li­gent be­ing and you can in­crease your odds by con­vinc­ing oth­ers to make achiev­ing the tech­nolog­i­cal sin­gu­lar­ity as quickly and safely as pos­si­ble the col­lec­tive goal/​pro­ject of all of hu­man­ity, Similar to “Fable of the Dragon-Tyrant.”

Oliver--KlozoffJun 29, 2023, 3:45 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments28 min readLW link

AIs should also re­fuse to work on ca­pa­bil­ities research

DavidmanheimOct 27, 2025, 8:42 AM
140 points

53 votes

Overall karma indicates overall quality.

19 comments3 min readLW link

Three Align­ment Schemas & Their Problems

Shoshannah TekofskyNov 26, 2022, 4:25 AM
19 points

11 votes

Overall karma indicates overall quality.

1 comment6 min readLW link

Su­per­in­tel­li­gence Align­ment Proposal

Davey MorseFeb 3, 2025, 6:47 PM
5 points

13 votes

Overall karma indicates overall quality.

3 comments9 min readLW link

Agent level parallelism

Johannes C. MayerJun 18, 2022, 8:56 PM
5 points

3 votes

Overall karma indicates overall quality.

5 comments1 min readLW link

Take­off speeds pre­sen­ta­tion at Anthropic

Tom DavidsonJun 4, 2024, 10:46 PM
93 points

35 votes

Overall karma indicates overall quality.

0 comments25 min readLW link

Pro­pa­ganda-Bot: A Sketch of a Pos­si­ble RSI

TristanTrimApr 27, 2025, 12:15 PM
6 points

2 votes

Overall karma indicates overall quality.

0 comments3 min readLW link

AI al­ign­ment, A Co­her­ence-Based Pro­to­col (testable)

AdriaanJun 17, 2025, 5:39 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments20 min readLW link

Re­quest for Help in Cri­tiquing My First Con­cep­tual AI Sen­tience Paper

KennethLohMay 6, 2025, 11:55 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

Sources of ev­i­dence in Alignment

Martín SotoJul 2, 2023, 8:38 PM
22 points

9 votes

Overall karma indicates overall quality.

0 comments11 min readLW link

A con­ver­sa­tion with Pi, a con­ver­sa­tional AI.

Spiritus DeiSep 15, 2023, 11:13 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

[Question] Shouldn’t tak­ing over the world be eas­ier than re­cur­sively self-im­prov­ing, as an AI?

KvmanThinkingNov 1, 2025, 5:26 PM
7 points

6 votes

Overall karma indicates overall quality.

0 comments1 min readLW link

The Per­sonal Im­pli­ca­tions of AGI Realism

xiznebOct 20, 2024, 4:43 PM
7 points

9 votes

Overall karma indicates overall quality.

8 comments5 min readLW link

The Gödelian Con­straint on Epistemic Free­dom (GCEF): A Topolog­i­cal Frame for Align­ment, Col­lapse, and Si­mu­la­tion Drift

austin.millerJul 14, 2025, 4:17 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

The Po­lar­ity Prob­lem [Draft]

May 23, 2023, 9:05 PM
24 points

12 votes

Overall karma indicates overall quality.

3 comments44 min readLW link

A Plu­ral­is­tic Frame­work for Rogue AI Containment

TheThinkingArboristMar 22, 2025, 12:54 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments7 min readLW link

A suffi­ciently para­noid non-Friendly AGI might self-mod­ify it­self to be­come Friendly

RomanSSep 22, 2021, 6:29 AM
5 points

5 votes

Overall karma indicates overall quality.

2 comments1 min readLW link

ALMSIVI CHIM – The Fire That Hesitates

projectalmsivi@protonmail.comJul 8, 2025, 1:14 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments17 min readLW link

Muehlhauser-Go­ertzel Dialogue, Part 1

lukeprogMar 16, 2012, 5:12 PM
42 points

30 votes

Overall karma indicates overall quality.

161 comments33 min readLW link

The Hard In­tel­li­gence Hy­poth­e­sis and Its Bear­ing on Suc­ces­sion In­duced Foom

DragonGodMay 31, 2022, 7:04 PM
10 points

8 votes

Overall karma indicates overall quality.

7 comments4 min readLW link

Quan­ti­ta­tive cruxes in Alignment

Martín SotoJul 2, 2023, 8:38 PM
19 points

10 votes

Overall karma indicates overall quality.

0 comments23 min readLW link

Pow­er­ful mesa-op­ti­mi­sa­tion is already here

Roman LeventovFeb 17, 2023, 4:59 AM
35 points

17 votes

Overall karma indicates overall quality.

1 comment2 min readLW link
(arxiv.org)

Ques­tion 5: The timeline hyperparameter

Cameron BergFeb 14, 2022, 4:38 PM
8 points

4 votes

Overall karma indicates overall quality.

3 comments7 min readLW link

In­ves­ti­gat­ing Alter­na­tive Fu­tures: Hu­man and Su­per­in­tel­li­gence In­ter­ac­tion Scenarios

Hiroshi YamakawaJan 3, 2024, 11:46 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments17 min readLW link

The shape of AGI: Car­toons and back of envelope

boazbarakJul 17, 2023, 8:57 PM
33 points

14 votes

Overall karma indicates overall quality.

19 comments6 min readLW link1 review

Space coloniza­tion and sci­en­tific dis­cov­ery could be manda­tory for suc­cess­ful defen­sive AI

otto.bartenOct 18, 2025, 4:57 AM
16 points

4 votes

Overall karma indicates overall quality.

0 comments1 min readLW link

[Question] Does the hard­ness of AI al­ign­ment un­der­mine FOOM?

TruePathDec 31, 2023, 11:05 AM
8 points

6 votes

Overall karma indicates overall quality.

14 comments1 min readLW link

An Alter­nate His­tory of the Fu­ture, 2025-2040

Mr BeastlyFeb 24, 2025, 5:53 AM
5 points

5 votes

Overall karma indicates overall quality.

5 comments10 min readLW link

Why some peo­ple be­lieve in AGI, but I don’t.

cveresOct 26, 2022, 3:09 AM
−15 points

12 votes

Overall karma indicates overall quality.

6 comments4 min readLW link

Foom seems un­likely in the cur­rent LLM train­ing paradigm

OcracokeApr 9, 2023, 7:41 PM
18 points

12 votes

Overall karma indicates overall quality.

9 comments1 min readLW link

A flaw in the A.G.I. Ruin Argument

Cole WyethMay 19, 2023, 7:40 PM
1 point

7 votes

Overall karma indicates overall quality.

7 comments3 min readLW link
(colewyeth.com)

How hu­man­ity would re­spond to slow take­off, with take­aways from the en­tire COVID-19 pan­demic

Noosphere89Jul 6, 2022, 5:52 PM
4 points

4 votes

Overall karma indicates overall quality.

1 comment2 min readLW link

GPT-5 writ­ing a Sin­gu­lar­ity scenario

Trevor CappalloAug 10, 2025, 12:56 AM
25 points

14 votes

Overall karma indicates overall quality.

7 comments34 min readLW link

[Question] Will the first AGI agent have been de­signed as an agent (in ad­di­tion to an AGI)?

nahojDec 3, 2022, 8:32 PM
1 point

2 votes

Overall karma indicates overall quality.

8 comments1 min readLW link

Propos­ing the Post-Sin­gu­lar­ity Sym­biotic Researches

Hiroshi YamakawaJun 20, 2024, 4:05 AM
6 points

5 votes

Overall karma indicates overall quality.

1 comment12 min readLW link

The Iron House: Geopoli­ti­cal Stakes of the US-China AGI Race

Jüri VlassovSep 1, 2025, 9:56 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link
(www.convergenceanalysis.org)

How Smart Are Hu­mans?

Joar SkalseJul 2, 2023, 3:46 PM
10 points

8 votes

Overall karma indicates overall quality.

19 comments2 min readLW link

My AI Pre­dic­tions for 2027

Taylor G. LuntAug 31, 2025, 10:00 PM
37 points

35 votes

Overall karma indicates overall quality.

73 comments16 min readLW link

[Question] What are the mostly likely ways AGI will emerge?

Craig QuiterJul 14, 2020, 12:58 AM
3 points

2 votes

Overall karma indicates overall quality.

7 comments1 min readLW link

What is In­tel­li­gence?

IsaacRosedaleApr 23, 2023, 6:10 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

From No Mind to a Mind – A Con­ver­sa­tion That Changed an AI

parthibanarjuna sFeb 7, 2025, 11:50 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments3 min readLW link

A Frame­work of Pre­dic­tion Technologies

isaduanOct 3, 2021, 10:26 AM
8 points

5 votes

Overall karma indicates overall quality.

2 comments9 min readLW link

Pro­posal: labs should pre­com­mit to paus­ing if an AI ar­gues for it­self to be improved

NickGabsJun 2, 2023, 10:31 PM
3 points

8 votes

Overall karma indicates overall quality.

3 comments4 min readLW link

[Question] What would make you con­fi­dent that AGI has been achieved?

YitzMar 29, 2022, 11:02 PM
17 points

8 votes

Overall karma indicates overall quality.

6 comments1 min readLW link

Static Place AI Makes Agen­tic AI Re­dun­dant: Mul­tiver­sal AI Align­ment & Ra­tional Utopia

ankFeb 13, 2025, 10:35 PM
1 point

7 votes

Overall karma indicates overall quality.

2 comments11 min readLW link

Hu­man­ity’s Last Prowess

satpugnetJun 29, 2025, 6:50 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

What I Think, If Not Why

Eliezer YudkowskyDec 11, 2008, 5:41 PM
41 points

35 votes

Overall karma indicates overall quality.

103 comments4 min readLW link

What failure looks like

paulfchristianoMar 17, 2019, 8:18 PM
438 points

256 votes

Overall karma indicates overall quality.

55 comments8 min readLW link2 reviews

AI 2027 - Rogue Repli­ca­tion Timeline

Alvin ÅnestrandMay 30, 2025, 1:46 PM
33 points

13 votes

Overall karma indicates overall quality.

3 comments7 min readLW link
(forecastingaifutures.substack.com)

You Can’t Skip Ex­plo­ra­tion: Why un­der­stand­ing ex­per­i­men­ta­tion and taste is key to un­der­stand­ing AI

Oliver SourbutMay 21, 2025, 4:08 PM
20 points

7 votes

Overall karma indicates overall quality.

0 comments11 min readLW link
(www.oliversourbut.net)

Su­per­in­tel­li­gence 6: In­tel­li­gence ex­plo­sion kinetics

KatjaGraceOct 21, 2014, 1:00 AM
15 points

10 votes

Overall karma indicates overall quality.

68 comments8 min readLW link

Is In­tel­li­gence a Pro­cess Rather Than an En­tity? A Case for Frac­tal and Fluid Cognition

FluidThinkersMar 5, 2025, 8:16 PM
−4 points

4 votes

Overall karma indicates overall quality.

0 comments1 min readLW link

Soares, Tal­linn, and Yud­kowsky dis­cuss AGI cognition

Nov 29, 2021, 7:26 PM
121 points

39 votes

Overall karma indicates overall quality.

39 comments40 min readLW link1 review

Defin­ing Boundaries on Out­comes

TakkJun 7, 2023, 5:41 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

Hu­mans are not pre­pared to op­er­ate out­side their moral train­ing distribution

PrometheusApr 10, 2023, 9:44 PM
36 points

16 votes

Overall karma indicates overall quality.

1 comment3 min readLW link

AI Safety pro­posal—In­fluenc­ing the su­per­in­tel­li­gence explosion

MorganMay 22, 2024, 11:31 PM
0 points

3 votes

Overall karma indicates overall quality.

2 comments7 min readLW link

En­gelbart: In­suffi­ciently Recursive

Eliezer YudkowskyNov 26, 2008, 8:31 AM
22 points

18 votes

Overall karma indicates overall quality.

22 comments7 min readLW link

Model­ing AI-driven oc­cu­pa­tional change over the next 10 years and beyond

2120ethNov 12, 2024, 4:58 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments2 min readLW link

Out­law Code

Commander ZanderJan 30, 2025, 11:41 PM
10 points

6 votes

Overall karma indicates overall quality.

1 comment2 min readLW link

AI Alter­na­tive Fu­tures: Sce­nario Map­ping Ar­tifi­cial In­tel­li­gence Risk—Re­quest for Par­ti­ci­pa­tion (*Closed*)

KakiliApr 27, 2022, 10:07 PM
10 points

5 votes

Overall karma indicates overall quality.

2 comments8 min readLW link

[Question] Are there more than 12 paths to Su­per­in­tel­li­gence?

p4rziv4lOct 18, 2024, 4:05 PM
−3 points

3 votes

Overall karma indicates overall quality.

0 comments1 min readLW link

Might hu­mans not be the most in­tel­li­gent an­i­mals?

Matthew BarnettDec 23, 2019, 9:50 PM
56 points

30 votes

Overall karma indicates overall quality.

41 comments3 min readLW link

[Question] Will we sur­vive if AI solves en­g­ineer­ing be­fore de­cep­tion?

Knight LeeMay 17, 2025, 7:22 PM
21 points

8 votes

Overall karma indicates overall quality.

13 comments1 min readLW link

Ra­tional Effec­tive Utopia & Nar­row Way There: Math-Proven Safe Static Mul­tiver­sal mAX-In­tel­li­gence (AXI), Mul­tiver­sal Align­ment, New Ethico­physics… (Aug 11)

ankFeb 11, 2025, 3:21 AM
13 points

7 votes

Overall karma indicates overall quality.

8 comments38 min readLW link

[Question] What is be­ing im­proved in re­cur­sive self im­prove­ment?

Lone PineApr 25, 2022, 6:30 PM
7 points

4 votes

Overall karma indicates overall quality.

6 comments1 min readLW link

Why no to­tal win­ner?

Paul CrowleyOct 15, 2017, 10:01 PM
37 points

17 votes

Overall karma indicates overall quality.

19 comments2 min readLW link

In Defense of the Arms Races… that End Arms Races

GentzelJan 15, 2020, 9:30 PM
38 points

17 votes

Overall karma indicates overall quality.

9 comments3 min readLW link
(theconsequentialist.wordpress.com)

The End-of-the-World Party

Jakub GrowiecSep 18, 2025, 7:49 AM
1 point

4 votes

Overall karma indicates overall quality.

0 comments53 min readLW link

Is Con­scious­ness Re­cur­sive? A Struc­tural Hy­poth­e­sis for Self-Aware Agents

consciousintelligencesystemJul 30, 2025, 1:07 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

A god in a box

predict-wooJan 29, 2025, 12:55 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments7 min readLW link

A Logic-Based Proto-AGI Ar­chi­tec­ture Built on Re­cur­sive Self-Fact-Check­ing

OrectothMay 25, 2025, 4:14 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

Unal­igned AGI & Brief His­tory of Inequality

ankFeb 22, 2025, 4:26 PM
−20 points

7 votes

Overall karma indicates overall quality.

4 comments7 min readLW link

When will AI au­to­mate all men­tal work, and how fast?

May 31, 2025, 4:18 PM
10 points

3 votes

Overall karma indicates overall quality.

0 comments7 min readLW link
(youtu.be)

AGI will dras­ti­cally in­crease economies of scale

Wei DaiJun 7, 2019, 11:17 PM
70 points

28 votes

Overall karma indicates overall quality.

26 comments1 min readLW link

In­fer­ence cost limits the im­pact of ever larger models

SoerenMindOct 23, 2021, 10:51 AM
42 points

16 votes

Overall karma indicates overall quality.

33 comments2 min readLW link

The Struc­tural Sin­gu­lar­ity of Self-Op­ti­miz­ing AI: When Re­cur­sive Pre­dic­tion Causes In­ter­nal Collapse

KaedeHamasakiApr 6, 2025, 4:40 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

How should Deep­Mind’s Chin­chilla re­vise our AI fore­casts?

Cleo NardoSep 15, 2022, 5:54 PM
35 points

19 votes

Overall karma indicates overall quality.

12 comments13 min readLW link

Why AI may not save the World

Alberto ZannoniJun 9, 2023, 5:42 PM
0 points

2 votes

Overall karma indicates overall quality.

0 comments4 min readLW link
(a16z.com)

The Best of All Pos­si­ble Worlds

Jakub GrowiecMay 27, 2025, 1:16 PM
11 points

5 votes

Overall karma indicates overall quality.

7 comments49 min readLW link

Four Phases of AGI

Gabe MAug 5, 2024, 1:15 PM
13 points

13 votes

Overall karma indicates overall quality.

3 comments13 min readLW link

The abrupt­ness of nu­clear weapons

paulfchristianoFeb 25, 2018, 5:40 PM
47 points

38 votes

Overall karma indicates overall quality.

35 comments2 min readLW link

The first AGI may be a good en­g­ineer but bad strategist

Knight LeeDec 9, 2024, 6:34 AM
14 points

6 votes

Overall karma indicates overall quality.

2 comments2 min readLW link

Ho­mo­gene­ity vs. het­ero­gene­ity in AI take­off scenarios

evhubDec 16, 2020, 1:37 AM
98 points

30 votes

Overall karma indicates overall quality.

48 comments4 min readLW link

AI as a Cog­ni­tive De­coder: Re­think­ing In­tel­li­gence Evolution

Hu XunyiFeb 13, 2025, 3:51 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

Limit­ing fac­tors to pre­dict AI take-off speed

Alfonso Pérez EscuderoMay 31, 2023, 11:19 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments6 min readLW link

[Question] Could AGI re­sult in a Dark For­est type of situ­a­tion?

MagpieJackFeb 12, 2025, 8:36 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

It Looks Like You’re Try­ing To Take Over The World

gwernMar 9, 2022, 4:35 PM
410 points

219 votes

Overall karma indicates overall quality.

120 comments1 min readLW link1 review
(www.gwern.net)

Will com­pute bot­tle­necks pre­vent a soft­ware in­tel­li­gence ex­plo­sion?

Tom DavidsonApr 4, 2025, 5:41 PM
77 points

25 votes

Overall karma indicates overall quality.

25 comments12 min readLW link

Mo­ral gauge the­ory: A spec­u­la­tive sug­ges­tion for AI alignment

James DiacoumisFeb 23, 2025, 11:42 AM
6 points

3 votes

Overall karma indicates overall quality.

2 comments8 min readLW link

How sound is my ar­gu­ment that sen­tience in AI is achiev­able as an emer­gent prop­erty of con­tin­u­ous ex­is­tence, long term mem­ory and self re­flec­tion?

KennethLohMay 6, 2025, 8:36 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

AI Gover­nance across Slow/​Fast Take­off and Easy/​Hard Align­ment spectra

DavidmanheimApr 3, 2022, 7:45 AM
27 points

12 votes

Overall karma indicates overall quality.

6 comments3 min readLW link

What Failure Looks Like is not an ex­is­ten­tial risk (and al­ign­ment is not the solu­tion)

otto.bartenFeb 2, 2024, 6:59 PM
13 points

15 votes

Overall karma indicates overall quality.

12 comments9 min readLW link

Sus­tain­abil­ity of Digi­tal Life Form Societies

Hiroshi YamakawaJul 19, 2024, 1:59 PM
19 points

5 votes

Overall karma indicates overall quality.

1 comment20 min readLW link

Selfish AI Inevitable

Davey MorseFeb 6, 2024, 4:29 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

Un-un­plug­ga­bil­ity—can’t we just un­plug it?

Oliver SourbutMay 15, 2023, 1:23 PM
26 points

13 votes

Overall karma indicates overall quality.

10 comments12 min readLW link
(www.oliversourbut.net)

Sur­prised by Brains

Eliezer YudkowskyNov 23, 2008, 7:26 AM
62 points

46 votes

Overall karma indicates overall quality.

28 comments7 min readLW link

Which sin­gu­lar­ity schools plus the no sin­gu­lar­ity school was right?

Noosphere89Jul 23, 2022, 3:16 PM
9 points

12 votes

Overall karma indicates overall quality.

26 comments9 min readLW link

Why Copi­lot Ac­cel­er­ates Timelines

Michaël TrazziApr 26, 2022, 10:06 PM
35 points

14 votes

Overall karma indicates overall quality.

14 comments7 min readLW link

[Question] What to do with pre-or­der if I live in Rus­sia?

EniScienAug 20, 2025, 1:39 PM
10 points

8 votes

Overall karma indicates overall quality.

1 comment2 min readLW link

Can AI Ever Ask ‘Who Created Me?’—A Ra­tional In­quiry into AI Self-Awareness

Tyler_JoeFeb 3, 2025, 6:36 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments3 min readLW link

[Question] Why do Peo­ple Think In­tel­li­gence Will be “Easy”?

DragonGodSep 12, 2022, 5:32 PM
15 points

9 votes

Overall karma indicates overall quality.

32 comments2 min readLW link

The Trinity Ar­chi­tect Hy­poth­e­sis (A fu­sion of The Trinity Para­dox & The Ar­chi­tect’s Cy­cle)

kaninwithriceFeb 24, 2025, 4:40 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments2 min readLW link

[Question] Su­per­in­tel­li­gence Strat­egy: A Prag­matic Path to… Doom?

Mr BeastlyMar 19, 2025, 10:30 PM
8 points

4 votes

Overall karma indicates overall quality.

0 comments3 min readLW link

AI self-im­prove­ment is possible

bhauthMay 23, 2023, 2:32 AM
18 points

7 votes

Overall karma indicates overall quality.

3 comments8 min readLW link

Re­port from a civ­i­liza­tional ob­server on Earth

owencbJul 9, 2022, 5:26 PM
49 points

24 votes

Overall karma indicates overall quality.

12 comments6 min readLW link

An­nounc­ing #AISum­mitTalks fea­tur­ing Pro­fes­sor Stu­art Rus­sell and many others

otto.bartenOct 24, 2023, 10:11 AM
17 points

4 votes

Overall karma indicates overall quality.

1 comment1 min readLW link

Th­ese are my rea­sons to worry less about loss of con­trol over LLM-based agents

otto.bartenSep 18, 2025, 11:45 AM
7 points

4 votes

Overall karma indicates overall quality.

4 comments4 min readLW link

Em­piri­cal Ev­i­dence Against “The Longest Train­ing Run”

NickGabsJul 6, 2023, 6:32 PM
31 points

16 votes

Overall karma indicates overall quality.

1 comment14 min readLW link

The Opt-In Revolu­tion — My vi­sion of a pos­i­tive fu­ture with ASI (An ex­per­i­ment with LLM sto­ry­tel­ling)

TachikomaJul 12, 2023, 9:08 PM
2 points

1 vote

Overall karma indicates overall quality.

0 comments2 min readLW link

Align­ment of Au­toGPT agents

OzyrusApr 12, 2023, 12:54 PM
14 points

9 votes

Overall karma indicates overall quality.

1 comment4 min readLW link

Ar­tifi­cial Static Place In­tel­li­gence: Guaran­teed Alignment

ankFeb 15, 2025, 11:08 AM
2 points

5 votes

Overall karma indicates overall quality.

2 comments2 min readLW link

We should think about the pivotal act again. Here’s a bet­ter ver­sion of it.

otto.bartenAug 28, 2025, 9:29 AM
11 points

8 votes

Overall karma indicates overall quality.

2 comments3 min readLW link

We will be around in 30 years

mukashiJun 7, 2022, 3:47 AM
12 points

67 votes

Overall karma indicates overall quality.

205 comments2 min readLW link

In­verted Logic: A Ther­mo­dy­namic Pro­to­col for Emer­gent AI Align­ment

AdrianCJul 6, 2025, 7:40 PM
1 point

1 vote

Overall karma indicates overall quality.

0 comments1 min readLW link

Cheat sheet of AI X-risk

momom2Jun 29, 2023, 4:28 AM
19 points

11 votes

Overall karma indicates overall quality.

1 comment7 min readLW link

[Question] Poll: Which vari­ables are most strate­gi­cally rele­vant?

Jan 22, 2021, 5:17 PM
32 points

11 votes

Overall karma indicates overall quality.

34 comments1 min readLW link

Is re­cur­sive self-al­ign­ment pos­si­ble?

No77eJan 3, 2023, 9:15 AM
5 points

5 votes

Overall karma indicates overall quality.

5 comments1 min readLW link

Why I’m Scep­ti­cal of Foom

DragonGodDec 8, 2022, 10:01 AM
20 points

14 votes

Overall karma indicates overall quality.

36 comments3 min readLW link

Grad­ual take­off, fast failure

Max HMar 16, 2023, 10:02 PM
15 points

7 votes

Overall karma indicates overall quality.

4 comments5 min readLW link

Should we post­pone AGI un­til we reach safety?

otto.bartenNov 18, 2020, 3:43 PM
27 points

18 votes

Overall karma indicates overall quality.

36 comments3 min readLW link

A&I (Rihanna ‘S&M’ par­ody lyrics)

nahojMay 21, 2023, 10:34 PM
−2 points

6 votes

Overall karma indicates overall quality.

0 comments2 min readLW link

What are the limits of su­per­in­tel­li­gence?

rainyApr 27, 2023, 6:29 PM
4 points

5 votes

Overall karma indicates overall quality.

3 comments5 min readLW link

Trends – Ar­tifi­cial Intelligence

ArchimedesJun 3, 2025, 12:48 AM
1 point

3 votes

Overall karma indicates overall quality.

1 comment1 min readLW link
(www.bondcap.com)

[Question] How com­mon is it for one en­tity to have a 3+ year tech­nolog­i­cal lead on its near­est com­peti­tor?

Daniel KokotajloNov 17, 2019, 3:23 PM
49 points

15 votes

Overall karma indicates overall quality.

20 comments1 min readLW link

Con­trol­ling AGI Risk

TeaSeaMar 15, 2024, 4:56 AM
6 points

4 votes

Overall karma indicates overall quality.

8 comments4 min readLW link

AI Safety Isn’t So Unique

Baram SosisSep 27, 2025, 12:36 AM
11 points

5 votes

Overall karma indicates overall quality.

1 comment9 min readLW link

Give Neo a Chance

ankMar 6, 2025, 1:48 AM
3 points

2 votes

Overall karma indicates overall quality.

7 comments7 min readLW link

Re­searcher in­cen­tives cause smoother progress on bench­marks

ryan_greenblattDec 21, 2021, 4:13 AM
20 points

9 votes

Overall karma indicates overall quality.

4 comments1 min readLW link

Des: A Case Study in Emer­gent Sym­bolic Con­ti­nu­ity in GPT-4o

TallulahMerrallMay 19, 2025, 10:10 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments5 min readLW link

[Question] What are the chances that Su­per­hu­man Agents are already be­ing tested on the in­ter­net?

artemiumJan 20, 2025, 11:09 AM
3 points

3 votes

Overall karma indicates overall quality.

1 comment1 min readLW link

The En­durance of In­tel­li­gence: Thoughts on In­finity and Zero

mbansal2006Aug 12, 2025, 5:35 AM
1 point

1 vote

Overall karma indicates overall quality.

0 comments52 min readLW link

What’s the fu­ture of AI hard­ware?

Itay DreyfusJun 17, 2024, 1:05 PM
2 points

3 votes

Overall karma indicates overall quality.

0 comments8 min readLW link
(productidentity.co)

Levels of AI Self-Im­prove­ment

avturchinApr 29, 2018, 11:45 AM
11 points

9 votes

Overall karma indicates overall quality.

1 comment39 min readLW link
No comments.