Limit­ing fac­tors to pre­dict AI take-off speed

Alfonso Pérez Escudero31 May 2023 23:19 UTC
1 point
0 comments6 min readLW link

The challenge of ar­tic­u­lat­ing tacit knowledge

Nina Rimsky31 May 2023 23:10 UTC
48 points
4 comments5 min readLW link
(ninarimsky.substack.com)

Un­pre­dictabil­ity and the In­creas­ing Difficulty of AI Align­ment for In­creas­ingly In­tel­li­gent AI

Max_He-Ho31 May 2023 22:25 UTC
5 points
2 comments20 min readLW link

Shut­down-Seek­ing AI

Simon Goldstein31 May 2023 22:19 UTC
48 points
31 comments15 min readLW link

Full Au­toma­tion is Un­likely and Un­nec­es­sary for Ex­plo­sive Growth

aogara31 May 2023 21:55 UTC
28 points
3 comments5 min readLW link

LessWrong Com­mu­nity Week­end 2023 Up­dates: Keynote Speaker Mal­colm Ocean, Re­main­ing Tick­ets and More

Henry Prowbell31 May 2023 21:53 UTC
23 points
0 comments2 min readLW link

The Div­ine Move Para­dox & Think­ing as a Species

Christopher James Hart31 May 2023 21:38 UTC
9 points
8 comments3 min readLW link

In­tent-al­igned AI sys­tems de­plete hu­man agency: the need for agency foun­da­tions re­search in AI safety

catubc31 May 2023 21:18 UTC
24 points
4 comments11 min readLW link

[Question] How much over­lap is there be­tween the util­ity func­tion of GPT-n and GPT-(n+1), as­sum­ing both are near AGI?

Phosphorous31 May 2023 20:28 UTC
2 points
0 comments2 min readLW link

My AI-risk cartoon

pre31 May 2023 19:46 UTC
6 points
0 comments1 min readLW link

Eval­u­a­tion Ev­i­dence Re­con­struc­tions of Mock Crimes Sub­mis­sion 3

Alan E Dunne31 May 2023 19:03 UTC
−1 points
0 comments3 min readLW link

Im­prov­ing Math­e­mat­i­cal Rea­son­ing with-Pro­cess Supervision

p.b.31 May 2023 19:00 UTC
12 points
3 comments1 min readLW link
(openai.com)

The Crux List

Zvi31 May 2023 18:30 UTC
72 points
19 comments33 min readLW link
(thezvi.wordpress.com)

Stages of Survival

Zvi31 May 2023 18:30 UTC
44 points
0 comments17 min readLW link
(thezvi.wordpress.com)

Types and De­grees of Alignment

Zvi31 May 2023 18:30 UTC
34 points
10 comments8 min readLW link
(thezvi.wordpress.com)

To Pre­dict What Hap­pens, Ask What Happens

Zvi31 May 2023 18:30 UTC
79 points
0 comments9 min readLW link
(thezvi.wordpress.com)

A push to­wards in­ter­ac­tive trans­former decoding

R0bk31 May 2023 17:56 UTC
3 points
0 comments2 min readLW link
(github.com)

Neu­roevolu­tion, So­cial In­tel­li­gence, and Logic

vinnik.dmitry0731 May 2023 17:54 UTC
1 point
0 comments10 min readLW link

Con­trast Pairs Drive the Em­piri­cal Perfor­mance of Con­trast Con­sis­tent Search (CCS)

Scott Emmons31 May 2023 17:09 UTC
97 points
0 comments6 min readLW link

Cos­mopoli­tan val­ues don’t come free

So8res31 May 2023 15:58 UTC
124 points
82 comments1 min readLW link

[Question] Ar­gu­ments Against Fos­sil Fu­ture?

Sable31 May 2023 13:41 UTC
11 points
29 comments1 min readLW link

On Ob­jec­tive Ethics, and a bit about boats

EndlessBlue31 May 2023 11:40 UTC
−7 points
3 comments2 min readLW link

Against Con­flat­ing Ex­per­tise: Dist­in­guish­ing AI Devel­op­ment from AI Im­pli­ca­tion Analysis

Ratios31 May 2023 9:50 UTC
13 points
4 comments1 min readLW link

A rough model for P(AI doom)

Michael Tontchev31 May 2023 8:58 UTC
0 points
1 comment2 min readLW link

[Question] What’s the con­sen­sus on porn?

FinalFormal231 May 2023 3:15 UTC
4 points
19 comments1 min readLW link

Product En­dorse­ment: Food for sleep interruptions

Elizabeth31 May 2023 1:50 UTC
45 points
7 comments1 min readLW link
(acesounderglass.com)

Op­ti­mal Clothing

Gordon Seidoh Worley31 May 2023 1:00 UTC
30 points
8 comments6 min readLW link

Ab­strac­tion is Big­ger than Nat­u­ral Abstraction

NicholasKross31 May 2023 0:00 UTC
18 points
0 comments5 min readLW link
(www.thinkingmuchbetter.com)

Hu­mans, chim­panzees and other animals

gjm30 May 2023 23:53 UTC
21 points
18 comments1 min readLW link

The case for re­mov­ing al­ign­ment and ML re­search from the train­ing dataset

beren30 May 2023 20:54 UTC
48 points
8 comments5 min readLW link

Why Job Dis­place­ment Pre­dic­tions are Wrong: Ex­pla­na­tions of Cog­ni­tive Automation

Moritz Wallawitsch30 May 2023 20:43 UTC
−4 points
0 comments8 min readLW link

PaLM-2 & GPT-4 in “Ex­trap­o­lat­ing GPT-N perfor­mance”

Lukas Finnveden30 May 2023 18:33 UTC
55 points
6 comments6 min readLW link

RoboNet—A new in­ter­net pro­to­col for AI

antoniomax30 May 2023 17:55 UTC
−13 points
1 comment18 min readLW link

Why I don’t think that the prob­a­bil­ity that AGI kills ev­ery­one is roughly 1 (but rather around 0.995).

Bastumannen30 May 2023 17:54 UTC
−6 points
0 comments2 min readLW link

AI X-risk is a pos­si­ble solu­tion to the Fermi Paradox

magic9mushroom30 May 2023 17:42 UTC
11 points
20 comments2 min readLW link

LIMA: Less Is More for Alignment

Ulisse Mini30 May 2023 17:10 UTC
16 points
6 comments1 min readLW link
(arxiv.org)

Boomerang—pro­to­col to dis­solve some com­mit­ment races

Filip Sondej30 May 2023 16:21 UTC
37 points
10 comments8 min readLW link

An­nounc­ing Apollo Research

30 May 2023 16:17 UTC
215 points
11 comments8 min readLW link

Ad­vice for new al­ign­ment peo­ple: Info Max

Jonas Hallgren30 May 2023 15:42 UTC
27 points
4 comments5 min readLW link

[Question] Who is li­able for AI?

jmh30 May 2023 13:54 UTC
14 points
4 comments1 min readLW link

AI Safety Newslet­ter #8: Rogue AIs, how to screen for AI risks, and grants for re­search on demo­cratic gov­er­nance of AI

30 May 2023 11:52 UTC
20 points
0 comments6 min readLW link
(newsletter.safe.ai)

The bul­ls­eye frame­work: My case against AI doom

titotal30 May 2023 11:52 UTC
89 points
35 comments1 min readLW link

State­ment on AI Ex­tinc­tion—Signed by AGI Labs, Top Aca­demics, and Many Other Notable Figures

Dan H30 May 2023 9:05 UTC
372 points
77 comments1 min readLW link
(www.safe.ai)

The­o­ret­i­cal Limi­ta­tions of Au­tore­gres­sive Models

Gabriel Wu30 May 2023 2:37 UTC
20 points
1 comment10 min readLW link
(gabrieldwu.github.io)

A book re­view for “An­i­mal Weapons” and cross-ap­ply­ing the les­sons to x-risk

Habeeb Abdulfatah30 May 2023 0:58 UTC
−6 points
1 comment1 min readLW link
(www.super-linear.org)

Without a tra­jec­tory change, the de­vel­op­ment of AGI is likely to go badly

Max H29 May 2023 23:42 UTC
16 points
2 comments13 min readLW link

Win­ners-take-how-much?

YonatanK29 May 2023 21:56 UTC
1 point
2 comments3 min readLW link

Re­ply to a fer­til­ity doc­tor con­cern­ing poly­genic em­bryo screening

GeneSmith29 May 2023 21:50 UTC
58 points
6 comments8 min readLW link

Sen­tience matters

So8res29 May 2023 21:25 UTC
141 points
96 comments2 min readLW link

Wikipe­dia as an in­tro­duc­tion to the al­ign­ment problem

SoerenMind29 May 2023 18:43 UTC
83 points
10 comments1 min readLW link
(en.wikipedia.org)