Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down by Eliezer Yudkowsky

jacquesthibs29 Mar 2023 23:16 UTC
298 points
296 comments3 min readLW link
(time.com)

Othello-GPT: Reflec­tions on the Re­search Process

Neel Nanda29 Mar 2023 22:13 UTC
36 points
0 comments15 min readLW link
(neelnanda.io)

Othello-GPT: Fu­ture Work I Am Ex­cited About

Neel Nanda29 Mar 2023 22:13 UTC
48 points
2 comments33 min readLW link
(neelnanda.io)

Ac­tu­ally, Othello-GPT Has A Lin­ear Emer­gent World Representation

Neel Nanda29 Mar 2023 22:13 UTC
211 points
24 comments19 min readLW link
(neelnanda.io)

Draft: De­tect­ing optimization

Alex_Altair29 Mar 2023 20:17 UTC
23 points
2 comments6 min readLW link

“Sorcerer’s Ap­pren­tice” from Fan­ta­sia as an anal­ogy for alignment

awg29 Mar 2023 18:21 UTC
7 points
4 comments1 min readLW link
(video.disney.com)

The Chang­ing Face of Twitter

Zvi29 Mar 2023 17:50 UTC
23 points
8 comments26 min readLW link
(thezvi.wordpress.com)

No­body’s on the ball on AGI alignment

leopold29 Mar 2023 17:40 UTC
95 points
37 comments9 min readLW link
(www.forourposterity.com)

Want to win the AGI race? Solve al­ign­ment.

leopold29 Mar 2023 17:40 UTC
21 points
3 comments5 min readLW link
(www.forourposterity.com)

ChatGPT and Bing Chat can’t play Botticelli

Asha Saavoss29 Mar 2023 17:39 UTC
11 points
0 comments6 min readLW link

The Ra­tion­al­ist Guide to Hinduism

Harsha G.29 Mar 2023 17:03 UTC
22 points
12 comments9 min readLW link
(somestrangeloops.substack.com)

“Un­in­ten­tional AI safety re­search”: Why not sys­tem­at­i­cally mine AI tech­ni­cal re­search for safety pur­poses?

ghostwheel29 Mar 2023 15:56 UTC
27 points
3 comments6 min readLW link

The open letter

kornai29 Mar 2023 15:09 UTC
−21 points
2 comments1 min readLW link

I made AI Risk Propaganda

monkymind29 Mar 2023 14:26 UTC
−3 points
0 comments1 min readLW link

Strong Cheap Signals

trevor29 Mar 2023 14:18 UTC
29 points
3 comments2 min readLW link
(betonit.substack.com)

Miss­ing fore­cast­ing tools: from cat­a­logs to a new kind of pre­dic­tion market

MichaelLatowicki29 Mar 2023 9:55 UTC
14 points
0 comments5 min readLW link

Spread­sheet for 200 Con­crete Prob­lems In Interpretability

Jay Bailey29 Mar 2023 6:51 UTC
12 points
0 comments1 min readLW link

[Question] Which parts of the ex­ist­ing in­ter­net are already likely to be in (GPT-5/​other soon-to-be-trained LLMs)’s train­ing cor­pus?

AnnaSalamon29 Mar 2023 5:17 UTC
49 points
2 comments1 min readLW link

[Question] Are there spe­cific books that it might slightly help al­ign­ment to have on the in­ter­net?

AnnaSalamon29 Mar 2023 5:08 UTC
78 points
25 comments1 min readLW link

FLI open let­ter: Pause gi­ant AI experiments

Zach Stein-Perlman29 Mar 2023 4:04 UTC
126 points
123 comments2 min readLW link
(futureoflife.org)

Run Posts By Orgs

jefftk29 Mar 2023 2:40 UTC
16 points
74 comments3 min readLW link
(www.jefftk.com)

De­sen­si­tiz­ing Deepfakes

Phib29 Mar 2023 1:20 UTC
1 point
0 comments1 min readLW link

Large lan­guage mod­els aren’t trained enough

sanxiyn29 Mar 2023 0:56 UTC
16 points
4 comments1 min readLW link
(finbarr.ca)

Job Board (28 March 2033)

dr_s28 Mar 2023 22:44 UTC
20 points
1 comment3 min readLW link

Four lenses on AI risks

jasoncrawford28 Mar 2023 21:52 UTC
23 points
5 comments3 min readLW link
(rootsofprogress.org)

Some com­mon con­fu­sion about in­duc­tion heads

Alexandre Variengien28 Mar 2023 21:51 UTC
46 points
4 comments5 min readLW link

Draft: The op­ti­miza­tion toolbox

Alex_Altair28 Mar 2023 20:40 UTC
13 points
1 comment7 min readLW link

Inch­ing “Kubla Khan” and GPT into the same in­tel­lec­tual frame­work @ 3 Quarks Daily

Bill Benzon28 Mar 2023 19:50 UTC
5 points
0 comments3 min readLW link

A rough and in­com­plete re­view of some of John Went­worth’s research

So8res28 Mar 2023 18:52 UTC
175 points
17 comments18 min readLW link

[Question] How do you man­age your in­puts?

Mateusz Bagiński28 Mar 2023 18:26 UTC
15 points
3 comments1 min readLW link

Chat­bot con­vinces Bel­gian to com­mit suicide

Jeroen De Ryck28 Mar 2023 18:14 UTC
60 points
18 comments3 min readLW link
(www.standaard.be)

A Primer On Chaos

johnswentworth28 Mar 2023 18:01 UTC
53 points
9 comments9 min readLW link

[Question] How likely are sce­nar­ios where AGI ends up overtly or de facto tor­tur­ing us? How likely are sce­nar­ios where AGI pre­vents us from com­mit­ting suicide or dy­ing?

JohnGreer28 Mar 2023 18:00 UTC
11 points
4 comments1 min readLW link

How do we al­ign hu­mans and what does it mean for the new Con­jec­ture’s strategy

Igor Ivanov28 Mar 2023 17:54 UTC
7 points
4 comments7 min readLW link

Govern­ing High-Im­pact AI Sys­tems: Un­der­stand­ing Canada’s Pro­posed AI Bill. April 15, Car­leton Univer­sity, Ottawa

Liav Koren28 Mar 2023 17:48 UTC
11 points
1 comment1 min readLW link
(forum.effectivealtruism.org)

I had a chat with GPT-4 on the fu­ture of AI and AI safety

Kristian Freed28 Mar 2023 17:47 UTC
1 point
0 comments8 min readLW link

LessWrong Hangout

Raymond Koopmanschap28 Mar 2023 17:47 UTC
0 points
0 comments1 min readLW link

Half-baked al­ign­ment idea

ozb28 Mar 2023 17:47 UTC
6 points
27 comments1 min readLW link

[Question] Solv­ing Mys­ter­ies -

Phib28 Mar 2023 17:46 UTC
1 point
0 comments1 min readLW link

Some of My Cur­rent Im­pres­sions En­ter­ing AI Safety

Phib28 Mar 2023 17:46 UTC
2 points
0 comments2 min readLW link

[Question] Why do the Se­quences say that “Löb’s The­o­rem shows that a math­e­mat­i­cal sys­tem can­not as­sert its own sound­ness with­out be­com­ing in­con­sis­tent.”?

Thoth Hermes28 Mar 2023 17:19 UTC
12 points
30 comments1 min readLW link

Cor­rigi­bil­ity, Self-Dele­tion, and Iden­ti­cal Strawberries

Robert_AIZI28 Mar 2023 16:54 UTC
8 points
2 comments6 min readLW link
(aizi.substack.com)

[Question] Why no ma­jor LLMs with mem­ory?

Kaj_Sotala28 Mar 2023 16:34 UTC
41 points
15 comments1 min readLW link

Re­sponse to Tyler Cowen’s Ex­is­ten­tial risk, AI, and the in­evitable turn in hu­man history

Zvi28 Mar 2023 16:00 UTC
72 points
27 comments20 min readLW link
(thezvi.wordpress.com)

Adapt­ing to Change: Over­com­ing Chronos­ta­sis in AI Lan­guage Models

RationalMindset28 Mar 2023 14:32 UTC
−1 points
0 comments6 min readLW link

Feel­ing Progress as Motivation

Sable28 Mar 2023 9:11 UTC
4 points
1 comment3 min readLW link
(affablyevil.substack.com)

Be Not Afraid

Alex Beyman28 Mar 2023 8:12 UTC
−12 points
0 comments6 min readLW link

Creat­ing a fam­ily with GPT-4

Kaj_Sotala28 Mar 2023 6:40 UTC
23 points
3 comments10 min readLW link
(kajsotala.fi)

Some 2-4-6 problems

abstractapplic28 Mar 2023 6:32 UTC
28 points
9 comments1 min readLW link
(h-b-p.github.io)

[Question] Deep fold­ing docs site?

mcint28 Mar 2023 6:01 UTC
−1 points
2 comments1 min readLW link