AI x-risk, ap­prox­i­mately or­dered by embarrassment

Alex Lawsen 12 Apr 2023 23:01 UTC
140 points
7 comments19 min readLW link

AXRP Epi­sode 20 - ‘Re­form’ AI Align­ment with Scott Aaronson

DanielFilan12 Apr 2023 21:30 UTC
22 points
2 comments68 min readLW link

Ap­ply to >30 AI safety fun­ders in one ap­pli­ca­tion with the Non­lin­ear Network

12 Apr 2023 21:23 UTC
65 points
12 comments2 min readLW link

LW Ac­count Restricted: OK for me, but not sure about LessWrong

amelia12 Apr 2023 19:45 UTC
69 points
19 comments4 min readLW link

AGI goal space is big, but nar­row­ing might not be as hard as it seems.

Jacy Reese Anthis12 Apr 2023 19:03 UTC
15 points
0 comments3 min readLW link

Nat­u­ral lan­guage alignment

Jacy Reese Anthis12 Apr 2023 19:02 UTC
30 points
2 comments2 min readLW link

Repug­nant lev­els of violins

Solenoid_Entity12 Apr 2023 17:11 UTC
64 points
10 comments12 min readLW link

Progress links and tweets, 2023-04-12

jasoncrawford12 Apr 2023 16:52 UTC
8 points
2 comments1 min readLW link
(rootsofprogress.org)

A ba­sic math­e­mat­i­cal struc­ture of intelligence

Golol12 Apr 2023 16:49 UTC
4 points
6 comments4 min readLW link

[Question] Should Au­toGPT up­date us to­wards re­search­ing IDA?

Michaël Trazzi12 Apr 2023 16:41 UTC
15 points
5 comments1 min readLW link

Box­ing lessons

yakimoff12 Apr 2023 16:19 UTC
1 point
0 comments1 min readLW link

Dazed and con­fused: Good olde’ walk around the Marin Head­lands

yakimoff12 Apr 2023 16:09 UTC
1 point
0 comments1 min readLW link

Towards a solu­tion to the al­ign­ment prob­lem via ob­jec­tive de­tec­tion and eval­u­a­tion

Paul Colognese12 Apr 2023 15:39 UTC
9 points
7 comments12 min readLW link

Ar­tifi­cial In­tel­li­gence as exit strat­egy from the age of acute ex­is­ten­tial risk

Arturo Macias12 Apr 2023 14:48 UTC
−7 points
15 comments7 min readLW link

The UBI dystopia: a glimpse into the fu­ture via pre­sent-day abuses

Solenoid_Entity12 Apr 2023 14:44 UTC
52 points
73 comments4 min readLW link

[Question] Goals of model vs. goals of simu­lacra?

dr_s12 Apr 2023 13:02 UTC
5 points
7 comments1 min readLW link

Align­ment of Au­toGPT agents

Ozyrus12 Apr 2023 12:54 UTC
14 points
1 comment4 min readLW link

Boundaries-based se­cu­rity and AI safety approaches

Allison Duettmann12 Apr 2023 12:36 UTC
42 points
2 comments6 min readLW link

Scaf­folded LLMs as nat­u­ral lan­guage computers

beren12 Apr 2023 10:47 UTC
93 points
10 comments11 min readLW link

LW is prob­a­bly not the place for “I asked this LLM (x) and here’s what it said!”, but where is?

lillybaeum12 Apr 2023 10:12 UTC
21 points
3 comments1 min readLW link

No con­vinc­ing ev­i­dence for gra­di­ent de­scent in ac­ti­va­tion space

Blaine12 Apr 2023 4:48 UTC
76 points
8 comments20 min readLW link

A Brief In­tro­duc­tion to ACI, 2: An Event-Cen­tric View

Akira Pyinya12 Apr 2023 3:23 UTC
1 point
0 comments2 min readLW link

Bos­ton So­cial Dance Covid Requirements

jefftk12 Apr 2023 2:30 UTC
7 points
2 comments1 min readLW link
(www.jefftk.com)

[Link] Sarah Con­stantin: “Why I am Not An AI Doomer”

lbThingrb12 Apr 2023 1:52 UTC
61 points
13 comments1 min readLW link
(sarahconstantin.substack.com)

[Question] Ra­tion­al­ist po­si­tion to­wards ly­ing?

WilliamTerry12 Apr 2023 1:21 UTC
−2 points
4 comments1 min readLW link

Na­tional Telecom­mu­ni­ca­tions and In­for­ma­tion Ad­minis­tra­tion: AI Ac­countabil­ity Policy Re­quest for Comment

sanxiyn11 Apr 2023 22:59 UTC
9 points
0 comments1 min readLW link
(ntia.gov)

Bi­nar­is­tic Bifur­ca­tion: How Real­ity Splits Into Two Separate Binaries

Thoth Hermes11 Apr 2023 21:19 UTC
−25 points
0 comments3 min readLW link
(thothhermes.substack.com)

Bryan Bishop AMA on the Progress Forum

jasoncrawford11 Apr 2023 21:05 UTC
8 points
0 comments1 min readLW link
(rootsofprogress.org)

AI Risk US Pres­i­den­tial Candidate

Simon Berens11 Apr 2023 19:31 UTC
5 points
3 comments1 min readLW link

Evolu­tion pro­vides no ev­i­dence for the sharp left turn

Quintin Pope11 Apr 2023 18:43 UTC
193 points
62 comments15 min readLW link

On “aiming for con­ver­gence on truth”

gjm11 Apr 2023 18:19 UTC
67 points
55 comments13 min readLW link

In fa­vor of ac­cel­er­at­ing prob­lems you’re try­ing to solve

Christopher King11 Apr 2023 18:15 UTC
2 points
2 comments4 min readLW link

[In­ter­view w/​ Jeffrey Ladish] Ap­ply­ing the ‘se­cu­rity mind­set’ to AI and x-risk

fowlertm11 Apr 2023 18:14 UTC
12 points
0 comments1 min readLW link

Re­quest to AGI or­ga­ni­za­tions: Share your views on paus­ing AI progress

11 Apr 2023 17:30 UTC
141 points
11 comments1 min readLW link

Core of AI pro­jec­tions from first prin­ci­ples: At­tempt 1

tailcalled11 Apr 2023 17:24 UTC
21 points
3 comments3 min readLW link

What Ja­son has been read­ing, April 2023

jasoncrawford11 Apr 2023 16:29 UTC
18 points
0 comments5 min readLW link
(rootsofprogress.org)

What about an AI that’s SUPPOSED to kill us (not ChaosGPT; only on pa­per)?

False Name11 Apr 2023 16:09 UTC
−13 points
1 comment3 min readLW link

Con­tra-Berkeley

False Name11 Apr 2023 16:06 UTC
0 points
0 comments4 min readLW link

Con­tra-Wittgen­stein; no postmodernism

False Name11 Apr 2023 16:05 UTC
−17 points
1 comment5 min readLW link

[MLSN #9] Ver­ify­ing large train­ing runs, se­cu­rity risks from LLM ac­cess to APIs, why nat­u­ral se­lec­tion may fa­vor AIs over humans

11 Apr 2023 16:03 UTC
11 points
0 comments6 min readLW link
(newsletter.mlsafety.org)

Where’s the foom?

Fergus Fettes11 Apr 2023 15:50 UTC
34 points
27 comments2 min readLW link

“The Need for Long-term Re­search”—Seeds of Science call for reviewers

rogersbacon11 Apr 2023 15:37 UTC
15 points
0 comments1 min readLW link

NTIA—AI Ac­countabil­ity Announcement

samshap11 Apr 2023 15:03 UTC
7 points
0 comments1 min readLW link
(www.ntia.doc.gov)

A cou­ple of ques­tions about Con­jec­ture’s Cog­ni­tive Emu­la­tion proposal

Igor Ivanov11 Apr 2023 14:05 UTC
25 points
1 comment3 min readLW link

Child­hood Roundup #2

Zvi11 Apr 2023 13:50 UTC
31 points
4 comments19 min readLW link
(thezvi.wordpress.com)

Mea­sur­ing ar­tifi­cial in­tel­li­gence on hu­man bench­marks is naive

Anomalous11 Apr 2023 11:34 UTC
11 points
4 comments1 min readLW link
(forum.effectivealtruism.org)

Killing Socrates

[DEACTIVATED] Duncan Sabien11 Apr 2023 10:28 UTC
173 points
144 comments8 min readLW link

Cy­berspace Ad­minis­tra­tion of China: Draft of “Reg­u­la­tion for Gen­er­a­tive Ar­tifi­cial In­tel­li­gence Ser­vices” is open for comments

sanxiyn11 Apr 2023 9:32 UTC
7 points
2 comments1 min readLW link
(archive.is)

[Question] Is “Strong Co­her­ence” Anti-Nat­u­ral?

DragonGod11 Apr 2023 6:22 UTC
23 points
25 comments2 min readLW link

Four mind­set dis­agree­ments be­hind ex­is­ten­tial risk dis­agree­ments in ML

Rob Bensinger11 Apr 2023 4:53 UTC
136 points
12 comments1 min readLW link