[Question] Is there a fun­da­men­tal dis­tinc­tion be­tween simu­lat­ing a mind and simu­lat­ing *be­ing* a mind? Is this a use­ful and im­por­tant dis­tinc­tion?

Thoth Hermes8 Apr 2023 23:44 UTC
−17 points
8 comments2 min readLW link

“warn­ing about ai doom” is also “an­nounc­ing ca­pa­bil­ities progress to noobs”

the gears to ascension8 Apr 2023 23:42 UTC
23 points
5 comments3 min readLW link

Fea­ture Re­quest: Right Click to Copy LaTeX

DragonGod8 Apr 2023 23:27 UTC
18 points
4 comments1 min readLW link

ELCK might re­quire non­triv­ial scal­able al­ign­ment progress, and seems tractable enough to try

Alex Lawsen 8 Apr 2023 21:49 UTC
17 points
0 comments2 min readLW link

GPTs are Pre­dic­tors, not Imitators

Eliezer Yudkowsky8 Apr 2023 19:59 UTC
376 points
90 comments3 min readLW link

4 gen­er­a­tions of alignment

qbolec8 Apr 2023 19:59 UTC
1 point
0 comments3 min readLW link

The sur­pris­ing pa­ram­e­ter effi­ciency of vi­sion models

beren8 Apr 2023 19:44 UTC
77 points
28 comments4 min readLW link

Ran­dom Ob­ser­va­tion on AI goals

FTPickle8 Apr 2023 19:28 UTC
−11 points
2 comments1 min readLW link

Can we eval­u­ate the “tool ver­sus agent” AGI pre­dic­tion?

Xodarap8 Apr 2023 18:40 UTC
16 points
7 comments1 min readLW link

Rel­a­tive Ab­stracted Agency

Audere8 Apr 2023 16:57 UTC
14 points
6 comments5 min readLW link

The benev­olence of the butcher

dr_s8 Apr 2023 16:29 UTC
53 points
30 comments6 min readLW link

SERI MATS—Sum­mer 2023 Cohort

8 Apr 2023 15:32 UTC
71 points
25 comments4 min readLW link

AI Pro­pos­als at ‘Two Ses­sions’: AGI as ‘Two Bombs, One Satel­lite’?

Derek M. Jones8 Apr 2023 11:31 UTC
5 points
0 comments1 min readLW link
(www.chinatalk.media)

All images from the WaitButWhy se­quence on AI

trevor8 Apr 2023 7:36 UTC
72 points
5 comments2 min readLW link

Guidelines for pro­duc­tive discussions

ambigram8 Apr 2023 6:00 UTC
37 points
0 comments5 min readLW link

All AGI Safety ques­tions wel­come (es­pe­cially ba­sic ones) [April 2023]

steven04618 Apr 2023 4:21 UTC
57 points
88 comments2 min readLW link

Bring­ing Agency Into AGI Ex­tinc­tion Is Superfluous

George3d68 Apr 2023 4:02 UTC
28 points
18 comments5 min readLW link

La­gos, Nige­ria—ACX Mee­tups Every­where 2023

damola8 Apr 2023 3:55 UTC
1 point
0 comments1 min readLW link

Up­com­ing Changes in Large Lan­guage Models

Andrew Keenan Richardson8 Apr 2023 3:41 UTC
43 points
8 comments4 min readLW link
(mechanisticmind.com)

Con­sider The Hand Axe

ymeskhout8 Apr 2023 1:31 UTC
142 points
16 comments6 min readLW link

AGI as a new data point

Will Rodgers8 Apr 2023 1:01 UTC
−1 points
0 comments1 min readLW link

Parametrize Pri­or­ity Evaluations

SilverFlame8 Apr 2023 0:39 UTC
2 points
2 comments6 min readLW link

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down

Eliezer Yudkowsky8 Apr 2023 0:36 UTC
246 points
39 comments12 min readLW link

Hu­man­i­tar­ian Phase Tran­si­tion needed be­fore Tech­nolog­i­cal Sin­gu­lar­ity

Dr_What7 Apr 2023 23:17 UTC
−9 points
5 comments2 min readLW link

[Question] Thoughts about Hug­ging Face?

Ariel Kwiatkowski7 Apr 2023 23:17 UTC
7 points
0 comments1 min readLW link

[Question] Is it cor­rect to frame al­ign­ment as “pro­gram­ming a good philos­o­phy of mean­ing”?

Util7 Apr 2023 23:16 UTC
2 points
3 comments1 min readLW link

Select Agent Speci­fi­ca­tions as Nat­u­ral Abstractions

lukemarks7 Apr 2023 23:16 UTC
19 points
3 comments5 min readLW link

n=3 AI Risk Quick Math and Reasoning

lionhearted (Sebastian Marshall)7 Apr 2023 20:27 UTC
6 points
3 comments4 min readLW link

[Question] What are good al­ter­na­tives to Pre­dic­tion­book for per­sonal pre­dic­tion track­ing? Edited: I origi­nally thought it was down but it was just 500 un­til I though of clear­ing cook­ies.

sortega7 Apr 2023 19:18 UTC
4 points
4 comments1 min readLW link

En­vi­ron­ments for Mea­sur­ing De­cep­tion, Re­source Ac­qui­si­tion, and Eth­i­cal Violations

Dan H7 Apr 2023 18:40 UTC
51 points
2 comments2 min readLW link
(arxiv.org)

Su­per­in­tel­li­gence Is Not Omniscience

Jeffrey Heninger7 Apr 2023 16:30 UTC
15 points
20 comments8 min readLW link
(aiimpacts.org)

An ‘AGI Emer­gency Eject Cri­te­ria’ con­sen­sus could be re­ally use­ful.

tcelferact7 Apr 2023 16:21 UTC
5 points
0 comments1 min readLW link

Reli­a­bil­ity, Se­cu­rity, and AI risk: Notes from in­fosec text­book chap­ter 1

Akash7 Apr 2023 15:47 UTC
34 points
1 comment4 min readLW link

Pre-reg­is­ter­ing a study

Robert_AIZI7 Apr 2023 15:46 UTC
10 points
0 comments6 min readLW link
(aizi.substack.com)

Live dis­cus­sion at Eastercon

Douglas_Reay7 Apr 2023 15:25 UTC
5 points
0 comments1 min readLW link

[Question] ChatGTP “Writ­ing ” News Sto­ries for The Guardian?

jmh7 Apr 2023 12:16 UTC
1 point
4 comments1 min readLW link

Sto­ry­tel­ler’s con­ven­tion, 2223 A.D.

plex7 Apr 2023 11:54 UTC
8 points
0 comments2 min readLW link

Stampy’s AI Safety Info—New Distil­la­tions #1 [March 2023]

markov7 Apr 2023 11:06 UTC
42 points
0 comments2 min readLW link
(aisafety.info)

Beren’s “De­con­fus­ing Direct vs Amor­tised Op­ti­mi­sa­tion”

DragonGod7 Apr 2023 8:57 UTC
51 points
10 comments3 min readLW link

Goal al­ign­ment with­out al­ign­ment on episte­mol­ogy, ethics, and sci­ence is futile

Roman Leventov7 Apr 2023 8:22 UTC
20 points
2 comments2 min readLW link

Po­lio Lab Leak Caught with Wastew­a­ter Sampling

Cullen7 Apr 2023 1:06 UTC
82 points
3 comments1 min readLW link

Catch­ing the Eye of Sauron

Casey B.7 Apr 2023 0:40 UTC
221 points
68 comments4 min readLW link

[Question] How to par­allelize “in­her­ently” se­rial the­ory work?

NicholasKross7 Apr 2023 0:08 UTC
16 points
6 comments1 min readLW link

If Align­ment is Hard, then so is Self-Improvement

PavleMiha7 Apr 2023 0:08 UTC
21 points
20 comments1 min readLW link

Risks from GPT-4 Byproduct of Re­cur­sively Op­ti­miz­ing AIs

ben hayum7 Apr 2023 0:02 UTC
73 points
9 comments10 min readLW link
(forum.effectivealtruism.org)

An­thropic is fur­ther ac­cel­er­at­ing the Arms Race?

sapphire6 Apr 2023 23:29 UTC
82 points
22 comments1 min readLW link
(techcrunch.com)

Sugges­tion for safe AI struc­ture (Cu­rated Trans­par­ent De­ci­sions)

Kane Gregory6 Apr 2023 22:00 UTC
5 points
6 comments3 min readLW link

10 rea­sons why lists of 10 rea­sons might be a win­ning strategy

trevor6 Apr 2023 21:24 UTC
101 points
7 comments1 min readLW link

A Defense of Utilitarianism

Pareto Optimal6 Apr 2023 21:09 UTC
−3 points
2 comments5 min readLW link
(paretooptimal.substack.com)

One Does Not Sim­ply Re­place the Hu­mans

JerkyTreats6 Apr 2023 20:56 UTC
9 points
3 comments4 min readLW link
(www.lesswrong.com)