In­creased Scam Qual­ity/​Quan­tity (Hy­poth­e­sis in need of data)?

BeeblebroxJan 9, 2023, 10:57 PM
9 points
6 comments1 min readLW link

Went­worth and Larsen on buy­ing time

Jan 9, 2023, 9:31 PM
74 points
6 comments12 min readLW link

EA & LW Fo­rum Sum­maries—Holi­day Edi­tion (19th Dec − 8th Jan)

Zoe WilliamsJan 9, 2023, 9:06 PM
11 points
0 commentsLW link

GWWC Should Re­quire Public Char­ity Evaluations

jefftkJan 9, 2023, 8:10 PM
28 points
0 comments4 min readLW link
(www.jefftk.com)

[MLSN #7]: an ex­am­ple of an emer­gent in­ter­nal optimizer

Jan 9, 2023, 7:39 PM
28 points
0 comments6 min readLW link

Try­ing to iso­late ob­jec­tives: ap­proaches to­ward high-level interpretability

JozdienJan 9, 2023, 6:33 PM
49 points
14 comments8 min readLW link

The spe­cial na­ture of spe­cial relativity

adamShimiJan 9, 2023, 5:30 PM
37 points
1 comment3 min readLW link
(epistemologicalvigilance.substack.com)

Pierre Me­nard, pixel art, and entropy

Joey MarcellinoJan 9, 2023, 4:34 PM
1 point
1 comment6 min readLW link

Fore­cast­ing ex­treme outcomes

AidanGothJan 9, 2023, 4:34 PM
4 points
1 comment2 min readLW link
(docs.google.com)

Ev­i­dence un­der Ad­ver­sar­ial Conditions

PeterMcCluskeyJan 9, 2023, 4:21 PM
57 points
1 comment3 min readLW link
(bayesianinvestor.com)

How to Bounded Distrust

ZviJan 9, 2023, 1:10 PM
122 points
17 comments4 min readLW link1 review
(thezvi.wordpress.com)

Reifi­ca­tion bias

Jan 9, 2023, 12:22 PM
25 points
6 comments2 min readLW link

Big list of AI safety videos

JakubKJan 9, 2023, 6:12 AM
11 points
2 comments1 min readLW link
(docs.google.com)

Ra­tion­al­ity Prac­tice: Self-Deception

DarmaniJan 9, 2023, 4:07 AM
6 points
0 comments1 min readLW link

Wolf In­ci­dent Postmortem

jefftkJan 9, 2023, 3:20 AM
137 points
13 comments1 min readLW link
(www.jefftk.com)

You’re Not One “You”—How De­ci­sion The­o­ries Are Talk­ing Past Each Other

keith_wynroeJan 9, 2023, 1:21 AM
28 points
11 comments8 min readLW link

On Blog­ging and Podcasting

DanielFilanJan 9, 2023, 12:40 AM
18 points
6 comments11 min readLW link
(danielfilan.com)

ChatGPT tells sto­ries about XP-708-DQ, Eliezer, drag­ons, dark sor­cer­esses, and un­al­igned robots be­com­ing aligned

Bill BenzonJan 8, 2023, 11:21 PM
6 points
2 comments18 min readLW link

Si­mu­lacra are Things

janusJan 8, 2023, 11:03 PM
63 points
7 comments2 min readLW link

[Question] GPT learn­ing from smarter texts?

ViliamJan 8, 2023, 10:23 PM
26 points
7 comments1 min readLW link

La­tent vari­able pre­dic­tion mar­kets mockup + de­signer request

tailcalledJan 8, 2023, 10:18 PM
25 points
4 comments1 min readLW link

Cita­bil­ity of Less­wrong and the Align­ment Forum

Leon LangJan 8, 2023, 10:12 PM
48 points
2 comments1 min readLW link

I tried to learn as much Deep Learn­ing math as I could in 24 hours

PhosphorousJan 8, 2023, 9:07 PM
31 points
2 comments7 min readLW link

[Question] What spe­cific thing would you do with AI Align­ment Re­search As­sis­tant GPT?

quetzal_rainbowJan 8, 2023, 7:24 PM
47 points
9 comments1 min readLW link

[Question] Re­search ideas (AI In­ter­pretabil­ity & Neu­ro­sciences) for a 2-months project

fluxJan 8, 2023, 3:36 PM
3 points
1 comment1 min readLW link

200 COP in MI: Image Model Interpretability

Neel NandaJan 8, 2023, 2:53 PM
18 points
3 comments6 min readLW link

Hal­i­fax Monthly Meetup: Moloch in the HRM

IdeopunkJan 8, 2023, 2:49 PM
10 points
0 comments1 min readLW link

Dangers of deference

TsviBTJan 8, 2023, 2:36 PM
62 points
5 comments2 min readLW link

Could evolu­tion pro­duce some­thing truly al­igned with its own op­ti­miza­tion stan­dards? What would an an­swer to this mean for AI al­ign­ment?

No77eJan 8, 2023, 11:04 AM
3 points
4 comments1 min readLW link

AI psy­chol­ogy should ground the the­o­ries of AI con­scious­ness and in­form hu­man-AI eth­i­cal in­ter­ac­tion design

Roman LeventovJan 8, 2023, 6:37 AM
20 points
8 comments2 min readLW link

Stop Talk­ing to Each Other and Start Buy­ing Things: Three Decades of Sur­vival in the Desert of So­cial Media

the gears to ascensionJan 8, 2023, 4:45 AM
1 point
14 comments1 min readLW link
(catvalente.substack.com)

Can Ads be GDPR Com­pli­ant?

jefftkJan 8, 2023, 2:50 AM
39 points
10 comments7 min readLW link
(www.jefftk.com)

Fea­ture sug­ges­tion: add a ‘clar­ity score’ to posts

LVSNJan 8, 2023, 1:00 AM
17 points
5 comments1 min readLW link

[Question] How do I bet­ter stick to a morn­ing sched­ule?

Randomized, ControlledJan 8, 2023, 12:52 AM
8 points
8 comments1 min readLW link

Pro­tec­tion­ism will Slow the De­ploy­ment of AI

Ben GoldhaberJan 7, 2023, 8:57 PM
30 points
6 comments2 min readLW link

David Krueger on AI Align­ment in Academia, Co­or­di­na­tion and Test­ing Intuitions

Michaël TrazziJan 7, 2023, 7:59 PM
13 points
0 comments4 min readLW link
(theinsideview.ai)

Look­ing for Span­ish AI Align­ment Researchers

AntbJan 7, 2023, 6:52 PM
7 points
3 comments1 min readLW link

Noth­ing New: Pro­duc­tive Reframing

adamShimiJan 7, 2023, 6:43 PM
44 points
7 comments3 min readLW link
(epistemologicalvigilance.substack.com)

[Question] Ask­ing for a name for a symp­tom of rationalization

metachiralityJan 7, 2023, 6:34 PM
6 points
5 comments1 min readLW link

The Foun­tain of Health: a First Prin­ci­ples Guide to Rejuvenation

PhilJacksonJan 7, 2023, 6:34 PM
115 points
39 comments41 min readLW link

What’s wrong with the pa­per­clips sce­nario?

No77eJan 7, 2023, 5:58 PM
31 points
11 comments1 min readLW link

Build­ing a Rosetta stone for re­duc­tion­ism and telism (WIP)

mrcbarbierJan 7, 2023, 4:22 PM
5 points
0 comments8 min readLW link

What should a telic sci­ence look like?

mrcbarbierJan 7, 2023, 4:13 PM
10 points
0 comments11 min readLW link

Open & Wel­come Thread—Jan­uary 2023

DragonGodJan 7, 2023, 11:16 AM
15 points
37 comments1 min readLW link

An­chor­ing fo­cal­ism and the Iden­ti­fi­able vic­tim effect: Bias in Eval­u­at­ing AGI X-Risks

RemmeltJan 7, 2023, 9:59 AM
1 point
2 commentsLW link

Can ChatGPT count?

p.b.Jan 7, 2023, 7:57 AM
13 points
11 comments2 min readLW link

Benev­olent AI and men­tal health

peter schwarzJan 7, 2023, 1:30 AM
−31 points
2 comments1 min readLW link

An Ig­no­rant View on Ineffec­tive­ness of AI Safety

IknownothingJan 7, 2023, 1:29 AM
14 points
7 comments3 min readLW link

Op­ti­miz­ing Hu­man Col­lec­tive In­tel­li­gence to Align AI

Shoshannah TekofskyJan 7, 2023, 1:21 AM
12 points
5 comments6 min readLW link

[Question] [Dis­cus­sion] How Broad is the Hu­man Cog­ni­tive Spec­trum?

DragonGodJan 7, 2023, 12:56 AM
29 points
51 comments2 min readLW link