AI Align­ment Break­throughs this week (10/​08/​23)

Logan Zoellner8 Oct 2023 23:30 UTC
30 points
14 comments6 min readLW link

“The Heart of Gam­ing is the Power Fan­tasy”, and Co­hab­itive Games

Raemon8 Oct 2023 21:02 UTC
81 points
49 comments4 min readLW link
(bottomfeeder.substack.com)

FAQ: What the heck is goal ag­nos­ti­cism?

porby8 Oct 2023 19:11 UTC
66 points
36 comments28 min readLW link

Time is ho­mo­ge­neous se­quen­tially-com­pos­able determination

TsviBT8 Oct 2023 14:58 UTC
14 points
0 comments21 min readLW link

Linkpost: Are Emer­gent Abil­ities in Large Lan­guage Models just In-Con­text Learn­ing?

Erich_Grunewald8 Oct 2023 12:14 UTC
12 points
6 comments2 min readLW link
(arxiv.org)

Bird-eye view vi­su­al­iza­tion of LLM activations

Sergii8 Oct 2023 12:12 UTC
11 points
2 comments1 min readLW link
(grgv.xyz)

Per­spec­tive Based Rea­son­ing Could Ab­solve CDT

dadadarren8 Oct 2023 11:22 UTC
4 points
5 comments5 min readLW link

The Gra­di­ent – The Ar­tifi­cial­ity of Alignment

mic8 Oct 2023 4:06 UTC
12 points
1 comment5 min readLW link
(thegradient.pub)

Com­par­ing An­thropic’s Dic­tionary Learn­ing to Ours

Robert_AIZI7 Oct 2023 23:30 UTC
136 points
8 comments4 min readLW link

A thought about the con­straints of debtless­ness in on­line communities

mako yass7 Oct 2023 21:26 UTC
57 points
23 comments1 min readLW link

Ar­gu­ments for util­i­tar­i­anism are im­pos­si­bil­ity ar­gu­ments un­der un­bounded prospects

MichaelStJules7 Oct 2023 21:08 UTC
7 points
7 comments21 min readLW link

Sam Alt­man’s sister, An­nie Alt­man, claims Sam has severely abused her

pl50157 Oct 2023 21:06 UTC
86 points
105 comments28 min readLW link

Griffin Island

jefftk7 Oct 2023 18:40 UTC
14 points
3 comments1 min readLW link
(www.jefftk.com)

Every Men­tion of EA in “Go­ing In­finite”

KirstenH7 Oct 2023 14:42 UTC
48 points
0 comments8 min readLW link
(open.substack.com)

Fix­ing In­sider Threats in the AI Sup­ply Chain

Madhav Malhotra7 Oct 2023 13:19 UTC
20 points
2 comments5 min readLW link

Con­tra Nora Belrose on Orthog­o­nal­ity Th­e­sis Be­ing Trivial

tailcalled7 Oct 2023 11:47 UTC
18 points
21 comments1 min readLW link

Re­lated Dis­cus­sion from Thomas Kwa’s MIRI Re­search Experience

Raemon7 Oct 2023 6:25 UTC
71 points
140 comments1 min readLW link

[Question] Cur­rent State of Prob­a­bil­is­tic Logic

lunatic_at_large7 Oct 2023 5:06 UTC
3 points
2 comments1 min readLW link

On the Re­la­tion­ship Between Vari­abil­ity and the Evolu­tion­ary Out­comes of Sys­tems in Nature

Artyom Shaposhnikov7 Oct 2023 3:06 UTC
2 points
0 comments1 min readLW link

An­nounc­ing Dialogues

Ben Pace7 Oct 2023 2:57 UTC
154 points
51 comments4 min readLW link

Don’t Dis­miss Sim­ple Align­ment Approaches

Chris_Leong7 Oct 2023 0:35 UTC
128 points
9 comments4 min readLW link

Link­ing Alt Accounts

jefftk6 Oct 2023 17:00 UTC
70 points
33 comments1 min readLW link
(www.jefftk.com)

Su­per-Ex­po­nen­tial ver­sus Ex­po­nen­tial Growth in Com­pute Price-Performance

moridinamael6 Oct 2023 16:23 UTC
37 points
21 comments2 min readLW link

A per­sonal ex­pla­na­tion of ELK con­cept and task.

Zeyu Qin6 Oct 2023 3:55 UTC
1 point
0 comments1 min readLW link

The Long-Term Fu­ture Fund is look­ing for a full-time fund chair

5 Oct 2023 22:18 UTC
52 points
0 comments7 min readLW link
(forum.effectivealtruism.org)

Prov­ably Safe AI

PeterMcCluskey5 Oct 2023 22:18 UTC
31 points
15 comments4 min readLW link
(bayesianinvestor.com)

Stampy’s AI Safety Info soft launch

5 Oct 2023 22:13 UTC
120 points
9 comments2 min readLW link

Im­pacts of AI on the hous­ing markets

PottedRosePetal5 Oct 2023 21:24 UTC
8 points
0 comments5 min readLW link

Towards Monose­man­tic­ity: De­com­pos­ing Lan­guage Models With Dic­tionary Learning

Zac Hatfield-Dodds5 Oct 2023 21:01 UTC
286 points
21 comments2 min readLW link
(transformer-circuits.pub)

Ideation and Tra­jec­tory Model­ling in Lan­guage Models

NickyP5 Oct 2023 19:21 UTC
15 points
2 comments10 min readLW link

A well-defined his­tory in mea­surable fac­tor spaces

Matthias G. Mayer5 Oct 2023 18:36 UTC
22 points
0 comments2 min readLW link

Eval­u­at­ing the his­tor­i­cal value mis­speci­fi­ca­tion argument

Matthew Barnett5 Oct 2023 18:34 UTC
162 points
140 comments7 min readLW link

Trans­la­tions Should Invert

abramdemski5 Oct 2023 17:44 UTC
46 points
19 comments3 min readLW link

Cen­sor­ship in LLMs is here to stay be­cause it mir­rors how our own in­tel­li­gence is structured

mnvr5 Oct 2023 17:37 UTC
3 points
0 comments1 min readLW link

Twin Cities ACX Meetup Oc­to­ber 2023

Timothy M.5 Oct 2023 16:29 UTC
1 point
2 comments1 min readLW link

This anime sto­ry­board doesn’t ex­ist: a graphic novel writ­ten and illus­trated by GPT4

RomanS5 Oct 2023 14:01 UTC
12 points
7 comments55 min readLW link

AI #32: Lie Detector

Zvi5 Oct 2023 13:50 UTC
45 points
19 comments44 min readLW link
(thezvi.wordpress.com)

Can the House Leg­is­late?

jefftk5 Oct 2023 13:40 UTC
26 points
6 comments2 min readLW link
(www.jefftk.com)

Mak­ing progress on the ``what al­ign­ment tar­get should be aimed at?″ ques­tion, is urgent

ThomasCederborg5 Oct 2023 12:55 UTC
2 points
0 comments18 min readLW link

Re­sponse to Quintin Pope’s Evolu­tion Pro­vides No Ev­i­dence For the Sharp Left Turn

Zvi5 Oct 2023 11:39 UTC
129 points
29 comments9 min readLW link

How to Get Ra­tion­al­ist Feedback

NicholasKross5 Oct 2023 2:03 UTC
13 points
0 comments2 min readLW link

On my AI Fable, and the im­por­tance of de re, de dicto, and de se refer­ence for AI alignment

PhilGoetz5 Oct 2023 0:50 UTC
9 points
4 comments1 min readLW link

Un­der­speci­fied Prob­a­bil­ities: A Thought Ex­per­i­ment

lunatic_at_large4 Oct 2023 22:25 UTC
8 points
4 comments2 min readLW link

Fra­ter­nal Birth Order Effect and the Ma­ter­nal Im­mune Hypothesis

Bucky4 Oct 2023 21:18 UTC
19 points
0 comments2 min readLW link

How to solve de­cep­tion and still fail.

Charlie Steiner4 Oct 2023 19:56 UTC
36 points
7 comments6 min readLW link

PortAu­dio M1 Latency

jefftk4 Oct 2023 19:10 UTC
8 points
5 comments1 min readLW link
(www.jefftk.com)

Open Philan­thropy is hiring for mul­ti­ple roles across our Global Catas­trophic Risks teams

aarongertler4 Oct 2023 18:04 UTC
6 points
0 comments3 min readLW link
(forum.effectivealtruism.org)

Safe­guard­ing Hu­man­ity: En­sur­ing AI Re­mains a Ser­vant, Not a Master

kgldeshapriya4 Oct 2023 17:52 UTC
−20 points
2 comments2 min readLW link

The 5 Pillars of Happiness

Gabi QUENE4 Oct 2023 17:50 UTC
−24 points
5 comments5 min readLW link

[Question] Us­ing Re­in­force­ment Learn­ing to try to con­trol the heat­ing of a build­ing (dis­trict heat­ing)

Tony Karlsson4 Oct 2023 17:47 UTC
3 points
5 comments1 min readLW link