Test­ing for par­allel rea­son­ing in LLMs

19 May 2024 15:28 UTC
9 points
7 comments9 min readLW link

Hot take: The AI safety move­ment is way too sec­tar­ian and this is greatly in­creas­ing p(doom)

O O19 May 2024 2:18 UTC
14 points
15 comments2 min readLW link

Some “meta-cruxes” for AI x-risk debates

Aryeh Englander19 May 2024 0:21 UTC
20 points
2 comments3 min readLW link

On Privilege

Shmi18 May 2024 22:36 UTC
16 points
10 comments2 min readLW link

Fund me please—I Work so Hard that my Feet start Bleed­ing and I Need to In­fil­trate University

Johannes C. Mayer18 May 2024 19:53 UTC
22 points
37 comments6 min readLW link

To Limit Im­pact, Limit KL-Divergence

J Bostock18 May 2024 18:52 UTC
10 points
1 comment5 min readLW link

[Question] Are There Other Ideas as Gen­er­ally Ap­pli­ca­ble as Nat­u­ral Selection

Amin Sennour18 May 2024 16:37 UTC
1 point
1 comment1 min readLW link

Scien­tific No­ta­tion Options

jefftk18 May 2024 15:10 UTC
27 points
13 comments1 min readLW link
(www.jefftk.com)

“If we go ex­tinct due to mis­al­igned AI, at least na­ture will con­tinue, right? … right?”

plex18 May 2024 14:09 UTC
54 points
23 comments2 min readLW link
(aisafety.info)

What Are Non-Zero-Sum Games?—A Primer

James Stephen Brown18 May 2024 9:19 UTC
4 points
7 comments3 min readLW link

Deep­Mind’s “​​Fron­tier Safety Frame­work” is weak and unambitious

Zach Stein-Perlman18 May 2024 3:00 UTC
159 points
14 comments4 min readLW link

In­ter­na­tional Scien­tific Re­port on the Safety of Ad­vanced AI: Key Information

Aryeh Englander18 May 2024 1:45 UTC
39 points
0 comments13 min readLW link

Good­hart in RL with KL: Appendix

Thomas Kwa18 May 2024 0:40 UTC
12 points
0 comments6 min readLW link

AI 2030 – AI Policy Roadmap

LTM17 May 2024 23:29 UTC
8 points
0 comments1 min readLW link

MIT Fu­tureTech are hiring for an Oper­a­tions and Pro­ject Man­age­ment role.

peterslattery17 May 2024 23:21 UTC
2 points
0 comments3 min readLW link

Lan­guage Models Model Us

eggsyntax17 May 2024 21:00 UTC
159 points
55 comments7 min readLW link

Towards Guaran­teed Safe AI: A Frame­work for En­sur­ing Ro­bust and Reli­able AI Systems

Joar Skalse17 May 2024 19:13 UTC
67 points
10 comments2 min readLW link

Agency

A*17 May 2024 19:11 UTC
8 points
0 comments1 min readLW link

Deep­Mind: Fron­tier Safety Framework

Zach Stein-Perlman17 May 2024 17:30 UTC
64 points
0 comments3 min readLW link
(deepmind.google)

Iden­ti­fy­ing Func­tion­ally Im­por­tant Fea­tures with End-to-End Sparse Dic­tionary Learning

17 May 2024 16:25 UTC
57 points
20 comments4 min readLW link
(arxiv.org)

AISafety.com – Re­sources for AI Safety

17 May 2024 15:57 UTC
83 points
3 comments1 min readLW link

Is There Really a Child Penalty in the Long Run?

Maxwell Tabarrok17 May 2024 11:56 UTC
23 points
6 comments5 min readLW link
(www.maximum-progress.com)

My Ham­mer Time Fi­nal Exam

adios17 May 2024 9:28 UTC
10 points
3 comments3 min readLW link

[Question] Is there a place to find the most cited LW ar­ti­cles of all time?

keltan17 May 2024 1:20 UTC
4 points
3 comments1 min readLW link

D&D.Sci (Easy Mode): On The Con­struc­tion Of Im­pos­si­ble Structures

abstractapplic17 May 2024 0:25 UTC
34 points
12 comments2 min readLW link

To an LLM, ev­ery­thing looks like a logic puzzle

Jesse Richardson16 May 2024 22:21 UTC
14 points
2 comments2 min readLW link

AI Safety In­sti­tute’s In­spect hello world ex­am­ple for AI evals

TheManxLoiner16 May 2024 20:47 UTC
3 points
0 comments1 min readLW link
(lovkush.medium.com)

Feel­ing (in­stru­men­tally) Rational

Morphism16 May 2024 18:56 UTC
14 points
5 comments1 min readLW link

Ad­vice for Ac­tivists from the His­tory of Environmentalism

Jeffrey Heninger16 May 2024 18:40 UTC
100 points
10 comments6 min readLW link
(blog.aiimpacts.org)

Ninety-five the­ses on AI

hamandcheese16 May 2024 17:51 UTC
21 points
0 comments7 min readLW link

GPT-4o My and Google I/​O Day

Zvi16 May 2024 17:50 UTC
41 points
2 comments37 min readLW link
(thezvi.wordpress.com)

AI #64: Feel the Mun­dane Utility

Zvi16 May 2024 15:20 UTC
28 points
11 comments47 min readLW link
(thezvi.wordpress.com)

AISN #35: Lob­by­ing on AI Reg­u­la­tion Plus, New Models from OpenAI and Google, and Le­gal Regimes for Train­ing on Copy­righted Data

16 May 2024 14:29 UTC
2 points
3 comments6 min readLW link
(newsletter.safe.ai)

FMT: a great op­por­tu­nity for (soon-to-be) parents

EternallyBlissful16 May 2024 13:24 UTC
13 points
1 comment18 min readLW link

Towards Guaran­teed Safe AI: A Frame­work for En­sur­ing Ro­bust and Reli­able AI Systems

Gunnar_Zarncke16 May 2024 13:09 UTC
51 points
20 comments1 min readLW link
(arxiv.org)

The Dun­ning-Kruger of dis­prov­ing Dun­ning-Kruger

kromem16 May 2024 10:11 UTC
56 points
0 comments5 min readLW link

A case for fair­ness-en­forc­ing ir­ra­tional behavior

cousin_it16 May 2024 9:41 UTC
16 points
3 comments2 min readLW link

Pod­cast: Eye4AI on 2023 Survey

KatjaGrace16 May 2024 7:40 UTC
8 points
0 comments1 min readLW link
(worldspiritsockpuppet.com)

Against “ar­gu­ment from over­hang risk”

RobertM16 May 2024 4:44 UTC
31 points
11 comments5 min readLW link

[Question] How can I make the most of Less On­line/​Camp/​Man­i­fest?

ErioirE16 May 2024 2:19 UTC
13 points
2 comments1 min readLW link

Do you be­lieve in hun­dred dol­lar bills ly­ing on the ground? Con­sider humming

Elizabeth16 May 2024 0:00 UTC
122 points
22 comments6 min readLW link
(acesounderglass.com)

Tack­ling Moloch: How YouCongress Offers a Novel Co­or­di­na­tion Mechanism

Hector Perez Arenas15 May 2024 23:13 UTC
29 points
9 comments6 min readLW link

In­tro­duc­ing Statis­ti­cal Utility Me­chan­ics: A Frame­work for Utility Maximizers

J Bostock15 May 2024 21:56 UTC
12 points
0 comments7 min readLW link

Tam­ing In­finity (Stat Mech Part 3)

J Bostock15 May 2024 21:43 UTC
11 points
0 comments7 min readLW link

Let’s De­sign A School, Part 2.3 School as Ed­u­ca­tion—The Cur­ricu­lum (Phase 2, Spe­cific)

Sable15 May 2024 20:58 UTC
24 points
0 comments14 min readLW link
(affablyevil.substack.com)

“A Paradigm for AI Con­scious­ness”—Seeds of Science call for reviewers

rogersbacon15 May 2024 20:55 UTC
7 points
0 comments1 min readLW link

Why you should learn a mu­si­cal instrument

cata15 May 2024 20:36 UTC
51 points
23 comments3 min readLW link

Con­tra Caller Gen­der III

jefftk15 May 2024 19:40 UTC
8 points
0 comments2 min readLW link
(www.jefftk.com)

In­struc­tion-fol­low­ing AGI is eas­ier and more likely than value al­igned AGI

Seth Herd15 May 2024 19:38 UTC
80 points
28 comments12 min readLW link

Why I’ll Keep My Crummy Draw­ings—How Gen­er­a­tive AI Art Won’t Sup­plant… Art.

James Stephen Brown15 May 2024 19:30 UTC
18 points
7 comments4 min readLW link