Product safety is a poor model for AI governance

Richard Korzekwa 1 Feb 2023 22:40 UTC
36 points
0 comments5 min readLW link
(aiimpacts.org)

Hin­ton: “mor­tal” effi­cient ana­log hard­ware may be learned-in-place, uncopyable

the gears to ascension1 Feb 2023 22:19 UTC
10 points
3 comments1 min readLW link

Can we “cure” can­cer?

jasoncrawford1 Feb 2023 22:03 UTC
41 points
31 comments2 min readLW link
(rootsofprogress.org)

Eli Lifland on Nav­i­gat­ing the AI Align­ment Landscape

ozziegooen1 Feb 2023 21:17 UTC
9 points
1 comment31 min readLW link
(quri.substack.com)

Schizophre­nia as a defi­ciency in long-range cor­tex-to-cor­tex communication

Steven Byrnes1 Feb 2023 19:32 UTC
34 points
29 comments11 min readLW link

AI Safety Ar­gu­ments: An In­ter­ac­tive Guide

Lukas Trötzmüller1 Feb 2023 19:26 UTC
20 points
0 comments3 min readLW link

More find­ings on Me­moriza­tion and dou­ble descent

Marius Hobbhahn1 Feb 2023 18:26 UTC
53 points
2 comments19 min readLW link

Lan­guage Models can be Utility-Max­imis­ing Agents

Raymond D1 Feb 2023 18:13 UTC
22 points
1 comment2 min readLW link

Trends in the dol­lar train­ing cost of ma­chine learn­ing systems

Ben Cottier1 Feb 2023 14:48 UTC
23 points
0 comments2 min readLW link
(epochai.org)

Po­lis: Why and How to Use it

brook1 Feb 2023 14:03 UTC
3 points
0 comments1 min readLW link

Su­biti­sa­tion of Self

vitaliya1 Feb 2023 9:18 UTC
14 points
4 comments2 min readLW link

Directed Babbling

Yudhister Kumar1 Feb 2023 9:10 UTC
20 points
1 comment3 min readLW link
(www.ykumar.org)

Vot­ing Re­sults for the 2021 Review

Raemon1 Feb 2023 8:02 UTC
66 points
10 comments38 min readLW link

Ab­strac­tion As Sym­me­try and Other Thoughts

Numendil1 Feb 2023 6:25 UTC
28 points
9 comments2 min readLW link

The effect of hori­zon length on scal­ing laws

Jacob_Hilton1 Feb 2023 3:59 UTC
23 points
2 comments1 min readLW link
(arxiv.org)

Con­tra Dance Lengths

jefftk1 Feb 2023 3:30 UTC
9 points
0 comments1 min readLW link
(www.jefftk.com)

Aiming for Con­ver­gence Is Like Dis­cour­ag­ing Betting

Zack_M_Davis1 Feb 2023 0:03 UTC
60 points
17 comments11 min readLW link

On value in hu­mans, other an­i­mals, and AI

Michele Campolo31 Jan 2023 23:33 UTC
3 points
17 comments5 min readLW link

Crit­i­cism of the main frame­work in AI alignment

Michele Campolo31 Jan 2023 23:01 UTC
19 points
2 comments6 min readLW link

Nice Clothes are Good, Actually

Gordon Seidoh Worley31 Jan 2023 19:22 UTC
62 points
28 comments4 min readLW link

[Linkpost] Hu­man-nar­rated au­dio ver­sion of “Is Power-Seek­ing AI an Ex­is­ten­tial Risk?”

Joe Carlsmith31 Jan 2023 19:21 UTC
12 points
1 comment1 min readLW link

No Really, At­ten­tion is ALL You Need—At­ten­tion can do feed­for­ward networks

Robert_AIZI31 Jan 2023 18:48 UTC
29 points
7 comments6 min readLW link
(aizi.substack.com)

Talk to me about your sum­mer/​ca­reer plans

Akash31 Jan 2023 18:29 UTC
31 points
3 comments2 min readLW link

Mechanis­tic In­ter­pretabil­ity Quick­start Guide

Neel Nanda31 Jan 2023 16:35 UTC
42 points
3 comments6 min readLW link
(www.neelnanda.io)

New Hackathon: Ro­bust­ness to dis­tri­bu­tion changes and ambiguity

Charbel-Raphaël31 Jan 2023 12:50 UTC
11 points
3 comments1 min readLW link

Squig­gle: Why and how to use it

brook31 Jan 2023 12:37 UTC
3 points
0 comments1 min readLW link

Be­ware of Fake Alternatives

silentbob31 Jan 2023 10:21 UTC
50 points
10 comments4 min readLW link

In­ner Misal­ign­ment in “Si­mu­la­tor” LLMs

Adam Scherlis31 Jan 2023 8:33 UTC
84 points
11 comments4 min readLW link

Why AI ex­perts’ jobs are always decades from be­ing automated

Allen Hoskins31 Jan 2023 3:01 UTC
0 points
1 comment5 min readLW link
(open.substack.com)

Ap­ply to HAIST/​MAIA’s AI Gover­nance Work­shop in DC (Feb 17-20)

31 Jan 2023 2:06 UTC
28 points
0 comments2 min readLW link

EA & LW Fo­rum Weekly Sum­mary (23rd − 29th Jan ’23)

Zoe Williams31 Jan 2023 0:36 UTC
12 points
0 comments1 min readLW link

Say­ing things be­cause they sound good

Adam Zerner31 Jan 2023 0:17 UTC
23 points
6 comments2 min readLW link

South Bay Meetup

DavidFriedman30 Jan 2023 23:35 UTC
2 points
0 comments1 min readLW link

Peter Thiel’s speech at Oxford De­bat­ing Union on tech­nolog­i­cal stag­na­tion, Nu­clear weapons, COVID, En­vi­ron­ment, Align­ment, ‘anti-anti anti-anti-clas­si­cal liber­al­ism’, Bostrom, LW, etc.

M. Y. Zuo30 Jan 2023 23:31 UTC
8 points
33 comments1 min readLW link

Med­i­cal Image Regis­tra­tion: The ob­scure field where Deep Me­saop­ti­miz­ers are already at the top of the bench­marks. (post + co­lab note­book)

Hastings30 Jan 2023 22:46 UTC
23 points
0 comments3 min readLW link

Hu­mans Can Be Man­u­ally Strategic

Screwtape30 Jan 2023 22:35 UTC
13 points
0 comments3 min readLW link

Why I hate the “ac­ci­dent vs. mi­suse” AI x-risk di­chotomy (quick thoughts on “struc­tural risk”)

David Scott Krueger (formerly: capybaralet)30 Jan 2023 18:50 UTC
32 points
41 comments2 min readLW link

2022 Unoffi­cial LessWrong Gen­eral Cen­sus

Screwtape30 Jan 2023 18:36 UTC
97 points
33 comments2 min readLW link

Call for sub­mis­sions: “(In)hu­man Values and Ar­tifi­cial Agency”, ALIFE 2023

the gears to ascension30 Jan 2023 17:37 UTC
29 points
4 comments1 min readLW link
(humanvaluesandartificialagency.com)

What I mean by “al­ign­ment is in large part about mak­ing cog­ni­tion aimable at all”

So8res30 Jan 2023 15:22 UTC
163 points
24 comments2 min readLW link

The En­ergy Re­quire­ments and Fea­si­bil­ity of Off-World Mining

clans30 Jan 2023 15:07 UTC
31 points
1 comment8 min readLW link
(locationtbd.home.blog)

What­ever their ar­gu­ments, Covid vac­cine scep­tics will prob­a­bly never con­vince me

contrarianbrit30 Jan 2023 13:42 UTC
8 points
10 comments3 min readLW link
(thomasprosser.substack.com)

Si­mu­lacra Levels Summary

Zvi30 Jan 2023 13:40 UTC
71 points
12 comments7 min readLW link
(thezvi.wordpress.com)

A Few Prin­ci­ples of Suc­cess­ful AI Design

Vestozia30 Jan 2023 10:42 UTC
1 point
0 comments8 min readLW link

Against Boltz­mann mesaoptimizers

porby30 Jan 2023 2:55 UTC
76 points
6 comments4 min readLW link

How Likely is Los­ing a Google Ac­count?

jefftk30 Jan 2023 0:20 UTC
52 points
11 comments3 min readLW link
(www.jefftk.com)

Model-driven feed­back could am­plify al­ign­ment failures

aogara30 Jan 2023 0:00 UTC
21 points
1 comment2 min readLW link

Take­aways from cal­ibra­tion training

Olli Järviniemi29 Jan 2023 19:09 UTC
38 points
1 comment3 min readLW link

Struc­ture, cre­ativity, and nov­elty

TsviBT29 Jan 2023 14:30 UTC
18 points
4 comments7 min readLW link

What is the ground re­al­ity of coun­tries tak­ing steps to re­cal­ibrate AI de­vel­op­ment to­wards Align­ment first?

Nebuch29 Jan 2023 13:26 UTC
8 points
6 comments3 min readLW link