Utility is not the se­lec­tion target

tailcalledNov 4, 2023, 10:48 PM
24 points
1 comment1 min readLW link

Stuxnet, not Skynet: Hu­man­ity’s dis­em­pow­er­ment by AI

RokoNov 4, 2023, 10:23 PM
107 points
24 comments6 min readLW link

The 6D effect: When com­pa­nies take risks, one email can be very pow­er­ful.

scasperNov 4, 2023, 8:08 PM
279 points
42 comments3 min readLW link

Ge­netic fit­ness is a mea­sure of se­lec­tion strength, not the se­lec­tion target

Kaj_SotalaNov 4, 2023, 7:02 PM
58 points
44 comments18 min readLW link

The Soul Key

Richard_NgoNov 4, 2023, 5:51 PM
112 points
10 comments8 min readLW link1 review
(www.narrativeark.xyz)

[Linkpost] Con­cept Align­ment as a Pr­ereq­ui­site for Value Alignment

Bogdan Ionut CirsteaNov 4, 2023, 5:34 PM
27 points
0 comments1 min readLW link
(arxiv.org)

We are already in a per­sua­sion-trans­formed world and must take precautions

trevorNov 4, 2023, 3:53 PM
37 points
14 comments6 min readLW link

Be­ing good at the basics

dominicqNov 4, 2023, 2:18 PM
33 points
1 comment3 min readLW link

If a lit­tle is good, is more bet­ter?

DanielFilanNov 4, 2023, 7:10 AM
25 points
16 comments2 min readLW link
(danielfilan.com)

Un­trusted smart mod­els and trusted dumb models

BuckNov 4, 2023, 3:06 AM
87 points
17 comments6 min readLW link1 review

As Many Ideas

ScrewtapeNov 3, 2023, 10:47 PM
11 points
0 comments4 min readLW link

Paul Chris­ti­ano on Dwarkesh Podcast

ESRogsNov 3, 2023, 10:13 PM
19 points
0 comments1 min readLW link
(www.dwarkeshpatel.com)

De­cep­tion Chess: Game #1

Nov 3, 2023, 9:13 PM
111 points
22 comments8 min readLW link1 review

8 ex­am­ples in­form­ing my pes­simism on up­load­ing with­out re­verse engineering

Steven ByrnesNov 3, 2023, 8:03 PM
118 points
12 comments12 min readLW link

In­tegrity in AI Gover­nance and Advocacy

Nov 3, 2023, 7:52 PM
134 points
57 comments23 min readLW link

Aver­ag­ing sam­ples from a pop­u­la­tion with log-nor­mal distribution

CrimsonChinNov 3, 2023, 7:42 PM
8 points
2 comments1 min readLW link

Se­cur­ing Civ­i­liza­tion Against Catas­trophic Pandemics

jefftkNov 3, 2023, 7:33 PM
13 points
0 comments1 min readLW link
(dam.gcsp.ch)

The Unavoid­able Ex­pe­rience of Free Will in a Deter­minis­tic World

gmaxNov 3, 2023, 5:55 PM
−12 points
0 comments3 min readLW link

Thoughts on open source AI

Sam MarksNov 3, 2023, 3:35 PM
62 points
17 comments10 min readLW link

Knowl­edge Base 6: Con­sen­sus the­ory of truth

iwisNov 3, 2023, 1:56 PM
−8 points
0 comments1 min readLW link

[Question] Shouldn’t we ‘Just’ Su­per­im­i­tate Low-Res Uploads?

lukemarksNov 3, 2023, 7:42 AM
15 points
2 comments2 min readLW link

The other side of the tidal wave

KatjaGraceNov 3, 2023, 5:40 AM
189 points
86 comments1 min readLW link
(worldspiritsockpuppet.com)

Does davi­dad’s up­load­ing moon­shot work?

Nov 3, 2023, 2:21 AM
146 points
35 comments25 min readLW link

Twin Cities ACX Meetup—Novem­ber 2023

Timothy M.Nov 3, 2023, 12:47 AM
1 point
1 comment1 min readLW link

San Fran­cisco ACX Meetup “First Satur­day”

guenaelNov 3, 2023, 12:10 AM
4 points
0 comments1 min readLW link

[Question] What are your fa­vorite posts, pod­cast epi­sodes, and recorded talks, on AI timelines, or fac­tors that would in­fluence AI timelines?

nonzerosumNov 2, 2023, 10:42 PM
2 points
0 comments1 min readLW link

One Day Sooner

ScrewtapeNov 2, 2023, 7:00 PM
122 points
8 comments8 min readLW link1 review

Pro­pa­ganda or Science: A Look at Open Source AI and Bioter­ror­ism Risk

1a3ornNov 2, 2023, 6:20 PM
193 points
79 comments23 min readLW link

AI #36: In the Background

ZviNov 2, 2023, 6:00 PM
45 points
5 comments37 min readLW link
(thezvi.wordpress.com)

Doubt Certainty

RationalDinoNov 2, 2023, 5:43 PM
4 points
13 comments3 min readLW link

Say­ing the quiet part out loud: trad­ing off x-risk for per­sonal immortality

disturbanceNov 2, 2023, 5:43 PM
84 points
89 comments5 min readLW link

Mech In­terp Challenge: Novem­ber—De­ci­pher­ing the Cu­mu­la­tive Sum Model

CallumMcDougallNov 2, 2023, 5:10 PM
18 points
2 comments2 min readLW link

Es­ti­mat­ing effec­tive di­men­sion­al­ity of MNIST models

Arjun PanicksseryNov 2, 2023, 2:13 PM
41 points
3 comments1 min readLW link

Aver­ages and sam­ple sizes

mruwnikNov 2, 2023, 9:52 AM
15 points
2 comments8 min readLW link

ACX/​LW/​EA crossover meetup

RasmusHBNov 2, 2023, 5:57 AM
2 points
0 comments1 min readLW link

Up­com­ing Feed­back Op­por­tu­nity on Dual-Use Foun­da­tion Models

Chris_LeongNov 2, 2023, 4:28 AM
3 points
0 comments1 min readLW link

Public Weights?

jefftkNov 2, 2023, 2:50 AM
49 points
19 comments3 min readLW link
(www.jefftk.com)

[Question] Should peo­ple build pro­duc­ti­za­tions of open source AI mod­els?

lcNov 2, 2023, 1:26 AM
23 points
0 comments1 min readLW link

Sin­gu­lar learn­ing the­ory and bridg­ing from ML to brain emulations

Nov 1, 2023, 9:31 PM
26 points
16 comments29 min readLW link

My thoughts on the so­cial re­sponse to AI risk

Matthew BarnettNov 1, 2023, 9:17 PM
157 points
37 comments10 min readLW link

Re­ac­tions to the Ex­ec­u­tive Order

ZviNov 1, 2023, 8:40 PM
77 points
4 comments29 min readLW link
(thezvi.wordpress.com)

Dario Amodei’s pre­pared re­marks from the UK AI Safety Sum­mit, on An­thropic’s Re­spon­si­ble Scal­ing Policy

Zac Hatfield-DoddsNov 1, 2023, 6:10 PM
85 points
1 comment4 min readLW link
(www.anthropic.com)

Book Re­view: Deter­mined by Sapolsky

Kailuo WangNov 1, 2023, 5:37 PM
1 point
0 comments7 min readLW link

AI Align­ment: A Com­pre­hen­sive Survey

Stephen McAleerNov 1, 2023, 5:35 PM
20 points
1 comment1 min readLW link
(arxiv.org)

A list of all the dead­lines in Bi­den’s Ex­ec­u­tive Order on AI

Valentin BaltadzhievNov 1, 2023, 5:14 PM
26 points
2 comments11 min readLW link

2023 LessWrong Com­mu­nity Cen­sus, Re­quest for Comments

ScrewtapeNov 1, 2023, 4:32 PM
43 points
37 comments2 min readLW link

[Question] Snap­shot of nar­ra­tives and frames against reg­u­lat­ing AI

Jan_KulveitNov 1, 2023, 4:30 PM
36 points
19 comments3 min readLW link

Com­men­sal Institutions

SableNov 1, 2023, 4:01 PM
8 points
12 comments4 min readLW link
(affablyevil.substack.com)

ChatGPT’s On­tolog­i­cal Land­scape

Bill BenzonNov 1, 2023, 3:12 PM
7 points
0 comments4 min readLW link

On the Ex­ec­u­tive Order

ZviNov 1, 2023, 2:20 PM
100 points
4 comments30 min readLW link
(thezvi.wordpress.com)