An ML pa­per on data steal­ing pro­vides a con­struc­tion for “gra­di­ent hack­ing”

David Scott Krueger (formerly: capybaralet)Jul 30, 2024, 9:44 PM
21 points
1 comment1 min readLW link
(arxiv.org)

Open Source Au­to­mated In­ter­pretabil­ity for Sparse Au­toen­coder Features

Jul 30, 2024, 9:11 PM
67 points
1 comment13 min readLW link
(blog.eleuther.ai)

Cater­pillars and Philosophy

Zero ContradictionsJul 30, 2024, 8:54 PM
2 points
0 comments1 min readLW link
(thewaywardaxolotl.blogspot.com)

François Chol­let on the limi­ta­tions of LLMs in reasoning

2PuNCheeZJul 30, 2024, 8:04 PM
1 point
1 comment2 min readLW link
(x.com)

Against AI As An Ex­is­ten­tial Risk

Noah BirnbaumJul 30, 2024, 7:10 PM
6 points
13 comments1 min readLW link
(irrationalitycommunity.substack.com)

[Question] Is ob­jec­tive moral­ity self-defeat­ing?

dialecticaJul 30, 2024, 6:23 PM
−4 points
3 comments2 min readLW link

Limi­ta­tions on the In­ter­pretabil­ity of Learned Fea­tures from Sparse Dic­tionary Learning

Tom AngstenJul 30, 2024, 4:36 PM
6 points
0 comments9 min readLW link

Self-Other Over­lap: A Ne­glected Ap­proach to AI Alignment

Jul 30, 2024, 4:22 PM
223 points
51 comments12 min readLW link

In­ves­ti­gat­ing the Abil­ity of LLMs to Rec­og­nize Their Own Writing

Jul 30, 2024, 3:41 PM
32 points
0 comments15 min readLW link

Can Gen­er­al­ized Ad­ver­sar­ial Test­ing En­able More Ri­gor­ous LLM Safety Evals?

scasperJul 30, 2024, 2:57 PM
25 points
0 comments4 min readLW link

RTFB: Cal­ifor­nia’s AB 3211

ZviJul 30, 2024, 1:10 PM
62 points
2 comments11 min readLW link
(thezvi.wordpress.com)

If You Can Climb Up, You Can Climb Down

jefftkJul 30, 2024, 12:00 AM
34 points
9 comments1 min readLW link
(www.jefftk.com)

What is Mo­ral­ity?

Zero ContradictionsJul 29, 2024, 7:19 PM
−1 points
0 comments1 min readLW link
(thewaywardaxolotl.blogspot.com)

Arch-an­ar­chism and im­mor­tal­ity

Peter lawless Jul 29, 2024, 6:10 PM
−5 points
1 comment2 min readLW link

AI Safety Newslet­ter #39: Im­pli­ca­tions of a Trump Ad­minis­tra­tion for AI Policy Plus, Safety Engineering

Jul 29, 2024, 5:50 PM
17 points
1 comment6 min readLW link
(newsletter.safe.ai)

New Blog Post Against AI Doom

Noah BirnbaumJul 29, 2024, 5:21 PM
1 point
5 comments1 min readLW link
(substack.com)

An In­ter­pretabil­ity Illu­sion from Pop­u­la­tion Statis­tics in Causal Analysis

Daniel TanJul 29, 2024, 2:50 PM
9 points
3 comments1 min readLW link

[Question] How to­k­eniza­tion in­fluences prompt­ing?

Boris KashirinJul 29, 2024, 10:28 AM
9 points
4 comments1 min readLW link

Un­der­stand­ing Po­si­tional Fea­tures in Layer 0 SAEs

Jul 29, 2024, 9:36 AM
43 points
0 comments5 min readLW link

Pre­dic­tion Mar­kets Explained

Benjamin_SturiskyJul 29, 2024, 8:02 AM
8 points
0 comments9 min readLW link

Rel­a­tivity The­ory for What the Fu­ture ‘You’ Is and Isn’t

FlorianHJul 29, 2024, 2:01 AM
7 points
49 comments4 min readLW link

Wittgen­stein and Word2vec: Cap­tur­ing Re­la­tional Mean­ing in Lan­guage and Thought

cleanwhiteroomJul 28, 2024, 7:55 PM
2 points
2 comments2 min readLW link

Mak­ing Beliefs Pay Rent

Jul 28, 2024, 5:59 PM
7 points
2 comments1 min readLW link

This is already your sec­ond chance

MalmesburyJul 28, 2024, 5:13 PM
185 points
13 comments8 min readLW link

[Question] Has Eliezer pub­li­cly and satis­fac­to­rily re­sponded to at­tempted re­but­tals of the anal­ogy to evolu­tion?

kalerJul 28, 2024, 12:23 PM
10 points
14 comments1 min readLW link

Fam­ily and Society

Zero ContradictionsJul 28, 2024, 7:05 AM
1 point
0 comments1 min readLW link
(thewaywardaxolotl.blogspot.com)

[Question] What is AI Safety’s line of re­treat?

RemmeltJul 28, 2024, 5:43 AM
12 points
12 commentsLW link

AXRP Epi­sode 34 - AI Eval­u­a­tions with Beth Barnes

DanielFilanJul 28, 2024, 3:30 AM
23 points
0 comments69 min readLW link

Rats, Back a Candidate

BlakeJul 28, 2024, 3:19 AM
−40 points
19 comments1 min readLW link

AI ex­is­ten­tial risk prob­a­bil­ities are too un­re­li­able to in­form policy

Oleg TrottJul 28, 2024, 12:59 AM
18 points
5 comments1 min readLW link
(www.aisnakeoil.com)

Idle Spec­u­la­tions on Pipeline Parallelism

DaemonicSigilJul 27, 2024, 10:40 PM
1 point
0 comments4 min readLW link
(pbement.com)

Re: An­thropic’s sug­gested SB-1047 amendments

RobertMJul 27, 2024, 10:32 PM
87 points
13 comments9 min readLW link
(www.documentcloud.org)

The prob­lem with psy­chol­ogy is that it has no the­ory.

Nicholas D.Jul 27, 2024, 7:36 PM
2 points
7 comments4 min readLW link
(nicholasdecker.substack.com)

Bryan John­son and a search for healthy longevity

NancyLebovitzJul 27, 2024, 3:28 PM
18 points
17 comments1 min readLW link

What are match­ing mar­kets?

ohmurphyJul 27, 2024, 3:05 PM
12 points
0 comments8 min readLW link
(ohmurphy.substack.com)

Safety con­sul­ta­tions for AI lab employees

Zach Stein-PerlmanJul 27, 2024, 3:00 PM
181 points
4 comments1 min readLW link

The Case Against UBI

Zero ContradictionsJul 27, 2024, 6:36 AM
−1 points
2 comments2 min readLW link
(thewaywardaxolotl.blogspot.com)

Un­lock­ing Solu­tions—By Un­der­stand­ing Co­or­di­na­tion Problems

James Stephen BrownJul 27, 2024, 4:52 AM
56 points
4 comments5 min readLW link
(nonzerosum.games)

Utili­tar­i­anism and the re­place­abil­ity of de­sires and attachments

MichaelStJulesJul 27, 2024, 1:57 AM
5 points
2 commentsLW link

In­spired by: Failures in Kindness

X4vierJul 27, 2024, 1:21 AM
60 points
2 comments3 min readLW link

My Ex­pe­rience Us­ing Gam­ifi­ca­tion

Wyatt SJul 26, 2024, 11:06 PM
13 points
4 comments4 min readLW link

How the AI safety tech­ni­cal land­scape has changed in the last year, ac­cord­ing to some practitioners

tlevinJul 26, 2024, 7:06 PM
55 points
6 comments2 min readLW link

A Vi­sual Task that’s Hard for GPT-4o, but Doable for Pri­mary Schoolers

Lennart FinkeJul 26, 2024, 5:51 PM
25 points
6 comments2 min readLW link

Unal­igned AI is com­ing re­gard­less.

verbalshadowJul 26, 2024, 4:41 PM
−15 points
3 comments2 min readLW link

In­dex of ra­tio­nal­ist groups in the Bay Area July 2024

Jul 26, 2024, 4:32 PM
39 points
14 comments2 min readLW link

End Sin­gle Fam­ily Zon­ing by Over­turn­ing Eu­clid V Ambler

Maxwell TabarrokJul 26, 2024, 2:08 PM
32 points
1 comment7 min readLW link
(www.maximum-progress.com)

Com­mon Uses of “Ac­cep­tance”

Yi-YangJul 26, 2024, 11:18 AM
14 points
5 comments24 min readLW link

Univer­sal Ba­sic In­come and Poverty

Eliezer YudkowskyJul 26, 2024, 7:23 AM
328 points
141 comments9 min readLW link

A Solomonoff In­duc­tor Walks Into a Bar: Schel­ling Points for Communication

Jul 26, 2024, 12:33 AM
93 points
2 comments13 min readLW link

What does a Gam­bler’s Ver­ity world look like?

ErioirEJul 25, 2024, 10:03 PM
7 points
6 comments1 min readLW link