The Long-Term Fu­ture Fund is look­ing for a full-time fund chair

5 Oct 2023 22:18 UTC
52 points
0 comments7 min readLW link
(forum.effectivealtruism.org)

Prov­ably Safe AI

PeterMcCluskey5 Oct 2023 22:18 UTC
31 points
15 comments4 min readLW link
(bayesianinvestor.com)

Stampy’s AI Safety Info soft launch

5 Oct 2023 22:13 UTC
120 points
9 comments2 min readLW link

Im­pacts of AI on the hous­ing markets

PottedRosePetal5 Oct 2023 21:24 UTC
8 points
0 comments5 min readLW link

Towards Monose­man­tic­ity: De­com­pos­ing Lan­guage Models With Dic­tionary Learning

Zac Hatfield-Dodds5 Oct 2023 21:01 UTC
286 points
21 comments2 min readLW link
(transformer-circuits.pub)

Ideation and Tra­jec­tory Model­ling in Lan­guage Models

NickyP5 Oct 2023 19:21 UTC
15 points
2 comments10 min readLW link

A well-defined his­tory in mea­surable fac­tor spaces

Matthias G. Mayer5 Oct 2023 18:36 UTC
22 points
0 comments2 min readLW link

Eval­u­at­ing the his­tor­i­cal value mis­speci­fi­ca­tion argument

Matthew Barnett5 Oct 2023 18:34 UTC
162 points
140 comments7 min readLW link

Trans­la­tions Should Invert

abramdemski5 Oct 2023 17:44 UTC
46 points
19 comments3 min readLW link

Cen­sor­ship in LLMs is here to stay be­cause it mir­rors how our own in­tel­li­gence is structured

mnvr5 Oct 2023 17:37 UTC
3 points
0 comments1 min readLW link

Twin Cities ACX Meetup Oc­to­ber 2023

Timothy M.5 Oct 2023 16:29 UTC
1 point
2 comments1 min readLW link

This anime sto­ry­board doesn’t ex­ist: a graphic novel writ­ten and illus­trated by GPT4

RomanS5 Oct 2023 14:01 UTC
12 points
7 comments55 min readLW link

AI #32: Lie Detector

Zvi5 Oct 2023 13:50 UTC
45 points
19 comments44 min readLW link
(thezvi.wordpress.com)

Can the House Leg­is­late?

jefftk5 Oct 2023 13:40 UTC
26 points
6 comments2 min readLW link
(www.jefftk.com)

Mak­ing progress on the ``what al­ign­ment tar­get should be aimed at?″ ques­tion, is urgent

ThomasCederborg5 Oct 2023 12:55 UTC
2 points
0 comments18 min readLW link

Re­sponse to Quintin Pope’s Evolu­tion Pro­vides No Ev­i­dence For the Sharp Left Turn

Zvi5 Oct 2023 11:39 UTC
129 points
29 comments9 min readLW link

How to Get Ra­tion­al­ist Feedback

NicholasKross5 Oct 2023 2:03 UTC
13 points
0 comments2 min readLW link

On my AI Fable, and the im­por­tance of de re, de dicto, and de se refer­ence for AI alignment

PhilGoetz5 Oct 2023 0:50 UTC
9 points
4 comments1 min readLW link

Un­der­speci­fied Prob­a­bil­ities: A Thought Ex­per­i­ment

lunatic_at_large4 Oct 2023 22:25 UTC
8 points
4 comments2 min readLW link

Fra­ter­nal Birth Order Effect and the Ma­ter­nal Im­mune Hypothesis

Bucky4 Oct 2023 21:18 UTC
19 points
0 comments2 min readLW link

How to solve de­cep­tion and still fail.

Charlie Steiner4 Oct 2023 19:56 UTC
36 points
7 comments6 min readLW link

PortAu­dio M1 Latency

jefftk4 Oct 2023 19:10 UTC
8 points
5 comments1 min readLW link
(www.jefftk.com)

Open Philan­thropy is hiring for mul­ti­ple roles across our Global Catas­trophic Risks teams

aarongertler4 Oct 2023 18:04 UTC
6 points
0 comments3 min readLW link
(forum.effectivealtruism.org)

Safe­guard­ing Hu­man­ity: En­sur­ing AI Re­mains a Ser­vant, Not a Master

kgldeshapriya4 Oct 2023 17:52 UTC
−20 points
2 comments2 min readLW link

The 5 Pillars of Happiness

Gabi QUENE4 Oct 2023 17:50 UTC
−24 points
5 comments5 min readLW link

[Question] Us­ing Re­in­force­ment Learn­ing to try to con­trol the heat­ing of a build­ing (dis­trict heat­ing)

Tony Karlsson4 Oct 2023 17:47 UTC
3 points
5 comments1 min readLW link

ra­tio­nal­is­tic prob­a­bil­ity(lit­ter­ally just throw­ing shit out there)

NotaSprayer ASprayer4 Oct 2023 17:46 UTC
−30 points
8 comments2 min readLW link

AISN #23: New OpenAI Models, News from An­thropic, and Rep­re­sen­ta­tion Engineering

4 Oct 2023 17:37 UTC
15 points
2 comments5 min readLW link
(newsletter.safe.ai)

I don’t find the lie de­tec­tion re­sults that sur­pris­ing (by an au­thor of the pa­per)

JanB4 Oct 2023 17:10 UTC
97 points
8 comments3 min readLW link

[Question] What ev­i­dence is there of LLM’s con­tain­ing world mod­els?

Chris_Leong4 Oct 2023 14:33 UTC
17 points
17 comments1 min readLW link

En­tan­gle­ment and in­tu­ition about words and mean­ing

Bill Benzon4 Oct 2023 14:16 UTC
4 points
0 comments2 min readLW link

Why a Mars colony would lead to a first strike situation

Remmelt4 Oct 2023 11:29 UTC
−57 points
8 comments1 min readLW link
(mflb.com)

[Question] What are some ex­am­ples of AIs in­stan­ti­at­ing the ‘near­est un­blocked strat­egy prob­lem’?

EJT4 Oct 2023 11:05 UTC
6 points
4 comments1 min readLW link

Graph­i­cal ten­sor no­ta­tion for interpretability

Jordan Taylor4 Oct 2023 8:04 UTC
129 points
11 comments19 min readLW link

[Link] Bay Area Win­ter Sols­tice 2023

4 Oct 2023 2:19 UTC
18 points
3 comments1 min readLW link
(fb.me)

[Question] Who de­ter­mines whether an al­ign­ment pro­posal is the defini­tive al­ign­ment solu­tion?

MiguelDev3 Oct 2023 22:39 UTC
−1 points
6 comments1 min readLW link

AXRP Epi­sode 25 - Co­op­er­a­tive AI with Cas­par Oesterheld

DanielFilan3 Oct 2023 21:50 UTC
43 points
0 comments92 min readLW link

When to Get the Booster?

jefftk3 Oct 2023 21:00 UTC
50 points
15 comments2 min readLW link
(www.jefftk.com)

OpenAI-Microsoft partnership

Zach Stein-Perlman3 Oct 2023 20:01 UTC
51 points
18 comments1 min readLW link

[Question] Cur­rent AI safety tech­niques?

Zach Stein-Perlman3 Oct 2023 19:30 UTC
30 points
2 comments2 min readLW link

Test­ing and Au­toma­tion for In­tel­li­gent Sys­tems.

Sai Kiran Kammari3 Oct 2023 17:51 UTC
−13 points
0 comments1 min readLW link
(resource-cms.springernature.com)

Me­tac­u­lus An­nounces Fore­cast­ing Tour­na­ment to Eval­u­ate Fo­cused Re­search Or­ga­ni­za­tions, in Part­ner­ship With the Fed­er­a­tion of Amer­i­can Scien­tists

ChristianWilliams3 Oct 2023 16:44 UTC
13 points
0 comments1 min readLW link
(www.metaculus.com)

What would it mean to un­der­stand how a large lan­guage model (LLM) works? Some quick notes.

Bill Benzon3 Oct 2023 15:11 UTC
20 points
4 comments8 min readLW link

[Question] Po­ten­tial al­ign­ment tar­gets for a sovereign su­per­in­tel­li­gent AI

Paul Colognese3 Oct 2023 15:09 UTC
29 points
4 comments1 min readLW link

Monthly Roundup #11: Oc­to­ber 2023

Zvi3 Oct 2023 14:10 UTC
42 points
12 comments35 min readLW link
(thezvi.wordpress.com)

Why We Use Money? - A Walrasian View

Savio Coelho3 Oct 2023 12:02 UTC
4 points
3 comments8 min readLW link

Mech In­terp Challenge: Oc­to­ber—De­ci­pher­ing the Sorted List Model

CallumMcDougall3 Oct 2023 10:57 UTC
23 points
0 comments3 min readLW link

Early Ex­per­i­ments in Re­ward Model In­ter­pre­ta­tion Us­ing Sparse Autoencoders

3 Oct 2023 7:45 UTC
11 points
0 comments5 min readLW link

Some Quick Fol­low-Up Ex­per­i­ments to “Taken out of con­text: On mea­sur­ing situ­a­tional aware­ness in LLMs”

miles3 Oct 2023 2:22 UTC
31 points
0 comments9 min readLW link

My Mid-Ca­reer Tran­si­tion into Biosecurity

jefftk2 Oct 2023 21:20 UTC
26 points
4 comments2 min readLW link
(www.jefftk.com)