Ex­pert trap: Why is it hap­pen­ing? (Part 2 of 3) – how hind­sight, hi­er­ar­chy, and con­fir­ma­tion bi­ases break con­duc­tivity and ac­cu­racy of knowledge

Paweł Sysiak9 Jun 2023 23:00 UTC
3 points
0 comments7 min readLW link

Ex­pert trap: What is it? (Part 1 of 3) – how hind­sight, hi­er­ar­chy, and con­fir­ma­tion bi­ases break con­duc­tivity and ac­cu­racy of knowledge

Paweł Sysiak9 Jun 2023 23:00 UTC
6 points
2 comments8 min readLW link

[Question] How ac­cu­rate is data about past earth tem­per­a­tures?

tailcalled9 Jun 2023 21:29 UTC
10 points
2 comments1 min readLW link

Proxi-An­tipodes: A Geo­met­ri­cal In­tu­ition For The Difficulty Of Align­ing AI With Mul­ti­tudi­nous Hu­man Values

Matthew_Opitz9 Jun 2023 21:21 UTC
7 points
0 comments5 min readLW link

Why AI may not save the World

Alberto Zannoni9 Jun 2023 17:42 UTC
0 points
0 comments4 min readLW link
(a16z.com)

You can now listen to the “AI Safety Fun­da­men­tals” courses

PeterH9 Jun 2023 16:45 UTC
6 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

Ex­plor­ing Con­cept-Spe­cific Slices in Weight Ma­tri­ces for Net­work Interpretability

DuncanFowler9 Jun 2023 16:39 UTC
1 point
0 comments6 min readLW link

A plea for solu­tion­ism on AI safety

jasoncrawford9 Jun 2023 16:29 UTC
72 points
6 comments6 min readLW link
(rootsofprogress.org)

Michael Shel­len­berger: US Has 12 Or More Alien Space­craft, Say Mili­tary And In­tel­li­gence Contractors

lc9 Jun 2023 16:11 UTC
11 points
31 comments3 min readLW link
(public.substack.com)

Im­prove­ment on MIRI’s Corrigibility

9 Jun 2023 16:10 UTC
54 points
8 comments13 min readLW link

D&D.Sci 5E: Re­turn of the League of Defen­ders Eval­u­a­tion & Ruleset

aphyer9 Jun 2023 15:25 UTC
29 points
8 comments6 min readLW link

In­ternLM—China’s Best (Un­ver­ified)

Lao Mein9 Jun 2023 7:39 UTC
51 points
4 comments1 min readLW link

[Question] Mark for fol­low up?

JNS9 Jun 2023 5:59 UTC
5 points
4 comments2 min readLW link

Bring­ing Lit­tle Kids to Con­tra Dances

jefftk9 Jun 2023 2:20 UTC
22 points
0 comments2 min readLW link
(www.jefftk.com)

[Question] (solved) how do i find oth­ers’ short­form posts?

kuira9 Jun 2023 2:15 UTC
1 point
1 comment1 min readLW link

[Question] AI Rights: In your view, what would be re­quired for an AGI to gain rights and pro­tec­tions from the var­i­ous Govern­ments of the World?

Super AGI9 Jun 2023 1:24 UTC
10 points
26 comments1 min readLW link

A com­par­i­son of causal scrub­bing, causal ab­strac­tions, and re­lated methods

8 Jun 2023 23:40 UTC
72 points
3 comments22 min readLW link

Up­dates and Reflec­tions on Op­ti­mal Ex­er­cise af­ter Nearly a Decade

romeostevensit8 Jun 2023 23:02 UTC
187 points
55 comments2 min readLW link

Take­aways from the Mechanis­tic In­ter­pretabil­ity Challenges

scasper8 Jun 2023 18:56 UTC
93 points
5 comments6 min readLW link

Leave an Emo­tional Line of Retreat

Johannes C. Mayer8 Jun 2023 18:36 UTC
23 points
1 comment1 min readLW link

Cur­rent AI harms are also sci-fi

Christopher King8 Jun 2023 17:49 UTC
26 points
3 comments1 min readLW link

Two Ways To Re­duce Un­hap­piness That Comes From Dis­torted Views of Reality

Anne Hsu8 Jun 2023 17:43 UTC
3 points
0 comments7 min readLW link

Col­lab­o­ra­tion in Science: Hap­pier Peo­ple ↔ Bet­ter Research

nadinespy8 Jun 2023 17:42 UTC
3 points
0 comments32 min readLW link

Biomimetic al­ign­ment: Align­ment be­tween an­i­mal genes and an­i­mal brains as a model for al­ign­ment be­tween hu­mans and AI sys­tems

geoffreymiller8 Jun 2023 16:05 UTC
10 points
0 comments16 min readLW link

A po­ten­tially high im­pact differ­en­tial tech­nolog­i­cal de­vel­op­ment area

Noosphere898 Jun 2023 14:33 UTC
5 points
2 comments2 min readLW link

[Question] Ques­tion for Pre­dic­tion Mar­ket peo­ple: where is the money sup­posed to come from?

Robert_AIZI8 Jun 2023 13:58 UTC
25 points
26 comments1 min readLW link

AI #15: The Prin­ci­ple of Charity

Zvi8 Jun 2023 12:10 UTC
73 points
16 comments44 min readLW link
(thezvi.wordpress.com)

if you’re read­ing this it’s too late (a new the­ory on what is caus­ing the Great Stag­na­tion)

rogersbacon8 Jun 2023 11:49 UTC
−10 points
2 comments13 min readLW link
(www.secretorum.life)

[Linkpost] Scal­ing laws for lan­guage en­cod­ing mod­els in fMRI

Bogdan Ionut Cirstea8 Jun 2023 10:52 UTC
30 points
0 comments1 min readLW link

Trans­for­ma­tive AI is a pro­cess

meijer19738 Jun 2023 8:57 UTC
2 points
0 comments5 min readLW link

Cri­sis of Faith case study: be­yond re­duc­tion­ism?

MalcolmOcean8 Jun 2023 6:11 UTC
6 points
9 comments19 min readLW link

I wrote this be­cause of watermelon

Arti8 Jun 2023 3:55 UTC
2 points
2 comments1 min readLW link

Learn­ing Trans­former Pro­grams [Linkpost]

aogara8 Jun 2023 0:16 UTC
7 points
0 comments1 min readLW link
(arxiv.org)

What will GPT-2030 look like?

jsteinhardt7 Jun 2023 23:40 UTC
182 points
42 comments23 min readLW link
(bounded-regret.ghost.io)

Progress links and tweets, 2023-06-07

jasoncrawford7 Jun 2023 23:26 UTC
11 points
0 comments1 min readLW link
(rootsofprogress.org)

LEAst-squares Con­cept Era­sure (LEACE)

tricky_labyrinth7 Jun 2023 21:51 UTC
68 points
10 comments1 min readLW link
(twitter.com)

Pro­posal: Tune LLMs to Use Cal­ibrated Language

OneManyNone7 Jun 2023 21:05 UTC
9 points
0 comments5 min readLW link

A moral back­lash against AI will prob­a­bly slow down AGI development

geoffreymiller7 Jun 2023 20:39 UTC
49 points
10 comments14 min readLW link

RAMP—RoboNet Ar­tifi­cial Me­dia Protocol

antoniomax7 Jun 2023 19:01 UTC
−1 points
0 comments19 min readLW link
(antoniomax.substack.com)

An Ex­er­cise to Build In­tu­itions on AGI Risk

Lauro Langosco7 Jun 2023 18:35 UTC
52 points
3 comments8 min readLW link

Elon talked with se­nior Chi­nese lead­er­ship about AI X-risk

ChristianKl7 Jun 2023 15:02 UTC
47 points
2 comments1 min readLW link
(www.youtube.com)

Ar­ti­cle Sum­mary: Cur­rent and Near-Term AI as a Po­ten­tial Ex­is­ten­tial Risk Factor

André Ferretti7 Jun 2023 13:51 UTC
28 points
3 comments1 min readLW link
(dl.acm.org)

gamers be­ware: mod­ded Minecraft has new malware

the gears to ascension7 Jun 2023 13:49 UTC
14 points
5 comments1 min readLW link
(github.com)

Launch­ing Light­speed Grants (Ap­ply by July 6th)

habryka7 Jun 2023 2:53 UTC
211 points
41 comments5 min readLW link

Cul­ti­vate an ob­ses­sion with the ob­ject level

Richard_Ngo7 Jun 2023 1:39 UTC
70 points
4 comments3 min readLW link

How to Slow AI Development

PeterMcCluskey7 Jun 2023 0:29 UTC
20 points
0 comments5 min readLW link
(bayesianinvestor.com)

[Question] Killing Re­cur­rent Me­mory Over Self At­ten­tion?

Del Nobolo6 Jun 2023 23:02 UTC
3 points
0 comments1 min readLW link

[Job Ad] SERI MATS is (still) hiring for our sum­mer program

6 Jun 2023 21:07 UTC
12 points
0 comments7 min readLW link

Why I am not a longter­mist (May 2022)

boazbarak6 Jun 2023 20:36 UTC
39 points
18 comments9 min readLW link
(windowsontheory.org)

So­ciety Library seek­ing con­tri­bu­tions for canon­i­cal AI Safety de­bate map

Jarred Filmer6 Jun 2023 18:15 UTC
36 points
0 comments1 min readLW link
(www.societylibrary.org)