A bunch of videos in comments

the gears to ascension12 Jun 2023 22:31 UTC
10 points
62 comments1 min readLW link

[Linkpost] The neu­ro­con­nec­tion­ist re­search programme

Bogdan Ionut Cirstea12 Jun 2023 21:58 UTC
5 points
1 comment1 min readLW link

Contin­gency: A Con­cep­tual Tool from Evolu­tion­ary Biol­ogy for Alignment

clem_acs12 Jun 2023 20:54 UTC
51 points
2 comments14 min readLW link
(acsresearch.org)

Book Re­view: Autoheterosexuality

tailcalled12 Jun 2023 20:11 UTC
27 points
9 comments24 min readLW link

Aura as a pro­pri­o­cep­tive glitch

pchvykov12 Jun 2023 19:30 UTC
36 points
4 comments4 min readLW link

Align­ing Math­e­mat­i­cal No­tions of In­finity with Hu­man Intuition

London L.12 Jun 2023 19:19 UTC
1 point
10 comments9 min readLW link
(medium.com)

ARC is hiring the­o­ret­i­cal researchers

12 Jun 2023 18:50 UTC
126 points
12 comments4 min readLW link
(www.alignment.org)

In­tro­duc­tion to Towards Causal Foun­da­tions of Safe AGI

12 Jun 2023 17:55 UTC
67 points
6 comments4 min readLW link

Man­i­fold Pre­dicted the AI Ex­tinc­tion State­ment and CAIS Wanted it Deleted

David Chee12 Jun 2023 15:54 UTC
70 points
14 comments12 min readLW link

Explicitness

TsviBT12 Jun 2023 15:05 UTC
29 points
0 comments15 min readLW link

If you are too stressed, walk away from the front lines

Neil 12 Jun 2023 14:26 UTC
42 points
14 comments5 min readLW link

UK PM: $125M for AI safety

Hauke Hillebrandt12 Jun 2023 12:33 UTC
31 points
11 comments1 min readLW link
(twitter.com)

[Question] Could in­duced and sta­bi­lized hy­po­ma­nia be a de­sir­able men­tal state?

MvB12 Jun 2023 12:13 UTC
8 points
22 comments2 min readLW link

Non-loss of con­trol AGI-re­lated catas­tro­phes are out of con­trol too

12 Jun 2023 12:01 UTC
0 points
3 comments24 min readLW link

Cri­tiques of promi­nent AI safety labs: Conjecture

Omega.12 Jun 2023 1:32 UTC
14 points
32 comments33 min readLW link

why I’m anti-YIMBY

bhauth12 Jun 2023 0:19 UTC
20 points
45 comments2 min readLW link

ACX Brno meetup #2

adekcz11 Jun 2023 13:53 UTC
2 points
0 comments1 min readLW link

[Linkpost] Large Lan­guage Models Con­verge on Brain-Like Word Representations

Bogdan Ionut Cirstea11 Jun 2023 11:20 UTC
36 points
12 comments1 min readLW link

In­fer­ence-Time In­ter­ven­tion: Elic­it­ing Truth­ful An­swers from a Lan­guage Model

likenneth11 Jun 2023 5:38 UTC
195 points
4 comments1 min readLW link
(arxiv.org)

You Are a Com­puter, and No, That’s Not a Metaphor

jakej11 Jun 2023 5:38 UTC
12 points
1 comment22 min readLW link
(sigil.substack.com)

Snake Eyes Paradox

Martin Randall11 Jun 2023 4:10 UTC
22 points
25 comments6 min readLW link

[Question] [Mostly solved] I get dis­tracted while read­ing, but can eas­ily com­pre­hend au­dio text for 8+ hours per day. What are the best AI text-to-speech read­ers? Alter­na­tively, do you have other ideas for what I could do?

kuira11 Jun 2023 3:49 UTC
18 points
7 comments1 min readLW link

The Dic­ta­tor­ship Problem

alyssavance11 Jun 2023 2:45 UTC
32 points
143 comments11 min readLW link

Higher Di­men­sion Carte­sian Ob­jects and Align­ing ‘Tiling Si­mu­la­tors’

lukemarks11 Jun 2023 0:13 UTC
22 points
0 comments5 min readLW link

Us­ing Con­sen­sus Mechanisms as an ap­proach to Alignment

Prometheus10 Jun 2023 23:38 UTC
9 points
2 comments6 min readLW link

Hu­man­i­ties first math prob­lem, The shal­low gene pool.

archeon10 Jun 2023 23:09 UTC
−2 points
0 comments1 min readLW link

I can see how I am Dumb

Johannes C. Mayer10 Jun 2023 19:18 UTC
46 points
11 comments5 min readLW link

Etho­dy­nam­ics of Omelas

dr_s10 Jun 2023 16:24 UTC
78 points
16 comments9 min readLW link

Deal­ing with UFO claims

ChristianKl10 Jun 2023 15:45 UTC
3 points
32 comments1 min readLW link

A The­ory of Un­su­per­vised Trans­la­tion Mo­ti­vated by Un­der­stand­ing An­i­mal Communication

jsd10 Jun 2023 15:44 UTC
19 points
0 comments1 min readLW link
(arxiv.org)

[Question] What are brains?

Valentine10 Jun 2023 14:46 UTC
10 points
22 comments2 min readLW link

EY in the New York Times

Blueberry10 Jun 2023 12:21 UTC
6 points
14 comments1 min readLW link
(www.nytimes.com)

Goal-mis­gen­er­al­iza­tion is ELK-hard

rokosbasilisk10 Jun 2023 9:32 UTC
2 points
0 comments1 min readLW link

[Question] What do benefi­cial TDT trades for hu­man­ity con­cretely look like?

Stephen Fowler10 Jun 2023 6:50 UTC
4 points
0 comments1 min readLW link

cloud seed­ing doesn’t work

bhauth10 Jun 2023 5:14 UTC
7 points
2 comments1 min readLW link

[FICTION] Un­box­ing Ely­sium: An AI’S Escape

Super AGI10 Jun 2023 4:41 UTC
−14 points
4 comments14 min readLW link

[FICTION] Prometheus Ris­ing: The Emer­gence of an AI Consciousness

Super AGI10 Jun 2023 4:41 UTC
−13 points
0 comments9 min readLW link

an Evan­ge­lion di­alogue ex­plain­ing the QACI al­ign­ment plan

Tamsin Leake10 Jun 2023 3:28 UTC
45 points
15 comments43 min readLW link
(carado.moe)

for­mal­iz­ing the QACI al­ign­ment for­mal-goal

10 Jun 2023 3:28 UTC
53 points
6 comments14 min readLW link
(carado.moe)

Ex­pert trap: Why is it hap­pen­ing? (Part 2 of 3) – how hind­sight, hi­er­ar­chy, and con­fir­ma­tion bi­ases break con­duc­tivity and ac­cu­racy of knowledge

Paweł Sysiak9 Jun 2023 23:00 UTC
3 points
0 comments7 min readLW link

Ex­pert trap: What is it? (Part 1 of 3) – how hind­sight, hi­er­ar­chy, and con­fir­ma­tion bi­ases break con­duc­tivity and ac­cu­racy of knowledge

Paweł Sysiak9 Jun 2023 23:00 UTC
6 points
2 comments8 min readLW link

[Question] How ac­cu­rate is data about past earth tem­per­a­tures?

tailcalled9 Jun 2023 21:29 UTC
10 points
2 comments1 min readLW link

Proxi-An­tipodes: A Geo­met­ri­cal In­tu­ition For The Difficulty Of Align­ing AI With Mul­ti­tudi­nous Hu­man Values

Matthew_Opitz9 Jun 2023 21:21 UTC
7 points
0 comments5 min readLW link

Why AI may not save the World

Alberto Zannoni9 Jun 2023 17:42 UTC
0 points
0 comments4 min readLW link
(a16z.com)

You can now listen to the “AI Safety Fun­da­men­tals” courses

PeterH9 Jun 2023 16:45 UTC
6 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

Ex­plor­ing Con­cept-Spe­cific Slices in Weight Ma­tri­ces for Net­work Interpretability

DuncanFowler9 Jun 2023 16:39 UTC
1 point
0 comments6 min readLW link

A plea for solu­tion­ism on AI safety

jasoncrawford9 Jun 2023 16:29 UTC
72 points
6 comments6 min readLW link
(rootsofprogress.org)

Michael Shel­len­berger: US Has 12 Or More Alien Space­craft, Say Mili­tary And In­tel­li­gence Contractors

lc9 Jun 2023 16:11 UTC
11 points
31 comments3 min readLW link
(public.substack.com)

Im­prove­ment on MIRI’s Corrigibility

9 Jun 2023 16:10 UTC
54 points
8 comments13 min readLW link

D&D.Sci 5E: Re­turn of the League of Defen­ders Eval­u­a­tion & Ruleset

aphyer9 Jun 2023 15:25 UTC
29 points
8 comments6 min readLW link