Notes from the Qatar Cen­ter for Global Bank­ing and Fi­nance 3rd An­nual Conference

PixelatedPenguinJul 7, 2023, 11:48 PM
2 points
0 comments1 min readLW link

In­tro­duc­ing bayescalc.io

Adele LopezJul 7, 2023, 4:11 PM
115 points
29 comments1 min readLW link
(bayescalc.io)

Meetup Tip: Ask At­ten­dees To Ex­plain It

ScrewtapeJul 7, 2023, 4:08 PM
10 points
0 comments4 min readLW link

In­ter­pret­ing Mo­du­lar Ad­di­tion in MLPs

Bart BussmannJul 7, 2023, 9:22 AM
20 points
0 comments6 min readLW link

In­ter­nal in­de­pen­dent re­view for lan­guage model agent alignment

Seth HerdJul 7, 2023, 6:54 AM
55 points
30 comments11 min readLW link

[Question] Can LessWrong provide me with some­thing I find ob­vi­ously highly use­ful to my own prac­ti­cal life?

agrippaJul 7, 2023, 3:08 AM
32 points
4 comments1 min readLW link

ask me about technology

bhauthJul 7, 2023, 2:03 AM
23 points
42 comments1 min readLW link

Ap­par­ently, of the 195 Million the DoD al­lo­cated in Univer­sity Re­search Fund­ing Awards in 2022, more than half of them con­cerned AI or com­pute hard­ware research

mako yassJul 7, 2023, 1:20 AM
41 points
5 comments2 min readLW link
(www.defense.gov)

What are the best non-LW places to read on al­ign­ment progress?

RaemonJul 7, 2023, 12:57 AM
50 points
14 comments1 min readLW link

Two paths to win the AGI transition

Nathan Helm-BurgerJul 6, 2023, 9:59 PM
11 points
8 comments4 min readLW link

Em­piri­cal Ev­i­dence Against “The Longest Train­ing Run”

NickGabsJul 6, 2023, 6:32 PM
31 points
0 comments14 min readLW link

Progress Stud­ies Fel­low­ship look­ing for members

jay ramJul 6, 2023, 5:41 PM
3 points
0 comments1 min readLW link

BOUNTY AVAILABLE: AI ethi­cists, what are your ob­ject-level ar­gu­ments against AI notkil­lev­ery­oneism?

Peter BerggrenJul 6, 2023, 5:32 PM
18 points
6 comments2 min readLW link

Lay­er­ing and Tech­ni­cal Debt in the Global Wayfind­ing Model

herschelJul 6, 2023, 5:30 PM
14 points
0 comments3 min readLW link

Lo­cal­iz­ing goal mis­gen­er­al­iza­tion in a maze-solv­ing policy network

Jan BetleyJul 6, 2023, 4:21 PM
37 points
2 comments7 min readLW link

Jesse Hoogland on Devel­op­men­tal In­ter­pretabil­ity and Sin­gu­lar Learn­ing Theory

Michaël TrazziJul 6, 2023, 3:46 PM
42 points
2 comments4 min readLW link
(theinsideview.ai)

Progress links and tweets, 2023-07-06: Ter­raformer Mark One, Is­raeli wa­ter man­age­ment, & more

jasoncrawfordJul 6, 2023, 3:35 PM
18 points
4 comments2 min readLW link
(rootsofprogress.org)

Towards Non-Panop­ti­con AI Alignment

Logan ZoellnerJul 6, 2023, 3:29 PM
7 points
0 comments3 min readLW link

A Defense of Work on Math­e­mat­i­cal AI Safety

DavidmanheimJul 6, 2023, 2:15 PM
28 points
13 comments3 min readLW link
(forum.effectivealtruism.org)

Un­der­stand­ing the two most com­mon men­tal health prob­lems in the world

spencergJul 6, 2023, 2:06 PM
19 points
0 commentsLW link

An­nounc­ing the EA Archive

Aaron BergmanJul 6, 2023, 1:49 PM
13 points
2 commentsLW link

Agency begets agency

Richard_NgoJul 6, 2023, 1:08 PM
60 points
1 comment4 min readLW link

AI #19: Hofs­tadter, Sutskever, Leike

ZviJul 6, 2023, 12:50 PM
60 points
16 comments40 min readLW link
(thezvi.wordpress.com)

Do you feel that AGI Align­ment could be achieved in a Type 0 civ­i­liza­tion?

Super AGIJul 6, 2023, 4:52 AM
−2 points
1 comment1 min readLW link

Open Thread—July 2023

RubyJul 6, 2023, 4:50 AM
11 points
35 comments1 min readLW link

AI Intermediation

jefftkJul 6, 2023, 1:50 AM
12 points
0 comments1 min readLW link
(www.jefftk.com)

An­nounc­ing Man­i­fund Regrants

Austin ChenJul 5, 2023, 7:42 PM
74 points
8 commentsLW link

In­fra-Bayesian Logic

Jul 5, 2023, 7:16 PM
15 points
2 comments1 min readLW link

[Linkpost] In­tro­duc­ing Superalignment

berenJul 5, 2023, 6:23 PM
175 points
69 comments1 min readLW link
(openai.com)

If you wish to make an ap­ple pie, you must first be­come dic­ta­tor of the universe

jasoncrawfordJul 5, 2023, 6:14 PM
27 points
9 comments13 min readLW link
(rootsofprogress.org)

An AGI kill switch with defined se­cu­rity properties

PeterpiperJul 5, 2023, 5:40 PM
−5 points
6 comments1 min readLW link

The risk-re­ward trade­off of in­ter­pretabil­ity research

Jul 5, 2023, 5:05 PM
15 points
1 comment6 min readLW link

(ten­ta­tively) Found 600+ Monose­man­tic Fea­tures in a Small LM Us­ing Sparse Autoencoders

Logan RiggsJul 5, 2023, 4:49 PM
60 points
1 comment7 min readLW link

[Question] What did AI Safety’s spe­cific fund­ing of AGI R&D labs lead to?

RemmeltJul 5, 2023, 3:51 PM
3 points
0 commentsLW link

AISN #13: An in­ter­dis­ci­plinary per­spec­tive on AI proxy failures, new com­peti­tors to ChatGPT, and prompt­ing lan­guage mod­els to misbehave

Dan HJul 5, 2023, 3:33 PM
13 points
0 commentsLW link

Ex­plor­ing Func­tional De­ci­sion The­ory (FDT) and a mod­ified ver­sion (ModFDT)

MiguelDevJul 5, 2023, 2:06 PM
11 points
11 comments15 min readLW link

Op­ti­mized for Some­thing other than Win­ning or: How Cricket Re­sists Moloch and Good­hart’s Law

A.H.Jul 5, 2023, 12:33 PM
53 points
26 comments4 min readLW link

Puffer-pope re­al­ity check

Neil Jul 5, 2023, 9:27 AM
20 points
2 comments1 min readLW link

Fi­nal Light­speed Grants cowork­ing/​office hours be­fore the ap­pli­ca­tion deadline

habrykaJul 5, 2023, 6:03 AM
13 points
2 comments1 min readLW link

MXR Talk­box Cap?

jefftkJul 5, 2023, 1:50 AM
9 points
0 comments1 min readLW link
(www.jefftk.com)

“Reifi­ca­tion”

herschelJul 5, 2023, 12:53 AM
11 points
4 comments2 min readLW link

Dom­i­nant As­surance Con­tract Ex­per­i­ment #2: Berkeley House Dinners

Arjun PanicksseryJul 5, 2023, 12:13 AM
51 points
8 comments1 min readLW link
(arjunpanickssery.substack.com)

Three camps in AI x-risk dis­cus­sions: My per­sonal very over­sim­plified overview

Aryeh EnglanderJul 4, 2023, 8:42 PM
21 points
0 commentsLW link

Six (and a half) in­tu­itions for SVD

CallumMcDougall4 Jul 2023 19:23 UTC
71 points
1 comment1 min readLW link

An­i­mal Weapons: Les­sons for Hu­mans in the Age of X-Risk

Damin Curtis4 Jul 2023 18:14 UTC
4 points
0 comments10 min readLW link

Apoca­lypse Prep­ping—Con­cise SHTF guide to pre­pare for AGI doomsday

prepper4 Jul 2023 17:41 UTC
−7 points
9 comments1 min readLW link
(prepper.i2phides.me)

Ways I Ex­pect AI Reg­u­la­tion To In­crease Ex­tinc­tion Risk

1a3orn4 Jul 2023 17:32 UTC
226 points
32 comments7 min readLW link

AI labs’ state­ments on governance

Zach Stein-Perlman4 Jul 2023 16:30 UTC
30 points
0 comments36 min readLW link

AIs teams will prob­a­bly be more su­per­in­tel­li­gent than in­di­vi­d­ual AIs

Robert_AIZI4 Jul 2023 14:06 UTC
3 points
1 comment2 min readLW link
(aizi.substack.com)

What I Think About When I Think About History

Jacob G-W4 Jul 2023 14:02 UTC
3 points
4 comments3 min readLW link
(g-w1.github.io)