Fun math facts about 2023

Adam ScherlisJan 1, 2023, 11:38 PM
9 points
6 comments1 min readLW link

The Thing­ness of Things

TsviBTJan 1, 2023, 10:19 PM
51 points
35 comments10 min readLW link

Thoughts On Ex­pand­ing the AI Safety Com­mu­nity: Benefits and Challenges of Outreach to Non-Tech­ni­cal Professionals

Yashvardhan SharmaJan 1, 2023, 7:21 PM
4 points
4 comments7 min readLW link

[Question] Would it be good or bad for the US mil­i­tary to get in­volved in AI risk?

Grant DemareeJan 1, 2023, 7:02 PM
50 points
12 comments1 min readLW link

Bet­ter New Year’s Goals through Align­ing the Elephant and the Rider

moridinamaelJan 1, 2023, 5:54 PM
20 points
0 comments2 min readLW link
(guildoftherose.org)

A Löbian ar­gu­ment pat­tern for im­plicit rea­son­ing in nat­u­ral lan­guage: Löbian party invitations

Andrew_CritchJan 1, 2023, 5:39 PM
23 points
8 comments7 min readLW link

woke offline, anti-woke online

Yair HalberstadtJan 1, 2023, 8:24 AM
13 points
12 comments1 min readLW link

Sum­mary of 80k’s AI prob­lem profile

JakubKJan 1, 2023, 7:30 AM
7 points
0 comments5 min readLW link
(forum.effectivealtruism.org)

What per­cent of peo­ple work in moral mazes?

RaemonJan 1, 2023, 4:33 AM
21 points
9 comments4 min readLW link

Re­cur­sive Mid­dle Man­ager Hell

RaemonJan 1, 2023, 4:33 AM
224 points
46 comments11 min readLW link1 review

Challenge to the no­tion that any­thing is (maybe) pos­si­ble with AGI

Jan 1, 2023, 3:57 AM
−27 points
4 comments1 min readLW link
(mflb.com)

The Roots of Progress’s 2022 in review

jasoncrawfordJan 1, 2023, 2:54 AM
14 points
2 comments15 min readLW link
(rootsofprogress.org)

In­vest­ing for a World Trans­formed by AI

PeterMcCluskeyJan 1, 2023, 2:47 AM
70 points
24 comments6 min readLW link1 review
(bayesianinvestor.com)

Why Free Will is NOT an illusion

Akira PyinyaJan 1, 2023, 2:29 AM
0 points
16 comments1 min readLW link

Lo­calhost Se­cu­rity Messaging

jefftkJan 1, 2023, 2:20 AM
7 points
3 comments1 min readLW link
(www.jefftk.com)

0 and 1 aren’t probabilities

Alok SinghJan 1, 2023, 12:09 AM
2 points
4 comments2 min readLW link
(en.wikipedia.org)

‘simu­la­tor’ fram­ing and con­fu­sions about LLMs

Beth BarnesDec 31, 2022, 11:38 PM
104 points
11 comments4 min readLW link

Mon­i­tor­ing de­vices I have loved

ElizabethDec 31, 2022, 10:51 PM
62 points
13 comments3 min readLW link1 review

Slack mat­ters more than any outcome

ValentineDec 31, 2022, 8:11 PM
164 points
56 comments19 min readLW link1 review

To Be Par­tic­u­lar About Morality

AGODec 31, 2022, 7:58 PM
6 points
2 comments7 min readLW link

200 COP in MI: In­ter­pret­ing Al­gorith­mic Problems

Neel NandaDec 31, 2022, 7:55 PM
33 points
2 comments10 min readLW link

The Feel­ing of Idea Scarcity

johnswentworthDec 31, 2022, 5:34 PM
249 points
23 comments5 min readLW link1 review

Curse of knowl­edge and Naive re­al­ism: Bias in Eval­u­at­ing AGI X-Risks

Dec 31, 2022, 1:33 PM
−7 points
1 comment1 min readLW link
(www.lesswrong.com)

[Question] What ca­reer ad­vice do you give to soft­ware en­g­ineers?

AntbDec 31, 2022, 12:01 PM
15 points
4 comments1 min readLW link

[Question] Are Mix­ture-of-Ex­perts Trans­form­ers More In­ter­pretable Than Dense Trans­form­ers?

simeon_cDec 31, 2022, 11:34 AM
8 points
5 comments1 min readLW link

[Question] In which cases can ChatGPT be used as an aid for the­sis or sci­en­tific pa­per writ­ing?

Bob GuranDec 31, 2022, 10:50 AM
1 point
1 comment1 min readLW link

Two Is­sues with Play­ing Chicken with the Universe

Chris_LeongDec 31, 2022, 6:47 AM
4 points
4 comments2 min readLW link

Ex­treme risk neu­tral­ity isn’t always wrong

Grant DemareeDec 31, 2022, 4:05 AM
28 points
19 comments4 min readLW link

Ver­bal par­ity: What is it and how to mea­sure it? + an ed­ited ver­sion of “Against John Searle, Gary Mar­cus, the Chi­nese Room thought ex­per­i­ment and its world”

philosophybearDec 31, 2022, 3:46 AM
2 points
0 comments11 min readLW link

Should AI sys­tems have to iden­tify them­selves?

Darren McKeeDec 31, 2022, 2:57 AM
2 points
2 comments1 min readLW link

[Question] What do you imag­ine, when you imag­ine “tak­ing over the world”?

johnswentworthDec 31, 2022, 1:04 AM
22 points
16 comments1 min readLW link

A few thoughts on my self-study for al­ign­ment research

Thomas KehrenbergDec 30, 2022, 10:05 PM
6 points
0 comments2 min readLW link

Christ­mas Microscopy

jefftkDec 30, 2022, 9:10 PM
27 points
0 comments1 min readLW link
(www.jefftk.com)

What “up­side” of AI?

False NameDec 30, 2022, 8:58 PM
0 points
5 comments4 min readLW link

Ev­i­dence on re­cur­sive self-im­prove­ment from cur­rent ML

berenDec 30, 2022, 8:53 PM
31 points
12 comments6 min readLW link

[Question] Is ChatGPT TAI?

Amal Dec 30, 2022, 7:44 PM
14 points
5 comments1 min readLW link

My thoughts on OpenAI’s al­ign­ment plan

Orpheus16Dec 30, 2022, 7:33 PM
55 points
3 comments20 min readLW link

Beyond Re­wards and Values: A Non-du­al­is­tic Ap­proach to Univer­sal Intelligence

Akira PyinyaDec 30, 2022, 7:05 PM
10 points
4 comments14 min readLW link

10 Years of LessWrong

SebastianG Dec 30, 2022, 5:15 PM
73 points
2 comments4 min readLW link

Chat­bots as a Publi­ca­tion Format

derek shillerDec 30, 2022, 2:11 PM
6 points
6 comments4 min readLW link

Hu­man sex­u­al­ity as an in­ter­est­ing case study of alignment

berenDec 30, 2022, 1:37 PM
39 points
26 comments3 min readLW link

The Twit­ter Files: Covid Edition

ZviDec 30, 2022, 1:30 PM
32 points
2 comments10 min readLW link
(thezvi.wordpress.com)

Wor­ldly Po­si­tions archive, briefly with pri­vate drafts

KatjaGraceDec 30, 2022, 12:20 PM
11 points
0 comments1 min readLW link
(worldspiritsockpuppet.com)

Models Don’t “Get Re­ward”

Sam RingerDec 30, 2022, 10:37 AM
316 points
62 comments5 min readLW link1 review

The hy­per­finite timeline

Alok SinghDec 30, 2022, 9:30 AM
3 points
6 comments1 min readLW link
(alok.github.io)

Re­ac­tive de­val­u­a­tion: Bias in Eval­u­at­ing AGI X-Risks

Dec 30, 2022, 9:02 AM
−15 points
9 comments1 min readLW link

Things I carry al­most ev­ery day, as of late De­cem­ber 2022

DanielFilanDec 30, 2022, 7:40 AM
38 points
9 comments5 min readLW link
(danielfilan.com)

More ways to spot abysses

KatjaGraceDec 30, 2022, 6:30 AM
21 points
1 comment1 min readLW link
(worldspiritsockpuppet.com)

Lan­guage mod­els are nearly AGIs but we don’t no­tice it be­cause we keep shift­ing the bar

philosophybearDec 30, 2022, 5:15 AM
105 points
13 comments7 min readLW link

Progress links and tweets, 2022-12-29

jasoncrawfordDec 30, 2022, 4:54 AM
12 points
0 comments1 min readLW link
(rootsofprogress.org)