GTP4 ca­pa­ble of limited re­cur­sive im­prov­ing?

Boris KashirinApr 2, 2023, 9:38 PM
2 points
3 comments1 min readLW link

[Question] Scared about the fu­ture of AI

eitan weissApr 2, 2023, 8:37 PM
−1 points
0 comments1 min readLW link

“a di­alogue with my­self con­cern­ing eliezer yud­kowsky” (not au­thor)

the gears to ascensionApr 2, 2023, 8:12 PM
13 points
18 comments3 min readLW link

Fine-in­sured boun­ties as AI deterrent

Virtual InstinctApr 2, 2023, 7:44 PM
1 point
0 comments2 min readLW link

[Question] What could EA’s new name be?

trevorApr 2, 2023, 7:25 PM
17 points
20 comments2 min readLW link

Ex­po­sure to Lizard­man is Lethal

Duncan Sabien (Inactive)Apr 2, 2023, 6:57 PM
90 points
96 comments3 min readLW link

Talk­box Bag­pipe Drones

jefftkApr 2, 2023, 6:50 PM
5 points
0 comments1 min readLW link
(www.jefftk.com)

[Question] Is there a LessWrong-ad­ja­cent place to hire free­lancers/​seek free­lance work?

nonzerosumApr 2, 2023, 4:39 PM
5 points
3 comments1 min readLW link

AISC 2023, Progress Re­port for March: Team In­ter­pretable Architectures

Apr 2, 2023, 4:19 PM
14 points
0 comments14 min readLW link

Ul­ti­mate ends may be eas­ily hid­able be­hind con­ver­gent subgoals

TsviBTApr 2, 2023, 2:51 PM
59 points
4 comments22 min readLW link

Trans­parency for Gen­er­al­iz­ing Align­ment from Toy Models

Johannes C. MayerApr 2, 2023, 10:47 AM
13 points
3 comments4 min readLW link

Ask First

intellectronicaApr 2, 2023, 10:45 AM
3 points
1 comment1 min readLW link
(intellectronica.net)

[Question] When should a neu­ral net­work-based ap­proach for plant con­trol sys­tems be preferred over a tra­di­tional con­trol method?

Bob GuranApr 2, 2023, 10:18 AM
11 points
0 comments1 min readLW link

Pes­simism about AI Safety

Apr 2, 2023, 7:43 AM
4 points
1 comment25 min readLW link

Some lesser-known megapro­ject ideas

LinchApr 2, 2023, 1:14 AM
19 points
4 commentsLW link

Anal­y­sis of GPT-4 com­pe­tence in as­sess­ing com­plex le­gal lan­guage: Ex­am­ple of Bill C-11 of the Cana­dian Par­li­a­ment. - Part 1

M. Y. ZuoApr 2, 2023, 12:01 AM
12 points
2 comments14 min readLW link

Policy dis­cus­sions fol­low strong con­tex­tu­al­iz­ing norms

Richard_NgoApr 1, 2023, 11:51 PM
231 points
61 comments3 min readLW link

A re­port about LessWrong karma volatility from a differ­ent universe

Ben PaceApr 1, 2023, 9:48 PM
181 points
7 comments1 min readLW link

A Con­fes­sion about the LessWrong Team

RubyApr 1, 2023, 9:47 PM
87 points
5 comments2 min readLW link

Shut­ting down AI is not enough. We need to de­stroy all tech­nol­ogy.

Matthew BarnettApr 1, 2023, 9:03 PM
152 points
36 comments1 min readLW link

A policy guaran­teed to in­crease AI timelines

Richard Korzekwa Apr 1, 2023, 8:50 PM
46 points
1 comment2 min readLW link
(aiimpacts.org)

Why I Think the Cur­rent Tra­jec­tory of AI Re­search has Low P(doom) - LLMs

GaPaApr 1, 2023, 8:35 PM
2 points
1 comment10 min readLW link

Re­pairing the Effort Asymmetry

Duncan Sabien (Inactive)Apr 1, 2023, 8:23 PM
40 points
11 comments2 min readLW link

Draft: In­fer­ring minimizers

Alex_AltairApr 1, 2023, 8:20 PM
9 points
0 comments1 min readLW link

AI Safety via Luck

JozdienApr 1, 2023, 8:13 PM
82 points
7 comments11 min readLW link

Ho Chi Minh ACX Meetup

cygnusApr 1, 2023, 7:41 PM
1 point
0 comments1 min readLW link

Quaker Prac­tice for the Aspiring Rationalist

maiaApr 1, 2023, 7:32 PM
25 points
4 comments6 min readLW link

The Plan: Put ChatGPT in Charge

Sven NilsenApr 1, 2023, 5:23 PM
−5 points
3 comments1 min readLW link

AI in­fosec: first strikes, zero-day mar­kets, hard­ware sup­ply chains, adop­tion barriers

Allison DuettmannApr 1, 2023, 4:44 PM
41 points
0 comments9 min readLW link

In­tro­duc­ing Align­men­tSearch: An AI Align­ment-In­formed Con­ver­sional Agent

Apr 1, 2023, 4:39 PM
79 points
14 comments4 min readLW link

How se­cu­rity and cryp­tog­ra­phy can aid AI safety [se­quence]

Allison DuettmannApr 1, 2023, 4:28 PM
24 points
0 comments1 min readLW link

AI com­mu­nity build­ing: EliezerKart

Christopher KingApr 1, 2023, 3:25 PM
45 points
0 comments2 min readLW link

[Question] Trans­former trained on it’s own con­tent?

MicromegasApr 1, 2023, 3:08 PM
1 point
0 comments1 min readLW link

The frozen neutrality

ProgramCrafterApr 1, 2023, 12:58 PM
3 points
0 comments3 min readLW link

Pro­posal: Butt bumps as a de­fault for phys­i­cal greetings

Adam ZernerApr 1, 2023, 12:48 PM
53 points
23 comments2 min readLW link

[Question] Is this true? paulg: [One spe­cial thing about AI risk is that peo­ple who un­der­stand AI well are more wor­ried than peo­ple who un­der­stand it poorly]

tailcalledApr 1, 2023, 11:59 AM
25 points
5 comments1 min readLW link

Some thought ex­per­i­ments on digi­tal consciousness

rorygreigApr 1, 2023, 11:45 AM
22 points
13 comments6 min readLW link

Sin­gu­lar­i­ties against the Sin­gu­lar­ity: An­nounc­ing Work­shop on Sin­gu­lar Learn­ing The­ory and Alignment

Apr 1, 2023, 9:58 AM
87 points
0 comments1 min readLW link
(singularlearningtheory.com)

Cam­paign for AI Safety: Please join me

Nik SamoylovApr 1, 2023, 9:32 AM
18 points
9 comments1 min readLW link

New Align­ment Re­search Agenda: Mas­sive Mul­ti­player Or­ganism Oversight

TsviBTApr 1, 2023, 8:02 AM
17 points
1 comment2 min readLW link

[April Fools’] Defini­tive con­fir­ma­tion of shard theory

TurnTroutApr 1, 2023, 7:27 AM
170 points
8 comments2 min readLW link

[New LW Fea­ture] “De­bates”

Apr 1, 2023, 7:00 AM
121 points
35 comments1 min readLW link

Keep Mak­ing AI Safety News

RedFishBlueFishApr 1, 2023, 6:27 AM
1 point
6 comments1 min readLW link

[Question] What Are Your Prefer­ences Re­gard­ing The FLI Let­ter?

JenniferRMApr 1, 2023, 4:52 AM
−4 points
122 comments16 min readLW link

An Aver­age Dialogue

Nicholas / Heather KrossApr 1, 2023, 4:01 AM
4 points
0 commentsLW link

The Sig­nifi­cance of “Align­ment vs Progress: The AI Rap Show­down” in the AI Safety Discourse

Jonathan GrantApr 1, 2023, 3:26 AM
−1 points
0 comments1 min readLW link

My Model of Gen­der Identity

Iris of RosebloomApr 1, 2023, 3:03 AM
42 points
4 comments12 min readLW link

Hooray for step­ping out of the limelight

So8resApr 1, 2023, 2:45 AM
284 points
26 comments1 min readLW link

Kal­lipo­lis, USA

Daniel KokotajloApr 1, 2023, 2:06 AM
13 points
1 comment1 min readLW link
(docs.google.com)

[Question] Is there a man­i­fold plot of all peo­ple who had a say in AI al­ign­ment?

skulk-and-quarrelMar 31, 2023, 9:50 PM
8 points
0 comments1 min readLW link