GTP4 ca­pa­ble of limited re­cur­sive im­prov­ing?

Boris Kashirin2 Apr 2023 21:38 UTC
2 points
3 comments1 min readLW link

[Question] Scared about the fu­ture of AI

eitan weiss2 Apr 2023 20:37 UTC
−1 points
0 comments1 min readLW link

“a di­alogue with my­self con­cern­ing eliezer yud­kowsky” (not au­thor)

the gears to ascension2 Apr 2023 20:12 UTC
13 points
18 comments3 min readLW link

Fine-in­sured boun­ties as AI deterrent

Virtual Instinct2 Apr 2023 19:44 UTC
1 point
0 comments2 min readLW link

[Question] What could EA’s new name be?

trevor2 Apr 2023 19:25 UTC
17 points
20 comments2 min readLW link

Ex­po­sure to Lizard­man is Lethal

[DEACTIVATED] Duncan Sabien2 Apr 2023 18:57 UTC
70 points
96 comments3 min readLW link

Talk­box Bag­pipe Drones

jefftk2 Apr 2023 18:50 UTC
5 points
0 comments1 min readLW link
(www.jefftk.com)

[Question] Is there a LessWrong-ad­ja­cent place to hire free­lancers/​seek free­lance work?

nonzerosum2 Apr 2023 16:39 UTC
5 points
3 comments1 min readLW link

AISC 2023, Progress Re­port for March: Team In­ter­pretable Architectures

2 Apr 2023 16:19 UTC
14 points
0 comments14 min readLW link

Ul­ti­mate ends may be eas­ily hid­able be­hind con­ver­gent subgoals

TsviBT2 Apr 2023 14:51 UTC
57 points
4 comments22 min readLW link

Trans­parency for Gen­er­al­iz­ing Align­ment from Toy Models

Johannes C. Mayer2 Apr 2023 10:47 UTC
13 points
3 comments4 min readLW link

Ask First

intellectronica2 Apr 2023 10:45 UTC
3 points
1 comment1 min readLW link
(intellectronica.net)

[Question] When should a neu­ral net­work-based ap­proach for plant con­trol sys­tems be preferred over a tra­di­tional con­trol method?

Bob Guran2 Apr 2023 10:18 UTC
11 points
0 comments1 min readLW link

Pes­simism about AI Safety

2 Apr 2023 7:43 UTC
4 points
1 comment25 min readLW link

Some lesser-known megapro­ject ideas

Linch2 Apr 2023 1:14 UTC
19 points
4 comments1 min readLW link

Anal­y­sis of GPT-4 com­pe­tence in as­sess­ing com­plex le­gal lan­guage: Ex­am­ple of Bill C-11 of the Cana­dian Par­li­a­ment. - Part 1

M. Y. Zuo2 Apr 2023 0:01 UTC
12 points
2 comments14 min readLW link

Policy dis­cus­sions fol­low strong con­tex­tu­al­iz­ing norms

Richard_Ngo1 Apr 2023 23:51 UTC
230 points
61 comments3 min readLW link

A re­port about LessWrong karma volatility from a differ­ent universe

Ben Pace1 Apr 2023 21:48 UTC
176 points
7 comments1 min readLW link

A Con­fes­sion about the LessWrong Team

Ruby1 Apr 2023 21:47 UTC
87 points
5 comments2 min readLW link

Shut­ting down AI is not enough. We need to de­stroy all tech­nol­ogy.

Matthew Barnett1 Apr 2023 21:03 UTC
150 points
36 comments1 min readLW link

A policy guaran­teed to in­crease AI timelines

Richard Korzekwa 1 Apr 2023 20:50 UTC
46 points
1 comment2 min readLW link
(aiimpacts.org)

Why I Think the Cur­rent Tra­jec­tory of AI Re­search has Low P(doom) - LLMs

GaPa1 Apr 2023 20:35 UTC
2 points
1 comment10 min readLW link

Re­pairing the Effort Asymmetry

[DEACTIVATED] Duncan Sabien1 Apr 2023 20:23 UTC
41 points
11 comments2 min readLW link

Draft: In­fer­ring minimizers

Alex_Altair1 Apr 2023 20:20 UTC
9 points
0 comments1 min readLW link

AI Safety via Luck

Jozdien1 Apr 2023 20:13 UTC
76 points
7 comments11 min readLW link

Ho Chi Minh ACX Meetup

cygnus1 Apr 2023 19:41 UTC
1 point
0 comments1 min readLW link

Quaker Prac­tice for the Aspiring Rationalist

maia1 Apr 2023 19:32 UTC
25 points
4 comments6 min readLW link

The Plan: Put ChatGPT in Charge

Sven Nilsen1 Apr 2023 17:23 UTC
−5 points
3 comments1 min readLW link

AI in­fosec: first strikes, zero-day mar­kets, hard­ware sup­ply chains, adop­tion barriers

Allison Duettmann1 Apr 2023 16:44 UTC
39 points
0 comments9 min readLW link

In­tro­duc­ing Align­men­tSearch: An AI Align­ment-In­formed Con­ver­sional Agent

1 Apr 2023 16:39 UTC
79 points
14 comments4 min readLW link

How se­cu­rity and cryp­tog­ra­phy can aid AI safety [se­quence]

Allison Duettmann1 Apr 2023 16:28 UTC
22 points
0 comments1 min readLW link

AI com­mu­nity build­ing: EliezerKart

Christopher King1 Apr 2023 15:25 UTC
45 points
0 comments2 min readLW link

[Question] Trans­former trained on it’s own con­tent?

Micromegas1 Apr 2023 15:08 UTC
1 point
0 comments1 min readLW link

The frozen neutrality

ProgramCrafter1 Apr 2023 12:58 UTC
3 points
0 comments3 min readLW link

Pro­posal: Butt bumps as a de­fault for phys­i­cal greetings

Adam Zerner1 Apr 2023 12:48 UTC
53 points
23 comments2 min readLW link

[Question] Is this true? paulg: [One spe­cial thing about AI risk is that peo­ple who un­der­stand AI well are more wor­ried than peo­ple who un­der­stand it poorly]

tailcalled1 Apr 2023 11:59 UTC
25 points
5 comments1 min readLW link

Some thought ex­per­i­ments on digi­tal consciousness

rorygreig1 Apr 2023 11:45 UTC
22 points
13 comments6 min readLW link

Sin­gu­lar­i­ties against the Sin­gu­lar­ity: An­nounc­ing Work­shop on Sin­gu­lar Learn­ing The­ory and Alignment

1 Apr 2023 9:58 UTC
87 points
0 comments1 min readLW link
(singularlearningtheory.com)

Cam­paign for AI Safety: Please join me

Nik Samoylov1 Apr 2023 9:32 UTC
18 points
9 comments1 min readLW link

New Align­ment Re­search Agenda: Mas­sive Mul­ti­player Or­ganism Oversight

TsviBT1 Apr 2023 8:02 UTC
17 points
1 comment2 min readLW link

[April Fools’] Defini­tive con­fir­ma­tion of shard theory

TurnTrout1 Apr 2023 7:27 UTC
166 points
7 comments2 min readLW link

[New LW Fea­ture] “De­bates”

1 Apr 2023 7:00 UTC
113 points
34 comments1 min readLW link

Keep Mak­ing AI Safety News

RedFishBlueFish1 Apr 2023 6:27 UTC
1 point
6 comments1 min readLW link

[Question] What Are Your Prefer­ences Re­gard­ing The FLI Let­ter?

JenniferRM1 Apr 2023 4:52 UTC
−4 points
122 comments16 min readLW link

An Aver­age Dialogue

NicholasKross1 Apr 2023 4:01 UTC
4 points
0 comments1 min readLW link

The Sig­nifi­cance of “Align­ment vs Progress: The AI Rap Show­down” in the AI Safety Discourse

Jonathan Grant1 Apr 2023 3:26 UTC
−1 points
0 comments1 min readLW link

My Model of Gen­der Identity

Iris of Rosebloom1 Apr 2023 3:03 UTC
32 points
4 comments12 min readLW link

Hooray for step­ping out of the limelight

So8res1 Apr 2023 2:45 UTC
281 points
24 comments1 min readLW link

Kal­lipo­lis, USA

Daniel Kokotajlo1 Apr 2023 2:06 UTC
13 points
1 comment1 min readLW link
(docs.google.com)

[Question] Is there a man­i­fold plot of all peo­ple who had a say in AI al­ign­ment?

skulk-and-quarrel31 Mar 2023 21:50 UTC
8 points
0 comments1 min readLW link