Hu­man­i­tar­ian Phase Tran­si­tion needed be­fore Tech­nolog­i­cal Sin­gu­lar­ity

Dr_WhatApr 7, 2023, 11:17 PM
−9 points
5 comments2 min readLW link

[Question] Thoughts about Hug­ging Face?

kwiat.devApr 7, 2023, 11:17 PM
7 points
0 comments1 min readLW link

[Question] Is it cor­rect to frame al­ign­ment as “pro­gram­ming a good philos­o­phy of mean­ing”?

UtilApr 7, 2023, 11:16 PM
2 points
3 comments1 min readLW link

Select Agent Speci­fi­ca­tions as Nat­u­ral Abstractions

lukemarksApr 7, 2023, 11:16 PM
19 points
3 comments5 min readLW link

n=3 AI Risk Quick Math and Reasoning

lionhearted (Sebastian Marshall)Apr 7, 2023, 8:27 PM
6 points
3 comments4 min readLW link

[Question] What are good al­ter­na­tives to Pre­dic­tion­book for per­sonal pre­dic­tion track­ing? Edited: I origi­nally thought it was down but it was just 500 un­til I though of clear­ing cook­ies.

sortegaApr 7, 2023, 7:18 PM
4 points
4 comments1 min readLW link

En­vi­ron­ments for Mea­sur­ing De­cep­tion, Re­source Ac­qui­si­tion, and Eth­i­cal Violations

Dan HApr 7, 2023, 6:40 PM
51 points
2 comments2 min readLW link
(arxiv.org)

Su­per­in­tel­li­gence Is Not Omniscience

Jeffrey HeningerApr 7, 2023, 4:30 PM
16 points
21 comments7 min readLW link
(aiimpacts.org)

An ‘AGI Emer­gency Eject Cri­te­ria’ con­sen­sus could be re­ally use­ful.

tcelferactApr 7, 2023, 4:21 PM
5 points
0 commentsLW link

Reli­a­bil­ity, Se­cu­rity, and AI risk: Notes from in­fosec text­book chap­ter 1

Orpheus16Apr 7, 2023, 3:47 PM
34 points
1 comment4 min readLW link

Pre-reg­is­ter­ing a study

Robert_AIZIApr 7, 2023, 3:46 PM
10 points
0 comments6 min readLW link
(aizi.substack.com)

Live dis­cus­sion at Eastercon

Douglas_ReayApr 7, 2023, 3:25 PM
5 points
0 comments1 min readLW link

[Question] ChatGTP “Writ­ing ” News Sto­ries for The Guardian?

jmhApr 7, 2023, 12:16 PM
1 point
4 comments1 min readLW link

Sto­ry­tel­ler’s con­ven­tion, 2223 A.D.

plexApr 7, 2023, 11:54 AM
8 points
0 comments2 min readLW link

Stampy’s AI Safety Info—New Distil­la­tions #1 [March 2023]

markovApr 7, 2023, 11:06 AM
42 points
0 comments2 min readLW link
(aisafety.info)

Beren’s “De­con­fus­ing Direct vs Amor­tised Op­ti­mi­sa­tion”

DragonGodApr 7, 2023, 8:57 AM
52 points
10 comments3 min readLW link

Goal al­ign­ment with­out al­ign­ment on episte­mol­ogy, ethics, and sci­ence is futile

Roman LeventovApr 7, 2023, 8:22 AM
20 points
2 comments2 min readLW link

Po­lio Lab Leak Caught with Wastew­a­ter Sampling

CullenApr 7, 2023, 1:06 AM
82 points
3 commentsLW link

Catch­ing the Eye of Sauron

Casey_Apr 7, 2023, 12:40 AM
221 points
68 comments4 min readLW link

[Question] How to par­allelize “in­her­ently” se­rial the­ory work?

Nicholas / Heather KrossApr 7, 2023, 12:08 AM
16 points
6 comments1 min readLW link

If Align­ment is Hard, then so is Self-Improvement

PavleMihaApr 7, 2023, 12:08 AM
21 points
20 comments1 min readLW link

An­thropic is fur­ther ac­cel­er­at­ing the Arms Race?

sapphireApr 6, 2023, 11:29 PM
82 points
22 comments1 min readLW link
(techcrunch.com)

Sugges­tion for safe AI struc­ture (Cu­rated Trans­par­ent De­ci­sions)

Kane GregoryApr 6, 2023, 10:00 PM
5 points
6 comments3 min readLW link

10 rea­sons why lists of 10 rea­sons might be a win­ning strategy

trevorApr 6, 2023, 9:24 PM
110 points
7 comments1 min readLW link

A Defense of Utilitarianism

Pareto OptimalApr 6, 2023, 9:09 PM
−3 points
2 comments5 min readLW link
(paretooptimal.substack.com)

One Does Not Sim­ply Re­place the Hu­mans

JerkyTreatsApr 6, 2023, 8:56 PM
9 points
3 comments4 min readLW link
(www.lesswrong.com)

[Question] Where to be­gin in ML/​AI?

Jake the StudentApr 6, 2023, 8:45 PM
9 points
4 comments1 min readLW link

Mis­gen­er­al­iza­tion as a misnomer

So8resApr 6, 2023, 8:43 PM
129 points
22 comments4 min readLW link

You can use GPT-4 to cre­ate prompt in­jec­tions against GPT-4

WitchBOTApr 6, 2023, 8:39 PM
87 points
8 comments2 min readLW link

AI scares and chang­ing pub­lic beliefs

Seth HerdApr 6, 2023, 6:51 PM
46 points
21 comments6 min readLW link

AISafety.world is a map of the AIS ecosystem

Hamish DoodlesApr 6, 2023, 6:37 PM
80 points
0 comments1 min readLW link

I asked my sen­a­tor to slow AI

OmidApr 6, 2023, 6:18 PM
21 points
5 comments2 min readLW link

Pause AI Devel­op­ment?

PeterMcCluskeyApr 6, 2023, 5:23 PM
11 points
0 comments2 min readLW link
(bayesianinvestor.com)

Use these three heuris­tic im­per­a­tives to solve alignment

GApr 6, 2023, 4:20 PM
−17 points
4 comments1 min readLW link

Eliezer on The Lu­nar So­ciety podcast

Max HApr 6, 2023, 4:18 PM
40 points
5 comments1 min readLW link
(www.dwarkeshpatel.com)

Do we get bet­ter or worse at adapt­ing to change?

jasoncrawfordApr 6, 2023, 2:42 PM
12 points
2 comments3 min readLW link
(rootsofprogress.org)

Is it true that only a chat­bot en­couraged a man to com­mit suicide?

Jeroen De RyckApr 6, 2023, 2:10 PM
6 points
0 comments4 min readLW link
(www.vrt.be)

A Fresh FAQ on GiveWiki and Im­pact Mar­kets Generally

Dawn DrescherApr 6, 2023, 2:02 PM
−1 points
0 commentsLW link
(impactmarkets.substack.com)

AI #6: Agents of Change

ZviApr 6, 2023, 2:00 PM
79 points
13 comments47 min readLW link
(thezvi.wordpress.com)

Stupid Ques­tions—April 2023

ChristianKlApr 6, 2023, 1:07 PM
17 points
46 comments1 min readLW link

(Yet Another) Map for AI Risk Discussion

chronolitusApr 6, 2023, 11:55 AM
1 point
0 comments2 min readLW link

The Com­pu­ta­tional Anatomy of Hu­man Values

berenApr 6, 2023, 10:33 AM
74 points
30 comments30 min readLW link

[Question] Is “Re­cur­sive Self-Im­prove­ment” Rele­vant in the Deep Learn­ing Paradigm?

DragonGodApr 6, 2023, 7:13 AM
32 points
36 comments7 min readLW link

Re­vis­it­ing the Hori­zon Length Hypothesis

Pablo VillalobosApr 6, 2023, 6:39 AM
23 points
4 comments3 min readLW link

Monthly Shorts 3/​23

CelerApr 6, 2023, 6:20 AM
7 points
1 comment4 min readLW link
(keller.substack.com)

Dual-Use­ness is a Ratio

jimrandomhApr 6, 2023, 5:46 AM
35 points
2 comments1 min readLW link

[Question] What’s the deal with Effec­tive Ac­cel­er­a­tionism (e/​acc)?

RomanHaukssonApr 6, 2023, 4:03 AM
23 points
9 comments2 min readLW link

No Sum­mer Har­vest: Why AI Devel­op­ment Won’t Pause

Stephen FowlerApr 6, 2023, 3:53 AM
14 points
17 comments12 min readLW link

Yoshua Ben­gio: “Slow­ing down de­vel­op­ment of AI sys­tems pass­ing the Tur­ing test”

Roman LeventovApr 6, 2023, 3:31 AM
49 points
2 comments5 min readLW link
(yoshuabengio.org)

Unal­igned sta­ble loops emerge at scale

Michael TontchevApr 6, 2023, 2:15 AM
9 points
8 comments4 min readLW link