[Linkpost] The Story Of VaccinateCA

hath9 Dec 2022 23:54 UTC
103 points
4 comments10 min readLW link
(www.worksinprogress.co)

Pro­saic mis­al­ign­ment from the Solomonoff Predictor

Cleo Nardo9 Dec 2022 17:53 UTC
40 points
2 comments5 min readLW link

Take 8: Queer the in­ner/​outer al­ign­ment di­chotomy.

Charlie Steiner9 Dec 2022 17:46 UTC
28 points
2 comments2 min readLW link

[Question] Does a LLM have a util­ity func­tion?

Dagon9 Dec 2022 17:19 UTC
17 points
11 comments1 min readLW link

Monthly Roundup #1

Zvi9 Dec 2022 17:10 UTC
31 points
6 comments21 min readLW link
(thezvi.wordpress.com)

Work­ing to­wards AI al­ign­ment is better

Johannes C. Mayer9 Dec 2022 15:39 UTC
8 points
2 comments2 min readLW link

You can still fetch the coffee to­day if you’re dead tomorrow

davidad9 Dec 2022 14:06 UTC
84 points
19 comments5 min readLW link

ChatGPT’s Misal­ign­ment Isn’t What You Think

stavros9 Dec 2022 11:11 UTC
3 points
12 comments1 min readLW link

ML Safety at NeurIPS & Paradig­matic AI Safety? MLAISU W49

9 Dec 2022 10:38 UTC
19 points
0 comments4 min readLW link
(newsletter.apartresearch.com)

[Question] What are your thoughts on the fu­ture of AI-as­sisted soft­ware de­vel­op­ment?

RomanHauksson9 Dec 2022 10:04 UTC
4 points
4 comments1 min readLW link

Fear miti­gated the nu­clear threat, can it do the same to AGI risks?

Igor Ivanov9 Dec 2022 10:04 UTC
6 points
8 comments5 min readLW link

Set­ting the Zero Point

[DEACTIVATED] Duncan Sabien9 Dec 2022 6:06 UTC
90 points
43 comments20 min readLW link1 review

Sys­tems of Survival

Vaniver9 Dec 2022 5:13 UTC
63 points
5 comments5 min readLW link

[Question] Do You Have an In­ter­nal Monologue?

belkarx9 Dec 2022 3:04 UTC
23 points
7 comments1 min readLW link

[Question] How is the “sharp left turn defined”?

Chris_Leong9 Dec 2022 0:04 UTC
14 points
4 comments1 min readLW link

Linkpost for a gen­er­al­ist al­gorith­mic learner: ca­pa­ble of car­ry­ing out sort­ing, short­est paths, string match­ing, con­vex hull find­ing in one network

lovetheusers9 Dec 2022 0:02 UTC
7 points
1 comment1 min readLW link
(twitter.com)

[Question] Where’s the eco­nomic in­cen­tive for wok­ism com­ing from?

Valentine8 Dec 2022 23:28 UTC
12 points
105 comments1 min readLW link

I Believe we are in a Hard­ware Overhang

nem8 Dec 2022 23:18 UTC
8 points
0 comments1 min readLW link

Of pump­kins, the Fal­con Heavy, and Grou­cho Marx: High-Level dis­course struc­ture in ChatGPT

Bill Benzon8 Dec 2022 22:25 UTC
2 points
0 comments8 min readLW link

How Many Lives Does X-Risk Work Save From Nonex­is­tence On Aver­age?

Jordan Arel8 Dec 2022 21:57 UTC
4 points
5 comments14 min readLW link

AI Safety Seems Hard to Measure

HoldenKarnofsky8 Dec 2022 19:50 UTC
71 points
6 comments14 min readLW link
(www.cold-takes.com)

Play­ing shell games with definitions

weverka8 Dec 2022 19:35 UTC
9 points
3 comments1 min readLW link

Notes on OpenAI’s al­ign­ment plan

Alex Flint8 Dec 2022 19:13 UTC
40 points
5 comments7 min readLW link

Rele­vant to nat­u­ral ab­strac­tions: Eu­clidean Sym­me­try Equiv­ar­i­ant Ma­chine Learn­ing—Overview, Ap­pli­ca­tions, and Open Questions

the gears to ascension8 Dec 2022 18:01 UTC
8 points
0 comments1 min readLW link
(youtu.be)

I’ve started pub­lish­ing the novel I wrote to pro­mote EA

Timothy Underwood8 Dec 2022 17:30 UTC
10 points
3 comments1 min readLW link

Neu­ral net­works bi­ased to­wards ge­o­met­ri­cally sim­ple func­tions?

DavidHolmes8 Dec 2022 16:16 UTC
16 points
2 comments3 min readLW link

If Went­worth is right about nat­u­ral ab­strac­tions, it would be bad for alignment

Wuschel Schulz8 Dec 2022 15:19 UTC
28 points
5 comments4 min readLW link

Covid 12/​8/​22: Another Win­ter Wave

Zvi8 Dec 2022 14:40 UTC
23 points
8 comments11 min readLW link
(thezvi.wordpress.com)

Why I’m Scep­ti­cal of Foom

DragonGod8 Dec 2022 10:01 UTC
20 points
36 comments3 min readLW link

Take 7: You should talk about “the hu­man’s util­ity func­tion” less.

Charlie Steiner8 Dec 2022 8:14 UTC
50 points
22 comments2 min readLW link

Ma­chine Learn­ing Consent

jefftk8 Dec 2022 3:50 UTC
38 points
14 comments3 min readLW link
(www.jefftk.com)

Riffing on the agent type

Quinn8 Dec 2022 0:19 UTC
21 points
3 comments4 min readLW link

[Question] Look­ing for ideas of pub­lic as­sets (stocks, funds, ETFs) that I can in­vest in to have a chance at prof­it­ing from the mass adop­tion and com­mer­cial­iza­tion of AI technology

Annapurna7 Dec 2022 22:35 UTC
15 points
9 comments1 min readLW link

A Fal­li­bil­ist Wordview

Toni MUENDEL7 Dec 2022 20:59 UTC
−13 points
2 comments13 min readLW link

Thoughts on AGI or­ga­ni­za­tions and ca­pa­bil­ities work

7 Dec 2022 19:46 UTC
102 points
17 comments5 min readLW link

How to Think About Cli­mate Models and How to Im­prove Them

clans7 Dec 2022 19:37 UTC
7 points
0 comments2 min readLW link
(locationtbd.home.blog)

The nov­elty quotient

River Lewis7 Dec 2022 17:16 UTC
4 points
7 comments2 min readLW link
(heytraveler.substack.com)

ChatGPT: “An er­ror oc­curred. If this is­sue per­sists...”

Bill Benzon7 Dec 2022 15:41 UTC
5 points
11 comments3 min readLW link

Take 6: CAIS is ac­tu­ally Or­wellian.

Charlie Steiner7 Dec 2022 13:50 UTC
14 points
8 comments2 min readLW link

Peter Thiel on Tech­nolog­i­cal Stag­na­tion and Out of Touch Rationalists

Matt Goldenberg7 Dec 2022 13:15 UTC
9 points
26 comments1 min readLW link
(youtu.be)

[Link] Wave­func­tions: from Lin­ear Alge­bra to Spinors

sen7 Dec 2022 12:44 UTC
11 points
12 comments1 min readLW link
(paperclip.substack.com)

Why I like Zulip in­stead of Slack or Discord

Alok Singh7 Dec 2022 9:28 UTC
31 points
10 comments1 min readLW link

Bioweapons, and ChatGPT (an­other vuln­er­a­bil­ity story)

joshuatanderson7 Dec 2022 7:27 UTC
−5 points
0 comments2 min readLW link

Where to be an AI Safety Pro­fes­sor

scasper7 Dec 2022 7:09 UTC
30 points
12 comments2 min readLW link

[Question] Are there any tools to con­vert LW se­quences to PDF or any other file for­mat?

quetzal_rainbow7 Dec 2022 5:28 UTC
2 points
2 comments1 min readLW link

Man­i­fold Mar­kets com­mu­nity meetup

Sinclair Chen7 Dec 2022 3:25 UTC
4 points
0 comments1 min readLW link

“At­ten­tion Pas­sen­gers”: not for Signs

jefftk7 Dec 2022 2:00 UTC
27 points
10 comments1 min readLW link
(www.jefftk.com)

[ASoT] Prob­a­bil­ity In­fects Con­cepts it Touches

Ulisse Mini7 Dec 2022 1:48 UTC
10 points
4 comments1 min readLW link

Sim­ple Way to Prevent Power-Seek­ing AI

research_prime_space7 Dec 2022 0:26 UTC
12 points
1 comment1 min readLW link

In defense of prob­a­bly wrong mechanis­tic models

evhub6 Dec 2022 23:24 UTC
53 points
10 comments2 min readLW link