[Question] Would it be use­ful to col­lect the con­texts, where var­i­ous LLMs think the same?

Martin Vlach24 Aug 2023 22:01 UTC
6 points
1 comment1 min readLW link

[Question] Help Needed: Craft­ing a Bet­ter CFAR Fol­low-Up Survey

kotrfa24 Aug 2023 17:26 UTC
9 points
2 comments1 min readLW link

AI #26: Fine Tun­ing Time

Zvi24 Aug 2023 15:30 UTC
49 points
6 comments33 min readLW link
(thezvi.wordpress.com)

Is this the be­gin­ning of the end for LLMS [as the royal road to AGI, what­ever that is]?

Bill Benzon24 Aug 2023 14:50 UTC
3 points
15 comments3 min readLW link

AI Safety Bounties

PatrickL24 Aug 2023 14:29 UTC
11 points
0 comments7 min readLW link
(rethinkpriorities.org)

AI Reg­u­la­tion May Be More Im­por­tant Than AI Align­ment For Ex­is­ten­tial Safety

otto.barten24 Aug 2023 11:41 UTC
65 points
39 comments5 min readLW link

AI Prob­a­bil­ity Trees—Katja Grace

Nathan Young24 Aug 2023 9:45 UTC
8 points
3 comments7 min readLW link

[Question] What wiki-edit­ing fea­tures would make you use the LessWrong wiki more?

Nathan Young24 Aug 2023 9:22 UTC
21 points
27 comments1 min readLW link

On Lucidity

Leber24 Aug 2023 8:45 UTC
0 points
5 comments1 min readLW link
(leber.substack.com)

The God of Hu­man­ity, and the God of the Robot Utilitarians

Raemon24 Aug 2023 8:27 UTC
76 points
12 comments2 min readLW link

I mea­sure Google’s Mu­sicLM over 3 months as it ap­pears to go from jaw-drop­ping to em­bar­rass­ingly re­peat­ing itself

AttentionResearcher24 Aug 2023 4:20 UTC
19 points
4 comments4 min readLW link

En­hanc­ing Cor­rigi­bil­ity in AI Sys­tems through Ro­bust Feed­back Loops

Justausername24 Aug 2023 3:53 UTC
1 point
0 comments6 min readLW link

The lost millennium

Ege Erdil24 Aug 2023 3:48 UTC
51 points
14 comments3 min readLW link

Re­greas­ing a KitchenAid Mixer

jefftk24 Aug 2023 2:30 UTC
15 points
0 comments1 min readLW link
(www.jefftk.com)

Assess­ment of in­tel­li­gence agency func­tion­al­ity is difficult yet important

trevor24 Aug 2023 1:42 UTC
47 points
5 comments9 min readLW link

China’s po­si­tion on au­tonomous weapons

bhauth23 Aug 2023 22:20 UTC
17 points
2 comments1 min readLW link
(academic.oup.com)

Diet Ex­per­i­ment Pr­ereg­is­tra­tion: Long-term wa­ter fast­ing + seed oil re­moval

lc23 Aug 2023 22:08 UTC
56 points
17 comments1 min readLW link

The Low-Hang­ing Fruit Prior and sloped valleys in the loss landscape

23 Aug 2023 21:12 UTC
79 points
1 comment13 min readLW link

Govern­ing, Fast and Slow

Carson23 Aug 2023 20:01 UTC
3 points
0 comments3 min readLW link

A prob­lem with the most re­cently pub­lished ver­sion of CEV

ThomasCederborg23 Aug 2023 18:05 UTC
9 points
5 comments8 min readLW link

[Question] Which paths to pow­er­ful AI should be boosted?

Zach Stein-Perlman23 Aug 2023 16:00 UTC
1 point
0 comments1 min readLW link

A The­ory of Laughter

Steven Byrnes23 Aug 2023 15:05 UTC
101 points
13 comments22 min readLW link

Why Is No One Try­ing To Align Profit In­cen­tives With Align­ment Re­search?

Prometheus23 Aug 2023 13:16 UTC
51 points
11 comments4 min readLW link

Ex­plor­ing the Re­spon­si­ble Path to AI Re­search in the Philippines

MiguelDev23 Aug 2023 8:44 UTC
6 points
0 comments6 min readLW link

[Question] Do agents with (mu­tu­ally known) iden­ti­cal util­ity func­tions but ir­rec­on­cilable knowl­edge some­times fight?

mako yass23 Aug 2023 8:13 UTC
14 points
13 comments1 min readLW link

South Bay ACX/​SSC Fall Mee­tups Everywhere

allisona23 Aug 2023 3:00 UTC
3 points
0 comments1 min readLW link

Separate the truth from your wishes

Jacob G-W23 Aug 2023 0:52 UTC
6 points
3 comments1 min readLW link
(jacobgw.com)

Im­pli­ca­tions of ev­i­den­tial co­op­er­a­tion in large worlds

Lukas Finnveden23 Aug 2023 0:43 UTC
39 points
4 comments17 min readLW link
(lukasfinnveden.substack.com)

South Bay Ca­sual Group Walk

allisona22 Aug 2023 22:43 UTC
7 points
2 comments1 min readLW link

Walk while you talk: don’t balk at “no chalk”

dkl922 Aug 2023 21:27 UTC
41 points
9 comments2 min readLW link
(dkl9.net)

State of Gen­er­ally Available Self-Driving

jefftk22 Aug 2023 18:50 UTC
66 points
6 comments2 min readLW link
(www.jefftk.com)

Seth Ex­plains Consciousness

Jacob Falkovich22 Aug 2023 18:06 UTC
38 points
125 comments14 min readLW link
(putanumonit.com)

ChatGPT challenges the case for hu­man irrationality

Kevin Dorst22 Aug 2023 12:46 UTC
4 points
10 comments7 min readLW link
(kevindorst.substack.com)

[Question] Does one have rea­son to be­lieve the simu­la­tion hy­poth­e­sis is prob­a­bly true?

kuira22 Aug 2023 8:34 UTC
1 point
20 comments1 min readLW link

The Joan of Arc Challenge For Ob­jec­tive List The­ory

omnizoid22 Aug 2023 8:01 UTC
−2 points
4 comments10 min readLW link

The Lop­sided Lives Ar­gu­ment For He­donism About Well-being

omnizoid22 Aug 2023 7:59 UTC
−2 points
8 comments22 min readLW link

Causal­ity and a Cost Se­man­tics for Neu­ral Networks

scottviteri21 Aug 2023 21:02 UTC
22 points
1 comment9 min readLW link

Ideas for im­prov­ing epistemics in AI safety outreach

mic21 Aug 2023 19:55 UTC
64 points
6 comments3 min readLW link

Rice’s The­o­rem says that AIs can’t de­ter­mine much from study­ing AI source code

Michael Weiss-Malik21 Aug 2023 19:05 UTC
−11 points
4 comments1 min readLW link

Large Lan­guage Models will be Great for Censorship

Ethan Edwards21 Aug 2023 19:03 UTC
183 points
14 comments8 min readLW link
(ethanedwards.substack.com)

“Throw­ing Ex­cep­tions” Is A Strange Pro­gram­ming Pattern

Thoth Hermes21 Aug 2023 18:50 UTC
−2 points
13 comments6 min readLW link
(thothhermes.substack.com)

[Question] Which pos­si­ble AI sys­tems are rel­a­tively safe?

Zach Stein-Perlman21 Aug 2023 17:00 UTC
42 points
20 comments1 min readLW link

Self-shut­down AI

jan betley21 Aug 2023 16:48 UTC
13 points
2 comments2 min readLW link

Con­tex­tual Trans­la­tions—At­tempt 1

Varshul Gupta21 Aug 2023 14:30 UTC
−1 points
0 comments2 min readLW link
(dubverseblack.substack.com)

DIY De­liber­ate Practice

lynettebye21 Aug 2023 12:22 UTC
62 points
4 comments5 min readLW link
(lynettebye.com)

Down­stairs Open­ing: 2br Apartment

jefftk21 Aug 2023 0:50 UTC
8 points
2 comments3 min readLW link
(www.jefftk.com)

Effi­ciency and re­source use scal­ing parity

Ege Erdil21 Aug 2023 0:18 UTC
47 points
0 comments20 min readLW link

Ruin­ing an ex­pected-log-money maximizer

philh20 Aug 2023 21:20 UTC
27 points
32 comments1 min readLW link
(reasonableapproximation.net)

Steven Wolfram on AI Alignment

Bill Benzon20 Aug 2023 19:49 UTC
65 points
15 comments4 min readLW link

[Question] What value does per­sonal pre­dic­tion track­ing have?

fx20 Aug 2023 18:43 UTC
7 points
3 comments1 min readLW link