Spec­u­la­tion on Path-Depen­dance in Large Lan­guage Models.

NickyPJan 15, 2023, 8:42 PM
16 points
2 comments7 min readLW link

Un­der­speci­fi­ca­tion of Or­a­cle AI

Jan 15, 2023, 8:10 PM
30 points
12 comments19 min readLW link

[Question] How Does the Hu­man Brain Com­pare to Deep Learn­ing on Sam­ple Effi­ciency?

DragonGodJan 15, 2023, 7:49 PM
11 points
6 comments1 min readLW link

De­cep­tive failures short of full catas­tro­phe.

Alex Lawsen Jan 15, 2023, 7:28 PM
33 points
5 comments9 min readLW link

Non-di­rected con­cep­tual found­ing

TsviBTJan 15, 2023, 2:56 PM
12 points
3 comments1 min readLW link

Panop­ti­cons aren’t enough

Program DenJan 15, 2023, 12:55 PM
−10 points
7 comments1 min readLW link

[Question] Is this chat GPT rewrite of my post bet­ter?

Yair HalberstadtJan 15, 2023, 9:47 AM
2 points
5 comments1 min readLW link

A sim­ple pro­posal for pre­serv­ing free speech on twitter

Yair HalberstadtJan 15, 2023, 9:42 AM
−2 points
13 comments1 min readLW link

Core Con­cept Con­ver­sa­tion: What is tech­nol­ogy?

Adam ZernerJan 15, 2023, 9:40 AM
8 points
1 comment1 min readLW link

Lan­guage Ex Machina

janusJan 15, 2023, 9:19 AM
42 points
23 comments24 min readLW link
(generative.ink)

Core Con­cept Con­ver­sa­tion: What is wealth?

Adam ZernerJan 15, 2023, 9:07 AM
13 points
30 comments3 min readLW link

Core Con­cept Conversations

Adam ZernerJan 15, 2023, 7:17 AM
14 points
1 comment1 min readLW link

In­cen­tives con­sid­ered harmful

Ulisse MiniJan 15, 2023, 6:38 AM
6 points
0 comments1 min readLW link
(uli.rocks)

Con­sider pay­ing for liter­a­ture or book re­views us­ing boun­ties and dom­i­nant as­surance contracts

Arjun PanicksseryJan 15, 2023, 3:56 AM
57 points
7 comments2 min readLW link

Pod­cast with Divia Eden on op­er­ant conditioning

DanielFilanJan 15, 2023, 2:44 AM
14 points
0 comments1 min readLW link
(youtu.be)

We Need Holis­tic AI Macrostrategy

NickGabsJan 15, 2023, 2:13 AM
39 points
4 comments8 min readLW link

[Question] When to men­tion ir­rele­vant ac­cu­sa­tions?

philhJan 14, 2023, 9:58 PM
20 points
50 comments1 min readLW link

World-Model In­ter­pretabil­ity Is All We Need

Thane RuthenisJan 14, 2023, 7:37 PM
36 points
22 comments21 min readLW link

Cur­rent AI Models Seem Suffi­cient for Low-Risk, Benefi­cial AI

harsimonyJan 14, 2023, 6:55 PM
17 points
1 comment2 min readLW link

[Question] Ba­sic Ques­tion about LLMs: how do they know what task to perform

GarakJan 14, 2023, 1:13 PM
1 point
3 comments1 min readLW link

Aligned with what?

Program DenJan 14, 2023, 10:28 AM
3 points
41 comments1 min readLW link

Wok­ism, re­think­ing pri­ori­ties and the Bostrom case

Arturo MaciasJan 14, 2023, 2:27 AM
−25 points
2 comments4 min readLW link

A gen­eral com­ment on dis­cus­sions of ge­netic group differences

anonymous8101Jan 14, 2023, 2:11 AM
71 points
46 comments3 min readLW link

Ab­strac­tions as mor­phisms be­tween (co)algebras

Erik JennerJan 14, 2023, 1:51 AM
17 points
1 comment8 min readLW link

Con­crete Rea­sons for Hope about AI

Zac Hatfield-DoddsJan 14, 2023, 1:22 AM
94 points
13 comments1 min readLW link

Nega­tive Ex­per­tise

Jonas KgomoJan 14, 2023, 12:51 AM
4 points
0 comments1 min readLW link
(twitter.com)

Mid-At­lantic AI Align­ment Alli­ance Unconference

QuinnJan 13, 2023, 8:33 PM
7 points
2 comments1 min readLW link

Smal­lpox vac­cines are widely available, for now

David HornbeinJan 13, 2023, 8:02 PM
26 points
5 comments1 min readLW link

How does GPT-3 spend its 175B pa­ram­e­ters?

Robert_AIZIJan 13, 2023, 7:21 PM
41 points
14 comments6 min readLW link
(aizi.substack.com)

[ASoT] Si­mu­la­tors show us be­havi­oural prop­er­ties by default

JozdienJan 13, 2023, 6:42 PM
36 points
3 comments3 min readLW link

Wheel of Con­sent The­ory for Ra­tion­al­ists and Effec­tive Altruists

adamwilderJan 13, 2023, 5:59 PM
1 point
0 comments2 min readLW link

Money is a way of thank­ing strangers

DirectedEvolutionJan 13, 2023, 5:06 PM
13 points
5 comments4 min readLW link

Tracr: Com­piled Trans­form­ers as a Lab­o­ra­tory for In­ter­pretabil­ity | Deep­Mind

DragonGodJan 13, 2023, 4:53 PM
62 points
12 comments1 min readLW link
(arxiv.org)

How we could stum­ble into AI catastrophe

HoldenKarnofskyJan 13, 2023, 4:20 PM
71 points
18 comments18 min readLW link
(www.cold-takes.com)

Ro­bust­ness & Evolu­tion [MLAISU W02]

Esben KranJan 13, 2023, 3:47 PM
10 points
0 comments3 min readLW link
(newsletter.apartresearch.com)

On Cook­ing With Gas

ZviJan 13, 2023, 2:20 PM
38 points
60 comments6 min readLW link
(thezvi.wordpress.com)

Be­ware safety-washing

LizkaJan 13, 2023, 1:59 PM
51 points
2 comments4 min readLW link

Some Ar­gu­ments Against Strong Scaling

Joar SkalseJan 13, 2023, 12:04 PM
25 points
21 comments16 min readLW link

[Question] Where do you find peo­ple who ac­tu­ally do things?

Ulisse MiniJan 13, 2023, 6:57 AM
7 points
12 comments1 min readLW link

[Question] Could Si­mu­lat­ing an AGI Tak­ing Over the World Ac­tu­ally Lead to a LLM Tak­ing Over the World?

simeon_cJan 13, 2023, 6:33 AM
15 points
1 comment1 min readLW link

Burn­ing Up­time: When your Sand­box of Em­pa­thy is Leaky and also an Hourglass

CedarJan 13, 2023, 5:18 AM
13 points
2 comments3 min readLW link

Disen­tan­gling Shard The­ory into Atomic Claims

Leon LangJan 13, 2023, 4:23 AM
86 points
6 comments18 min readLW link

AGISF adap­ta­tion for in-per­son groups

Jan 13, 2023, 3:24 AM
44 points
2 comments3 min readLW link

Ac­tions and Flows

Alok SinghJan 13, 2023, 3:20 AM
5 points
0 comments1 min readLW link
(alok.github.io)

A Thor­ough In­tro­duc­tion to Abstraction

RohanSJan 13, 2023, 12:30 AM
9 points
1 comment18 min readLW link

The AI Con­trol Prob­lem in a wider in­tel­lec­tual context

philosophybearJan 13, 2023, 12:28 AM
11 points
3 comments12 min readLW link

The Align­ment Problems

Martín SotoJan 12, 2023, 10:29 PM
20 points
0 comments4 min readLW link

Pro­posal for In­duc­ing Steganog­ra­phy in LMs

Logan RiggsJan 12, 2023, 10:15 PM
22 points
3 comments2 min readLW link

An­nounc­ing the 2023 PIBBSS Sum­mer Re­search Fellowship

Jan 12, 2023, 9:31 PM
32 points
0 comments1 min readLW link

Vic­to­ria Krakovna on AGI Ruin, The Sharp Left Turn and Paradigms of AI Alignment

Michaël TrazziJan 12, 2023, 5:09 PM
40 points
3 comments4 min readLW link
(www.theinsideview.ai)