D&D.Sci Hyper­sphere Anal­y­sis Part 1: Datafields & Pre­limi­nary Analysis

aphyerJan 13, 2024, 8:16 PM
29 points
1 comment5 min readLW link

Some ad­di­tional SAE thoughts

HoagyJan 13, 2024, 7:31 PM
31 points
4 comments13 min readLW link

(4 min read) An in­tu­itive ex­pla­na­tion of the AI in­fluence situation

trevorJan 13, 2024, 5:34 PM
12 points
26 comments4 min readLW link

AI #47: Meet the New Year

ZviJan 13, 2024, 4:20 PM
36 points
7 comments57 min readLW link
(thezvi.wordpress.com)

Take­aways from the NeurIPS 2023 Tro­jan De­tec­tion Competition

mikesJan 13, 2024, 12:35 PM
20 points
2 comments1 min readLW link
(confirmlabs.org)

[Question] Why do so many think de­cep­tion in AI is im­por­tant?

PrometheusJan 13, 2024, 8:14 AM
24 points
12 comments1 min readLW link

Elimi­nat­ing Cookie Ban­ners is Hard

jefftkJan 13, 2024, 3:00 AM
23 points
15 comments3 min readLW link
(www.jefftk.com)

In­tro­duc­ing Align­ment Stress-Test­ing at Anthropic

evhubJan 12, 2024, 11:51 PM
182 points
23 comments2 min readLW link

D&D.Sci(-fi): Coloniz­ing the SuperHyperSphere

abstractapplicJan 12, 2024, 11:36 PM
48 points
23 comments2 min readLW link

Com­mon­wealth Fu­sion Sys­tems is the Same Scale as OpenAI

Jeffrey HeningerJan 12, 2024, 9:43 PM
22 points
13 comments2 min readLW link

Through­put vs. Latency

Jan 12, 2024, 9:37 PM
29 points
2 comments13 min readLW link

Sleeper Agents: Train­ing De­cep­tive LLMs that Per­sist Through Safety Training

Jan 12, 2024, 7:51 PM
305 points
95 comments3 min readLW link
(arxiv.org)

METAPHILOSOPHY—A Philoso­phiz­ing through log­i­cal consequences

SeremoniaJan 12, 2024, 6:47 PM
−7 points
7 comments1 min readLW link

Ideal­ism, Real­is­tic & Pragmatic

SeremoniaJan 12, 2024, 6:16 PM
−7 points
3 comments1 min readLW link

The ex­is­ten­tial threat of hu­mans.

Spiritus DeiJan 12, 2024, 5:50 PM
−24 points
0 comments3 min readLW link

[Question] Con­crete ex­am­ples of do­ing agen­tic things?

Jacob G-WJan 12, 2024, 3:59 PM
13 points
10 comments1 min readLW link

Land Recla­ma­tion is in the 9th Cir­cle of Stag­na­tion Hell

Maxwell TabarrokJan 12, 2024, 1:36 PM
54 points
6 comments2 min readLW link
(maximumprogress.substack.com)

What good is G-fac­tor if you’re dumped in the woods? A field re­port from a camp coun­selor.

HastingsJan 12, 2024, 1:17 PM
149 points
22 comments1 min readLW link

A Chi­nese Room Con­tain­ing a Stack of Stochas­tic Parrots

RogerDearnaleyJan 12, 2024, 6:29 AM
20 points
3 comments5 min readLW link

De­cent plan prize an­nounce­ment (1 para­graph, $1k)

lemonhopeJan 12, 2024, 6:27 AM
25 points
19 comments1 min readLW link

in­tro­duc­tion to solid ox­ide electrolytes

bhauthJan 12, 2024, 5:35 AM
17 points
0 comments4 min readLW link
(www.bhauth.com)

Ap­ply to the 2024 PIBBSS Sum­mer Re­search Fellowship

Jan 12, 2024, 4:06 AM
39 points
1 comment2 min readLW link

A Bench­mark for De­ci­sion Theories

StrivingForLegibilityJan 11, 2024, 6:54 PM
10 points
0 comments2 min readLW link

An even deeper atheism

Joe CarlsmithJan 11, 2024, 5:28 PM
125 points
47 comments15 min readLW link

Mo­ti­vat­ing Align­ment of LLM-Pow­ered Agents: Easy for AGI, Hard for ASI?

RogerDearnaleyJan 11, 2024, 12:56 PM
35 points
4 comments39 min readLW link

Re­pro­gram­ing the Mind: Med­i­ta­tion as a Tool for Cog­ni­tive Optimization

Jonas HallgrenJan 11, 2024, 12:03 PM
32 points
3 comments11 min readLW link

AI-Gen­er­ated Mu­sic for Learning

nomagicpillJan 11, 2024, 4:11 AM
9 points
1 comment1 min readLW link
(210ethan.github.io)

In­tro­duce a Speed Maximum

jefftkJan 11, 2024, 2:50 AM
36 points
28 comments2 min readLW link
(www.jefftk.com)

[Question] Pre­dic­tion mar­kets are con­sis­tently un­der­con­fi­dent. Why?

Sinclair ChenJan 11, 2024, 2:44 AM
11 points
4 comments1 min readLW link

Try­ing to al­ign hu­mans with in­clu­sive ge­netic fitness

peterbarnettJan 11, 2024, 12:13 AM
23 points
5 comments10 min readLW link

Univer­sal Love In­te­gra­tion Test: Hitler

RaemonJan 10, 2024, 11:55 PM
76 points
65 comments9 min readLW link

The Per­cep­tron Controversy

Yuxi_LiuJan 10, 2024, 11:07 PM
65 points
18 comments1 min readLW link
(yuxi-liu-wired.github.io)

The Aspiring Ra­tion­al­ist Congregation

maiaJan 10, 2024, 10:52 PM
86 points
23 comments10 min readLW link

An Ac­tu­ally In­tu­itive Ex­pla­na­tion of the Oberth Effect

Isaac KingJan 10, 2024, 8:23 PM
63 points
37 comments6 min readLW link

Be­ware the sub­op­ti­mal routine

jwfiredragonJan 10, 2024, 7:02 PM
13 points
3 comments3 min readLW link

The true cost of fences

pleiotrothJan 10, 2024, 7:01 PM
3 points
2 comments4 min readLW link

“Dark Con­sti­tu­tion” for con­strain­ing some superintelligences

ValentineJan 10, 2024, 4:02 PM
3 points
9 comments1 min readLW link
(www.anarchonomicon.com)

[Question] rab­bit (a new AI com­pany) and Large Ac­tion Model (LAM)

MiguelDevJan 10, 2024, 1:57 PM
17 points
3 comments1 min readLW link

Sav­ing the world sucks

Defective AltruismJan 10, 2024, 5:55 AM
50 points
29 comments3 min readLW link

[Question] Ques­tions about Solomonoff induction

mukashiJan 10, 2024, 1:16 AM
7 points
11 comments1 min readLW link

AI as a nat­u­ral disaster

Neil Jan 10, 2024, 12:42 AM
11 points
1 comment7 min readLW link

Stop be­ing sur­prised by the pas­sage of time

Jan 10, 2024, 12:36 AM
−2 points
1 comment3 min readLW link

A dis­cus­sion of nor­ma­tive ethics

Jan 9, 2024, 11:29 PM
10 points
6 comments25 min readLW link

On the Con­trary, Steel­man­ning Is Nor­mal; ITT-Pass­ing Is Niche

Zack_M_DavisJan 9, 2024, 11:12 PM
45 points
31 comments4 min readLW link

[Question] What’s the pro­to­col for if a novice has ML ideas that are un­likely to work, but might im­prove ca­pa­bil­ities if they do work?

droctaJan 9, 2024, 10:51 PM
6 points
2 comments2 min readLW link

Good­bye, Shog­goth: The Stage, its An­i­ma­tron­ics, & the Pup­peteer – a New Metaphor

RogerDearnaleyJan 9, 2024, 8:42 PM
47 points
8 comments36 min readLW link

Bent or Blunt Hoods?

jefftkJan 9, 2024, 8:10 PM
23 points
0 comments1 min readLW link
(www.jefftk.com)

2024 ACX Pre­dic­tions: Blind/​Buy/​Sell/​Hold

ZviJan 9, 2024, 7:30 PM
33 points
2 comments31 min readLW link
(thezvi.wordpress.com)

An­nounc­ing the Dou­ble Crux Bot

Jan 9, 2024, 6:54 PM
53 points
10 comments3 min readLW link

Does AI risk “other” the AIs?

Joe CarlsmithJan 9, 2024, 5:51 PM
60 points
3 comments8 min readLW link