Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
Eliminating Cookie Banners is Hard
jefftk
Jan 13, 2024, 3:00 AM
23
points
15
comments
3
min read
LW
link
(www.jefftk.com)
Introducing Alignment Stress-Testing at Anthropic
evhub
Jan 12, 2024, 11:51 PM
182
points
23
comments
2
min read
LW
link
D&D.Sci(-fi): Colonizing the SuperHyperSphere
abstractapplic
Jan 12, 2024, 11:36 PM
48
points
23
comments
2
min read
LW
link
Commonwealth Fusion Systems is the Same Scale as OpenAI
Jeffrey Heninger
Jan 12, 2024, 9:43 PM
22
points
13
comments
2
min read
LW
link
Throughput vs. Latency
alkjash
and
Ruby
Jan 12, 2024, 9:37 PM
29
points
2
comments
13
min read
LW
link
Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training
evhub
,
Carson Denison
,
Meg
,
Monte M
,
David Duvenaud
,
Nicholas Schiefer
and
Ethan Perez
Jan 12, 2024, 7:51 PM
305
points
95
comments
3
min read
LW
link
(arxiv.org)
METAPHILOSOPHY—A Philosophizing through logical consequences
Seremonia
Jan 12, 2024, 6:47 PM
−7
points
7
comments
1
min read
LW
link
Idealism, Realistic & Pragmatic
Seremonia
Jan 12, 2024, 6:16 PM
−7
points
3
comments
1
min read
LW
link
The existential threat of humans.
Spiritus Dei
Jan 12, 2024, 5:50 PM
−24
points
0
comments
3
min read
LW
link
[Question]
Concrete examples of doing agentic things?
Jacob G-W
Jan 12, 2024, 3:59 PM
13
points
10
comments
1
min read
LW
link
Land Reclamation is in the 9th Circle of Stagnation Hell
Maxwell Tabarrok
Jan 12, 2024, 1:36 PM
54
points
6
comments
2
min read
LW
link
(maximumprogress.substack.com)
What good is G-factor if you’re dumped in the woods? A field report from a camp counselor.
Hastings
Jan 12, 2024, 1:17 PM
149
points
22
comments
1
min read
LW
link
A Chinese Room Containing a Stack of Stochastic Parrots
RogerDearnaley
Jan 12, 2024, 6:29 AM
20
points
3
comments
5
min read
LW
link
Decent plan prize announcement (1 paragraph, $1k)
lemonhope
Jan 12, 2024, 6:27 AM
25
points
19
comments
1
min read
LW
link
introduction to solid oxide electrolytes
bhauth
Jan 12, 2024, 5:35 AM
17
points
0
comments
4
min read
LW
link
(www.bhauth.com)
Apply to the 2024 PIBBSS Summer Research Fellowship
Nora_Ammann
,
DusanDNesic
and
Lucas Teixeira
Jan 12, 2024, 4:06 AM
39
points
1
comment
2
min read
LW
link
A Benchmark for Decision Theories
StrivingForLegibility
Jan 11, 2024, 6:54 PM
10
points
0
comments
2
min read
LW
link
An even deeper atheism
Joe Carlsmith
Jan 11, 2024, 5:28 PM
125
points
47
comments
15
min read
LW
link
Motivating Alignment of LLM-Powered Agents: Easy for AGI, Hard for ASI?
RogerDearnaley
Jan 11, 2024, 12:56 PM
35
points
4
comments
39
min read
LW
link
Reprograming the Mind: Meditation as a Tool for Cognitive Optimization
Jonas Hallgren
Jan 11, 2024, 12:03 PM
32
points
3
comments
11
min read
LW
link
AI-Generated Music for Learning
nomagicpill
Jan 11, 2024, 4:11 AM
9
points
1
comment
1
min read
LW
link
(210ethan.github.io)
Introduce a Speed Maximum
jefftk
Jan 11, 2024, 2:50 AM
36
points
28
comments
2
min read
LW
link
(www.jefftk.com)
[Question]
Prediction markets are consistently underconfident. Why?
Sinclair Chen
Jan 11, 2024, 2:44 AM
11
points
4
comments
1
min read
LW
link
Trying to align humans with inclusive genetic fitness
peterbarnett
Jan 11, 2024, 12:13 AM
23
points
5
comments
10
min read
LW
link
Universal Love Integration Test: Hitler
Raemon
Jan 10, 2024, 11:55 PM
76
points
65
comments
9
min read
LW
link
The Perceptron Controversy
Yuxi_Liu
Jan 10, 2024, 11:07 PM
65
points
18
comments
1
min read
LW
link
(yuxi-liu-wired.github.io)
The Aspiring Rationalist Congregation
maia
Jan 10, 2024, 10:52 PM
86
points
23
comments
10
min read
LW
link
An Actually Intuitive Explanation of the Oberth Effect
Isaac King
Jan 10, 2024, 8:23 PM
63
points
37
comments
6
min read
LW
link
Beware the suboptimal routine
jwfiredragon
Jan 10, 2024, 7:02 PM
13
points
3
comments
3
min read
LW
link
The true cost of fences
pleiotroth
Jan 10, 2024, 7:01 PM
3
points
2
comments
4
min read
LW
link
“Dark Constitution” for constraining some superintelligences
Valentine
Jan 10, 2024, 4:02 PM
3
points
9
comments
1
min read
LW
link
(www.anarchonomicon.com)
[Question]
rabbit (a new AI company) and Large Action Model (LAM)
MiguelDev
Jan 10, 2024, 1:57 PM
17
points
3
comments
1
min read
LW
link
Saving the world sucks
Defective Altruism
Jan 10, 2024, 5:55 AM
50
points
29
comments
3
min read
LW
link
[Question]
Questions about Solomonoff induction
mukashi
Jan 10, 2024, 1:16 AM
7
points
11
comments
1
min read
LW
link
AI as a natural disaster
Neil
Jan 10, 2024, 12:42 AM
11
points
1
comment
7
min read
LW
link
Stop being surprised by the passage of time
duck_master
and
00aleae
Jan 10, 2024, 12:36 AM
−2
points
1
comment
3
min read
LW
link
A discussion of normative ethics
Gordon Seidoh Worley
and
Adam Zerner
Jan 9, 2024, 11:29 PM
10
points
6
comments
25
min read
LW
link
On the Contrary, Steelmanning Is Normal; ITT-Passing Is Niche
Zack_M_Davis
Jan 9, 2024, 11:12 PM
45
points
31
comments
4
min read
LW
link
[Question]
What’s the protocol for if a novice has ML ideas that are unlikely to work, but might improve capabilities if they do work?
drocta
Jan 9, 2024, 10:51 PM
6
points
2
comments
2
min read
LW
link
Goodbye, Shoggoth: The Stage, its Animatronics, & the Puppeteer – a New Metaphor
RogerDearnaley
Jan 9, 2024, 8:42 PM
48
points
8
comments
36
min read
LW
link
Bent or Blunt Hoods?
jefftk
Jan 9, 2024, 8:10 PM
23
points
0
comments
1
min read
LW
link
(www.jefftk.com)
2024 ACX Predictions: Blind/Buy/Sell/Hold
Zvi
Jan 9, 2024, 7:30 PM
33
points
2
comments
31
min read
LW
link
(thezvi.wordpress.com)
Announcing the Double Crux Bot
sanyer
,
Sofia Vanhanen
and
sarah.bluhm
9 Jan 2024 18:54 UTC
53
points
10
comments
3
min read
LW
link
Does AI risk “other” the AIs?
Joe Carlsmith
9 Jan 2024 17:51 UTC
60
points
3
comments
8
min read
LW
link
AI demands unprecedented reliability
Jono
9 Jan 2024 16:30 UTC
22
points
5
comments
2
min read
LW
link
Uncertainty in all its flavours
Cleo Nardo
9 Jan 2024 16:21 UTC
34
points
6
comments
35
min read
LW
link
Compensating for Life Biases
Jonathan Moregård
9 Jan 2024 14:39 UTC
24
points
6
comments
3
min read
LW
link
(honestliving.substack.com)
Can Morality Be Quantified?
Julius
9 Jan 2024 6:35 UTC
3
points
0
comments
5
min read
LW
link
Learning Math in Time for Alignment
Nicholas / Heather Kross
9 Jan 2024 1:02 UTC
32
points
5
comments
3
min read
LW
link
Brief Thoughts on Justifications for Paternalism
Srdjan Miletic
9 Jan 2024 0:36 UTC
4
points
0
comments
4
min read
LW
link
(dissent.blog)
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel