Archive
Sequences
About
Search
Log In
Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
RSS
New
Hot
Active
Old
Page
1
Deep Causal Transcoding: A Framework for Mechanistically Eliciting Latent Behaviors in Language Models
Andrew Mack
and
TurnTrout
3 Dec 2024 21:19 UTC
40
points
2
comments
41
min read
LW
link
Do simulacra dream of digital sheep?
EuanMcLean
3 Dec 2024 20:25 UTC
17
points
1
comment
10
min read
LW
link
(The) Lightcone is nothing without its people: LW + Lighthaven’s big fundraiser
habryka
30 Nov 2024 2:55 UTC
518
points
157
comments
41
min read
LW
link
Should there be just one western AGI project?
rosehadshar
and
Tom Davidson
3 Dec 2024 10:11 UTC
46
points
12
comments
15
min read
LW
link
“Alignment at Large”: Bending the Arc of History Towards Life-Affirming Futures
welfvh
3 Dec 2024 21:17 UTC
3
points
0
comments
4
min read
LW
link
Cognitive Biases Contributing to AI X-risk — a deleted excerpt from my 2018 ARCHES draft
Andrew_Critch
3 Dec 2024 9:29 UTC
31
points
2
comments
5
min read
LW
link
A case for donating to AI risk reduction (including if you work in AI)
tlevin
2 Dec 2024 19:05 UTC
60
points
2
comments
1
min read
LW
link
2024 Unofficial LessWrong Census/Survey
Screwtape
2 Dec 2024 5:30 UTC
81
points
34
comments
1
min read
LW
link
Drexler’s Nanotech Software
PeterMcCluskey
2 Dec 2024 4:55 UTC
59
points
4
comments
4
min read
LW
link
(bayesianinvestor.com)
Fertility Roundup #4
Zvi
2 Dec 2024 14:30 UTC
27
points
9
comments
49
min read
LW
link
(thezvi.wordpress.com)
Chemical Turing Machines
Yudhister Kumar
3 Dec 2024 5:26 UTC
10
points
1
comment
4
min read
LW
link
(www.yudhister.me)
You should consider applying to PhDs (soon!)
bilalchughtai
29 Nov 2024 20:33 UTC
109
points
19
comments
6
min read
LW
link
You are not too “irrational” to know your preferences.
DaystarEld
26 Nov 2024 15:01 UTC
201
points
46
comments
13
min read
LW
link
How to make evals for the AISI evals bounty
TheManxLoiner
3 Dec 2024 10:44 UTC
2
points
0
comments
5
min read
LW
link
[Question]
Who are the worthwhile non-European pre-Industrial thinkers?
Lorec
3 Dec 2024 1:45 UTC
5
points
2
comments
1
min read
LW
link
AXRP Episode 39 - Evan Hubinger on Model Organisms of Misalignment
DanielFilan
1 Dec 2024 6:00 UTC
39
points
0
comments
67
min read
LW
link
[Question]
Which Biases are most important to Overcome?
abstractapplic
1 Dec 2024 15:40 UTC
30
points
14
comments
1
min read
LW
link
Teaching My Younger Self to Program: A case study of how I’d pass on my skill at self-learning
Shoshannah Tekofsky
1 Dec 2024 21:05 UTC
22
points
1
comment
7
min read
LW
link
(thinkfeelplay.substack.com)
Hierarchical Agency: A Missing Piece in AI Alignment
Jan_Kulveit
27 Nov 2024 5:49 UTC
99
points
17
comments
11
min read
LW
link
Levels of Thought: from Points to Fields
HNX
2 Dec 2024 20:25 UTC
4
points
2
comments
23
min read
LW
link
Back to top
Next