Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
[Question]
Were there any ancient rationalists?
OliverHayman
May 3, 2024, 6:26 PM
11
points
3
comments
1
min read
LW
link
Key takeaways from our EA and alignment research surveys
Cameron Berg
,
Judd Rosenblatt
,
florin_pop
and
AE Studio
May 3, 2024, 6:10 PM
112
points
10
comments
21
min read
LW
link
“AI Safety for Fleshy Humans” an AI Safety explainer by Nicky Case
habryka
May 3, 2024, 6:10 PM
90
points
11
comments
4
min read
LW
link
(aisafety.dance)
AI Clarity: An Initial Research Agenda
Justin Bullock
,
Corin Katzke
,
Zershaaneh Qureshi
and
David_Kristoffersson
May 3, 2024, 1:54 PM
18
points
1
comment
8
min read
LW
link
Apply to ESPR & PAIR, Rationality and AI Camps for Ages 16-21
Anna Gajdova
May 3, 2024, 12:36 PM
58
points
5
comments
1
min read
LW
link
On precise out-of-context steering
Olli Järviniemi
May 3, 2024, 9:41 AM
9
points
6
comments
3
min read
LW
link
LLM+Planners hybridisation for friendly AGI
installgentoo
May 3, 2024, 8:40 AM
7
points
2
comments
1
min read
LW
link
Mechanistic Interpretability Workshop Happening at ICML 2024!
Neel Nanda
,
LawrenceC
and
Fazl
May 3, 2024, 1:18 AM
48
points
6
comments
1
min read
LW
link
Weekly newsletter for AI safety events and training programs
Bryce Robertson
May 3, 2024, 12:33 AM
29
points
0
comments
1
min read
LW
link
CCS: Counterfactual Civilization Simulation
Morphism
May 2, 2024, 10:54 PM
3
points
0
comments
2
min read
LW
link
Let’s Design A School, Part 2.1 School as Education—Structure
Sable
May 2, 2024, 10:04 PM
26
points
2
comments
10
min read
LW
link
(affablyevil.substack.com)
Why I’m not doing PauseAI
kwiat.dev
May 2, 2024, 10:00 PM
−8
points
5
comments
4
min read
LW
link
AI #61: Meta Trouble
Zvi
May 2, 2024, 6:40 PM
29
points
0
comments
52
min read
LW
link
(thezvi.wordpress.com)
Why is AGI/ASI Inevitable?
DeathlessAmaranth
May 2, 2024, 6:27 PM
14
points
6
comments
1
min read
LW
link
AISN #34: New Military AI Systems Plus, AI Labs Fail to Uphold Voluntary Commitments to UK AI Safety Institute, and New AI Policy Proposals in the US Senate
Corin Katzke
and
Dan H
May 2, 2024, 4:12 PM
6
points
0
comments
8
min read
LW
link
(newsletter.safe.ai)
Ai Salon: Trustworthy AI Futures #1
Ian Eisenberg
May 2, 2024, 4:07 PM
1
point
0
comments
1
min read
LW
link
How to write Pseudocode and why you should
Johannes C. Mayer
May 2, 2024, 3:53 PM
8
points
5
comments
3
min read
LW
link
AI #62: Too Soon to Tell
Zvi
May 2, 2024, 3:40 PM
30
points
8
comments
31
min read
LW
link
(thezvi.wordpress.com)
Whiteboard Program Traceing: Debug a Program Before you have the Code
Johannes C. Mayer
May 2, 2024, 3:30 PM
3
points
0
comments
1
min read
LW
link
[Question]
Which skincare products are evidence-based?
Vanessa Kosoy
May 2, 2024, 3:22 PM
120
points
48
comments
1
min read
LW
link
Q&A on Proposed SB 1047
Zvi
May 2, 2024, 3:10 PM
74
points
8
comments
44
min read
LW
link
(thezvi.wordpress.com)
[Question]
What are the Activities that make up your Research Process?
Johannes C. Mayer
May 2, 2024, 3:01 PM
4
points
0
comments
1
min read
LW
link
[Question]
How do you Select the Right Research Acitivity in the Right Moment?
Johannes C. Mayer
May 2, 2024, 2:45 PM
6
points
1
comment
1
min read
LW
link
[Question]
Can stealth aircraft be detected optically?
Yair Halberstadt
May 2, 2024, 7:47 AM
20
points
27
comments
1
min read
LW
link
An explanation of evil in an organized world
KatjaGrace
May 2, 2024, 5:20 AM
26
points
9
comments
2
min read
LW
link
(worldspiritsockpuppet.com)
Why I stopped working on AI safety
jbkjr
May 2, 2024, 5:08 AM
−5
points
0
comments
4
min read
LW
link
(jbkjr.me)
[Linkpost] Silver Bulletin: For most people, politics is about fitting in
Gunnar_Zarncke
May 1, 2024, 6:12 PM
18
points
4
comments
1
min read
LW
link
(www.natesilver.net)
Launching applications for AI Safety Careers Course India 2024
Axiom_Futures
May 1, 2024, 5:55 PM
4
points
1
comment
1
min read
LW
link
[Question]
Shane Legg’s necessary properties for every AGI Safety plan
jacquesthibs
May 1, 2024, 5:15 PM
58
points
12
comments
1
min read
LW
link
KAN: Kolmogorov-Arnold Networks
Gunnar_Zarncke
May 1, 2024, 4:50 PM
18
points
15
comments
1
min read
LW
link
(arxiv.org)
Manifund Q1 Retro: Learnings from impact certs
Austin Chen
May 1, 2024, 4:48 PM
40
points
1
comment
LW
link
ACX Covid Origins Post convinced readers
ErnestScribbler
May 1, 2024, 1:06 PM
77
points
7
comments
2
min read
LW
link
LessWrong Community Weekend 2024, open for applications
UnplannedCauliflower
and
jt
May 1, 2024, 10:18 AM
79
points
2
comments
7
min read
LW
link
Take SCIFs, it’s dangerous to go alone
latterframe
,
Jeffrey Ladish
and
schroederdewitt
May 1, 2024, 8:02 AM
42
points
1
comment
3
min read
LW
link
AXRP Episode 30 - AI Security with Jeffrey Ladish
DanielFilan
May 1, 2024, 2:50 AM
25
points
0
comments
79
min read
LW
link
Neuro/BCI/WBE for Safe AI Workshop
Allison Duettmann
May 1, 2024, 12:46 AM
3
points
0
comments
1
min read
LW
link
AGI: Cryptography, Security & Multipolar Scenarios Workshop
Allison Duettmann
May 1, 2024, 12:42 AM
8
points
1
comment
1
min read
LW
link
The formal goal is a pointer
Morphism
May 1, 2024, 12:27 AM
20
points
10
comments
1
min read
LW
link
Arch-anarchy:Theory and practice
Peter lawless
Apr 30, 2024, 11:20 PM
−6
points
0
comments
2
min read
LW
link
“Open Source AI” is a lie, but it doesn’t have to be
jacobhaimes
Apr 30, 2024, 11:10 PM
19
points
5
comments
6
min read
LW
link
(jacob-haimes.github.io)
Questions for labs
Zach Stein-Perlman
Apr 30, 2024, 10:15 PM
77
points
11
comments
8
min read
LW
link
Reality comprehensibility: are there illogical things in reality?
DDthinker
Apr 30, 2024, 9:30 PM
−3
points
0
comments
10
min read
LW
link
Mechanistically Eliciting Latent Behaviors in Language Models
Andrew Mack
and
TurnTrout
Apr 30, 2024, 6:51 PM
210
points
43
comments
45
min read
LW
link
[Question]
What is the easiest/funnest way to build up a comprehensive understanding of AI and AI Safety?
Jordan Arel
Apr 30, 2024, 6:41 PM
4
points
2
comments
1
min read
LW
link
Transcoders enable fine-grained interpretable circuit analysis for language models
Jacob Dunefsky
,
Philippe Chlenski
and
Neel Nanda
Apr 30, 2024, 5:58 PM
74
points
14
comments
17
min read
LW
link
Announcing the 2024 Roots of Progress Blog-Building Intensive
jasoncrawford
Apr 30, 2024, 5:37 PM
14
points
0
comments
2
min read
LW
link
(rootsofprogress.org)
The Intentional Stance, LLMs Edition
Eleni Angelou
Apr 30, 2024, 5:12 PM
30
points
3
comments
8
min read
LW
link
Introducing AI Lab Watch
Zach Stein-Perlman
Apr 30, 2024, 5:00 PM
225
points
30
comments
1
min read
LW
link
(ailabwatch.org)
Why I’m doing PauseAI
Joseph Miller
30 Apr 2024 16:21 UTC
108
points
16
comments
4
min read
LW
link
LLMs could be as conscious as human emulations, potentially
Canaletto
30 Apr 2024 11:36 UTC
15
points
15
comments
3
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel