Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
Project ideas: Epistemics
Lukas Finnveden
Jan 5, 2024, 11:41 PM
43
points
4
comments
LW
link
(www.forethought.org)
Almost everyone I’ve met would be well-served thinking more about what to focus on
Henrik Karlsson
Jan 5, 2024, 9:01 PM
96
points
8
comments
11
min read
LW
link
(www.henrikkarlsson.xyz)
The Next ChatGPT Moment: AI Avatars
kolmplex
and
southpaw
Jan 5, 2024, 8:14 PM
43
points
10
comments
1
min read
LW
link
AI Impacts 2023 Expert Survey on Progress in AI
habryka
Jan 5, 2024, 7:42 PM
28
points
2
comments
7
min read
LW
link
(wiki.aiimpacts.org)
Technology path dependence and evaluating expertise
bhauth
and
Muireall
Jan 5, 2024, 7:21 PM
25
points
2
comments
15
min read
LW
link
The Hippie Rabbit Hole -Nuggets of Gold in Rivers of Bullshit
Jonathan Moregård
Jan 5, 2024, 6:27 PM
39
points
20
comments
8
min read
LW
link
(honestliving.substack.com)
[Question]
What technical topics could help with boundaries/membranes?
Chipmonk
Jan 5, 2024, 6:14 PM
15
points
25
comments
1
min read
LW
link
Catching AIs red-handed
ryan_greenblatt
and
Buck
Jan 5, 2024, 5:43 PM
111
points
27
comments
17
min read
LW
link
AI Impacts Survey: December 2023 Edition
Zvi
Jan 5, 2024, 2:40 PM
34
points
6
comments
10
min read
LW
link
(thezvi.wordpress.com)
Forecast your 2024 with Fatebook
Sage Future
Jan 5, 2024, 2:07 PM
19
points
0
comments
1
min read
LW
link
(fatebook.io)
Predictive model agents are sort of corrigible
Raymond Douglas
Jan 5, 2024, 2:05 PM
35
points
6
comments
3
min read
LW
link
Striking Implications for Learning Theory, Interpretability — and Safety?
RogerDearnaley
Jan 5, 2024, 8:46 AM
37
points
4
comments
2
min read
LW
link
If I ran the zoo
Optimization Process
Jan 5, 2024, 5:14 AM
18
points
1
comment
2
min read
LW
link
Does AI care about reality or just its own perception?
RedFishBlueFish
Jan 5, 2024, 4:05 AM
−6
points
8
comments
1
min read
LW
link
MIRI 2024 Mission and Strategy Update
Malo
Jan 5, 2024, 12:20 AM
223
points
44
comments
8
min read
LW
link
Project ideas: Governance during explosive technological growth
Lukas Finnveden
Jan 4, 2024, 11:51 PM
14
points
0
comments
LW
link
(www.forethought.org)
Hello
S Benfield
Jan 4, 2024, 11:35 PM
6
points
0
comments
2
min read
LW
link
Using Threats to Achieve Socially Optimal Outcomes
StrivingForLegibility
Jan 4, 2024, 11:30 PM
8
points
0
comments
3
min read
LW
link
Best-Responding Is Not Always the Best Response
StrivingForLegibility
Jan 4, 2024, 11:30 PM
10
points
0
comments
3
min read
LW
link
Safety Data Sheets for Optimization Processes
StrivingForLegibility
Jan 4, 2024, 11:30 PM
15
points
1
comment
4
min read
LW
link
The Gears of Argmax
StrivingForLegibility
Jan 4, 2024, 11:30 PM
11
points
0
comments
3
min read
LW
link
Cellular reprogramming, pneumatic launch systems, and terraforming Mars: Some things I learned about at Foresight Vision Weekend
jasoncrawford
Jan 4, 2024, 7:33 PM
28
points
0
comments
8
min read
LW
link
(rootsofprogress.org)
Deep atheism and AI risk
Joe Carlsmith
Jan 4, 2024, 6:58 PM
153
points
22
comments
27
min read
LW
link
Some Vacation Photos
johnswentworth
Jan 4, 2024, 5:15 PM
83
points
0
comments
1
min read
LW
link
AISN #29: Progress on the EU AI Act Plus, the NY Times sues OpenAI for Copyright Infringement, and Congressional Questions about Research Standards in AI Safety
Dan H
and
Corin Katzke
Jan 4, 2024, 4:09 PM
8
points
0
comments
6
min read
LW
link
(newsletter.safe.ai)
EAG Bay Area Satellite event: AI Institution Design Hackathon 2024
beatrice@foresight.org
Jan 4, 2024, 3:02 PM
1
point
0
comments
1
min read
LW
link
AI #45: To Be Determined
Zvi
Jan 4, 2024, 3:00 PM
52
points
4
comments
31
min read
LW
link
(thezvi.wordpress.com)
Screen-supported Portable Monitor
jefftk
Jan 4, 2024, 1:50 PM
16
points
10
comments
1
min read
LW
link
(www.jefftk.com)
[Question]
Which investments for aligned-AI outcomes?
tailcalled
Jan 4, 2024, 1:28 PM
8
points
9
comments
2
min read
LW
link
Non-alignment project ideas for making transformative AI go well
Lukas Finnveden
Jan 4, 2024, 7:23 AM
44
points
1
comment
LW
link
(www.forethought.org)
Fact Checking and Retaliation Against Sources
jefftk
Jan 4, 2024, 12:41 AM
7
points
2
comments
4
min read
LW
link
(www.jefftk.com)
Investigating Alternative Futures: Human and Superintelligence Interaction Scenarios
Hiroshi Yamakawa
Jan 3, 2024, 11:46 PM
1
point
0
comments
17
min read
LW
link
“Attitudes Toward Artificial General Intelligence: Results from American Adults 2021 and 2023”—call for reviewers (Seeds of Science)
rogersbacon
Jan 3, 2024, 8:11 PM
4
points
0
comments
1
min read
LW
link
What’s up with LLMs representing XORs of arbitrary features?
Sam Marks
Jan 3, 2024, 7:44 PM
158
points
63
comments
16
min read
LW
link
Spirit Airlines Merger Play
sapphire
Jan 3, 2024, 7:25 PM
5
points
12
comments
1
min read
LW
link
$300 for the best sci-fi prompt: the results
RomanS
Jan 3, 2024, 7:10 PM
16
points
19
comments
7
min read
LW
link
Agent membranes/boundaries and formalizing “safety”
Chipmonk
3 Jan 2024 17:55 UTC
26
points
46
comments
3
min read
LW
link
Safety First: safety before full alignment. The deontic sufficiency hypothesis.
Chipmonk
3 Jan 2024 17:55 UTC
48
points
3
comments
3
min read
LW
link
Practically A Book Review: Appendix to “Nonlinear’s Evidence: Debunking False and Misleading Claims” (ThingOfThings)
tailcalled
3 Jan 2024 17:07 UTC
111
points
25
comments
2
min read
LW
link
(thingofthings.substack.com)
Trivial Mathematics as a Path Forward
ACrackedPot
3 Jan 2024 16:41 UTC
−4
points
2
comments
2
min read
LW
link
Copyright Confrontation #1
Zvi
3 Jan 2024 15:50 UTC
34
points
7
comments
18
min read
LW
link
(thezvi.wordpress.com)
[Question]
Theoretically, could we balance the budget painlessly?
Logan Zoellner
3 Jan 2024 14:46 UTC
4
points
12
comments
1
min read
LW
link
Johannes’ Biography
Johannes C. Mayer
3 Jan 2024 13:27 UTC
24
points
0
comments
10
min read
LW
link
What Helped Me—Kale, Blood, CPAP, X-tiamine, Methylphenidate
Johannes C. Mayer
3 Jan 2024 13:22 UTC
35
points
12
comments
2
min read
LW
link
[Question]
Does LessWrong make a difference when it comes to AI alignment?
PhilosophicalSoul
3 Jan 2024 12:21 UTC
18
points
13
comments
1
min read
LW
link
[Question]
Terminology: <something>-ware for ML?
Oliver Sourbut
3 Jan 2024 11:42 UTC
17
points
27
comments
1
min read
LW
link
Trading off Lives
jefftk
3 Jan 2024 3:40 UTC
53
points
12
comments
2
min read
LW
link
(www.jefftk.com)
MonoPoly Restricted Trust
ymeskhout
2 Jan 2024 23:02 UTC
42
points
37
comments
9
min read
LW
link
Agent membranes and causal distance
Chipmonk
2 Jan 2024 22:43 UTC
20
points
3
comments
3
min read
LW
link
Focusing on Mal-Alignment
John Fisher
2 Jan 2024 19:51 UTC
1
point
0
comments
1
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel