Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
Page
1
[Question]
Monotonous Work
Gideon Bauer
Feb 2, 2023, 9:35 PM
1
point
0
comments
1
min read
LW
link
Is AI risk assessment too anthropocentric?
Craig Mattson
Feb 2, 2023, 9:34 PM
3
points
6
comments
1
min read
LW
link
Halifax Monthly Meetup: Introduction to Effective Altruism
Ideopunk
Feb 2, 2023, 9:10 PM
10
points
0
comments
1
min read
LW
link
Conditioning Predictive Models: Outer alignment via careful conditioning
evhub
,
Adam Jermyn
,
Johannes Treutlein
,
Rubi J. Hudson
and
kcwoolverton
Feb 2, 2023, 8:28 PM
72
points
15
comments
57
min read
LW
link
Conditioning Predictive Models: Large language models as predictors
evhub
,
Adam Jermyn
,
Johannes Treutlein
,
Rubi J. Hudson
and
kcwoolverton
Feb 2, 2023, 8:28 PM
88
points
4
comments
13
min read
LW
link
Normative vs Descriptive Models of Agency
mattmacdermott
Feb 2, 2023, 8:28 PM
26
points
5
comments
4
min read
LW
link
Andrew Huberman on How to Optimize Sleep
Leon Lang
Feb 2, 2023, 8:17 PM
37
points
6
comments
6
min read
LW
link
[Question]
How can I help inflammation-based nerve damage be temporary?
Optimization Process
Feb 2, 2023, 7:20 PM
17
points
4
comments
1
min read
LW
link
More findings on maximal data dimension
Marius Hobbhahn
Feb 2, 2023, 6:33 PM
27
points
1
comment
11
min read
LW
link
Heritability, Behaviorism, and Within-Lifetime RL
Steven Byrnes
Feb 2, 2023, 4:34 PM
39
points
3
comments
4
min read
LW
link
Covid 2/2/23: The Emergency Ends on 5/11
Zvi
Feb 2, 2023, 2:00 PM
22
points
6
comments
7
min read
LW
link
(thezvi.wordpress.com)
You are probably not a good alignment researcher, and other blatant lies
junk heap homotopy
Feb 2, 2023, 1:55 PM
83
points
16
comments
2
min read
LW
link
Don’t Judge a Tool by its Average Output
silentbob
Feb 2, 2023, 1:42 PM
12
points
2
comments
4
min read
LW
link
Epoch Impact Report 2022
Jsevillamol
Feb 2, 2023, 1:09 PM
16
points
0
comments
LW
link
You Don’t Exist, Duncan
Duncan Sabien (Inactive)
Feb 2, 2023, 8:37 AM
252
points
107
comments
9
min read
LW
link
Temporally Layered Architecture for Adaptive, Distributed and Continuous Control
Roman Leventov
Feb 2, 2023, 6:29 AM
6
points
4
comments
1
min read
LW
link
(arxiv.org)
Research agenda: Formalizing abstractions of computations
Erik Jenner
Feb 2, 2023, 4:29 AM
93
points
10
comments
31
min read
LW
link
Progress links and tweets, 2023-02-01
jasoncrawford
Feb 2, 2023, 2:25 AM
10
points
0
comments
1
min read
LW
link
(rootsofprogress.org)
Retrospective on the AI Safety Field Building Hub
Vael Gates
Feb 2, 2023, 2:06 AM
30
points
0
comments
LW
link
How to export Android Chrome tabs to an HTML file in Linux (as of February 2023)
Adam Scherlis
Feb 2, 2023, 2:03 AM
7
points
3
comments
2
min read
LW
link
(adam.scherlis.com)
Hacked Account Spam
jefftk
Feb 2, 2023, 1:50 AM
13
points
5
comments
1
min read
LW
link
(www.jefftk.com)
A simple technique to reduce negative rumination
cranberry_bear
Feb 2, 2023, 1:33 AM
9
points
0
comments
1
min read
LW
link
A Brief Overview of AI Safety/Alignment Orgs, Fields, Researchers, and Resources for ML Researchers
Austin Witte
Feb 2, 2023, 1:02 AM
18
points
1
comment
2
min read
LW
link
Interviews with 97 AI Researchers: Quantitative Analysis
Maheen Shermohammed
and
Vael Gates
Feb 2, 2023, 1:01 AM
23
points
0
comments
7
min read
LW
link
“AI Risk Discussions” website: Exploring interviews from 97 AI Researchers
Vael Gates
,
Lukas Trötzmüller
,
Maheen Shermohammed
,
michaelkeenan
and
zchuang
Feb 2, 2023, 1:00 AM
43
points
1
comment
LW
link
Predicting researcher interest in AI alignment
Vael Gates
Feb 2, 2023, 12:58 AM
25
points
0
comments
LW
link
Focus on the places where you feel shocked everyone’s dropping the ball
So8res
Feb 2, 2023, 12:27 AM
463
points
64
comments
4
min read
LW
link
3
reviews
Exercise is Good, Actually
Gordon Seidoh Worley
Feb 2, 2023, 12:09 AM
91
points
27
comments
3
min read
LW
link
Product safety is a poor model for AI governance
Richard Korzekwa
Feb 1, 2023, 10:40 PM
36
points
0
comments
5
min read
LW
link
(aiimpacts.org)
Hinton: “mortal” efficient analog hardware may be learned-in-place, uncopyable
the gears to ascension
Feb 1, 2023, 10:19 PM
12
points
3
comments
1
min read
LW
link
Can we “cure” cancer?
jasoncrawford
Feb 1, 2023, 10:03 PM
41
points
31
comments
2
min read
LW
link
(rootsofprogress.org)
Eli Lifland on Navigating the AI Alignment Landscape
ozziegooen
Feb 1, 2023, 9:17 PM
9
points
1
comment
31
min read
LW
link
(quri.substack.com)
Schizophrenia as a deficiency in long-range cortex-to-cortex communication
Steven Byrnes
Feb 1, 2023, 7:32 PM
35
points
38
comments
11
min read
LW
link
AI Safety Arguments: An Interactive Guide
Lukas Trötzmüller
Feb 1, 2023, 7:26 PM
20
points
0
comments
3
min read
LW
link
More findings on Memorization and double descent
Marius Hobbhahn
Feb 1, 2023, 6:26 PM
53
points
2
comments
19
min read
LW
link
Language Models can be Utility-Maximising Agents
Raymond D
Feb 1, 2023, 6:13 PM
22
points
1
comment
2
min read
LW
link
Trends in the dollar training cost of machine learning systems
Ben Cottier
Feb 1, 2023, 2:48 PM
23
points
0
comments
2
min read
LW
link
(epochai.org)
Polis: Why and How to Use it
brook
Feb 1, 2023, 2:03 PM
5
points
0
comments
LW
link
Subitisation of Self
vitaliya
Feb 1, 2023, 9:18 AM
14
points
4
comments
2
min read
LW
link
Directed Babbling
Yudhister Kumar
Feb 1, 2023, 9:10 AM
20
points
1
comment
3
min read
LW
link
(www.ykumar.org)
Voting Results for the 2021 Review
Raemon
Feb 1, 2023, 8:02 AM
66
points
10
comments
38
min read
LW
link
Abstraction As Symmetry and Other Thoughts
Numendil
Feb 1, 2023, 6:25 AM
28
points
9
comments
2
min read
LW
link
The effect of horizon length on scaling laws
Jacob_Hilton
1 Feb 2023 3:59 UTC
23
points
2
comments
1
min read
LW
link
(arxiv.org)
Contra Dance Lengths
jefftk
1 Feb 2023 3:30 UTC
9
points
0
comments
1
min read
LW
link
(www.jefftk.com)
Aiming for Convergence Is Like Discouraging Betting
Zack_M_Davis
1 Feb 2023 0:03 UTC
62
points
18
comments
11
min read
LW
link
1
review
On value in humans, other animals, and AI
Michele Campolo
31 Jan 2023 23:33 UTC
3
points
17
comments
5
min read
LW
link
Criticism of the main framework in AI alignment
Michele Campolo
31 Jan 2023 23:01 UTC
19
points
2
comments
6
min read
LW
link
Nice Clothes are Good, Actually
Gordon Seidoh Worley
31 Jan 2023 19:22 UTC
72
points
28
comments
4
min read
LW
link
[Linkpost] Human-narrated audio version of “Is Power-Seeking AI an Existential Risk?”
Joe Carlsmith
31 Jan 2023 19:21 UTC
12
points
1
comment
1
min read
LW
link
No Really, Attention is ALL You Need—Attention can do feedforward networks
Robert_AIZI
31 Jan 2023 18:48 UTC
29
points
7
comments
6
min read
LW
link
(aizi.substack.com)
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel