Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
Page
2
A Brief Overview of AI Safety/Alignment Orgs, Fields, Researchers, and Resources for ML Researchers
Austin Witte
Feb 2, 2023, 1:02 AM
18
points
1
comment
2
min read
LW
link
Interviews with 97 AI Researchers: Quantitative Analysis
Maheen Shermohammed
and
Vael Gates
Feb 2, 2023, 1:01 AM
23
points
0
comments
7
min read
LW
link
“AI Risk Discussions” website: Exploring interviews from 97 AI Researchers
Vael Gates
,
Lukas Trötzmüller
,
Maheen Shermohammed
,
michaelkeenan
and
zchuang
Feb 2, 2023, 1:00 AM
43
points
1
comment
LW
link
Predicting researcher interest in AI alignment
Vael Gates
Feb 2, 2023, 12:58 AM
25
points
0
comments
LW
link
Focus on the places where you feel shocked everyone’s dropping the ball
So8res
Feb 2, 2023, 12:27 AM
466
points
64
comments
4
min read
LW
link
3
reviews
Exercise is Good, Actually
Gordon Seidoh Worley
Feb 2, 2023, 12:09 AM
91
points
27
comments
3
min read
LW
link
Product safety is a poor model for AI governance
Richard Korzekwa
Feb 1, 2023, 10:40 PM
36
points
0
comments
5
min read
LW
link
(aiimpacts.org)
Hinton: “mortal” efficient analog hardware may be learned-in-place, uncopyable
the gears to ascension
Feb 1, 2023, 10:19 PM
12
points
3
comments
1
min read
LW
link
Can we “cure” cancer?
jasoncrawford
Feb 1, 2023, 10:03 PM
41
points
31
comments
2
min read
LW
link
(rootsofprogress.org)
Eli Lifland on Navigating the AI Alignment Landscape
ozziegooen
Feb 1, 2023, 9:17 PM
9
points
1
comment
31
min read
LW
link
(quri.substack.com)
Schizophrenia as a deficiency in long-range cortex-to-cortex communication
Steven Byrnes
Feb 1, 2023, 7:32 PM
35
points
38
comments
11
min read
LW
link
AI Safety Arguments: An Interactive Guide
Lukas Trötzmüller
Feb 1, 2023, 7:26 PM
20
points
0
comments
3
min read
LW
link
More findings on Memorization and double descent
Marius Hobbhahn
Feb 1, 2023, 6:26 PM
53
points
2
comments
19
min read
LW
link
Language Models can be Utility-Maximising Agents
Raymond Douglas
Feb 1, 2023, 6:13 PM
22
points
1
comment
2
min read
LW
link
Trends in the dollar training cost of machine learning systems
Ben Cottier
Feb 1, 2023, 2:48 PM
23
points
0
comments
2
min read
LW
link
(epochai.org)
Polis: Why and How to Use it
brook
Feb 1, 2023, 2:03 PM
5
points
0
comments
LW
link
Subitisation of Self
vitaliya
Feb 1, 2023, 9:18 AM
14
points
4
comments
2
min read
LW
link
Directed Babbling
Yudhister Kumar
Feb 1, 2023, 9:10 AM
20
points
1
comment
3
min read
LW
link
(www.ykumar.org)
Voting Results for the 2021 Review
Raemon
Feb 1, 2023, 8:02 AM
66
points
10
comments
38
min read
LW
link
Abstraction As Symmetry and Other Thoughts
Numendil
Feb 1, 2023, 6:25 AM
28
points
9
comments
2
min read
LW
link
The effect of horizon length on scaling laws
Jacob_Hilton
Feb 1, 2023, 3:59 AM
23
points
2
comments
1
min read
LW
link
(arxiv.org)
Contra Dance Lengths
jefftk
Feb 1, 2023, 3:30 AM
9
points
0
comments
1
min read
LW
link
(www.jefftk.com)
Aiming for Convergence Is Like Discouraging Betting
Zack_M_Davis
Feb 1, 2023, 12:03 AM
62
points
18
comments
11
min read
LW
link
1
review
On value in humans, other animals, and AI
Michele Campolo
Jan 31, 2023, 11:33 PM
3
points
17
comments
5
min read
LW
link
Criticism of the main framework in AI alignment
Michele Campolo
Jan 31, 2023, 11:01 PM
19
points
2
comments
6
min read
LW
link
Nice Clothes are Good, Actually
Gordon Seidoh Worley
Jan 31, 2023, 7:22 PM
72
points
28
comments
4
min read
LW
link
[Linkpost] Human-narrated audio version of “Is Power-Seeking AI an Existential Risk?”
Joe Carlsmith
Jan 31, 2023, 7:21 PM
12
points
1
comment
1
min read
LW
link
No Really, Attention is ALL You Need—Attention can do feedforward networks
Robert_AIZI
Jan 31, 2023, 6:48 PM
29
points
7
comments
6
min read
LW
link
(aizi.substack.com)
Talk to me about your summer/career plans
Orpheus16
Jan 31, 2023, 6:29 PM
31
points
3
comments
2
min read
LW
link
Mechanistic Interpretability Quickstart Guide
Neel Nanda
Jan 31, 2023, 4:35 PM
42
points
3
comments
6
min read
LW
link
(www.neelnanda.io)
New Hackathon: Robustness to distribution changes and ambiguity
Charbel-Raphaël
Jan 31, 2023, 12:50 PM
12
points
3
comments
1
min read
LW
link
Squiggle: Why and how to use it
brook
Jan 31, 2023, 12:37 PM
3
points
0
comments
LW
link
Beware of Fake Alternatives
silentbob
Jan 31, 2023, 10:21 AM
57
points
11
comments
4
min read
LW
link
1
review
Inner Misalignment in “Simulator” LLMs
Adam Scherlis
Jan 31, 2023, 8:33 AM
84
points
12
comments
4
min read
LW
link
Why AI experts’ jobs are always decades from being automated
Allen Hoskins
Jan 31, 2023, 3:01 AM
0
points
1
comment
5
min read
LW
link
(open.substack.com)
Apply to HAIST/MAIA’s AI Governance Workshop in DC (Feb 17-20)
Phosphorous
,
Xander Davies
,
CMD
,
Paramedic
and
tlevin
Jan 31, 2023, 2:06 AM
28
points
0
comments
2
min read
LW
link
EA & LW Forum Weekly Summary (23rd − 29th Jan ’23)
Zoe Williams
Jan 31, 2023, 12:36 AM
12
points
0
comments
LW
link
Saying things because they sound good
Adam Zerner
Jan 31, 2023, 12:17 AM
23
points
6
comments
2
min read
LW
link
South Bay Meetup
DavidFriedman
Jan 30, 2023, 11:35 PM
2
points
0
comments
1
min read
LW
link
Peter Thiel’s speech at Oxford Debating Union on technological stagnation, Nuclear weapons, COVID, Environment, Alignment, ‘anti-anti anti-anti-classical liberalism’, Bostrom, LW, etc.
M. Y. Zuo
Jan 30, 2023, 11:31 PM
8
points
33
comments
1
min read
LW
link
Medical Image Registration: The obscure field where Deep Mesaoptimizers are already at the top of the benchmarks. (post + colab notebook)
Hastings
Jan 30, 2023, 10:46 PM
35
points
1
comment
3
min read
LW
link
Humans Can Be Manually Strategic
Screwtape
Jan 30, 2023, 10:35 PM
13
points
0
comments
3
min read
LW
link
Why I hate the “accident vs. misuse” AI x-risk dichotomy (quick thoughts on “structural risk”)
David Scott Krueger (formerly: capybaralet)
30 Jan 2023 18:50 UTC
34
points
41
comments
2
min read
LW
link
2022 Unofficial LessWrong General Census
Screwtape
30 Jan 2023 18:36 UTC
97
points
33
comments
2
min read
LW
link
Call for submissions: “(In)human Values and Artificial Agency”, ALIFE 2023
the gears to ascension
30 Jan 2023 17:37 UTC
29
points
4
comments
1
min read
LW
link
(humanvaluesandartificialagency.com)
What I mean by “alignment is in large part about making cognition aimable at all”
So8res
30 Jan 2023 15:22 UTC
171
points
25
comments
2
min read
LW
link
The Energy Requirements and Feasibility of Off-World Mining
clans
30 Jan 2023 15:07 UTC
31
points
1
comment
8
min read
LW
link
(locationtbd.home.blog)
Whatever their arguments, Covid vaccine sceptics will probably never convince me
contrarianbrit
30 Jan 2023 13:42 UTC
8
points
10
comments
3
min read
LW
link
(thomasprosser.substack.com)
Simulacra Levels Summary
Zvi
30 Jan 2023 13:40 UTC
77
points
14
comments
7
min read
LW
link
(thezvi.wordpress.com)
A Few Principles of Successful AI Design
Vestozia
30 Jan 2023 10:42 UTC
1
point
0
comments
8
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel