Archive
Sequences
About
Search
Log In
Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
RSS
New
Hot
Active
Old
Page
1
AI Safety proposal—Influencing the superintelligence explosion
Morgan
22 May 2024 23:31 UTC
0
points
1
comment
7
min read
LW
link
The Button (Short Comic)
milanrosko
22 May 2024 23:28 UTC
3
points
0
comments
1
min read
LW
link
Implementing Asimov’s Laws of Robotics—How I imagine alignment working.
Joshua Clancy
22 May 2024 23:15 UTC
2
points
0
comments
11
min read
LW
link
Higher-Order Forecasts
ozziegooen
22 May 2024 21:49 UTC
30
points
0
comments
1
min read
LW
link
A Positive Double Standard—Self-Help Principles Work For Individuals Not Populations
James Stephen Brown
22 May 2024 21:37 UTC
2
points
2
comments
5
min read
LW
link
A Bi-Modal Brain Model
Johannes C. Mayer
22 May 2024 20:10 UTC
9
points
1
comment
2
min read
LW
link
[Question]
Should we be concerned about eating too much soy?
ChristianKl
22 May 2024 12:53 UTC
20
points
2
comments
1
min read
LW
link
Procedural Executive Function, Part 3
DaystarEld
22 May 2024 11:58 UTC
15
points
2
comments
1
min read
LW
link
Cicadas, Anthropic, and the bilateral alignment problem
kromem
22 May 2024 11:09 UTC
17
points
0
comments
5
min read
LW
link
“Which chains-of-thought was that faster than?”
Emrik
22 May 2024 8:21 UTC
31
points
1
comment
4
min read
LW
link
ARIA’s Safeguarded AI grant program is accepting applications for Technical Area 1.1 until May 28th
Brendon_Wong
22 May 2024 6:54 UTC
10
points
0
comments
1
min read
LW
link
(www.aria.org.uk)
Anthropic announces interpretability advances. How much does this advance alignment?
Seth Herd
21 May 2024 22:30 UTC
48
points
4
comments
3
min read
LW
link
(www.anthropic.com)
EIS XIII: Reflections on Anthropic’s SAE Research Circa May 2024
scasper
21 May 2024 20:15 UTC
114
points
9
comments
3
min read
LW
link
Mitigating extreme AI risks amid rapid progress [Linkpost]
Akash
21 May 2024 19:59 UTC
18
points
5
comments
4
min read
LW
link
Helping loved ones with their finances: the why and how of an unusually impactful opportunity
Sam Anschell
21 May 2024 18:48 UTC
0
points
1
comment
1
min read
LW
link
(forum.effectivealtruism.org)
rough draft on what happens in the brain when you have an insight
Emrik
21 May 2024 18:02 UTC
9
points
2
comments
1
min read
LW
link
On Dwarkesh’s Podcast with OpenAI’s John Schulman
Zvi
21 May 2024 17:30 UTC
65
points
3
comments
20
min read
LW
link
(thezvi.wordpress.com)
[Question]
Is deleting capabilities still a relevant research question?
tailcalled
21 May 2024 13:24 UTC
15
points
1
comment
1
min read
LW
link
My Dating Heuristic
Declan Molony
21 May 2024 5:28 UTC
13
points
4
comments
2
min read
LW
link
Scorable Functions: A Format for Algorithmic Forecasting
ozziegooen
21 May 2024 4:14 UTC
26
points
0
comments
1
min read
LW
link
Back to top
Next