Archive
Sequences
About
Search
Log In
Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
RSS
catubc
Karma:
85
All
Posts
Comments
New
Top
Old
Intent-aligned AI systems deplete human agency: the need for agency foundations research in AI safety
catubc
31 May 2023 21:18 UTC
24
points
4
comments
11
min read
LW
link
Red-teaming AI-safety concepts that rely on science metaphors
catubc
16 Mar 2023 6:52 UTC
5
points
4
comments
5
min read
LW
link
AGIs may value intrinsic rewards more than extrinsic ones
catubc
17 Nov 2022 21:49 UTC
8
points
6
comments
4
min read
LW
link
LLMs may capture key components of human agency
catubc
17 Nov 2022 20:14 UTC
27
points
0
comments
4
min read
LW
link
Agency engineering: is AI-alignment “to human intent” enough?
catubc
2 Sep 2022 18:14 UTC
9
points
10
comments
6
min read
LW
link
Back to top