Archive
Sequences
About
Search
Log In
Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
RSS
Chris_Leong
Karma:
7,605
All
Posts
Comments
New
Top
Old
Page
2
Reverse Regulatory Capture
Chris_Leong
Apr 11, 2024, 2:40 AM
12
points
3
comments
1
min read
LW
link
On the Confusion between Inner and Outer Misalignment
Chris_Leong
Mar 25, 2024, 11:59 AM
17
points
10
comments
1
min read
LW
link
The Best Essay (Paul Graham)
Chris_Leong
Mar 11, 2024, 7:25 PM
25
points
2
comments
1
min read
LW
link
(paulgraham.com)
[Question]
Can we get an AI to “do our alignment homework for us”?
Chris_Leong
Feb 26, 2024, 7:56 AM
53
points
33
comments
1
min read
LW
link
[Question]
What’s the theory of impact for activation vectors?
Chris_Leong
Feb 11, 2024, 7:34 AM
61
points
12
comments
1
min read
LW
link
Notice When People Are Directionally Correct
Chris_Leong
Jan 14, 2024, 2:12 PM
136
points
8
comments
2
min read
LW
link
Are Metaculus AI Timelines Inconsistent?
Chris_Leong
Jan 2, 2024, 6:47 AM
17
points
7
comments
2
min read
LW
link
Random Musings on Theory of Impact for Activation Vectors
Chris_Leong
Dec 7, 2023, 1:07 PM
8
points
0
comments
1
min read
LW
link
Goodhart’s Law Example: Training Verifiers to Solve Math Word Problems
Chris_Leong
Nov 25, 2023, 12:53 AM
27
points
2
comments
1
min read
LW
link
(arxiv.org)
Upcoming Feedback Opportunity on Dual-Use Foundation Models
Chris_Leong
Nov 2, 2023, 4:28 AM
3
points
0
comments
1
min read
LW
link
On Having No Clue
Chris_Leong
Nov 1, 2023, 1:36 AM
20
points
11
comments
1
min read
LW
link
Is Yann LeCun strawmanning AI x-risks?
Chris_Leong
Oct 19, 2023, 11:35 AM
26
points
4
comments
1
min read
LW
link
Don’t Dismiss Simple Alignment Approaches
Chris_Leong
Oct 7, 2023, 12:35 AM
137
points
9
comments
4
min read
LW
link
[Question]
What evidence is there of LLM’s containing world models?
Chris_Leong
Oct 4, 2023, 2:33 PM
17
points
17
comments
1
min read
LW
link
The Role of Groups in the Progression of Human Understanding
Chris_Leong
Sep 27, 2023, 3:09 PM
11
points
0
comments
2
min read
LW
link
The Flow-Through Fallacy
Chris_Leong
Sep 13, 2023, 4:28 AM
21
points
7
comments
1
min read
LW
link
Chariots of Philosophical Fire
Chris_Leong
Aug 26, 2023, 12:52 AM
12
points
0
comments
1
min read
LW
link
(l.facebook.com)
Call for Papers on Global AI Governance from the UN
Chris_Leong
Aug 20, 2023, 8:56 AM
19
points
0
comments
LW
link
(www.linkedin.com)
Yann LeCun on AGI and AI Safety
Chris_Leong
Aug 6, 2023, 9:56 PM
37
points
13
comments
1
min read
LW
link
(drive.google.com)
A Naive Proposal for Constructing Interpretable AI
Chris_Leong
Aug 5, 2023, 10:32 AM
18
points
6
comments
2
min read
LW
link
Previous
Back to top
Next