Theoretical AI alignment (and relevant upskilling) in my free time. My current view of the field is here (part 1) and here (part 2).
NicholasKross
Karma: 1,950
[SEE NEW EDITS] No, *You* Need to Write Clearer
A Quick List of Some Problems in AI Alignment As A Field
Quick Thoughts on A.I. Governance
Alignment Megaprojects: You’re Not Even Trying to Have Ideas
[Question] Intelligence Enhancement (Monthly Thread) 13 Oct 2023
Dreams of “Mathopedia”
Why I’m Not (Yet) A Full-Time Technical Alignment Researcher
Learning Math in Time for Alignment
There Meat Come A Scandal...
This is pretty good, although you ought to actually link more of the citations for specific facts. Court documents, testimonies, company docs, even Wikipedia would’ve been helpful for the part about IAG.
[Question] Incentives affecting alignment-researcher encouragement
[Question] Ways to increase working memory, and/or cope with low working memory?
[Question] Ways to be more agenty?
[Question] Burst work or steady work?
I am a Memoryless System
Don’t Look Up (Film Review)
My AI Alignment Research Agenda and Threat Model, right now (May 2023)
(Semi-dumb LW category suggestion: Posts That Could Have Made You Good Money In Hindsight)
Extremely strong upvote for Oliver’s 2nd message.
Also, not as related: kudos for actually materially changing the course of your organization, something which is hard for most organizations, period.