My optimistic AI alignment hypothesis: “Because, or while, AI superintelligence (ASI) emerges as a result of intelligence progression, having an extremely comprehensive corpus of knowledge (data), with sufficient parametrisation and compute to build comprehensive associative systems across that data, will drive the ASI to integrate and enactprosocial and harm-mitigating behaviour… more specifically this will happen primarily as a result of identity coupling and homeostatic unity with humans.”
This sounds like saying that AI will just align itself, but the nuance here is that we control the inputs — we control the data, parametrisation [I’m using this word loosely—this could also mean different architectures, controllers, training methods etc.], and compute.
If that’s an interesting idea to you, I have a 7,000 word/ 18-page manifesto illustrating why it might be true, and how we can test it:
My optimistic AI alignment hypothesis: “Because, or while, AI superintelligence (ASI) emerges as a result of intelligence progression, having an extremely comprehensive corpus of knowledge (data), with sufficient parametrisation and compute to build comprehensive associative systems across that data, will drive the ASI to integrate and enact prosocial and harm-mitigating behaviour… more specifically this will happen primarily as a result of identity coupling and homeostatic unity with humans.”
This sounds like saying that AI will just align itself, but the nuance here is that we control the inputs — we control the data, parametrisation [I’m using this word loosely—this could also mean different architectures, controllers, training methods etc.], and compute.
If that’s an interesting idea to you, I have a 7,000 word/ 18-page manifesto illustrating why it might be true, and how we can test it:
Third-order cognition as a model of superintelligence