Something I’d like WBE researchers to keep in mind: It seems like, by default, the cortex is the easiest part to get a functionally working quasi-emulation of, because it’s relatively uniform (and because it’s relatively easier to tell whether problem solving works compared to whether you’re feeling angry at the right times). But if you get a quasi-cortex working and not all the other stuff, this actually does seem like an alignment issue. One of the main arguments for alignment of uploads would be “it has all the stuff that humans have that produces stuff like caring, love, wisdom, reflection”. But if you delete a bunch of stuff including presumably much of the steering systems, this argument would seem to go right out the window.
Something I’d like WBE researchers to keep in mind: It seems like, by default, the cortex is the easiest part to get a functionally working quasi-emulation of, because it’s relatively uniform (and because it’s relatively easier to tell whether problem solving works compared to whether you’re feeling angry at the right times). But if you get a quasi-cortex working and not all the other stuff, this actually does seem like an alignment issue. One of the main arguments for alignment of uploads would be “it has all the stuff that humans have that produces stuff like caring, love, wisdom, reflection”. But if you delete a bunch of stuff including presumably much of the steering systems, this argument would seem to go right out the window.
https://www.lesswrong.com/posts/jTiSWHKAtnyA723LE/overview-of-strong-human-intelligence-amplification-methods#Brain_emulation