My understanding is that at present there’s a great deal of uncertainty concerning how future advanced technologies are going to develop (I’ve gotten an impression that e.g. Nick Bostrom and Josh Tenenbaum hold this view). In view of such uncertainty, it’s easy to imagine new data emerging over the next decades that makes it clear that pursuit of whole-brain emulation (or some currently unimagined strategy) is a far more effective strategy for existential risk reduction than Friendly AI research.
At present it looks to me like a positive singularity is substantially more likely to occur starting with whole-brain emulation than with Friendly AI.
Various people have suggested to me that initially pursuing Friendly AI might have higher expected value on the chance that it turns out to be easy. So I could imagine that it’s rational for you personally to focus your efforts on Friendly AI research (EDIT: even if I’m correct in my estimation in the above point). My remarks in the grandparent above were not intended as a criticism of your strategy.
I would be interested in hearing more about your own thinking about the relative feasibility of Friendly AI vs. stable whole-brain emulation and current arbitrage opportunities for existential risk reduction, whether on or off the record.
At present it looks to me like a positive singularity is substantially more likely to occur starting with whole-brain emulation than with Friendly AI.
That’s an interesting claim, and you should post your analysis of it (e.g. the evidence and reasoning that you use to form the estimate that a positive singularity is “substantially more likely” given WBE).
You may want to read this paper I presented at FHI. Note that there’s a big difference between the probability of risk conditional on WBE coming first or AI coming first and marginal impact of effort. In particular some of our uncertainty is about logical facts about the space of algorithms and technology landscape, and some of it is about the extent and effectiveness of activism/intervention.
At present it looks to me like a positive singularity is substantially more likely to occur starting with whole-brain emulation than with Friendly AI.
Do you mean that the role of ems is in developing FAI faster (as opposed to biological-human-built FAI), or are you thinking of something else? If ems merely speed time up, they don’t change the shape of FAI challenge much, unless (and to the extent that) we leverage them in a way we can’t for the human society to reduce existential risk before FAI is complete (but this can turn out worse as well, ems can well launch the first arbitrary-goal AGI).
but this can turn out worse as well, ems can well launch the first arbitrary-goal AGI
That’s the main thing that’s worried me about the possibility of ems coming first. But it depends on who is able to upload and who wants to, I suppose. If an average FAI researcher is more likely to upload, increase their speed, and possibly make copies of themselves than an average non-FAI AGI researcher, then it seems like that would be a reduction in risk.
I’m not sure whether that would be the case — a person working on FAI is likely to consider their work to be a matter of life and death, and would want all the speed increases they could get, but an AGI researcher may feel the same way about the threat to their career and status posed by the possibility of someone else getting to AGI first. And if uploading is very expensive at first, it’ll only be the most well-funded AGI researchers (i.e. not SIAI and friends) who will have access to it early on and will be likely to attempt it (if it provides enough of a speed increase that they’d consider it to be worth it).
(I originally thought that uploading would be of little to no help in increasing one’s own intelligence (in ways aside from thinking the same way but faster), since an emulation of a brain isn’t automatically any more comprehensible than an actual brain, but now I can see a few ways it could help — the equivalent of any kind of brain surgery could be attempted quickly, freely, and reversibly, and the same could be said for experimenting with nootropic-type effects within the emulation. So it’s possible that uploaded people would get somewhat smarter and not just faster. Of course, that’s only soft self-improvement, nowhere near the ability to systematically change one’s cognition at the algorithmic level, so I’m not worried about an upload bootstrapping itself to superintelligence (as some people apparently are). Which is good, since humans are not Friendly.)
There’s a lot to respond to here. Some quick points:
It should be born in mind that greatly increased speed and memory may by themselves strongly affect a thinking entity. I imagine that if I could think a million times as fast I would think a lot more carefully about my interactions with the outside world than I do now.
I don’t see any reason to think that SIAI will continue to be the only group thinking about safety considerations. If nothing else, SIAI or FHI can raise awareness of the dangers of AI within the community of AI researchers.
Assuming that brain uploads precede superhuman artificial intelligence, it would obviously be very desirable to have the right sort of human uploaded first.
I presently have a very dim view as to the prospects for modern day humans developing Friendly AI. This skepticism is the main reason why I think that pursuing whole-brain emulations first is more promising. See the comment by Carl that I mentioned in response to Vladimir Nesov’s question. Of course, my attitude on this point is subject to change with incoming evidence.
Sped-up ems have slower computers relative to their thinking speed. If Moore’s Law of Mad Science means that increasing computing power allows researchers to build AI with less understanding (and thus more risk of UFAI), then a speedup of researchers relative to computing speed makes it more likely that the first non-WBE AIs will be the result of a theory-intensive approach with high understanding. Anders Sandberg of FHI and I are working on a paper exploring some of these issues.
This argument lowers the estimate of danger, but AIs developed on relatively slow computers are not necessarily theory-intense, could also be coding-intense, which leads to UFAI. And theory-intense doesn’t necessarily imply adequate concern about AI’s preference.
it’s easy to imagine new data emerging over the next decades that makes it clear that pursuit of whole-brain emulation (or some currently unimagined strategy) is a far more effective strategy for existential risk reduction than Friendly AI research.
Not with the whole-brain emulation for me. That would represent an incredible miracle, IMO.
At present it looks to me like a positive singularity is substantially more likely to occur starting with whole-brain emulation than with Friendly AI.
The other way around, surely! Machine intelligence is much more likely to come first, and be significant. Few carried on with bird scanning and emulation once we had working planes. Essentially, the whole project went nowhere.
In general, when one individual asserts that something seems very likely to them it isn’t helpful to simply assert that the opposite seems extremely likely without giving some minimal reasoning for why you think that will be the case.
Your comment is a puzzling one. You are apparently advising me to offer more assistance. Right—but isn’t that up to me? You can’t possibly analyse the budgets I have to allocate for such things remotely.
Let me put it this way then: Most LW readers don’t like reading unproductive conversations. And it is hard to get more unproductive than one person saying “I believe in X!” and another saying “Yeah, well I believe ~X, so there!” You are welcome to do that, but don’t be surprised if the rest of us decide to vote down such comments as things we don’t want.
Good to hear from you :-)
My understanding is that at present there’s a great deal of uncertainty concerning how future advanced technologies are going to develop (I’ve gotten an impression that e.g. Nick Bostrom and Josh Tenenbaum hold this view). In view of such uncertainty, it’s easy to imagine new data emerging over the next decades that makes it clear that pursuit of whole-brain emulation (or some currently unimagined strategy) is a far more effective strategy for existential risk reduction than Friendly AI research.
At present it looks to me like a positive singularity is substantially more likely to occur starting with whole-brain emulation than with Friendly AI.
Various people have suggested to me that initially pursuing Friendly AI might have higher expected value on the chance that it turns out to be easy. So I could imagine that it’s rational for you personally to focus your efforts on Friendly AI research (EDIT: even if I’m correct in my estimation in the above point). My remarks in the grandparent above were not intended as a criticism of your strategy.
I would be interested in hearing more about your own thinking about the relative feasibility of Friendly AI vs. stable whole-brain emulation and current arbitrage opportunities for existential risk reduction, whether on or off the record.
That’s an interesting claim, and you should post your analysis of it (e.g. the evidence and reasoning that you use to form the estimate that a positive singularity is “substantially more likely” given WBE).
There’s a thread with some relevant points (both for and against) titled Hedging our Bets: The Case for Pursuing Whole Brain Emulation to Safeguard Humanity’s Future. I hadn’t looked at the comments until just now and still have to read them all; but see in particular a comment by Carl Shulman.
After reading all of the comments I’ll think about whether I have something to add beyond them and get back to you.
You may want to read this paper I presented at FHI. Note that there’s a big difference between the probability of risk conditional on WBE coming first or AI coming first and marginal impact of effort. In particular some of our uncertainty is about logical facts about the space of algorithms and technology landscape, and some of it is about the extent and effectiveness of activism/intervention.
Thanks for the very interesting reference! Is it linked on the SIAI research papers page? I didn’t see it there.
I appreciate this point which you’ve made to me previously (and which appears in your comment that I linked above!).
Do you mean that the role of ems is in developing FAI faster (as opposed to biological-human-built FAI), or are you thinking of something else? If ems merely speed time up, they don’t change the shape of FAI challenge much, unless (and to the extent that) we leverage them in a way we can’t for the human society to reduce existential risk before FAI is complete (but this can turn out worse as well, ems can well launch the first arbitrary-goal AGI).
That’s the main thing that’s worried me about the possibility of ems coming first. But it depends on who is able to upload and who wants to, I suppose. If an average FAI researcher is more likely to upload, increase their speed, and possibly make copies of themselves than an average non-FAI AGI researcher, then it seems like that would be a reduction in risk.
I’m not sure whether that would be the case — a person working on FAI is likely to consider their work to be a matter of life and death, and would want all the speed increases they could get, but an AGI researcher may feel the same way about the threat to their career and status posed by the possibility of someone else getting to AGI first. And if uploading is very expensive at first, it’ll only be the most well-funded AGI researchers (i.e. not SIAI and friends) who will have access to it early on and will be likely to attempt it (if it provides enough of a speed increase that they’d consider it to be worth it).
(I originally thought that uploading would be of little to no help in increasing one’s own intelligence (in ways aside from thinking the same way but faster), since an emulation of a brain isn’t automatically any more comprehensible than an actual brain, but now I can see a few ways it could help — the equivalent of any kind of brain surgery could be attempted quickly, freely, and reversibly, and the same could be said for experimenting with nootropic-type effects within the emulation. So it’s possible that uploaded people would get somewhat smarter and not just faster. Of course, that’s only soft self-improvement, nowhere near the ability to systematically change one’s cognition at the algorithmic level, so I’m not worried about an upload bootstrapping itself to superintelligence (as some people apparently are). Which is good, since humans are not Friendly.)
There’s a lot to respond to here. Some quick points:
It should be born in mind that greatly increased speed and memory may by themselves strongly affect a thinking entity. I imagine that if I could think a million times as fast I would think a lot more carefully about my interactions with the outside world than I do now.
I don’t see any reason to think that SIAI will continue to be the only group thinking about safety considerations. If nothing else, SIAI or FHI can raise awareness of the dangers of AI within the community of AI researchers.
Assuming that brain uploads precede superhuman artificial intelligence, it would obviously be very desirable to have the right sort of human uploaded first.
I presently have a very dim view as to the prospects for modern day humans developing Friendly AI. This skepticism is the main reason why I think that pursuing whole-brain emulations first is more promising. See the comment by Carl that I mentioned in response to Vladimir Nesov’s question. Of course, my attitude on this point is subject to change with incoming evidence.
Sped-up ems have slower computers relative to their thinking speed. If Moore’s Law of Mad Science means that increasing computing power allows researchers to build AI with less understanding (and thus more risk of UFAI), then a speedup of researchers relative to computing speed makes it more likely that the first non-WBE AIs will be the result of a theory-intensive approach with high understanding. Anders Sandberg of FHI and I are working on a paper exploring some of these issues.
This argument lowers the estimate of danger, but AIs developed on relatively slow computers are not necessarily theory-intense, could also be coding-intense, which leads to UFAI. And theory-intense doesn’t necessarily imply adequate concern about AI’s preference.
My idea here is the same as the one that Carl Shulman mentioned in a response to one of your comments from nine months ago.
Not with the whole-brain emulation for me. That would represent an incredible miracle, IMO.
The other way around, surely! Machine intelligence is much more likely to come first, and be significant. Few carried on with bird scanning and emulation once we had working planes. Essentially, the whole project went nowhere.
In general, when one individual asserts that something seems very likely to them it isn’t helpful to simply assert that the opposite seems extremely likely without giving some minimal reasoning for why you think that will be the case.
For more, see: http://alife.co.uk/essays/against_whole_brain_emulation/
Your comment is a puzzling one. You are apparently advising me to offer more assistance. Right—but isn’t that up to me? You can’t possibly analyse the budgets I have to allocate for such things remotely.
Let me put it this way then: Most LW readers don’t like reading unproductive conversations. And it is hard to get more unproductive than one person saying “I believe in X!” and another saying “Yeah, well I believe ~X, so there!” You are welcome to do that, but don’t be surprised if the rest of us decide to vote down such comments as things we don’t want.