Thanks. To go back to my original point a bit, how useful is it to debate philosophers about this? (When debating AI researchers, given that they probably have a limited appetite for reading papers arguing that what they’re doing is dangerous, it seems like it would be better to skip this paper and give the practical arguments directly.)
Maybe I’ve spent too much time around philosophers—but there are some AI designers who seem to spout weak arguments like that, and this paper can’t hurt. When we get a round to writing a proper justification for AI researchers, having this paper to refer back to avoids going over the same points again.
Plus, it’s a lot easier to write this paper first, and was good practice.
Hum, good point; I’ll try and put in some disclaimer, emphasising that this is a partial result...
Thanks. To go back to my original point a bit, how useful is it to debate philosophers about this? (When debating AI researchers, given that they probably have a limited appetite for reading papers arguing that what they’re doing is dangerous, it seems like it would be better to skip this paper and give the practical arguments directly.)
Maybe I’ve spent too much time around philosophers—but there are some AI designers who seem to spout weak arguments like that, and this paper can’t hurt. When we get a round to writing a proper justification for AI researchers, having this paper to refer back to avoids going over the same points again.
Plus, it’s a lot easier to write this paper first, and was good practice.