I think the problem with the things you mention is that they are just super vague, where you don’t even know what is the thing that you are talking about. What does it mean that:
Most important of all, perhaps, is making such machines learn from their own experience.
Finally, we’ll get machines that think about themselves and make up theories, good or bad, of how they, themselves might work.
Also, all of this seems to be some sort of vague stuff about imagining how AI systems could be. I’m actually interested in just making the AI systems and making them in a very specific way such that they have good alignment properties and not vaguely philosophizing about what could happen. The whole point of writing down algorithms explicitly, which is one non-dumb way to build AGI, is that you can just see what’s going on in the algorithm and understand it and design the algorithm in such a way that it would think in a very particular way.
So it’s not like, oh yes, these machines will think for themselves some stuff and it will be good or bad, it’s more like, I make these machines think, how do I make them think, what’s the actual algorithm to make them think, how can I make this algorithm such that it will actually be aligned. And I am controlling what they are thinking, I am controlling if it’s good or bad, I am controlling if they are going to build a model of themselves, maybe that’s dangerous for alignment purposes in some context and then I would want the algorithm to not want the system to build a model of themselves.
For, at that point, they’ll probably object to being called machines.
I think it’s pretty accurate to say that I am a machine.
(Also, as a meta note, it would be very good, I think, if you do not break the lines as you did in this big text block because that’s pretty annoying to block quote.)
I think the problem with the things you mention is that they are just super vague, where you don’t even know what is the thing that you are talking about. What does it mean that:
Also, all of this seems to be some sort of vague stuff about imagining how AI systems could be. I’m actually interested in just making the AI systems and making them in a very specific way such that they have good alignment properties and not vaguely philosophizing about what could happen. The whole point of writing down algorithms explicitly, which is one non-dumb way to build AGI, is that you can just see what’s going on in the algorithm and understand it and design the algorithm in such a way that it would think in a very particular way.
So it’s not like, oh yes, these machines will think for themselves some stuff and it will be good or bad, it’s more like, I make these machines think, how do I make them think, what’s the actual algorithm to make them think, how can I make this algorithm such that it will actually be aligned. And I am controlling what they are thinking, I am controlling if it’s good or bad, I am controlling if they are going to build a model of themselves, maybe that’s dangerous for alignment purposes in some context and then I would want the algorithm to not want the system to build a model of themselves.
(Also, as a meta note, it would be very good, I think, if you do not break the lines as you did in this big text block because that’s pretty annoying to block quote.)