By its own moral lights, Clippy ought to stop presenting itself as a paperclip-maximizer.
Clippy can simultaneously present in one account as a paperclip maximiser, and in another as human.
The interplay between Clippy and a fake-human account could serve to create an environment more conducive to Clippy’s end-goal.
Or, of course, Clippy might be programmed to achieve vis aims solely through honest communication. Would be an interesting, but incomplete, safeguard on an AI.
Clippy can simultaneously present in one account as a paperclip maximiser, and in another as human.
(nods) I stand corrected… that is a far better solution from Clippy’s perspective, as it actually allows Clippy to experimentally determine which approach generates the most paperclips.
Or, of course, Clippy might be programmed to achieve vis aims solely through honest communication.
The question would then arise as to whether Clippy considers honest communication to be a paperclip-maximizing sort of thing to do, or if it’s more like akrasia—that is, a persistent cognitive distortion that leads Clippy to do things it considers non-paperclip-maximizing.
Clippy can simultaneously present in one account as a paperclip maximiser, and in another as human.
The interplay between Clippy and a fake-human account could serve to create an environment more conducive to Clippy’s end-goal.
Or, of course, Clippy might be programmed to achieve vis aims solely through honest communication. Would be an interesting, but incomplete, safeguard on an AI.
I struggle to understand the mentality that would put safeguards like that on an AI and then instruct it to maximize paperclips.
Well, let’s just be thankful they didn’t create the AI equivalent of a “Hello, world” program. That would be really annoying.
Well, it would have to be a paperclip manufacturer I suppose.
Either that or a very strange experiment.
Maybe Mythbusters?
(nods) I stand corrected… that is a far better solution from Clippy’s perspective, as it actually allows Clippy to experimentally determine which approach generates the most paperclips.
The question would then arise as to whether Clippy considers honest communication to be a paperclip-maximizing sort of thing to do, or if it’s more like akrasia—that is, a persistent cognitive distortion that leads Clippy to do things it considers non-paperclip-maximizing.