Yes, please try to clarify. In particular, I don’t understand your “|” notation (as in “S|Output”).
I realized that I was a bit confused in what I said earlier. I think it’s clear that (proposed) SHF schemes should be able to do at least as well as a human, given the same amount of time, because they have human “on top” (as “CEO”) who can merely ignore all the AI helpers(/underlings).
But now I can also see an argument for why SHF couldn’t do ALD, if it doesn’t have arbitrarily long to deliberate: there would need to be some parallelism/decomposition in SHF, and that might not work well/perfectly for all problems.
OK, so it sounds like your argument why SHF can’t do ALD is (a specific, technical version of) the same argument that I mentioned in my last response. Can you confirm?
Aha, OK. So I either misunderstand or disagree with that.
I think SHF (at least most examples) have the human as “CEO” with AIs as “advisers”, and thus the human can chose to ignore all of the advice and make the decision unaided.
Yes, please try to clarify. In particular, I don’t understand your “|” notation (as in “S|Output”).
I realized that I was a bit confused in what I said earlier. I think it’s clear that (proposed) SHF schemes should be able to do at least as well as a human, given the same amount of time, because they have human “on top” (as “CEO”) who can merely ignore all the AI helpers(/underlings).
But now I can also see an argument for why SHF couldn’t do ALD, if it doesn’t have arbitrarily long to deliberate: there would need to be some parallelism/decomposition in SHF, and that might not work well/perfectly for all problems.
“|” meant concatenation, so “S|Output := H(S|Input)” means you set S to the first half of H(S|Input), and Output to the second half of H(S|Input).
OK, so it sounds like your argument why SHF can’t do ALD is (a specific, technical version of) the same argument that I mentioned in my last response. Can you confirm?
I’m not sure. It seems like my argument applies even if SHF did have arbitrarily long to deliberate?
Aha, OK. So I either misunderstand or disagree with that.
I think SHF (at least most examples) have the human as “CEO” with AIs as “advisers”, and thus the human can chose to ignore all of the advice and make the decision unaided.