I’m assuming that it can function at high level despite/because of its biases. And the problem is not that it might not work effectively, but that our job of ensuring it behaves well just got harder, because we just got worse at predicting its decisions.
Does it matter if we aren’t able to recognize it’s biases? Humans are able to function with biases.
We are also are able to recognize and correct for their own biases. And we can’t even look at, let alone rewrite, our own source code.
I’m assuming that it can function at high level despite/because of its biases. And the problem is not that it might not work effectively, but that our job of ensuring it behaves well just got harder, because we just got worse at predicting its decisions.