I’m not making any claims about feasibility, I only dispute the claim that it’s known that permanently giving up the potential for human control is an acceptable thing to do, or that making such a call (epistemic call about what is known) is reasonable in the foreseeable future. To the extent it’s possible to defer this call, it should therefore be deferred (this is a normative claim, not a plan or a prediction of feasibility). If it’s not possible to keep the potential for human control despite this uncertainty, then it’s not possible, but that won’t be because the uncertainty got resolved to the extent that it could be humanly resolved.
I’m not making any claims about feasibility, I only dispute the claim that it’s known that permanently giving up the potential for human control is an acceptable thing to do, or that making such a call (epistemic call about what is known) is reasonable in the foreseeable future. To the extent it’s possible to defer this call, it should therefore be deferred (this is a normative claim, not a plan or a prediction of feasibility). If it’s not possible to keep the potential for human control despite this uncertainty, then it’s not possible, but that won’t be because the uncertainty got resolved to the extent that it could be humanly resolved.