The issue isn’t the AI/s, it’s the user. Ignoring issues like ‘where does this aligned AI come from, and how does this happen as a result of such negotiation’*, how is compliance proved? Seems like it’d work if there was a simple protocol, which can be shown, or the AI/s design a better tax code.
*The AI/s are all negotiating with each other. Might be risky if they’re not ‘aligned’.
**Whether or not it is useful to model them as one system, or multiple isn’t clear here. Also, some of these assistants are going to have similar code, if that world is similar to this one.
Seems like such restriction isn’t needed:
The AI/s** can provide it’s/their source code.
The issue isn’t the AI/s, it’s the user. Ignoring issues like ‘where does this aligned AI come from, and how does this happen as a result of such negotiation’*, how is compliance proved? Seems like it’d work if there was a simple protocol, which can be shown, or the AI/s design a better tax code.
*The AI/s are all negotiating with each other. Might be risky if they’re not ‘aligned’.
**Whether or not it is useful to model them as one system, or multiple isn’t clear here. Also, some of these assistants are going to have similar code, if that world is similar to this one.