Are smart contracts AI-complete?
Many people are probably aware of the hack at DAO, using a bug in their smart contract system to steal millions of dollars worth of the crypto currency Ethereum.
There’s various arguments as to whether this theft was technically allowed or not, and what should be done about it, and so on. Many people are arguing that the code is the contract, and that therefore no-one should be allowed to interfere with it—DAO just made a coding mistake, and are now being (deservedly?) punished for it.
That got me wondering whether its ever possible to make a smart contract without a full AI of some sort. For instance, if the contract is triggered by the delivery of physical goods—how can you define what the goods are, what constitutes delivery, what constitutes possession of them, and so on. You could have a human confirm delivery—but that’s precisely the kind of judgement call you want to avoid. You could have an automated delivery confirmation system—but what happens if someone hacks or triggers that? You could connect it automatically with scanning headlines of media reports, but again, this is relying on aggregated human judgement, which could be hacked or influenced.
Digital goods seem more secure, as you can automate confirmation of delivery/services rendered, and so on. But, again, this leaves the confirmation process open to hacking. Which would be illegal, if you’re going to profit from the hack. Hum...
This seems the most promising avenue for smart contracts that doesn’t involve full AI: clear out the bugs in the code, then ground the confirmation procedure in such a way that it can only be hacked in a way that’s already illegal. Sort of use the standard legal system as a backstop, fixing the basic assumptions, and then setting up the smart contracts on top of them (which is not the same as using the standard legal system within the contract).