AI demands unprecedented reliability

My AI pessimism is not addressed by optimistic arguments that do not promise the level of reliability required for a technology that our societies vital functions are expected to rely on.

Historically, useful improvements are hard to reverse and solutions can suddenly fail after we had become dependent on them. The potato seemed perfect for Ireland until they all blighted. Dutch dams seemed good, until they broke. Banks seem to keep suddenly failing on the regular.

AI is useful and we rely on it. Our culture is under heavy influence of search engines and algorithmic recommendations, people entrust their lives to self-driving cars and companies entrust sensitive or copy-righted data to generative models.

AI relies on AI. Some hypothetical ChatGPT call queried Bing which returned a page written by some human who, inspired by a Facebook recommendation, made Bart summarize the report of Covid prediction model that was co-written with Copilot. This low-stakes example fails all the time, but the state of affairs is still useful enough for people to build on top of and thereby increasingly rely upon.

The tool-AI future looks like one where man and machine rely on various interdependent models with medical diagnoses, mental help, bureaucracy, military protection, news, matchmaking, electricity supply, medicine production, weather forecasting, forecasting and AI development. The agentic-AI future has us rely on AI to build all of these itself.

We should demand a level of reliability that is stricter than any level of reliability our past technologies have achieved because our reliance on AI will be unprecedentedly large.

Addendum: unpersuasive arguments

I can only guess at what requirements for AI are good enough. Having humanity at AI’s mercy and surviving is good enough, but not a situation we’d want to get into without knowing in advance that that is what will happen.
In the meantime, here is a list of argument types that should be insufficient for lowering anyone’s P(Doom):

Counterarguments to failure modes
Unless your technique counteracts all possible (not only all named) failure modes, there is likely a failure mode we both missed.

Optimistic arguments that work in most situations
We are going to see all the situations. If your technique mostly works, it will surely fail.

Arguments against the reliability of humans
Humans are reliable because we live at humanity’s mercy and are surviving. I don’t know why this is the case, and if I were a dumb alien, I should not conjure a human into existence.

Evidence of prosaic reliability
All our current observations are but a tiny subset of all the possible situations AI is going to find itself in. A perfect track record is necessary yet insufficient.

Arguments for wiping out humanity
No.

Discussions with these types of arguments can still be worthwhile, but can not be sufficient for the purpose of making a dent into anyone’s P(Doom).