RSS

Limits to Control

TagLast edit: 2 Jul 2025 7:39 UTC by Remmelt

Limits to Control is a field of research aiming to discover and verify:

  1. Fundamental limits to controlling fully autonomous AI using any method of causation.

  2. Threat models of AI convergent dynamics that cannot be sufficiently controlled (by 1.).

  3. Impossibility theorems, by contradiction of ‘long-term AI safety’ with convergence result (2.)

~ ~ ~

Definitions

‘Fundamental limit’

A limit to the capacity to reach some states over other states. This limit is absolute – it is not just based on practical impediments, but on the physics or information signalling of the system itself.

‘Control’

‘Long term’

‘AI safety’

Ambient conditions caused by AI’s operations fall within the environmental range needed for the survival of humans (a minimum-threshold definition of safety).

‘Fully autonomous AI’

Any assembly of artificial components that persistently learns new functions from (and thus maintains and adapts its components across) the environment, without the need for humans or other organic life.

‘AI convergent dynamic that cannot be sufficiently controlled’

The learning/​adapting components propagate new environmental effects, where a subset of the dynamically explored state space is both unsafe and falls outside even one fundamental limit to control.

‘Verify’

To check whether an argument’s premises are empirically sound and whether its reasoning steps are logically valid.

Lenses of Control

WillPetillo22 Oct 2024 7:51 UTC
14 points
0 comments9 min readLW link

De­con­fus­ing ‘AI’ and ‘evolu­tion’

Remmelt11 Jul 2025 1:44 UTC
12 points
11 comments28 min readLW link

Limits to Con­trol Workshop

18 May 2025 16:05 UTC
12 points
2 comments3 min readLW link

The Robot, the Pup­pet-mas­ter, and the Psychohistorian

WillPetillo28 Dec 2024 0:12 UTC
8 points
2 comments3 min readLW link

Pro­jects I would like to see (pos­si­bly at AI Safety Camp)

Linda Linsefors27 Sep 2023 21:27 UTC
22 points
12 comments4 min readLW link

De­sign­ing for per­pet­ual control

Remmelt12 Oct 2025 2:06 UTC
1 point
11 comments2 min readLW link

What if Align­ment is Not Enough?

WillPetillo7 Mar 2024 8:10 UTC
16 points
46 comments9 min readLW link

The Con­trol Prob­lem: Un­solved or Un­solv­able?

Remmelt2 Jun 2023 15:42 UTC
57 points
46 comments13 min readLW link

Pro­ject Moonbeam

WillPetillo27 Jun 2025 21:08 UTC
14 points
2 comments6 min readLW link

Prin­ci­ples of AI Uncontrollability

WillPetillo7 Aug 2025 21:10 UTC
1 point
0 comments7 min readLW link

Why Re­cur­sive Self-Im­prove­ment Might Not Be the Ex­is­ten­tial Risk We Fear

Nassim_A24 Nov 2024 17:17 UTC
1 point
0 comments9 min readLW link
No comments.