# Selec­tion Theorems

TagLast edit: 25 Dec 2022 20:56 UTC by

A Selection Theorem tells us something about what agent type signatures will be selected for in some broad class of environments. Two important points:

• The theorem need not directly talk about selection—e.g. it could state some general property of optima, of “broad” optima, of “most” optima, or of optima under a particular kind of selection pressure (like natural selection or financial profitability).

• Any given theorem need not address every question about agent type signatures; it just needs to tell us something about agent type signatures.

For instance, the subagents argument says that, when our “agents” have internal state in a coherence-theorem-like setup, the “goals” will be pareto optimality over multiple utilities, rather than optimality of a single utility function. This says very little about embeddedness or world models or internal architecture; it addresses only one narrow aspect of agent type signatures. And, like the coherence theorems, it doesn’t directly talk about selection; it just says that any strategy which doesn’t fit the pareto-optimal form is strictly dominated by some other strategy (and therefore we’d expect that other strategy to be selected, all else equal).

# Selec­tion The­o­rems: A Pro­gram For Un­der­stand­ing Agents

28 Sep 2021 5:03 UTC
124 points

# Epistemic Strate­gies of Selec­tion Theorems

18 Oct 2021 8:57 UTC
33 points

# Some Ex­ist­ing Selec­tion Theorems

30 Sep 2021 16:13 UTC
54 points

# Un­der­stand­ing Selec­tion Theorems

28 May 2022 1:49 UTC
41 points

# What Selec­tion The­o­rems Do We Ex­pect/​Want?

1 Oct 2021 16:03 UTC
65 points

# Les­sons from Con­ver­gent Evolu­tion for AI Alignment

27 Mar 2023 16:25 UTC
51 points

# [Question] Why The Fo­cus on Ex­pected Utility Max­imisers?

27 Dec 2022 15:49 UTC
116 points

# Selec­tion pro­cesses for subagents

30 Jun 2022 23:57 UTC
36 points

# How Do Selec­tion The­o­rems Re­late To In­ter­pretabil­ity?

9 Jun 2022 19:39 UTC
60 points

# Clar­ify­ing the Agent-Like Struc­ture Problem

29 Sep 2022 21:28 UTC
58 points

# Riffing on the agent type

8 Dec 2022 0:19 UTC
21 points

# Pro­ject In­tro: Selec­tion The­o­rems for Modularity

4 Apr 2022 12:59 UTC
71 points

# AXRP Epi­sode 15 - Nat­u­ral Ab­strac­tions with John Wentworth

23 May 2022 5:40 UTC
34 points