The Alignment Project Research Agenda

The Alignment Project is a global fund of over £15 million, dedicated to accelerating progress in AI control and alignment research. It is backed by an international coalition of governments, industry, venture capital and philanthropic funders.

This sequence sets out the research areas we are excited to fund – we hope this list of research ideas presents a novel contribution to the alignment field. We have deliberately focused on areas that we think the AI safety community currently underrates.

Apply now to join researchers worldwide in advancing AI safety.

For those with experience scaling and running ambitious projects, apply to our Strategy & Operations role here.

The Align­ment Pro­ject by UK AISI

Re­search Areas in In­for­ma­tion The­ory and Cryp­tog­ra­phy (The Align­ment Pro­ject by UK AISI)

Re­search Areas in Com­pu­ta­tional Com­plex­ity The­ory (The Align­ment Pro­ject by UK AISI)

Re­search Areas in Eco­nomic The­ory and Game The­ory (The Align­ment Pro­ject by UK AISI)

Re­search Areas in Prob­a­bil­is­tic Meth­ods (The Align­ment Pro­ject by UK AISI)

Re­search Areas in Learn­ing The­ory (The Align­ment Pro­ject by UK AISI)

Re­search Areas in Eval­u­a­tion and Guaran­tees in Re­in­force­ment Learn­ing (The Align­ment Pro­ject by UK AISI)

Re­search Areas in Cog­ni­tive Science (The Align­ment Pro­ject by UK AISI)

Re­search Areas in In­ter­pretabil­ity (The Align­ment Pro­ject by UK AISI)

Re­search Areas in Bench­mark De­sign and Eval­u­a­tion (The Align­ment Pro­ject by UK AISI)

Re­search Areas in Meth­ods for Post-train­ing and Elic­i­ta­tion (The Align­ment Pro­ject by UK AISI)

Re­search Areas in AI Con­trol (The Align­ment Pro­ject by UK AISI)