RSS

AE Studio

Karma: 1,261

AE Studio is a team of 160+ programmers, product designers, and data scientists focused on increasing human agency through neglected high-impact approaches. Originally successful in BCI development and consulting, we’re now applying our expertise to AI alignment research, believing that the space of plausible alignment solutions is vast and under-explored.

Our alignment work includes prosociality research on self-modeling in neural systems, with attention schema theory in particular, self-other overlap mechanisms, and various neglected technical and policy approaches. We maintain a profitable consulting business that allows us to fund and pursue promising but overlooked research directions without pressure to expedite AGI development.

Learn more about us and our mission here:
https://​​ae.studio/​​ai-alignment

AE Stu­dio is hiring!

AE Studio21 Apr 2025 20:35 UTC
29 points
2 comments2 min readLW link

Mis­tral Large 2 (123B) seems to ex­hibit al­ign­ment faking

27 Mar 2025 15:39 UTC
80 points
4 comments13 min readLW link

Re­duc­ing LLM de­cep­tion at scale with self-other over­lap fine-tuning

13 Mar 2025 19:09 UTC
155 points
41 comments6 min readLW link

Align­ment can be the ‘clean en­ergy’ of AI

22 Feb 2025 0:08 UTC
67 points
8 comments8 min readLW link

Mak­ing a con­ser­va­tive case for alignment

15 Nov 2024 18:55 UTC
208 points
67 comments7 min readLW link

Science ad­vances one funeral at a time

1 Nov 2024 23:06 UTC
100 points
9 comments2 min readLW link

Self-pre­dic­tion acts as an emer­gent regularizer

23 Oct 2024 22:27 UTC
91 points
9 comments4 min readLW link

The case for a nega­tive al­ign­ment tax

18 Sep 2024 18:33 UTC
76 points
20 comments7 min readLW link

Self-Other Over­lap: A Ne­glected Ap­proach to AI Alignment

30 Jul 2024 16:22 UTC
223 points
51 comments12 min readLW link

Video In­tro to Guaran­teed Safe AI

11 Jul 2024 17:53 UTC
27 points
0 comments1 min readLW link
(youtu.be)

DIY RLHF: A sim­ple im­ple­men­ta­tion for hands on experience

10 Jul 2024 12:07 UTC
29 points
0 comments6 min readLW link

Key take­aways from our EA and al­ign­ment re­search sur­veys

3 May 2024 18:10 UTC
112 points
10 comments21 min readLW link

AE Stu­dio @ SXSW: We need more AI con­scious­ness re­search (and fur­ther re­sources)

26 Mar 2024 20:59 UTC
67 points
8 comments3 min readLW link

Sur­vey for al­ign­ment re­searchers!

2 Feb 2024 20:41 UTC
71 points
11 comments1 min readLW link

The ‘Ne­glected Ap­proaches’ Ap­proach: AE Stu­dio’s Align­ment Agenda

18 Dec 2023 20:35 UTC
178 points
23 comments12 min readLW link1 review