RSS

Jan_Kulveit

Karma: 3,482

My current research interests:
- alignment in systems which are complex and messy, composed of both humans and AIs?
- actually good mathematized theories of cooperation and coordination
- active inference
- bounded rationality

Research at Alignment of Complex Systems Research Group (acsresearch.org), Centre for Theoretical Studies, Charles University in Prague. Formerly research fellow Future of Humanity Institute, Oxford University

Previously I was a researcher in physics, studying phase transitions, network science and complex systems.

Talk­ing pub­li­cly about AI risk

Jan_Kulveit21 Apr 2023 11:28 UTC
173 points
8 comments6 min readLW link

The self-un­al­ign­ment problem

14 Apr 2023 12:10 UTC
144 points
22 comments10 min readLW link

Limits to Legibility

Jan_Kulveit29 Jun 2022 17:42 UTC
137 points
11 comments5 min readLW link1 review

Cy­borg Pe­ri­ods: There will be mul­ti­ple AI transitions

22 Feb 2023 16:09 UTC
103 points
9 comments6 min readLW link

An­nounc­ing the Align­ment of Com­plex Sys­tems Re­search Group

4 Jun 2022 4:10 UTC
91 points
20 comments5 min readLW link

Why Si­mu­la­tor AIs want to be Ac­tive In­fer­ence AIs

10 Apr 2023 18:23 UTC
86 points
8 comments8 min readLW link

We don’t un­der­stand what hap­pened with cul­ture enough

Jan_Kulveit9 Oct 2023 9:54 UTC
86 points
21 comments6 min readLW link

The Cave Alle­gory Re­vis­ited: Un­der­stand­ing GPT’s Worldview

Jan_Kulveit14 Feb 2023 16:00 UTC
81 points
5 comments3 min readLW link

Why it took so long to do the Fermi calcu­la­tion right?

Jan_Kulveit2 Jul 2018 20:29 UTC
76 points
20 comments1 min readLW link

Elon Musk an­nounces xAI

Jan_Kulveit13 Jul 2023 9:01 UTC
75 points
35 comments1 min readLW link
(www.ft.com)

Multi-agent pre­dic­tive minds and AI alignment

Jan_Kulveit12 Dec 2018 23:48 UTC
63 points
18 comments10 min readLW link

Deon­tol­ogy and virtue ethics as “effec­tive the­o­ries” of con­se­quen­tial­ist ethics

Jan_Kulveit17 Nov 2022 14:11 UTC
63 points
9 comments1 min readLW link1 review

Les­sons from Con­ver­gent Evolu­tion for AI Alignment

27 Mar 2023 16:25 UTC
53 points
9 comments8 min readLW link

Ex­per­i­men­tal longter­mism: the­ory needs data

24 Mar 2022 8:23 UTC
52 points
0 comments4 min readLW link
(forum.effectivealtruism.org)

We can do bet­ter than argmax

Jan_Kulveit10 Oct 2022 10:32 UTC
48 points
4 comments1 min readLW link