RSS

Robert Kralisch

Karma: 190

I am Robert Kralisch, one of the organizers for the AI Safety Camp, working as a research coordinator by evaluating and supporting research projects that fit under the umbrella of “technical AI Safety research” and “conceptually sound approaches to AI Alignment”.

I’m also an independent conceptual/​theoretical Alignment Researcher. I have a background in Cognitive Science and I am interested in collaborating on an end-to-end strategy for AGI alignment.

The three main branches that I aim to contribute to are conceptual clarity (what should we mean by agency, intelligence, embodiment, etc), the exploration of more inherently interpretable cognitive architectures, and Simulator theory.

One of my concrete goals is to figure out how to design a cognitively powerful agent such that it does not become a Superoptimiser in the limit.

AI Safety Camp 11

Nov 4, 2025, 2:56 PM
5 points

2 votes

Overall karma indicates overall quality.

0 comments15 min readLW link

In­vi­ta­tion to lead a pro­ject at AI Safety Camp (Vir­tual Edi­tion, 2026)

Sep 6, 2025, 1:17 PM
7 points

4 votes

Overall karma indicates overall quality.

0 comments4 min readLW link

AI Safety Camp 10 Outputs

Sep 5, 2025, 8:27 AM
18 points

6 votes

Overall karma indicates overall quality.

0 comments17 min readLW link

We don’t want to post again “This might be the last AI Safety Camp”

Jan 21, 2025, 12:03 PM
36 points

28 votes

Overall karma indicates overall quality.

17 comments1 min readLW link
(manifund.org)

Fund­ing Case: AI Safety Camp 11

Dec 23, 2024, 8:51 AM
60 points

27 votes

Overall karma indicates overall quality.

4 comments6 min readLW link
(manifund.org)

AI Safety Camp 10

Oct 26, 2024, 11:08 AM
38 points

17 votes

Overall karma indicates overall quality.

9 comments18 min readLW link

In­vi­ta­tion to lead a pro­ject at AI Safety Camp (Vir­tual Edi­tion, 2025)

Aug 23, 2024, 2:18 PM
17 points

6 votes

Overall karma indicates overall quality.

2 comments4 min readLW link

Re­search Dis­cus­sion on PSCA with Claude Son­net 3.5

Robert KralischJul 24, 2024, 4:53 PM
−2 points

4 votes

Overall karma indicates overall quality.

0 comments25 min readLW link

The Prop-room and Stage Cog­ni­tive Architecture

Robert KralischApr 29, 2024, 12:48 AM
14 points

5 votes

Overall karma indicates overall quality.

4 comments14 min readLW link

How are Si­mu­la­tors and Agents re­lated?

Robert KralischApr 29, 2024, 12:22 AM
6 points

3 votes

Overall karma indicates overall quality.

0 comments7 min readLW link

Ex­tended Embodiment

Robert KralischApr 29, 2024, 12:18 AM
8 points

6 votes

Overall karma indicates overall quality.

1 comment3 min readLW link

Refer­en­tial Containment

Robert KralischApr 29, 2024, 12:16 AM
2 points

1 vote

Overall karma indicates overall quality.

4 comments3 min readLW link

Disen­tan­gling Com­pe­tence and Intelligence

Robert KralischApr 29, 2024, 12:12 AM
23 points

8 votes

Overall karma indicates overall quality.

7 comments6 min readLW link

Introduction

Jun 30, 2023, 8:45 PM
8 points

6 votes

Overall karma indicates overall quality.

0 comments2 min readLW link

In­her­ently In­ter­pretable Architectures

Jun 30, 2023, 8:43 PM
4 points

3 votes

Overall karma indicates overall quality.

0 comments7 min readLW link

Pos­i­tive Attractors

Jun 30, 2023, 8:43 PM
6 points

4 votes

Overall karma indicates overall quality.

0 comments13 min readLW link

AISC 2023, Progress Re­port for March: Team In­ter­pretable Architectures

Apr 2, 2023, 4:19 PM
14 points

8 votes

Overall karma indicates overall quality.

0 comments14 min readLW link

Com­men­tary on “AGI Safety From First Prin­ci­ples by Richard Ngo, Septem­ber 2020”

Robert KralischOct 14, 2021, 3:11 PM
3 points

2 votes

Overall karma indicates overall quality.

0 comments19 min readLW link