RSS

Remmelt

Karma: 456

Research Coordinator of area “Do Not Build Uncontrollable AI” for AI Safety Camp.

See explainer on why AGI could not be controlled enough to stay safe:
https://​​www.lesswrong.com/​​posts/​​xp6n2MG5vQkPpFEBH/​​the-control-problem-unsolved-or-unsolvable

Twelve Law­suits against OpenAI

Remmelt9 Mar 2024 12:22 UTC
31 points
4 comments1 min readLW link

Why I think it’s net harm­ful to do tech­ni­cal safety re­search at AGI labs

Remmelt7 Feb 2024 4:17 UTC
26 points
24 comments1 min readLW link

This might be the last AI Safety Camp

24 Jan 2024 9:33 UTC
180 points
33 comments1 min readLW link

The con­ver­gent dy­namic we missed

Remmelt12 Dec 2023 23:19 UTC
2 points
2 comments1 min readLW link

Fund­ing case: AI Safety Camp

12 Dec 2023 9:08 UTC
65 points
5 comments5 min readLW link
(manifund.org)

My first con­ver­sa­tion with An­nie Altman

Remmelt21 Nov 2023 21:58 UTC
8 points
3 comments1 min readLW link
(open.spotify.com)

Why a Mars colony would lead to a first strike situation

Remmelt4 Oct 2023 11:29 UTC
−57 points
8 comments1 min readLW link
(mflb.com)

Ap­ply to lead a pro­ject dur­ing the next vir­tual AI Safety Camp

13 Sep 2023 13:29 UTC
19 points
0 comments5 min readLW link
(aisafety.camp)

How teams went about their re­search at AI Safety Camp edi­tion 8

9 Sep 2023 16:34 UTC
28 points
0 comments13 min readLW link

4 types of AGI se­lec­tion, and how to con­strain them

Remmelt8 Aug 2023 10:02 UTC
−4 points
3 comments3 min readLW link

[Question] What did AI Safety’s spe­cific fund­ing of AGI R&D labs lead to?

Remmelt5 Jul 2023 15:51 UTC
3 points
0 comments1 min readLW link

AISC end of pro­gram presentations

6 Jun 2023 15:45 UTC
18 points
0 comments1 min readLW link

The Con­trol Prob­lem: Un­solved or Un­solv­able?

Remmelt2 Jun 2023 15:42 UTC
48 points
46 comments14 min readLW link

An­chor­ing fo­cal­ism and the Iden­ti­fi­able vic­tim effect: Bias in Eval­u­at­ing AGI X-Risks

Remmelt7 Jan 2023 9:59 UTC
1 point
2 comments1 min readLW link

Illu­sion of truth effect and Am­bi­guity effect: Bias in Eval­u­at­ing AGI X-Risks

Remmelt5 Jan 2023 4:05 UTC
−13 points
2 comments1 min readLW link

Nor­malcy bias and Base rate ne­glect: Bias in Eval­u­at­ing AGI X-Risks

Remmelt4 Jan 2023 3:16 UTC
−16 points
0 comments1 min readLW link

Sta­tus quo bias; Sys­tem jus­tifi­ca­tion: Bias in Eval­u­at­ing AGI X-Risks

3 Jan 2023 2:50 UTC
−11 points
0 comments1 min readLW link

Belief Bias: Bias in Eval­u­at­ing AGI X-Risks

2 Jan 2023 8:59 UTC
−10 points
1 comment1 min readLW link

Challenge to the no­tion that any­thing is (maybe) pos­si­ble with AGI

1 Jan 2023 3:57 UTC
−27 points
4 comments1 min readLW link
(mflb.com)

Curse of knowl­edge and Naive re­al­ism: Bias in Eval­u­at­ing AGI X-Risks

31 Dec 2022 13:33 UTC
−7 points
1 comment1 min readLW link
(www.lesswrong.com)