RSS

Max H

Karma: 2,209

Most of my posts and comments are about AI and alignment. Posts I’m most proud of, which also provide a good introduction to my worldview:

I also created Forum Karma, and wrote a longer self-introduction here.

PMs and private feedback are always welcome.

NOTE: I am not Max Harms, author of Crystal Society. I’d prefer for now that my LW postings not be attached to my full name when people Google me for other reasons, but you can PM me here or on Discord (m4xed) if you want to know who I am.

Grad­ual take­off, fast failure

Max H16 Mar 2023 22:02 UTC
15 points
4 comments5 min readLW link

In­stan­ti­at­ing an agent with GPT-4 and text-davinci-003

Max H19 Mar 2023 23:57 UTC
13 points
3 comments32 min readLW link

Grind­ing slimes in the dun­geon of AI al­ign­ment research

Max H24 Mar 2023 4:51 UTC
10 points
2 comments4 min readLW link

Steer­ing systems

Max H4 Apr 2023 0:56 UTC
50 points
1 comment15 min readLW link

Eliezer on The Lu­nar So­ciety podcast

Max H6 Apr 2023 16:18 UTC
40 points
5 comments1 min readLW link
(www.dwarkeshpatel.com)

A decade of lurk­ing, a month of posting

Max H9 Apr 2023 0:21 UTC
70 points
4 comments5 min readLW link

“Aligned” foun­da­tion mod­els don’t im­ply al­igned systems

Max H13 Apr 2023 4:13 UTC
39 points
10 comments5 min readLW link

Pay­ing the cor­rigi­bil­ity tax

Max H19 Apr 2023 1:57 UTC
14 points
1 comment13 min readLW link

A test of your ra­tio­nal­ity skills

Max H20 Apr 2023 1:19 UTC
11 points
11 comments4 min readLW link

LLM cog­ni­tion is prob­a­bly not hu­man-like

Max H8 May 2023 1:22 UTC
26 points
14 comments7 min readLW link

Gra­di­ent hack­ing via ac­tual hacking

Max H10 May 2023 1:57 UTC
12 points
7 comments3 min readLW link

Re­ward is the op­ti­miza­tion tar­get (of ca­pa­bil­ities re­searchers)

Max H15 May 2023 3:22 UTC
32 points
4 comments5 min readLW link

Where do you lie on two axes of world ma­nipu­la­bil­ity?

Max H26 May 2023 3:04 UTC
30 points
15 comments3 min readLW link

Without a tra­jec­tory change, the de­vel­op­ment of AGI is likely to go badly

Max H29 May 2023 23:42 UTC
16 points
2 comments13 min readLW link

Four lev­els of un­der­stand­ing de­ci­sion theory

Max H1 Jun 2023 20:55 UTC
12 points
11 comments4 min readLW link

10 quick takes about AGI

Max H20 Jun 2023 2:22 UTC
35 points
17 comments7 min readLW link

Fo­rum Karma: view stats and find highly-rated com­ments for any LW user

Max H1 Jul 2023 15:36 UTC
58 points
16 comments2 min readLW link
(forumkarma.com)

Ac­tu­ally, “per­sonal at­tacks af­ter ob­ject-level ar­gu­ments” is a pretty good rule of epistemic conduct

Max H17 Sep 2023 20:25 UTC
37 points
15 comments7 min readLW link

An ex­pla­na­tion for ev­ery to­ken: us­ing an LLM to sam­ple an­other LLM

Max H11 Oct 2023 0:53 UTC
34 points
4 comments11 min readLW link

Con­crete pos­i­tive vi­sions for a fu­ture with­out AGI

Max H8 Nov 2023 3:12 UTC
41 points
28 comments8 min readLW link