RSS

Has Diagram

TagLast edit: 29 Apr 2023 22:52 UTC by Gunnar_Zarncke

This tag is used to indicate that the post contains diagrams. This may be useful to quickly find such posts, or to exclude them in case you are visually impaired.

Be­ing the (Pareto) Best in the World

johnswentworth24 Jun 2019 18:36 UTC
359 points
56 comments3 min readLW link3 reviews

[In­tro to brain-like-AGI safety] 10. The al­ign­ment problem

Steven Byrnes30 Mar 2022 13:24 UTC
48 points
4 comments19 min readLW link

[In­tro to brain-like-AGI safety] 12. Two paths for­ward: “Con­trol­led AGI” and “So­cial-in­stinct AGI”

Steven Byrnes20 Apr 2022 12:58 UTC
42 points
10 comments16 min readLW link

Draw­ing Less Wrong: Tech­ni­cal Skill

Raemon5 Dec 2011 5:12 UTC
37 points
36 comments9 min readLW link

All images from the WaitButWhy se­quence on AI

trevor8 Apr 2023 7:36 UTC
72 points
5 comments2 min readLW link

The Nat­u­ral Ab­strac­tion Hy­poth­e­sis: Im­pli­ca­tions and Evidence

TheMcDouglas14 Dec 2021 23:14 UTC
34 points
8 comments19 min readLW link

Test­ing The Nat­u­ral Ab­strac­tion Hy­poth­e­sis: Pro­ject Update

johnswentworth20 Sep 2021 3:44 UTC
86 points
17 comments8 min readLW link1 review

Open tech­ni­cal prob­lem: A Quinean proof of Löb’s the­o­rem, for an eas­ier car­toon guide

Andrew_Critch24 Nov 2022 21:16 UTC
52 points
34 comments3 min readLW link

[In­tro to brain-like-AGI safety] 5. The “long-term pre­dic­tor”, and TD learning

Steven Byrnes23 Feb 2022 14:44 UTC
50 points
25 comments21 min readLW link

[In­tro to brain-like-AGI safety] 6. Big pic­ture of mo­ti­va­tion, de­ci­sion-mak­ing, and RL

Steven Byrnes2 Mar 2022 15:26 UTC
56 points
13 comments15 min readLW link

[In­tro to brain-like-AGI safety] 7. From hard­coded drives to fore­sighted plans: A worked example

Steven Byrnes9 Mar 2022 14:28 UTC
71 points
0 comments9 min readLW link

[In­tro to brain-like-AGI safety] 8. Take­aways from neuro 1/​2: On AGI development

Steven Byrnes16 Mar 2022 13:59 UTC
50 points
2 comments14 min readLW link

[In­tro to brain-like-AGI safety] 9. Take­aways from neuro 2/​2: On AGI motivation

Steven Byrnes23 Mar 2022 12:48 UTC
40 points
10 comments21 min readLW link

[In­tro to brain-like-AGI safety] 13. Sym­bol ground­ing & hu­man so­cial instincts

Steven Byrnes27 Apr 2022 13:30 UTC
65 points
15 comments14 min readLW link

[In­tro to brain-like-AGI safety] 14. Con­trol­led AGI

Steven Byrnes11 May 2022 13:17 UTC
35 points
25 comments19 min readLW link

[In­tro to brain-like-AGI safety] 1. What’s the prob­lem & Why work on it now?

Steven Byrnes26 Jan 2022 15:23 UTC
139 points
19 comments23 min readLW link

[In­tro to brain-like-AGI safety] 2. “Learn­ing from scratch” in the brain

Steven Byrnes2 Feb 2022 13:22 UTC
53 points
12 comments23 min readLW link

[In­tro to brain-like-AGI safety] 3. Two sub­sys­tems: Learn­ing & Steering

Steven Byrnes9 Feb 2022 13:09 UTC
74 points
3 comments24 min readLW link

[In­tro to brain-like-AGI safety] 4. The “short-term pre­dic­tor”

Steven Byrnes16 Feb 2022 13:12 UTC
60 points
11 comments13 min readLW link

Bayes’ The­o­rem Illus­trated (My Way)

komponisto3 Jun 2010 4:40 UTC
166 points
195 comments9 min readLW link

In­duc­tion heads—illustrated

TheMcDouglas2 Jan 2023 15:35 UTC
70 points
4 comments3 min readLW link

An Illus­trated Proof of the No Free Lunch Theorem

lifelonglearner8 Jun 2020 1:54 UTC
19 points
0 comments1 min readLW link
(mlu.red)

How much do you be­lieve your re­sults?

Eric Neyman6 May 2023 20:31 UTC
396 points
12 comments15 min readLW link
(ericneyman.wordpress.com)

Resi­d­ual stream norms grow ex­po­nen­tially over the for­ward pass

7 May 2023 0:46 UTC
65 points
17 comments11 min readLW link

The Car­toon Guide to Löb’s Theorem

Eliezer Yudkowsky17 Aug 2008 20:35 UTC
32 points
103 comments1 min readLW link

Cor­rigi­bil­ity, Much more de­tail than any­one wants to Read

Logan Zoellner7 May 2023 1:02 UTC
23 points
2 comments7 min readLW link

A new­comer’s guide to the tech­ni­cal AI safety field

zeshen4 Nov 2022 14:29 UTC
35 points
3 comments10 min readLW link

Embed­ding safety in ML development

zeshen31 Oct 2022 12:27 UTC
24 points
1 comment18 min readLW link

Levels of goals and alignment

zeshen16 Sep 2022 16:44 UTC
27 points
4 comments6 min readLW link