RSS

DragonGod

Karma: 2,385

Theoretical Computer Science Msc student at the University of [Redacted] in the United Kingdom.

I’m an aspiring alignment theorist; my research vibes are descriptive formal theories of intelligent systems (and their safety properties) with a bias towards constructive theories.

I think it’s important that our theories of intelligent systems remain rooted in the characteristics of real world intelligent systems; we cannot develop adequate theory from the null string as input.

Con­se­quen­tial­ism is in the Stars not Ourselves

DragonGod24 Apr 2023 0:02 UTC
7 points
19 comments5 min readLW link

[Question] Is “Strong Co­her­ence” Anti-Nat­u­ral?

DragonGod11 Apr 2023 6:22 UTC
23 points
25 comments2 min readLW link

Fea­ture Re­quest: Right Click to Copy LaTeX

DragonGod8 Apr 2023 23:27 UTC
18 points
4 comments1 min readLW link

Beren’s “De­con­fus­ing Direct vs Amor­tised Op­ti­mi­sa­tion”

DragonGod7 Apr 2023 8:57 UTC
51 points
10 comments3 min readLW link

[Question] Is “Re­cur­sive Self-Im­prove­ment” Rele­vant in the Deep Learn­ing Paradigm?

DragonGod6 Apr 2023 7:13 UTC
32 points
36 comments7 min readLW link

Orthog­o­nal­ity is Expensive

DragonGod3 Apr 2023 0:43 UTC
21 points
3 comments1 min readLW link
(www.beren.io)

“Dangers of AI and the End of Hu­man Civ­i­liza­tion” Yud­kowsky on Lex Fridman

DragonGod30 Mar 2023 15:43 UTC
38 points
32 comments1 min readLW link
(www.youtube.com)

Sparks of Ar­tifi­cial Gen­eral In­tel­li­gence: Early ex­per­i­ments with GPT-4 | Microsoft Research

DragonGod23 Mar 2023 5:45 UTC
68 points
23 comments1 min readLW link
(arxiv.org)

Con­tra “Strong Co­her­ence”

DragonGod4 Mar 2023 20:05 UTC
39 points
24 comments1 min readLW link

In­cen­tives and Selec­tion: A Miss­ing Frame From AI Threat Dis­cus­sions?

DragonGod26 Feb 2023 1:18 UTC
11 points
16 comments2 min readLW link

[Question] Is In­struc­tGPT Fol­low­ing In­struc­tions in Other Lan­guages Sur­pris­ing?

DragonGod13 Feb 2023 23:26 UTC
39 points
15 comments1 min readLW link

[Question] Do the Safety Prop­er­ties of Pow­er­ful AI Sys­tems Need to be Ad­ver­sar­i­ally Ro­bust? Why?

DragonGod9 Feb 2023 13:36 UTC
22 points
42 comments2 min readLW link

[About Me] Cin­era’s Home Page

DragonGod7 Feb 2023 12:56 UTC
30 points
2 comments9 min readLW link

[Question] What Are The Pre­con­di­tions/​Pr­ereq­ui­sites for Asymp­totic Anal­y­sis?

DragonGod3 Feb 2023 21:26 UTC
8 points
2 comments1 min readLW link

AI Risk Man­age­ment Frame­work | NIST

DragonGod26 Jan 2023 15:27 UTC
36 points
4 comments2 min readLW link
(www.nist.gov)

“Hereti­cal Thoughts on AI” by Eli Dourado

DragonGod19 Jan 2023 16:11 UTC
145 points
38 comments3 min readLW link
(www.elidourado.com)

[Question] How Does the Hu­man Brain Com­pare to Deep Learn­ing on Sam­ple Effi­ciency?

DragonGod15 Jan 2023 19:49 UTC
10 points
6 comments1 min readLW link

Tracr: Com­piled Trans­form­ers as a Lab­o­ra­tory for In­ter­pretabil­ity | Deep­Mind

DragonGod13 Jan 2023 16:53 UTC
62 points
12 comments1 min readLW link
(arxiv.org)

Microsoft Plans to In­vest $10B in OpenAI; $3B In­vested to Date | For­tune

DragonGod12 Jan 2023 3:55 UTC
23 points
10 comments2 min readLW link
(fortune.com)

Open & Wel­come Thread—Jan­uary 2023

DragonGod7 Jan 2023 11:16 UTC
15 points
37 comments1 min readLW link