RSS

Andrew_Critch

Karma: 4,852

This is Dr. Andrew Critch’s professional LessWrong account. Andrew is the CEO of Encultured AI, and works for ~1 day/​week as a Research Scientist at the Center for Human-Compatible AI (CHAI) at UC Berkeley. He also spends around a ½ day per week volunteering for other projects like the Berkeley Existential Risk initiative and the Survival and Flourishing Fund. Andrew earned his Ph.D. in mathematics at UC Berkeley studying applications of algebraic geometry to machine learning models. During that time, he cofounded the Center for Applied Rationality and SPARC. Dr. Critch has been offered university faculty and research positions in mathematics, mathematical biosciences, and philosophy, worked as an algorithmic stock trader at Jane Street Capital’s New York City office, and as a Research Fellow at the Machine Intelligence Research Institute. His current research interests include logical uncertainty, open source game theory, and mitigating race dynamics between companies and nations in AI development.

The Onion Test for Per­sonal and In­sti­tu­tional Honesty

Sep 27, 2022, 3:26 PM
163 points
31 comments3 min readLW link3 reviews

An­nounc­ing En­cul­tured AI: Build­ing a Video Game

Aug 18, 2022, 2:16 AM
103 points
26 comments4 min readLW link

En­cul­tured AI Pre-plan­ning, Part 2: Pro­vid­ing a Service

Aug 11, 2022, 8:11 PM
33 points
4 comments3 min readLW link

En­cul­tured AI, Part 1 Ap­pendix: Rele­vant Re­search Examples

Aug 8, 2022, 10:44 PM
11 points
1 comment7 min readLW link

En­cul­tured AI Pre-plan­ning, Part 1: En­abling New Benchmarks

Aug 8, 2022, 10:44 PM
63 points
2 comments6 min readLW link

«Boundaries», Part 2: trends in EA’s han­dling of boundaries

Andrew_CritchAug 6, 2022, 12:42 AM
81 points
15 comments7 min readLW link

«Boundaries», Part 1: a key miss­ing con­cept from util­ity theory

Andrew_CritchJul 26, 2022, 11:03 PM
158 points
33 comments7 min readLW link

«Boundaries» Se­quence (In­dex Post)

Andrew_CritchJul 26, 2022, 7:12 PM
25 points
1 comment1 min readLW link

What’s next for in­stru­men­tal ra­tio­nal­ity?

Andrew_CritchJul 23, 2022, 10:55 PM
63 points
7 comments1 min readLW link

Cu­rat­ing “The Epistemic Se­quences” (list v.0.1)

Andrew_CritchJul 23, 2022, 10:17 PM
65 points
12 comments7 min readLW link

Pivotal out­comes and pivotal processes

Andrew_CritchJun 17, 2022, 11:43 PM
97 points
31 comments4 min readLW link

Slow mo­tion videos as AI risk in­tu­ition pumps

Andrew_CritchJun 14, 2022, 7:31 PM
241 points
41 comments2 min readLW link1 review

In­ter­gen­er­a­tional trauma im­ped­ing co­op­er­a­tive ex­is­ten­tial safety efforts

Andrew_CritchJun 3, 2022, 8:13 AM
129 points
29 comments3 min readLW link

“Tech com­pany sin­gu­lar­i­ties”, and steer­ing them to re­duce x-risk

Andrew_CritchMay 13, 2022, 5:24 PM
75 points
11 comments4 min readLW link

“Pivotal Act” In­ten­tions: Nega­tive Con­se­quences and Fal­la­cious Arguments

Andrew_CritchApr 19, 2022, 8:25 PM
139 points
55 comments7 min readLW link1 review

Power dy­nam­ics as a blind spot or blurry spot in our col­lec­tive world-mod­el­ing, es­pe­cially around AI

Andrew_CritchJun 1, 2021, 6:45 PM
189 points
26 comments6 min readLW link

What Mul­tipo­lar Failure Looks Like, and Ro­bust Agent-Ag­nos­tic Pro­cesses (RAAPs)

Andrew_CritchMar 31, 2021, 11:50 PM
285 points
65 comments22 min readLW link1 review

Some AI re­search ar­eas and their rele­vance to ex­is­ten­tial safety

Andrew_CritchNov 19, 2020, 3:18 AM
205 points
37 comments50 min readLW link2 reviews