RSS

Ro­bust Agents

TagLast edit: 14 Sep 2020 23:17 UTC by Ruby

Robust Agents are decision-makers who can perform well in a variety of situations. Whereas some humans rely on folk wisdom or instinct, and some AIs might be designed to achieve a narrow set of goals, a Robust Agent has a coherent set of values and decision-procedures. This enables them to adapt to new circumstances (such as succeeding in a new environment, or responding to a new strategy by a competitor).

See also

Be­ing a Ro­bust Agent

Raemon11 Jan 2020 2:06 UTC
107 points
28 comments7 min readLW link

On Be­ing Robust

TurnTrout10 Jan 2020 3:51 UTC
45 points
7 comments2 min readLW link

Embed­ded Agency (full-text ver­sion)

15 Nov 2018 19:49 UTC
115 points
11 comments54 min readLW link

Ro­bust Delegation

4 Nov 2018 16:38 UTC
108 points
10 comments1 min readLW link

Subagents, akra­sia, and co­her­ence in humans

Kaj_Sotala25 Mar 2019 14:24 UTC
98 points
31 comments16 min readLW link

Ro­bust Agency for Peo­ple and Organizations

Raemon19 Jul 2019 1:18 UTC
58 points
10 comments12 min readLW link

“Can you keep this con­fi­den­tial? How do you know?”

Raemon21 Jul 2020 0:33 UTC
128 points
40 comments3 min readLW link

[Question] What if memes are com­mon in highly ca­pa­ble minds?

Daniel Kokotajlo30 Jul 2020 20:45 UTC
32 points
8 comments2 min readLW link

The Power of Agency

lukeprog7 May 2011 1:38 UTC
86 points
78 comments1 min readLW link

Up­com­ing sta­bil­ity of values

Stuart_Armstrong15 Mar 2018 11:36 UTC
13 points
15 comments2 min readLW link

Ro­bust­ness to Scale

Scott Garrabrant21 Feb 2018 22:55 UTC
99 points
21 comments2 min readLW link

Se­cu­rity Mind­set and Or­di­nary Paranoia

Eliezer Yudkowsky25 Nov 2017 17:53 UTC
74 points
21 comments29 min readLW link

Se­cu­rity Mind­set and the Lo­gis­tic Suc­cess Curve

Eliezer Yudkowsky26 Nov 2017 15:58 UTC
67 points
45 comments20 min readLW link

Reflec­tion in Prob­a­bil­is­tic Logic

Eliezer Yudkowsky24 Mar 2013 16:37 UTC
101 points
171 comments3 min readLW link

Tiling Agents for Self-Mod­ify­ing AI (OPFAI #2)

Eliezer Yudkowsky6 Jun 2013 20:24 UTC
84 points
259 comments3 min readLW link

2-D Robustness

vlad_m30 Aug 2019 20:27 UTC
67 points
1 comment2 min readLW link

Me­taphilo­soph­i­cal com­pe­tence can’t be dis­en­tan­gled from alignment

zhukeepa1 Apr 2018 0:38 UTC
32 points
39 comments3 min readLW link

An an­gle of at­tack on Open Prob­lem #1

Benya18 Aug 2012 12:08 UTC
47 points
85 comments7 min readLW link

Vingean Reflec­tion: Reli­able Rea­son­ing for Self-Im­prov­ing Agents

So8res15 Jan 2015 22:47 UTC
36 points
5 comments9 min readLW link

Thoughts on the 5-10 Problem

Tofly18 Jul 2019 18:56 UTC
18 points
11 comments1 min readLW link