RSS

Ro­bust Agents

TagLast edit: 14 Sep 2020 23:17 UTC by Ruby

Robust Agents are decision-makers who can perform well in a variety of situations. Whereas some humans rely on folk wisdom or instinct, and some AIs might be designed to achieve a narrow set of goals, a Robust Agent has a coherent set of values and decision-procedures. This enables them to adapt to new circumstances (such as succeeding in a new environment, or responding to a new strategy by a competitor).

See also

Be­ing a Ro­bust Agent

Raemon18 Oct 2018 7:00 UTC
129 points
32 comments7 min readLW link2 reviews

On Be­ing Robust

TurnTrout10 Jan 2020 3:51 UTC
45 points
7 comments2 min readLW link

Embed­ded Agency (full-text ver­sion)

15 Nov 2018 19:49 UTC
143 points
15 comments54 min readLW link

Ro­bust Delegation

4 Nov 2018 16:38 UTC
110 points
10 comments1 min readLW link

Subagents, akra­sia, and co­her­ence in humans

Kaj_Sotala25 Mar 2019 14:24 UTC
122 points
31 comments16 min readLW link

Ro­bust Agency for Peo­ple and Organizations

Raemon19 Jul 2019 1:18 UTC
59 points
10 comments12 min readLW link

“Can you keep this con­fi­den­tial? How do you know?”

Raemon21 Jul 2020 0:33 UTC
151 points
44 comments3 min readLW link2 reviews

Se­cu­rity Mind­set and Or­di­nary Paranoia

Eliezer Yudkowsky25 Nov 2017 17:53 UTC
98 points
24 comments29 min readLW link

[Question] What if memes are com­mon in highly ca­pa­ble minds?

Daniel Kokotajlo30 Jul 2020 20:45 UTC
36 points
15 comments2 min readLW link

The Power of Agency

lukeprog7 May 2011 1:38 UTC
103 points
78 comments1 min readLW link

Up­com­ing sta­bil­ity of values

Stuart_Armstrong15 Mar 2018 11:36 UTC
15 points
15 comments2 min readLW link

Gra­da­tions of Agency

Daniel Kokotajlo23 May 2022 1:10 UTC
40 points
6 comments5 min readLW link

Hu­mans are very re­li­able agents

alyssavance16 Jun 2022 22:02 UTC
248 points
35 comments3 min readLW link

Ro­bust­ness to Scale

Scott Garrabrant21 Feb 2018 22:55 UTC
109 points
22 comments2 min readLW link1 review

Se­cu­rity Mind­set and the Lo­gis­tic Suc­cess Curve

Eliezer Yudkowsky26 Nov 2017 15:58 UTC
76 points
48 comments20 min readLW link

Reflec­tion in Prob­a­bil­is­tic Logic

Eliezer Yudkowsky24 Mar 2013 16:37 UTC
108 points
172 comments3 min readLW link

Tiling Agents for Self-Mod­ify­ing AI (OPFAI #2)

Eliezer Yudkowsky6 Jun 2013 20:24 UTC
84 points
259 comments3 min readLW link

2-D Robustness

vlad_m30 Aug 2019 20:27 UTC
77 points
8 comments2 min readLW link

Me­taphilo­soph­i­cal com­pe­tence can’t be dis­en­tan­gled from alignment

zhukeepa1 Apr 2018 0:38 UTC
32 points
39 comments3 min readLW link

An an­gle of at­tack on Open Prob­lem #1

Benya18 Aug 2012 12:08 UTC
47 points
85 comments7 min readLW link

Vingean Reflec­tion: Reli­able Rea­son­ing for Self-Im­prov­ing Agents

So8res15 Jan 2015 22:47 UTC
37 points
5 comments9 min readLW link

Thoughts on the 5-10 Problem

Tofly18 Jul 2019 18:56 UTC
18 points
11 comments1 min readLW link

Can we achieve AGI Align­ment by bal­anc­ing mul­ti­ple hu­man ob­jec­tives?

Ben Smith3 Jul 2022 2:51 UTC
11 points
1 comment4 min readLW link

Sets of ob­jec­tives for a multi-ob­jec­tive RL agent to optimize

23 Nov 2022 6:49 UTC
4 points
0 comments8 min readLW link