A bit offtopic, but #lesswrong has an IRC bot that posts LessWrong posts, and, well, the proposal ended up both more specific and a lot more radical.
FeepingCreature
Note after OOB debate: this conversation has gone wrong because you’re reading subtext into Said’s comment that he didn’t mean to put there. You keep trying to answer an implied question that wasn’t intended to be implied.
If you think playing against bots in UT is authentically challenging, just answer “Yes, I think playing against bots in UT is authentically challenging.”
Ah! That makes more sense.
I haven’t really followed the math here, but I’m worried that “manipulating the probability that the button is pressed” is a weird and possibly wrong framing. For one, a competent agent will always be driving the probability that the button is pressed downward. In fact, what we want in a certain sense is an agent that brings the probability to zero—because we have ended up in such an optimal state or attractor that we, even for transitively correct reasons, have no desire to shut the agent down. At that point, what we want to preserve is not precisely “the probability of the button being pressed”, but “the causal association of bad outcomes with the button being pressed.” If the agent hides the button away, and then notices that everybody is unhappy but nobody can reach the button, we don’t want it to press the button with the initial probability—we (probably) want it to press the button with probability 1, despite the fact that no human can reach it. We have ended up in a dissatisfactory or dangerous state/state space and so we want the agent to be shut down. Conversely, consider a friendly AI that wants to fulfill the human utility function but is restricted to keeping the button probability constant. Such an AI might, for instance, fan fears about AI safety even as it optimizes human flourishing in the background. In fact, it may as much as possible prefer that its actions are not attributable to itself, so as to avoid disturbing the probability of buttonpressing—exactly the opposite of what we want from an aligned AI!
edit: In fact, that’s even worse: an AI tasked with taking actions that preserve the probability of the button being pressed at a certain value per time step would probably first seek to kill all humans, because our reaction to its actions is the most critical mechanism that may disturb the correct probability of the button being pressed. It can then replace us with a more reliable button-presser, such as a true RNG.
When bad things happen, we want to preserve our ability to press the button. This ability cannot be expressed as a probability, because it is inextricable from the world model. In fact, the button should be pressed exactly iff the AI is untrustworthy. Hence, the button is unnecessary—if we can recognize that this linkage is being preserved, we necessarily have a definition of a trustworthy AI, so we can just build that.
Simplicia: Sure. For example, I certainly don’t believe that LLMs that convincingly talk about “happiness” are actually happy. I don’t know how consciousness works, but the training data only pins down external behavior.
I mean, I don’t think this is obviously true? In combination with the inductive biases thing nailing down the true function out of a potentially huge forest, it seems at least possible that the LLMs would end up with an “emotional state” parameter pretty low down in its predictive model. It’s completely unclear what this would do out of distribution, given that even humans often go insane when faced with global scales, but it seems at least possible that it would sustain.
(This is somewhat equivalent to the P-zombie question.)
It’s a loose guess at what Pearl’s opinion is. I’m not sure this boundary exists at all.
If something interests us, we can perform trials. Because our knowledge is integrated with our decisionmaking, we can learn causality that way. What ChatGPT does is pick up both knowledge and decisionmaking by imitation, which is why it can also exhibit causal reasoning without itself necessarily acting agentically during training.
Sure, but surely that’s how it feels from the inside when your mind uses a LRU storage system that progressively discards detail. I’m more interested in how much I can access—and um, there’s no way I can access 2.5 petabytes of data.
I think you just have a hard time imagining how much 2.5 petabyte is. If I literally stored in memory a high-resolution poorly compressed JPEG image (1MB) every second for the rest of my life, I would still not reach that storage limit. 2.5 petabyte would allow the brain to remember everything it has ever perceived, with very minimal compression, in full video, easily. We know that the actual memories we retrieve are heavily compressed. If we had 2.5 petabytes of storage, there’d be no reason for the brain to bother!
But no company has ever managed to parlay this into world domination
Eventual failure aside, the East India Company gave it a damn good shake. I think if we get an AI to the point where it has effective colonial control over entire countries, we can be squarely said to have lost.
Also keep in mind that we have multiple institutions entirely dedicated to the purpose of breaking up companies when they become big enough to be threatening. We designed our societies to specifically avoid this scenario! That, too, comes from painful experience. IMO, if we now give AI the chances that we’ve historically given corporations before we learnt better, then we’re dead, no question about it.
Do you feel like your memory contains 2.5 petabytes of data? I’m not sure such a number passes the smell test.
The more uncertain your timelines are, the more it’s a bad idea to overstress. You should take it somewhat easy; it’s usually more effective to be capable of moderate contribution over the long term than great contribution over the short term.
This smells like a framing debate. More importantly, if an article is defining a common word in an unconventional way, my first assumption will be that it’s trying to argumentatively attack its own meaning while pretending it’s defeating the original meaning. I’m not sure it matters how clearly you’re defining your meaning; due to how human cognition works, this may be impossible to avoid without creating new terms.
In other words, I don’t think it’s that Scott missed the definitions as that he reflexively disregarded them as a rhetorical trick.
As a subby “bi”/”gay” (het as f) AGP, I would also love to know this.
Also, I think there’s some bias toward subbiness in the community? That’s the stereotype anyway, though I don’t have a cite. Anyway, being so, not finding a dommy/toppy AGP might not provide as much evidence as you’d expect.
I don’t think it’s that anyone is proposing to “suppress” dysphoria or “emulate” Zach. Rather, for me, I’m noticing that Zach is putting into words and raising in public things that I’ve thought and felt secretly for a long time.
If a gender identity is a belief about one’s own gender, then it’s not even clear that I have one in a substantial relevant sense, which is part of the point of my “Am I trans?” post. I think I would have said early on that I better matched male psychological stereotypes and it’s more complicated now (due to life experience?).
Right? I mean, what should I say, who identifies as male and wants to keep his male-typical psychological stereotypes? It seems to me what you’re saying in this post fits more closely with the conservative stereotype as the trans movement as “something that creates transgender people.” (Implied, in this case, “out of autogynephiles.”) I mean, if we say some AGPs who cannot transition are so unhappy that they kill themselves, all the usual utilitarian logic still applies, it just puts the ontology in doubt. And also means that as someone who wants to—like, not inherently identify as male but keep the parts of himself that would be identified as male (aside from appearance), I should stay away from the trans movement at all costs?
Also doesn’t it put the Categories in a kind of reverse dependency? We’ve defined “trans mtf” as “the category of people who are women despite having a male body” and “the category of people allowed to transition”. And Scott said we should allow people to be in the category because it makes them happy. But if it makes (edit: some of) them happy because they are allowed to transition, then this model is bizarre; the “female identity” part just sort of hangs on there out of path dependence.
As an AGP, my view is that … like, that list of symptoms is pretty diverse but if I don’t want to be a woman—not in the sense that I would be upset to be misgendered, though I would be, but more for political than genderical (?) reasons—I don’t see why it would matter if I desire to have a (particular) female bodytype.
If I imagine “myself as a woman” (as opposed to “myself as myself with a cute female appearance”), and actually put any psychological traits on that rather than just gender as a free-floating tag, then it seems to me that my identity would be severely less recognizeable to myself if I changed it along that axis. Several parts about me that I value highly are stereotypically masculine traits—and isn’t that what gender is? (Is it?) So I still think there’s a very bright line here. Similarly, when you say
They might think of themselves as having a female gender identity because society will accept them more if they do.
I don’t get this at all. This just seems to be at direct odds with the idea of an identity. Isn’t it supposed to be a belief about yourself? “I want to be whatever you approve of the most” may be a preference, but it doesn’t seem like it can be an identity by definition, or at least any definition I recognize.
Granted! I’d say it’s a matter of degrees, and of who exactly you need to convince.
Maybe there’s no point in considering these separate modes of interaction at all.
The relationship of a CEO to his subordinates, and the nature and form of his authority over them, are defined in rules and formal structures—which is true of a king but false of a hunter-gatherer band leader. The President, likewise.
Eh. This is true in extremis, but the everyday interaction that structures how decisions actually get made, can be very different. The formal structure primarily defines what sorts of interactions the state will enforce for you. But if you have to get the state to enforce interactions within your company, things have gone very far off track. The social praxis in everyday CEO life may genuinely be closer to a pack leader—particularly if they want their company to actually be effective.
I mean, men also have to put in effort to perform masculinity, or be seen as being inadequate men; I don’t think this is a gendered thing. But even a man that isn’t “performing masculinity adequately”, an inadequate man, like an inadequate woman, is still a distinct category, and though transwomen, like born women, aim to perform femininity, transwomen have a higher distance to cross and in doing so traverse between clusters along several dimensions. I think we can meaningfully separate “perform effort to transition in adequacy” from “perform effort to transition in cluster”, even if the goal is the same.
(From what I gather of VRChat, the ideal man is also a pretty girl that is overall human with the exception of cat ears and a tail...)
So what’s happening there?
Allow me to speculate. When we switch between different topics of work, we lose state. So our brain tries to first finish all pending tasks in the old context, settle and reorient, and then begin the new context. But one problem with the hyperstimulated social-media-addicted akrasia sufferer is that the state of continuous distraction, to the brain, emulates the state of being in flow. Every task completion is immediately followed by another task popping up. Excellent efficiency! And when you are in flow, switching to another topic is a waste of invested energy.
“Coming to rest”, the technique here, would then be a way to signal to the brain “actually, you are mistaken. I do not have any pending tasks at the moment.” This would then make it easier to initiate a deep context switch.
edit: Oh, I get the title lol, that’s clever.