I can corroborate; I also somehow read that as 1M instead of 1.4M the first time.
kyleherndon
It is quite strange to me that you would describe this story as “light and pleasant.” I can understand funny/absurd for the right sense of humor that is not mine.
With such a disclaimer, I think I would have been less confused but still not positive on the text. Maybe I’d be less negative? Neutral, even? A lack of disclaimer is not really the core of it for me though.
I have strange tastes in fiction, I think, and normally after reading something that anti-resonated with me so much, I’d just stop reading. But, I read to the end because a bunch of upvotes signaled to me that I might otherwise miss out on something later in the text.
I live in the Bay Area and work in tech and follow AI news. For me, I already knew that there are traits I don’t like in others, and had approximate models of their frequency in and correlations to the AI scene. Each character struck me as plausible and not really meaningfully more concrete to me. Concrete in the sense of “Oh, this behavior is warning sign that this is a morally bankrupt person I should stay away from” or “I might accidentally be falling into this pattern if I find this character relatable in some way.” I can definitely see this being useful to others who do not live in the bay area/work in tech/do not follow AI news, but I mostly expected people on lesswrong to match enough of that description that they wouldn’t find themselves learning from this “concretely describing a vibe or something.” And indeed, some of the people liked it for reasons other than that, which I failed to predict!
A big part of it for me is that I just cannot stand the first person narration of this main character. An example line, pulled at random: “But regardless, she makes very good coffee and is very nice and beautiful which sometimes amounts to the same thing.” I find this thoroughly jarring. I read it as the character’s internal monologue, but this is just so thoroughly outside the domain of what would go in my own internal monologue it feels… rude? an empty attempt at caricature? a failure to understand what internal thoughts look like or otherwise an indication of a mind horribly alien to my own such that it bears no resemblance to real humans? I do find the characters unappealing, but this just seemed unnecessary to me.
As an update, I’ve now gotten a bunch of useful responses of why people liked this! I was worried people would interpret me as being needlessly negative, but it seems mostly not so. I think I’m still far from enjoying this myself, but I do think I’ve learned a good amount from the answers I got, and I got some interesting things to think about.
I just went and read that one and found it interesting, yes!
That said, even if I’m unlikely to get a satisfactory response, I still want to ask. I would like to be able to better predict what other people like/think. And sometimes, I’ve even found that understanding someone else can help find new dimensions to appreciate :)
I personally don’t believe “writing quality” can be divorced from content, and if you shove a bunch of words together in a “masterful way” but don’t say anything, I don’t want to read that and therefore would call that bad writing.
Upon slightly more reflection, I think I can appreciate on an intellectual level the quality of the writing if the goal is to evoke disgust. People (myself included, on occasion) partake in Horror content, which is also traditionally a negative emotion. I haven’t heard of a Revulsion genre before, so I didn’t really consider that this might be a thing people pursue, but I would still be a little surprised if that was what most people got out of the post.
I would also be surprised if the source of all the upvotes was just that it is “”“high quality””” writing. I usually find LessWrong to be more focused on content, and I still want to know what other people see in this post.
I did not enjoy this. I did not feel like I got anything out of reading this.
However, this got curated and >500 karma, so I feel like I must be missing something. Can anyone inform me: Why did other people enjoy this? The best theory I can scrounge together is that this is “relatable” in some way to people in SF, like it conveys a vibe they are feeling?
If the goal is to evoke a sensation of disgust with the characters, then I guess you’ve succeeded for me. I already knew I would not like the sort of person described in this story, though, so I didn’t feel like I learned anything as a result, but I could see how something like that could be useful for others. I essentially just felt disgust the entire time reading this.
Artificial General Intelligence, AGI, is an AI, that can do anything an individual human can do, (especially on economically productive metrics).
Artificial Superintelligence, ASI, is an AI, that can do much more than all of humanity working together.
These definitions wouldn’t be suitable for a legal purpose, I imagine, in that they lack a “technical” precision. However, in my mind, there is a very big difference between the two, and an observer wouldn’t be likely to mislabel a system as ASI when it is actually AGI, or vice versa.
Yet, in my mind, one of the biggest risks of AGI is that it is used to build ASI, which is why I still agree with your post.
I like most of this post, but:
AGI != ASI. Defining AGI, and only post ante-fining a company into oblivion that makes AGI may be enough to prevent the death of humanity. I would put pretty good odds on it being enough, as long as it was strongly enforced and detected.
I would still support regulation preventing AGI, I just want the terminology to be straight. ASI is the thing that IABIED.
I would love an excuse to go back and learn QFT. Looking forward to your QFT AI insights :D
Although as I note elsewhere I’m starting to have some ideas of how something with elements of this might have a chance of working.
I’ve missed where you discussed this. Does anyone have a link or can anyone expound?
I think the problem of “actually specifying to an AI to do something physical, in reality, like ‘create a copy of strawberry down to the cellular but not molecular level’, and not just manipulate its own sensors to believe it perceives itself achieving that even if it accomplishes real things in the world to do that” is a problem that is very deeply related to physics, and is almost certainly dependent on the physical laws of world more than some abstract disembodied notion of an agent.
You’re thinking much too small, this only stops things occurring that are causally *downstream* of us. Things will still occur in other timelines, and we should prevent though things from happening too. I propose we create a “hyperintelligence” that acausally trades across timelines or invents time travel to prevent anything from happening in any other universe or timeline as well. Then we’ll be safe from AI ruin.
Thanks for the great link. Fine-tuning leading to mode collapse wasn’t the core issue underlying my main concern/confusion (intuitively that makes sense). paulfchristiano’s reply leaves me now mostly completely unconfused, especially with the additional clarification from you. That said I am still concerned; this makes RLHF seem very ‘flimsy’ to me.
I was also thinking the same thing as you, but after reading paulfchristiano’s reply, I now think it’s that you can use the model to use generate probabilities of next tokens, and that those next tokens are correct as often as those probabilities. This is to say it’s not referring to the main way of interfacing with GPT-n (wherein a temperature schedule determines how often it picks something other than the option with the highest probability assigned; i.e. not asking the model “in words” for its predicted probabilities).
GPT-4 can also be confidently wrong in its predictions, not taking care to double-check work when it’s likely to make a mistake. Interestingly, the base pre-trained model is highly calibrated (its predicted confidence in an answer generally matches the probability of being correct). However, through our current post-training process, the calibration is reduced.
What??? This is so weird and concerning.
I graduated college in four years with two bachelors and a masters. Some additions:
AP Tests:
You don’t need to take the AP course to take the test at all. This is NOT a requirement. If your high school doesn’t offer the test you may need to take it at another school, though. Also unfortunate is that if it is the same as when I did this, your school probably gets test fees waived for students who took the course and thus you may need to pay for the test. https://apstudents.collegeboard.org/faqs/can-i-register-ap-exam-if-my-school-doesnt-offer-ap-courses-or-administer-ap-exams
Proficiency Tests:
The college I went to offered “Proficiency Tests” for many courses (mostly freshman targeted) which were effectively final exams for courses that you could take, and if you satisfied with some grade you got credit for the course. If you are good at studying on your own, this will probably be significantly less work than taking the course and it is an especially effective for courses that you are not interested in.
Taking More Classes:
I literally planned my entire course load for all four years way before I got on campus (with built in flexibility for when courses were full or if I wanted to leave a couple of wildcards in for fun or whatever). This is important because if you’re planning something like what I was doing, it’s important not to have all your hard classes in the same semester and then burn out.
The big accusation, I think, is of sub-maximal procreation. If we cared at all about the genetic proliferation that natural selection wanted for us, then this time of riches would be a time of fifty-child families, not one of coddled dogs and state-of-the-art sitting rooms.
Natural selection, in its broadest, truest, (most idiolectic?) sense, doesn’t care about genes.
So what did natural selection want for us? What were we selected for? Existence.
I think there might be a meaningful way to salvage the colloquial concept of “humans have overthrown natural selection.”
Let [natural selection] refer to the concept of trying to maximize genetic fitness and specifically refer to maximizing the spread of genes. Let [evolution] refer to the concept of trying to maximize ‘existence’ or persistence. There’s sort of a hierarchy of optimizers where [evolution] > [natural selection] > humanity where you could make the claim that humanity has “overthrown our boss and took their position” such that humanity reports directly to [evolution] now instead of having [natural selection] as our middle manager boss. One can make the argument that ideas in brains are the preferred substrate over DNA now, as an example of this model.
This description also makes the warning with respect to AI a little more clear: any box or “boss” is at risk of being overthrown.
(This critique contains not only my own critiques, but also critiques I would expect others on this site to have)
First, I don’t think that you’ve added anything new to the conversation. Second, I don’t think what you have mentioned even provides a useful summary of the current state of the conversation: it is neither comprehensive, nor the strongest version of various arguments already made. Also, I would prefer to see less of this sort of content on LessWrong. Part of that might be because it is written for a general audience, and LessWrong is not very like the general audience.
This is an example of something that seems to push the conversation forward slightly, by collecting all the evidence for a particular argument and by reframing the problem as different, specific, answerable questions. While I don’t think this actually “solves the hard problem of consciousness as Halberstadt notes in the comments, I think it could help clear up some confusions for you. Namely, I think it is most meaningful to start from a vaguely panpsychist model of “everything is conscious,” what we mean by consciousness is “the feeling of what it is like to be” and the move on to talk about what sorts of consciousness we care about: namely consciousness that looks remotely similar to ours. In this framework, AI is already conscious, but I don’t think there’s any reason to care about that.
More specifics:
Consciousness is not, contrary to the popular imagination, the same thing as intelligence.
I don’t think that’s a popular opinion here. And while I think some people might just have a cluster of “brain/thinky” words in their head when they don’t think about the meaning of things closely, I don’t think this is a popular opinion of people in general unless they’re really not thinking about it.
But there’s nothing that it’s like to be a rock
Citation needed.
But that could be very bad, because it would mean we wouldn’t be able to tell whether or not the system deserves any kind of moral concern.
Assuming we make an AI conscious, and that consciousness is actually something like what we mean by it more colloquially (human-like, not just panpsychistly), it isn’t clear that this makes it a moral concern.
There should be significantly more research on the nature of consciousness.
I think there shouldn’t. At least not yet. The average intelligent person thrown at this problem produces effectively nothing useful, in my opinion. Meanwhile, I feel like there is a lot of lower hanging fruit in neuroscience that would also help solve this problem more easily later in addition to actually being useful now.
In my opinion, you choose to push for more research when you have questions you want answered. I do not consider humanity to have actually phrased the hard problem of consciousness as a question, nor do I think we currently have the tools to notice an answer if we were given one. I think there is potentially useful philosophy to do around but not on the hard problem of consciousness in terms of actually asking a question or learning how we could recognize an answer
Researchers should not create conscious AI systems until we fully understand what giving those systems rights would mean for us.
They cannot choose not to because they don’t know what it is, so this is unactionable and useless advice.
AI companies should wait to proliferate AI systems that have a substantial chance of being conscious until they have more information about whether they are or not.
Same thing as above, and also the prevailing view here is that it is much more important that AI will kill us, and if we’re theoretically spending (social) capital to make these people care about things, the not killing us is astronomically more important.
AI researchers should continue to build connections with philosophers and cognitive scientists to better understand the nature of consciousness
I don’t think you’ve made strong enough arguments to support this claim given the opportunity costs. I don’t have an opinion on whether or not you are right here.
Philosophers and cognitive scientists who study consciousness should make more of their work accessible to the public
Same thing as above.
Nitpick: there’s something weird going on with your formatting because some of your recommendations show up on the table of contents and I don’t think that’s intended.
Besides cryonics, if you can “revive” someone by figuring out the exact parameters of our universe, simulating history, and thusly finding the people who died and saving them, this changes the calculus further, as now we need to consider extending the lifespan of everyone who ever lived.
Mind uploading, besides directly increasing lifespan, could also result in copies and branches of selves. I find it extremely difficult to reason about the utility of having N copies of myself.
It’s unclear to me if this possible, but there’s at least uncertainty for me here, when we talk about superintelligence.