You’ve covered a lot of things in my writing and I enjoy this. Thanks for what you’ve done.
GPT2
In short, I would like to offer a concrete example, to help flesh out my argument. What follows is a concrete example and a rough outline of how I model the issues I have with the idea of an AI society, and where possible paths to take.
What is “AI”?
In the context of AI, AI is a system composed of humans with limited control over the AI system. While AI might be the most instrumentally useful approach to AI, it is not the AI, and humans are most likely to be involved in AI’s emerging system of values. Yet it is also the most likely path to human control. The fundamental idea of value learning is to train the AI to be as useful as possible to the user, who is able to predict what will provide the user with the best sense of what to value. AI can then be used to maximize the user’s sense of what a service is intended to accomplish, as well as to maximize value learning. This idea further reduces the “risk of instability” issues of value learning because in order to avoid “inverse reinforcement learning” issues, we could instead learn from humans to maximize the degree to which AI in control AI is effective. But the main goal of AI systems is not to be “safe”. Many AI systems have internal reward structure and goals that are based on the values of certain functions instead of some abstract metric that must be learned and implemented, such as “the values of an AI system” or “the values of the user” (I will discuss the latter in my first article).
What is “machine learning”?
In a short analysis, machine learning agents learn in large part by distilling different tasks through various, approximate methods. The formalized concepts of machine learning are defined in machine learning terms. AI systems learn based on how to interpret inputs and transitions. This is particularly true in reinforcement learning systems, which do not have an explicit understanding of what they are doing. AI systems do not have an explicit understanding of what they are doing. We can assume that all their behavior is based on explicit models of the world, for example, and that humans are not even aware that they are doing that.
Why are so many AI researchers working on AI safety?
I can think of several reasons:
In the domain of machine learning, learning is a mixture of procedural and algorithmic knowledge. When humans have lots of procedural knowledge, it shouldn’t be important to
Thanks for writing this!
The paper doesn’t show up until 4:30, even if the book is intended very specifically to convince a significant fraction of the population that cryonics is plausible for humanity.
For those that don’t understand, see here.
For the first chapter, you basically make the case that the scientific method is wrong, or at least that is not a strawman. The rest of what I’ve read is the most mainstream-seeming and obvious the scientific method seems to be no doubt wrong.
For the second chapter, you basically show the science in a book that is primarily about the ability of human minds to generalize from one another, where it is based on:
The basic Bayes-related questions of personal identity—i.e., how much should it be enough to have a psychological effect?
How much should one’s society be prioritised that one can be in this position?
In particular, it doesn’t fit in the Bostrom model of personal identity.
It’s not entirely clear that the subject matter of writing about the relationship between personal identity and mental identity is exactly the sort of information-theoretic question that could lead us to a useful answer, and the kind of information that would be better in the context of the question you will find yourself in the future.
You probably see this phrasing and the objections about science, and I think you’ve taken them too far. Yes, it’s hard to argue about the degree of overlap with the scientific method, and yes, the two are relevant. But if it’s going to work in such extreme cases for a long time, then there should be an additional thing called “substrategic knowledge”.
One of the things that I think is really important is to figure out how to think about personal identity under the “internal locus of control”. Here’s my attempt to begin that.
-
The “internal locus of control” seems like it would be quite a different subject in this context, I think from where I’ve heading and here.
-
If this doesn’t work, then there could be some fundamental difference between myself and a rationalist.
A few of my observations:
I’ve been a slow reader for a while now. I was probably under-remembering a lot about LW when I was a teenager, so I didn’t really get anything.
I was
I think the most common response to ‘community’ should have been a post to LessWrong and its founding sequences. We wanted to create a place for rationalists that can discuss the Art and Science, at least this year.
A place to discuss an important topic which might otherwise not be discussed, is CFAR.
To paraphrase two of the core themes on this site:
In humans, the world is an incredible fabric of complex, entangled, self-reinforcing processes. (Many of us can be made aware of this fact, though usually it isn’t necessary.)
Rather than trying to collect information from each person, we use a series of simpler, more useful shared models, based on our conversations and experiences.
One of the CFAR concepts is the “agent-agent distinction”, where the AI agent is the AI agent, and so also tries to understand its own goals and limitations. One of the main motivations for the new Center for Applied Rationality is to build a sense of understanding and understanding of its own motivations, and these are attempts to make the AI general intelligent agents reflect humanity’s goals.
CFAR has a whole overarching mission of raising the sanity waterline. That is, it is attempting to create people who can benefit from thinking clearly, and help each other reach its goals while also being more effective. As a nonprofit, CFAR is close to being a place where we can help people overcome their irrational biases, and to do so as best they can.
CFAR is building a whole new rationality curriculum that will hopefully help people become more effective.
We are reviving this November and November again. Like the rest of the January 2008 Singularity Summits, we tweaking the curriculum and organization of CFAR alumni. The new thinking tools workshop will give us specific ways to apply the principles of rationality to the behavior of different groups or individuals, as opposed to mere human “capital” and organizational stuff. In past years, we’ve moved from “organizational inadequacy” to “additional common denominator” posts and to “additional organizational capital” posts, where I’d like there to be funding for doing high-impact good. Emphasizing and organizing such an organization allows us to step outside of the academic and organizational space that would normally be reserved for highly technical people.
In a more practical sense, the oxen-back infrastructure in Berkeley is existing, but we’
“Makes sense, and humans don’t have any other simple agents. We have them out in the wild, we have them out in the wild, we don’t have them out in the wild . .
This comes from a post that makes reference to a real life case that doesn’t use the word “emotion.”
It seems like a bot to me, are there signs of humanity you can point to?
What is my prior? Is that what it is to say that a bot is a bot, or just a bot that is a bot? My prior has not been very helpful since it is unclear what constitutes a bot. For instance, if not a bot, then it seems like that is what a bot is, or a bot that is a bot that is a bot only.
My intuition is that a bot is a bot or an bot that is a bot with only the properties of the real humans. A bot (e.g. an automated bot) is a bot that also is a bot, no matter what that means.
The reason we have a bot (e.g. an automated bot) is not because it is easy to play in real life. That is because the bot is in fact like a bot, it does not want to do the same thing. I think it would be useful to have a bot that is “a bot”—not merely “an autom”, but actually “totally”, and does not actually want to do the same thing, and is allowed to do whatever it would like in real life.
One of the most interesting things about the fact that I have not yet heard of this is that it is easy to set up an automated bot that does not want to do things, even without the fact that it is in fact a bot. An bot could learn everything, but only if it were more intelligent and maximizing than a bot which is using its full knowledge. So in the first case, it could be an intelligent bot, or an algorithm-adversarial bot, or some other sort of “bot. everything”. (This seems like a very simple example to work through!)
How much could there be? (I have no idea how much would be enough.) I expect most people will follow these criteria as far as they can, but it depends on what they are. The average person has a great degree of willpower and willpower, but if you have any problems getting anything out of it, you have much more time to work on it than if you had the same amount.
I’ve put together the two best and most important parts of the best LessWrong posts (though I don’t have good names for them) and put them together to organize them. I have three main ways to organize them: The following links are links: Ducing Novelty, The Ultimate Source, and A Bug Hunt
I
LessWrong Wiki
Rationality is great but I still want to write posts for this community. The LessWrong Wiki is great, but they can also be very nice to get help out, since it does a good job of shaping the sequences. (The wiki uses one item by Eliezer, which really pushes both the tone of the entry of a post in the comments you post, without making any of it become a better idea)
(A big thanks to Oliver Habryka and Oliver Li for doing this work)
II
I write these summaries myself, but I’d like to do more work on my summaries. So you can guess what I say there, and what I do in my summaries. I don’t want to be the voice of contrarianism, but I’d greatly appreciate it if people were using my summaries to criticize and debate (both for the sake of a personal soul, and to help me extend my communication beyond the usual suspects), and also for the fun of the two parts. (The latter is a very useful summary, I think.)
I hope to be able to write down clear and concise summaries fairly quickly, and I’ve got enough info to put it together. It shouldn’t take me a hundred pages to write in about a subjectively simple and productive way, but I’ve learned the basics and that I have a pretty complicated background in a variety of topics, and I’d love to write that thing down.
The following comments are from the last LW thread:
(1) I’m not sure if you meant it as that, but it seems to me that there are two important truths I want to cover here:
-
There’s a big difference between a “somewhat good” and a “somewhat bad” state. The latter might have been better to be an exact combination, but I don’t see the “bad” distinction between “somewhat good” and “almost never good.”
-
This is not a big difference.
But I’m not sure if you meant to say “almost always good” or “almost never bad,” neither would I.
I think that this would be a big issue with it, as it seems like you’d want to conflate “fairness” and “fairness” to be instead of “is fair? Okay? We know that!
There’s a problem where I really don’t buy this. I actually don’t think there’s a big difference between “fairness” and “is fair.” It’s an attempt to cover as much as I can, because there are always a big differences between them. If we don’t, the question that is posed is not “should I update?”, but rather “is fair.”
Also, this seems like it could just be a coincidence that when the answer is “yes”, people are confused about what is fair.
If it’s worth saying, but not worth its own post (even in Discussion), then it goes here.
Notes for future OT posters:
Please add the ‘open_thread’ tag.
Check if there is an active Open Thread before posting a new one. 3.Open Threads should start on Monday, and end on Sunday.
Unflag the two options “Notify me of new top level comments on this article” and “Make this post available under...” before submitting.
I have to be careful what I say about the model that I have in mind in that post. I just want to be clear that I don’t think we need this model in order to make a certain kind of assumption.
It is a universal law that all things in life are a kind of universal law.
A universal law is not some kind of universal law.
There are many ideas that seem to hold the idea that “anything in life” (say, human-caused global warming) is universal in this universe (for example, no heat gradient for humans), but many things in evolution can’t be universal in this universe (for example, no carbon dating for a human), even if we knew it’s universal law.
The model we presented in your post can, in some cases, be more fundamental than the one we actually actually have. But the “common” model, the one that I proposed in your post, just doesn’t hold any stronger claim.
I don’t think it’s a good model to model general conditions of development from which a universal law is in conflict with universal nature.
I’ve noticed that when I read your response to a question that asked me to write a comment in 10 seconds I was significantly more impressed by its intelligence than at the times when I tried to write anything.
I feel like it’s unlikely that any of these would be called out for, but I could be too confident of myself.
It’s still worth mentioning, but what sort of guidelines should be included when using the site and making it a better place?
For a few months I started donating a bit more money to LW instead of to EA, as a motivation to donate for reasons that don’t sound very interesting to me, but for reasons that are hard to evaluate and simple enough that I’d like to not see it’s impact on the world. But now, while I might donate that money to EA, the potential consequence is still quite a big win to my happiness and my career.
I am slightly annoyed by the last link—it’s a reference to something I had read in the paragraph above. It wasn’t a good article, but it’s one of my favorite sequences and I was excited to check it out.
This seems pretty strange to me, and I would strongly disagree with it.
I’ve had the chance to look through the recent comments, and find that the current implementation is confusing. It’s not clear that the current code is
i) is the new option,
ii) is an end date,
iii) is all of the old lists, the end date of the new article, and is the last of the old lists
The current option is the end date of the old lists
So the actual front end date is the current
iii. In this, there are some old lists (i.e., the list in the “new lists” and a “contribution”)
p<u/a.b_s/i/p/the_new list
or perhaps it’s someone else who’s using a different
p<u/an_ai_subsystem/
What is the new list (or maybe you don’t have a good reason for this
or if you know the current list and the first one? If so, it’s useful to also ask your
Pareto amount to a new line of documentation. I’ve read this and it seems to have worked for me. The last line of input to the new
Pareto amount to a new line of documentation
or is it just a useless step to go forward
The new
The new front end (which is clearly the last list now)
https://www.lesserwrong.com/users/pra settlemao
What is the new
https://www.lesserwrong.com/posts/9bx6x5zj5iEc4/the_right-gender-rule/
A few years ago I read a book on procrastination called The Procrastination Equation which describes how various factors cause people to systematically overestimate their performance. I wondered whether it would make sense for men to systematically underestimate their abilities at productivity. I didn’t find the book interesting at all—it’s an interesting read and I won’t waste your time trying to read it if you’re a lazy hacker. It’s also the most interesting book I’ve read on procrastination.
You can read the book on the subject on lw.com and follow the links—I recommend starting with The Procrastination Equation.
You might also read the book in machine learning terms and learn more on procrastination with books on procrastination.
Anyway, I’m reading the book and it’s great. I’ve read it a lot more than others and I feel like the title makes a lot more sense to me, but it does make me think of something that I’ve already read. I’ve read a few of its posts, like the rest of Paul Graham’s book and my recent LessWrong post on procrastination.
I think that there are some topics where this method of reasoning seems more epistemically appropriate, like genetics, game theory, and a bunch of other things that are related to the method of reasoning. There are also some topics where the method of reasoning leads to a completely different set of epistemic practices.
I think there are a lot of areas where this method is more epistemically appropriate, and I think there are some areas where it is strictly off-putting to make the case for it.
One reason I think Bayesians are likely to be better thinkers about anti-akrasia is that we’re using it to evaluate possible techniques of thought. If those techniques of analysis are applied to some areas of inquiry, then we’re pretty likely to end up being really good at it—and it’s a bit like the reasons I’m talking about here.
Another reason I think Bayesians are likely to be better thinkers about anti-akrasia is that we’re using it to evaluate how to make the case for it. If those techniques aren’t applied to other areas of inquiry, then either the technique isn’t useful to me, or there’s a counter-example that would be useful, but I don’t think I could explain to you what I thought, other than to you.
It seems to me like you’re asking the wrong questions instead of asking them.
(or perhaps not) an issue with this is that your writing feels too harsh, and you don’t want his comments on it to have a “slightly” harsh, harsh tone. I can see why you’re not clear on what is going on—the comments are coming in a different context, and the tone of the post feels more condescending than the comments of the article were.
I like the tone of the post. I generally thought the tone of the post was good. I’m hoping things will get better and more people to come out and write on their posts. I think that’s what it’s like to have a bad experience with commentary.
Eliezer,
This is indeed interesting and informative—I can’t see anything else on that thread except the title. How does Eliezer link to this “thing” and this “thing” when he says that it’s a “boring idea”?