I think the questions of the next survey should be a superset of those on the last survey. Maybe not strictly, but it’s too interesting to track year-on-year changes to remove questions unless it’s really unquestionably obvious that they’re superfluous.
c0rw1n
IEEE released the first draft of their AI ethics guide
Where do hypotheses come from?
The evolution of superstitious and superstition-like behaviour
Note, last year’s survey was also run by /u/ingres
This means you are trying to Procustes the human squishiness into legibility, with consistent values. You should, instead, be trying to make pragmatic AIs that would frame the world for the humans, in the ways that the humans would approve*, taking into account their objectively stupid incoherence. Because that would be Friendly and parsed as such by the humans.
*=this doesn’t mean that such human preferences as those that violate meta-universalizability from behind the veil of ignorance should not be factored out of the calculation of what is ethically relevant; but it means that the states of the world that violate those preferences should still be hidden from the humans who have those preferences. This obviously results in humans being allowed to more accurately see the states of the world, the more their preferences are tolerant of other people’s preferences; there is absolutely nothing that could possibly ever go wrong from this, considering that the AIs, being Friendly, would simply prevent them from sociopathically exploiting that information asymmetry since that would violate the ethical principle.
I am confused about what sort of blog post you are requesting people write. I assume you don’t mean that people should list off a variety of interesting facts about the Bay Area, e.g. “the public transit system, while deeply inadequate, is one of the best in the country,” “UCSF is one of the top hospitals in the United States for labor and delivery,” “everything in San Francisco smells like urine,” “adult night at the Exploratorium is awesome,” “there are multiple socialist pizza places in Berkeley which schismed over whether to serve beer.” This seems of limited interest to people who don’t live in the Bay Area and of equally limited interest to people who live in the Bay Area (they know). If people are moving to the Bay Area I assume they’d discuss this with their friends who already live here.
Those posts would be useful indeed.
Specifically because :
If you are ignorant of verifiable facts about the Bay Area like “it is relatively uncommon to commute by car”, I have to question your ability to make accurate statements about more subtle and hard-to-check issues of social dynamics. If you have not talked to enough Berkeley residents to hear “gotta go, train’s here”, then I doubt your ability to comment on why our social demographics are the way they are.
You literally just pointed out that people are not writing about the practical details of living there, and you’re dismissing Ben’s points that reference the things that he ostensibly knows because people are writing about them, because he does not know the things that you yourself pointed out that people are not writing about!
Well so far it’s … a group house; with long late night conversations, also running self-experiments (currently measuring results of a low-carb diet), and organizing the local monthly rationalist meetup.
We are developing social tech solutions such as a database of competing access needs and a formal system for dealing with house logistics.
Why yes, there should be such a list; I don’t know of any existing one.
welp, 2⁄4 current residents and the next one planned to come there are trans women so um, what gender ratio issue again?
I think the lesswrong wiki was supposed to be that repository of the interesting/important things that were posted to the community blog.
It could be a good idea to make a wiki in lw2.0 and award site karma to people contributing to it.
I’m thinking of something like a section on the main lesserwrong.com page showing the latest edits to the wiki, so that the users of the site could see them and choose to go look at whether what changed in the article is worth points.
Is this enlightenment anything like described in https://aellagirl.com/2017/07/07/the-abyss-of-want/ ?
Also possibly related : http://nonsymbolic.org/wp-content/uploads/2014/02/PNSE-Article.pdf (can you point on that map where you think you found yourself)
Not easy, no. But there is a shorter version here : http://nonsymbolic.org/PNSE-Summary-2013.pdf
No. Correct models are good. Or rather, more correct models, applied properly, are better than less correct models, or models applied wrongly.
All those examples, however, are bad:
Calories in / Calories out is a bad model because different sources of calories are metabolized differently and have different effects on the organism. It is bad because it is incomplete and used improperly for things that it is bad at. It stays true that to get output from a mechanism, you do have to input some fuel into it; CICO is good enough to calculate, for example, how many calories one should eat in, say polar climates where one needs many thousands of them to survive the cold; that is not an argument against using a Correct model, it is an argument for using a model that outputs the correct type of data for the problem domain.
Being unwilling to update is bad, yes, that is a problem with the user of the model, not a problem with the model one is using. Do you mean that knowing and using the best known model makes one unwilling to update on the subsequent better model? Because that is still not a problem with using a Correct model.
That is an entirely different definition of Bad. “Bad for the person holding the model (provided they can’t pretend to hold the more socially-accepted model when that is more relevant to staying alive and non-ostracized)” is absolutely not the same thing as “Bad at actually predicting how reality will react on inputs”.
That is also a problem with the users of the model, both those making the prediction and those accepting the predictions, in domains that the model is not good at predicting.
Still not a problem due to using a Correct model. Also, bad outcomes for who? If immigration is good for the immigrants in some respsect and bad for them in some other respect, that is neither all good or all bad; if it is good for some citizens already there in some ways but bad in others (with some things being a lot of good for some citizens and some other things being a little bad, while also some things being a little good and other things being a lot of bad for other citizens), then the Correct model is the one that takes into account the entirety of all the distributions of the good and the bad things, across all citizens, and predict them Correctly.
( and now please in the name of Truth let this thread not devolve into an object-level discussion of that specific example >_< )
The AIs still have to make atoms move for anything Actually Bad to happen.
It may be that the person keeps expounding their reasons for their wanting you to do the thing because it feels aversive to them to stop infodumping, and/or because they expect you to respond with your reasons for doing the thing so that they know whether your doing the thing is an instance of 2 or of 3.
New users need 2 points to vote.