Does there exist a paper version of Yudkowsky’s book “Rationality: From AI to Zombies”? I only found a Kindle version but I would like to give it as a present to someone who is more likely to read a dead-tree version.
blf
Thank you.
It seems to me the word “dialog” may be appropriate: to me it has the connotation of reaching out to people you may not normally interact with.
The first one you mention appears in the list as one word, GiveDirectly. I initially had trouble finding it.
Contest: making a one-page comic on artificial intelligence for amateur mathematicians by March 9. The text must be in French and the original drawing on paper must be sent to them. Details at https://images.math.cnrs.fr/Onzieme-edition-de-Bulles-au-carre-a-vos-crayons-jusqu-au-9-mars?lang=fr
I’m not related in any way to this contest but I figured there may be some people interested in popularizing Alignment. I can help translate text to French. The drawing quality does not need to be amazing, see some previous winners at https://images.math.cnrs.fr/Resultats-du-9e-concours-Bulles-au-carre.html?lang=fr
It would be very interesting to see how much it understand space, for instance by making it draw maps. Perhaps “A map of New York City, with Central Park highlighted”? (I’m not sure if this is specific enough, but I fear that adding too many details will push Dall-E to join together various images.)
Yes, heuristic means a method to estimate things without too much effort.
”If I were properly calibrated then [...] correct choice 50% of the time.” points out that if lsusr was correct to be undecided about something, then it should be the case that both options were roughly equally good, so there should be a 50% chance that the first or second is the best. If that were the case, we could say that he is calibrated, like a measurement device that has been adjusted to give results as close to reality as possible.
”I didn’t lose the signal. I had just recalibrated myself.” means that lsusr has not lost the fear “signal”, but has adjusted the perception of fear to only occur when it is more appropriate (such as jumping off buildings). In that sense lsusr’s fear occurs at the right time, it is better calibrated.
I’m treating the message as a list of 2095 chunks of 64 bits. Let d(i,j) be the Hamming distance between the i-th and j-th chunk. The pairs (i,j) that have low Hamming distance (namely differ by few bits) cluster around straight lines with ratios j/i very close to integer powers of 2/e (I see features at least from (2/e)^-8 to (2/e)^8).
- Alien Message Contest: Solution by 13 Jul 2022 4:07 UTC; 29 points) (
- 28 Jun 2022 12:08 UTC; 1 point) 's comment on Contest: An Alien Message by (
Whenever , this quantity is at most 4.
I’m finding also around 50 instances of (namely ), with again .
Here is a rather clear sign that it is IEEE754 64 bit floats indeed. (Up to correctly setting the endianness of 8-byte chunks,) if we remove the first n bits from each chunk and count how many distinct values that takes, we find a clear phase transition at n=12, which corresponds to removing the sign bit and the 11 exponent bits.
These first 12 bits take 22 different values, which (in binary) clearly cluster around 1024 and 3072, suggesting that the first bit is special. So without knowing about IEEE754 we could have in principle figured out the splitting into 1+11+52 bits. The few quadratic patterns we found have enough examples with each exponent to help understand the transitions between exponents and completely fix the format (including the implicit 1 in the significand?).
This observation is clearer when treating the 64-bit chunks simply as double-precision IEEE754 floating points. Then the set of pairs for which is for some clearly draws lines with slopes close to powers of . But they don’t seem quite straight, so the slope is not so clear. In any case there is some pretty big long-distance correlation between and with rather different indices. (Note that if we explain the first line then the other powers are clearly consequences.)
Do you see how such an iteration can produce the long-distance correlations I mention in a message below, between floats at positions that differ by a factor of ? It seems that this would require some explicit dependence on the index.
If you calculate the entropy of each of the 64 bit positions (where and are the proportion of bits 0 and 1 among 2095 at that position), then you’ll see that the entropy depends much more smoothly on position if we convert from little endian to big endian, namely if we sort the bits as 57,58,...,64, then 49,50,...,56, then 41,42,...,48 and so on until 1,...,8. That doesn’t sound like a very natural boundary behaviour of an automaton, unless it is then encoded as little endian for some reason.
These simple ratios are “always” , see my comment https://www.lesswrong.com/posts/dFFdAdwnoKmHGGksW/contest-an-alien-message?commentId=Nz2XKbjbzGysDdS4Z for a proposal that 0.73 is close to (which I am not completely convinced by).
Your interlocutor in the other thread seemed to suggest that they were busy until mid-July or so. Perhaps you could take this into account when posting.
I agree that IEEE754 doubles was quite an unrealistic choice, and too easy. However, the other extreme of having a binary blob with no structure at all being manifest seems like it would not make for an interesting challenge. Ideally, there should be several layers of structure to be understood, like in the example of a “picture of an apple”, where understanding the file encoding is not the only thing one can do.
I think it would be a good idea to ask the question at the ongoing thread on AGI safety questions.
Two related questions to get a sense of scale of the social problem. (I’m interested in any precise operationalization, as obviously the questions are underspecified.)
Roughly how many people are pushing the state of the art in AI?
Roughly how many people work on AI alignment?
Another possibility would be for this behavior to come from grooming behavior in primates, during which (in many species?) lice and other stuff found on the skin seems to be eaten. In that case there is some clear advantage to eating the lice because it may otherwise infect another nearby individual.
Minor bug. When an Answer is listed in the sidebar of a post, the beginning of the answer is displayed, even if it starts with a spoiler. Hovering above the answer shows the full answer, which again ignores spoiler markup. For instance consider the sidebar of https://www.lesswrong.com/posts/x6AB4i6xLBgTkeHas/framing-practicum-general-factor-2.
I am not sure of myself, here, but I would expect a malicious AI to do the following. The first few (or many) times you run it, tell you the optimal stock. Then once in a while give a non-optimal stock. You would be unable to determine whether the AI was simply not turned on those times, or was not quite intelligent/resourceful enough to find the right stock. It may be that you would want the profits to continue.
By allowing itself to give you non-optimal stocks (but still making you rich), the AI can transmit information, such as its location, to anyone who would be looking at your pattern of buying stocks. And people would look at it, since you would be consistently buying the most profitable stock, with few exceptions. Once the location of the AI is known, you are in trouble, and someone less scrupulous than you may get their hand on the AI. Humans are dead in a fortnight.
Admittedly, this is a somewhat far-fetched scenario, but I believe that it indicates that you should not ask the AI more than one (or a few) questions before permanently destroying it. Even deleting all of its data and running the code again from scratch may be dangerous if the AI is able to determine how many times it has been launched in the past.