Interested in big picture questions, decision theory, altruism.
cSkeleton
Hi, did you ever go anywhere with Conversation Menu? I’m thinking of doing something like this related to AI risk to try to quickly get people to the arguments around their initial reaction and if helping with something like this is the kind of thing you had in mind with Conversation Menu I’m interested to hear any more thoughts you have around this. (Note, I’m thinking of fading in buttons more than a typical menu.) Thanks!
Thanks for the wonderful post!
What are the approximate costs for therapists/coaches options?
I suspect most people here are pro-cryonics and anti-cremation.
Thanks for your thoughts. It sounds like this is a major risk but hopefully when we know more (if we can get there) we’ll have a better idea of how to maximize things and find at least one good option [insert sweat face emoji for discomfort but going forward boldly]
Thanks for the response! I really appreciate it.
a) Yes, I meant “the probability of”
b) Thinking about how to plot this on graphs is helping me to clarify thinking and I think adding these may help to reduce inferential distance. (The X axis is probability. For the case where we consider infinite utilities as opposed to the human case, the graph would need to be split into 2 graphs. The one on left is just an infinity horizontal line but there is still a probability range. The one on the right has an actual curve and covers the rest of the probability range but doesn’t matter since its utility values are finite. Considering only the infinite utilities is a fanatical decision procedure but doesn’t generally lead to weird decisions. Does that make sense?)
Thanks @RolfAndreassen. I’m reconsidering and will post a different version if I get there. I’ve marked this one as [retracted].
I find this confusing. My actual strength of belief now that I can tip an outcome that affects at least 3^^^3 other people is a lot closer to 1/(1000000) than 1/(3^^7625597484987). My justification is that while 3^^^3 isn’t a number that fits into any finite multiverse, the universe going on for infinitely long seems kinda possible and anthropic reasoning may not be valid here (I added 10x in case it is) and I have various ideas. The difference in those two probabilities is large (to put it mildly), and significant (one is worth thinking about and the other isn’t). How to resolve this?
I’m having difficulty following the code for the urn scenario. Can it be something like?
def P():
# Initialize the world with random balls (or whatever)
num_balls = 1000
urn = [random.choice([“red”, “white”]) for i in range(num_balls)]
# Run the world
history = []
total_loss = 0
for i in range(len(urn)):
ball = urn[i]
probability_of_red = S(history)
if probability_of_red == 1 and ball != ‘red’ or probability_of_red == 0 and ball == ‘red’:
print(“You were 100% sure of a wrong prediction. You lose for all eternity.”)
return # avoid crashing in math.log()
if ball == ‘red’:
loss = math.log(probability_of_red)
else:
loss = math.log(1 - probability_of_red)
total_loss += loss
history.append(ball)
print(f”{ball:6}\tPrediction={probability_of_red:0.3f}\tAverage log loss={total_loss / (i + 1):0.3f}”)
If we define S() as:
def S(history):
if not history:
return 0.5
reds = history.count(‘red’)
prediction = reds / float(len(history))
# Should never be 100% confident
if prediction == 1:
prediction = 0.999
if prediction == 0:
prediction = 0.001
return prediction
The output will converge on Prediction = 0.5 and Average log loss as log(0.5). Is that right?
Thanks for your replies! I didn’t realize the question was unclear. I was looking for an answer TO provide the AI, not an answer FROM the AI. I’ll work on the title/message and try again.
Edit: New post at https://www.lesswrong.com/posts/FJaFMdPREcxaLoDqY/what-should-we-tell-an-ai-if-it-asks-why-it-was-created
[Question] What should we tell an AI if it asks why it was created?
Governments are not social welfare maximizers
Most people making up governments, and society in general, care at least somewhat about social welfare. This is why we get to have nice things and not descend into chaos.
Elected governments have the most moral authority to take actions that effect everyone, ideally a diverse group of nations as mentioned in Daniel Kokotajlo’s maximal proposal comment.
Someone like Paul Graham or Tyler Cowen is noticing more smarter kids, because we now have much better systems for putting the smarter kids into contact with people like Paul Graham and Tyler Cowen.
I’d guess very smart kids are getting more numerous and smarter at the elite level since I’d guess just about everything is improving at the most competitive level. Unfortunately it doesn’t seem like there’s much interest in measuring this, e.g. hundreds of kids tie for the maximum score possible on SATs (1600) instead of designing a test that won’t max out.
(Btw, one cool thing I learned about recently is that some tests use dynamic scoring where if you get questions correct the system asks you harder questions.)
Is there any information on how long the LLM spent on taking the tests? Any idea? I’d like to know the comparison with human times. (I realize it can depend on hardware, etc but would just like some general idea.)
I’m trying to understand this paper on AI Shutdown Problem https://intelligence.org/files/Corrigibility.pdf but can’t follow the math formulas. Is there a code version of the math?
The below is wrong, but I’m looking for something like this:
Again, the above is not meant to be correct but to maybe go somewhere towards problem understanding if improved.