I really like learning new things!
Jacob G-W
The Puritans would one-box: evidential decision theory in the 17th century
Google Gemini Announced
Why it’s necessary to shoot yourself in the foot
From the outside, American schooling is weird
I made a manifold market for if this will replicate: https://manifold.markets/g_w1/will-george3d6s-increasing-iq-is-tr I’m not really sure what the resolution criteria should be, so I just made some that sounded reasonable, but feel free to give suggestions.
Noticing confusion in physics
A question I have for the subjects in the experimental group:
Do they feel any different? Surely being +0.67 std will make someone feel different. Do they feel faster, smoother, or really anything different? Both physically and especially mentally? I’m curious if this is just helping for the IQ test or if they can notice (not rigorously ofc) a difference in their life. Of course, this could be placebo, but it would still be interesting, especially if they work at a cognitively demanding job (like are they doing work faster/better?).
Here’s a market if you want to predict if this will replicate: https://manifold.markets/g_w1/will-george3d6s-increasing-iq-is-tr
[linkpost] Self-Rewarding Language Models
Taking into account preferences of past selves
[Question] Concrete examples of doing agentic things?
@lsusr recently did a video about this. Interestingly, he thought that the hardest people to love were not actually the Hitler type (they are still hard), but the people that you are actively hurting.
Sure, but they only use 16 frames, which doesn’t really seem like it’s “video” to me.
Understanding video input is an important step towards a useful generalist agent. We measure the video understanding capability across several established benchmarks that are held-out from training. These tasks measure whether the model is able to understand and reason over a temporally-related sequence of frames. For each video task, we sample 16 equally-spaced frames from each video clip and feed them to the Gemini models. For the YouTube video datasets (all datasets except NextQA and the Perception test), we evaluate the Gemini models on videos that were still publicly available in the month of November, 2023
I think the idea actually works pretty well with superintelligence (with one big exception if you assume we all die). Lots of people don’t understand how/why superintelligence could kill us all. They naively think that creating a superintelligence would be a great idea. If we all died, then they would understand why alignment is a necessary complexity. The only problem with this is that we are all dead.
In set theory, everything is a set
XAI releases Grok base model
Do you think this is permanent? Or will you have to keep up all of the interventions for it to stay +13points indefinitely?
This makes sense. I’ve changed my mind, thanks!
It seems to do something similar to Gato where everything is just serialized into tokens, which is pretty cool
I wonder if they are just doing a standard transformer for everything, or doing some sort of diffusion model for the images inside the model?
I really enjoyed this post. Thank you for writing it!
I also have no clue what is going to happen. I predict that it will be wild, and I also predict that it will happen in <=10 years. Let’s fight for the future we want!