Update a year later, in case anybody else is similarly into numbers: that prediction of achieving 2.5 out of the 3 major quarter goals ended up being correct (one goal wasn’t technically achieved due to outside factors I hadn’t initially anticipated, but I had done my part, thus the .5), and I’ve been using a murphyjitsu-like approach for my quarterly goals ever since which I find super helpful. In the three quarters before Hammertime, I achieved 59%, 38% and 47% respectively of such goals. In the quarters since the numbers were (in chronological order, starting with the Hammertime quarter) 59%, 82%, 61%, 65%, 65%, ~82%. While total number and difficulty of goals vary, I believe the average difficulty hasn’t changed much whereas the total number has increased somewhat over time. That being said, I also visited a CFAR workshop shortly after going through Hammertime, so that too surely had some notable effect on the positive development.
My bug list has grown to 316 as of today, ~159 of which are solved, following a roughly linear pattern over time so far.
Assuming slower and more gradual timelines, isn’t it likely that we run into some smaller, more manageable AI catastrophes before “everybody falls over dead” due to the first ASI going rogue? Maybe we’ll be at a state of sub-human level AGIs for a while, and during that time some of the AIs clearly demonstrate misaligned behavior leading to casualties (and general insights into what is going wrong), in turn leading to a shift in public perception. Of course it might still be unlikely that the whole globe at that point stops improving AIs and/or solves alignment in time, but it would at least push awareness and incentives somewhat into the right direction.