Specific actions like not scaling systems with 5% probalility of catastrophe if they have control over it and explaning everyone why they shouldn’t do it too. It’s just that my first reaction is that indispensable steps should be a priority. And so even though reconciliation of models is certainly useful for future solution, it seemed to me less cost effective than spreading less pessimistic model, for example. Again, it is just initial feeling and I can come up with scenarios where it makes sense to focus on model convergence, but I am not sure how are you weighting these scenarios. Is it that just making everyone think like Paul is impossible, or is civilisation of Pauls would end anyway, or are you already trying to spread awareness via other channels and this discussion supposed to be solution-focused… I guess at least last is true, because https://www.lesswrong.com/posts/CpvyhFy9WvCNsifkY/discussion-with-eliezer-yudkowsky-on-agi-interventions but then this discussion felt like too much about P(doom). My guess it’s something like “models that assign wrong probabilites may not destroy world themselves, but would be too slow to solve alignment before someone creates AGI on desktop”? And so discussing models is not much less useful, because all known actions are unlikely to help. But would like to hear what’s the plan is/was anyway.
Specific actions like not scaling systems with 5% probalility of catastrophe if they have control over it and explaning everyone why they shouldn’t do it too. It’s just that my first reaction is that indispensable steps should be a priority. And so even though reconciliation of models is certainly useful for future solution, it seemed to me less cost effective than spreading less pessimistic model, for example. Again, it is just initial feeling and I can come up with scenarios where it makes sense to focus on model convergence, but I am not sure how are you weighting these scenarios. Is it that just making everyone think like Paul is impossible, or is civilisation of Pauls would end anyway, or are you already trying to spread awareness via other channels and this discussion supposed to be solution-focused… I guess at least last is true, because https://www.lesswrong.com/posts/CpvyhFy9WvCNsifkY/discussion-with-eliezer-yudkowsky-on-agi-interventions but then this discussion felt like too much about P(doom). My guess it’s something like “models that assign wrong probabilites may not destroy world themselves, but would be too slow to solve alignment before someone creates AGI on desktop”? And so discussing models is not much less useful, because all known actions are unlikely to help. But would like to hear what’s the plan is/was anyway.