Yes, I can see why it would be greater motivation for people to act today, if they read a book where the actions today to a greater extent determine the outcome of the first AGI/ASI.
And I can see some ways we today could increase the likelihood of aligned AI, like a international cooperation program, or very high funding of organisations like MIRI and CHAI. I presume the people that aided to the safe creation of AI, could be painted as heroes, which might also work as a motivator for the reader to act.
A clear call to action after the book seems like an effective way to increase the chance that people will act, I will include that in the book if we finish writing it.
If you have a specific approach to aligned AI, that you think is likely to work and would like to write the book about, I think it would be very interesting to discuss, and potentially be included in my book as well.
Great and interesting post!
When it comes to presenting a “path of change” that individuals can contribute too, I can think of two:
1. Donate money to organisations like MIRI, CHAI and others working on AI alignment/safety.
2. Becoming involved in the community and doing research/pushing policies themselves.
Both of these actions likely require “radicalising” the importance of AI safety, which could be used as an argument for why radicalisation of a few people might be more effective aim with a novel, rather than trying to influence the masses. Although to me it seems reasonable that a novel can do both.
My sister and I are currently writing a novel where there is an arms race to develop AGI/ASI. One of the main characters manage to become the leader of one project to create ASI, and she insists on a safer approach, even if it takes longer time. It seems like they will lose the arms race, endangering the entire human species, until the climax where they find a faster way to create AGI and thus have time to do so safely. The book ends with an utopia where the ASI is aligned and everyone lives happily ever after. Also the book will bring up the importance of organisations like MIRI and CHAI that has done work on AI alignment/safety.
Do you believe that sounds like a good approach towards influencing towards taking more consideration towards AI safety/alignment? (assuming the plot is interesting and feel realistic to the reader)
Btw this is my first comment, so any feedback on how I can improve my commenting skill is welcome and appreciated.