Thanks for your reply.
To the point of contention, I believe it is actually fairly well illustrated on the website, that it is a fallacy in itself to make it up to such nuances and to demand proof of any particular of the possible outcomes in the future to shape one’s actions, if it pertains to the question of whether or not you should do your bests to mitigate the risk in ensuring your basic survival.
An unsurvivable AGI outcome is just one of the many possible scenarios. Although you can speculate about the details of how it could play out (partial extermination, full extermination, no extermination) and what means AGI might use, the whole point of thinking logically about the issue in terms of ensuring your survival is to recognize that those things are ultimately unknowable. What you have to do is to simply develop a strategy that deals with all eventualities as best as possible.
I don’t know if it is clear to you, but those are the basic scenarios of AGI development taken into consideration on the website, which are basically everything you can consider:
AGI will be docile, that is super-intelligent but fully obedient to humans
-> society can collapse due to abuse or misuse (ranging from mild to severe)
-> society can collapse due to system shock (ranging from mild to severe)
-> society can flourish due to benefitsAGI will become fully independent
-> accidental extinction due to insanity (ranging from mild to severe)
-> deliberate extinction to eliminate humans as a threat (ranging from mild to severe)
-> coincidental extinction due to disregard for human life (ranging from mild to severe)
-> AGI will become god or a guardian angle and help or digitalize human existence
-> AGI will leave on a space ship and simply leave us alone, because we are insignificant and there is an infinite amount of planets to exploit resources fromExperimental stages of AGI lead to some kind of destruction
AGI will not develop in the near future, because it is too complicated or outlawed
AGIs will fight against each other and may cause some kind of detrimental chaos
Obviously you can also mix those scenarios, as they all derive from each other, and they all can be rather permanent or temporary in nature, mild or severe as well. For example you could have a collapsing stock market, due to system shock of hyper-intelligent but fully human-controlled AI systems, all the while independent AGI systems are on the rise 2 years later and then waging digital wars against each other and some of them try to exterminate humans, while others try to protect them or enslave them. While this seems somewhat ridiculous to consider, it is just to illustrate that there is a wide range of possible outcomes with a wide range of details, and no single definitive outcome (e.g. full extinction of every last human on earth from AGI) can be determined with any degree of certainty.
In the end in most scenarios, society will recover in time and some amount of human life will still be present afterwards. But even if there was just a single scenario with the off chance of human survival, then this would be enough to work towards by literally spending all your time and resources on it. Anything else can only be described as suicidal.
Personally though, I believe the chances of overall human survival are very high, and the chances of hostile AGI are rather low. This is exactly how it is reflected by expert opinion as well. But there is a lot of in-between risk that you need to consider, most of which concerns intermediary stages of AGI, and this is where spending reasonable amounts of money (e.g. 2000 Euros) come into play.
So I was hoping you can help me to improve the page and tell me how that was not clear from reading it through and maybe how I can write it differently, without adding thousands of words to the page.
This guide was not written for LessWrong btw., but for common people who are smart enough to follow the arguments and are willing to protect themselves if necessary in light of this new information and proper consideration and risk management.
Only live-attenuated vaccines may (sometimes) not need adjuvants. Plus you sometimes have other ingredients acting as adjuvants that are not declared as such. For example mercury is declared as a preservative, not adjuvant, but it performs the same function. Also as of recent they started removing constituents from the ingredient list, that were part of the manufacturing process (e.g. culture media), but are not “intended” part of the final product. If a food manufacturer washes potatoes with iodine for example in order to clean them, he is not required to list that as an ingredient, regardless of whether or not quantities in the final product are relevant.
To put simply without a live virus, the immune system recognizes the would-be antigens as simply garbage molecules, and not as a threat. In order for immunization to work, you need to inject something dangerous like a live virus, aluminium, some kind of toxic protein or cytokine alongside the antigen.