If I understood correctly, he mentions augmenting humans as a way out of the existential risk. At least I understood he has more faith in it than in making AI do our alignment homework. What does he mean by that? Increasing productivity? New drug development? Helping us get insights into new technology to develop? All of the above? I’d love to understand the ideas around that possible way out.
If I understood correctly, he mentions augmenting humans as a way out of the existential risk. At least I understood he has more faith in it than in making AI do our alignment homework. What does he mean by that? Increasing productivity? New drug development? Helping us get insights into new technology to develop? All of the above? I’d love to understand the ideas around that possible way out.