Even if I’ve done all I can directly for my own health, until we reach longevity escape velocity, pushing longevity technology, and supporting any direct (computers, genetic engineering) or indirect (cognitive enhancement, productivity enhancements) technologies would seem to give more egoistic any utilitarian bang for the buck, at least if you’re focused on the utility of actually existing people.
Hmm, that’s a tough call. I note however that at that point, where your marginal dollar goes is more a matter of a cost-benefit calculation than a real difference in preferences (I also mostly care about currently existing people).
The question is, which will maximize life expectancy ? If you estimate that existential risks are sufficiently near and high, you would reduce them. If they are sufficiently far and low, you’d go for life extension first.
I reckon It depends on a range of personal factors, not least of which your own age. You may very well estimate that if you where not egoist, you’d go for existential risks, but maximizing your own life expectancy calls for life extension. Even then, that shouldn’t be a big problem for altruists. Because at that point, you’re doing good for everyone anyway.
Even if I’ve done all I can directly for my own health, until we reach longevity escape velocity, pushing longevity technology, and supporting any direct (computers, genetic engineering) or indirect (cognitive enhancement, productivity enhancements) technologies would seem to give more egoistic any utilitarian bang for the buck, at least if you’re focused on the utility of actually existing people.
Hmm, that’s a tough call. I note however that at that point, where your marginal dollar goes is more a matter of a cost-benefit calculation than a real difference in preferences (I also mostly care about currently existing people).
The question is, which will maximize life expectancy ? If you estimate that existential risks are sufficiently near and high, you would reduce them. If they are sufficiently far and low, you’d go for life extension first.
I reckon It depends on a range of personal factors, not least of which your own age. You may very well estimate that if you where not egoist, you’d go for existential risks, but maximizing your own life expectancy calls for life extension. Even then, that shouldn’t be a big problem for altruists. Because at that point, you’re doing good for everyone anyway.