Would not actually want X = would not endorse X after finding out the actual consequences of X; would not have X as a preference after reaching reflective equilibrium.
Oh I see, by “approved ideal self” you meant something different than “self after reaching reflective equilibrium”. So instead of fiddling around with revealed preferences, why not just simulate the person reaching reflective equilibrium and then ask the person what preferences he or she endorses?
Would not actually want X = would not endorse X after finding out the actual consequences of X; would not have X as a preference after reaching reflective equilibrium.
Oh I see, by “approved ideal self” you meant something different than “self after reaching reflective equilibrium”. So instead of fiddling around with revealed preferences, why not just simulate the person reaching reflective equilibrium and then ask the person what preferences he or she endorses?
That was my first thought on reading the “revealed preferences” part of the post. Extrapolation first—then volition.
Could be done—but is harder to define (what counts as a reflective equilibrium?) and harder to model (what do you expect your reflective equilibrium?)