> My current (2025–2026) belief system says that strategic capabilities can be decoupled from moral capabilities, but that it turns out in practice that the most efficient way to get strategic capabilities involves learning basically all human concepts and “correcting” them (finding more coherent explanations)
(Possibly this is addressed somewhere in that dialogue, but anyway:)
Wouldn’t this imply that frontier LLMs are better than humans at ~[(legible) moral philosophy]?
(Possibly this is addressed somewhere in that dialogue, but anyway:)
Wouldn’t this imply that frontier LLMs are better than humans at ~[(legible) moral philosophy]?