Demystifying “Alignment” through a Comic

I am keeping this comic online as a record of an earlier perspective. My current work in fixed point logic and proof theory has led me to reject several foundational assumptions in mainstream alignment and interpretability research. This does not imply that AI poses no risks (it implies nothing about AI).