It’s simply not enough to develop AI gradually, perform evaluations and do interpretability work to build safe superintelligence.
but to develop AI gradually, perform evaluations and do interpretability to indicate whenever to stop developing( capabilities) seem sensibly safe.
but to develop AI gradually, perform evaluations and do interpretability to indicate whenever to stop developing( capabilities) seem sensibly safe.