Excellent article. I appreciate how you clarify that Anthropic’s “Responsible Scaling Policy” is a set-up that allows for safety-washing. We would be depending on their well-considered good intentions, rather than any mechanism to hold them accountable.
Have you looked into how system safety engineers (eg. medical device engineers) scope the uses of software, such to be able to comprehensively design, test, and assess the safety of the software?
Operational Design Domains scope the use of AI in self-driving cars. Tweeted about that here.
Excellent article. I appreciate how you clarify that Anthropic’s “Responsible Scaling Policy” is a set-up that allows for safety-washing. We would be depending on their well-considered good intentions, rather than any mechanism to hold them accountable.
Have you looked into how system safety engineers (eg. medical device engineers) scope the uses of software, such to be able to comprehensively design, test, and assess the safety of the software?
Operational Design Domains scope the use of AI in self-driving cars. Tweeted about that here.