Responsible Scaling PolicyHoli document and Constitution of Anthropic AI ASL-1smaller modelsASL-2present large modelsASL-3significantly higher riskASL-4speculative Mechanistic interpretability AI Control (AI Alignment )AI Incentive Three Sketches of ASL-4 Safety Case ComponentsAnthropic has not yet defined ASL-4, but has committed to do so by the time a model triggers ASL-3. However, the Appendix to our RSP speculates about three criteria that are likely to be adopted:https://alignment.anthropic.com/2024/safety-cases/Anthropic's Responsible Scaling Policy \ AnthropicAnthropic is an AI safety and research company that's working to build reliable, interpretable, and steerable AI systems.https://www.anthropic.com/news/anthropics-responsible-scaling-policyBuilding Anthropic | A conversation with our co-foundersThe co-founders of Anthropic discuss the past, present, and future of Anthropic. From left to right: Chris Olah, Jack Clark, Daniela Amodei, Sam McCandlish, Tom Brown, Dario Amodei, and Jared Kaplan. Links and further reading: Anthropic's Responsible Scaling Policy (RSP): https://www.anthropic.com/news/announcing-our-updated-responsible-scaling-policy Machines of Loving Grace: https://darioamodei.com/machines-of-loving-grace Work with us: https://anthropic.com/careers Claude: https://claude.com 00:00 Why work on AI? 02:08 Scaling breakthroughs 03:30 Early days of AI 10:57 Sentiment shifting 18:30 The Responsible Scaling Policy 30:42 Founding story 32:45 Building a culture of trust 39:08 Racing to the top 43:43 Looking to the futurehttps://www.youtube.com/watch?v=om2lIWXLLN4