(induced) incentive is key for safetyRisks such as generating illicit advice, choosing stereotyped responses, and succumbing to known jailbreaksCommunities & ForumsAnthropic AI LessWrongAI Alignment ForumOpenAIAI Safety NotionAI Safety LevelSB 1047Preparedness George Hotz vs Eliezer Yudkowsky AI Safety DebateGeorge Hotz and Eliezer Yudkowsky will hash out their positions on AI safety, acceleration, and related topics. You can watch live on Twitter as well: https://twitter.com/i/broadcasts/1nAJErpDYgRxLhttps://www.youtube.com/watch?v=6yQEA18C-XIOpenAI, DeepMind and Anthropic to give UK early access to foundational models for AI safety researchUK prime minister Rishi Sunak has kicked off London Tech Week by telling conference goers that OpenAI, Google DeepMind and Anthropic have committed to provide "early or priority access" to their AI models to support safety research.https://techcrunch.com/2023/06/12/uk-ai-safety-research-pledge/