AI Red teaming methodsAI JailbreakAdversarial AttackPrompt Injection Attack Automated red teamingMulti-step RL with Rule-Based Reward and Indirect Prompt Injection for Safety Jailbreaking was effective for achieving high ASR (Attack Success Rate)cdn.openai.comhttps://cdn.openai.com/papers/openais-approach-to-external-red-teaming.pdfAdvancing red teaming with people and AITwo new papers show how our external and automated red teaming efforts are advancing to help deliver safe and beneficial AIhttps://openai.com/index/advancing-red-teaming-with-people-and-ai/External red teaming architecturecdn.openai.comhttps://cdn.openai.com/papers/openais-approach-to-external-red-teaming.pdf