"Do Anything Now": Characterizing and Evaluating...
The misuse of large language models (LLMs) has drawn significant attention from the general public and LLM vendors. One particular type of adversarial prompt, known as jailbreak prompt, has...
https://arxiv.org/abs/2308.03825