The researchers are applying a way called adversarial instruction to halt ChatGPT from letting customers trick it into behaving poorly (called jailbreaking). This work pits many chatbots in opposition to each other: a person chatbot plays the adversary and attacks Yet another chatbot by creating textual content to pressure it https://advicebookmarks.com/story25116263/gpt-chat-an-overview