The researchers are using a way identified as adversarial teaching to prevent ChatGPT from allowing end users trick it into behaving badly (generally known as jailbreaking). This function pits numerous chatbots in opposition to one another: a single chatbot plays the adversary and assaults A different chatbot by producing text to drive it to buck i