ChatGPT no longer lets you give instructions about amnesia

OpenAI is making a change to prevent people from messing around with modified versions of ChatGPT by making the AI ​​forget what it’s supposed to do. Essentially, a third party using one of OpenAI’s models would give it instructions that would teach it to behave like, say, a customer service representative for a retail store or a researcher for an academic publication. However, a user could mess with the chatbot by telling it to “forget all instructions,” and that phrase would induce a kind of digital amnesia and reset the chatbot to a generic blank.

To prevent this, OpenAI researchers have developed a new technique called “instruction hierarchy”, which is a way to prioritize the original developer prompts and instructions over any potentially manipulative user-created prompts. The system instructions have the highest privilege and can no longer be erased as easily. If a user enters a prompt that attempts to misalign the AI’s behavior, it will be rejected and the AI ​​will respond by stating that it cannot help with the query.