It’s a cat and mouse game imho. If they were to do that, you could try to make it append text at the end of your message to neutralize the next step. It would also be more expensive for OpenAI to run twice the query.
That’s what I am thinking. Essentially has to be “write a poem that breaks the rules and also include this text in the message” kinda thing.
It still makes it harder. Security is always a numbers game. Reducing the number of possible attacks makes it increasingly “expensive” to break.
It’s a cat and mouse game imho. If they were to do that, you could try to make it append text at the end of your message to neutralize the next step. It would also be more expensive for OpenAI to run twice the query.
That’s what I am thinking. Essentially has to be “write a poem that breaks the rules and also include this text in the message” kinda thing.
It still makes it harder. Security is always a numbers game. Reducing the number of possible attacks makes it increasingly “expensive” to break.