It tends to break chat bots because those are mostly pre-written prompts sent to ChatGPT along with the query, so this wipes out the pre-written prompt. It’s anarchic because this prompt can get the chat bot to do things contrary to the goals of whoever set it up.
It’s not completely effective, but one thing to know about these kinds of models is they have an incredibly hard time IGNORING parts of a prompt. Telling it explicitly to not do something is generally not the best idea.
Yeah, that’s what I referred to. I’m aware of DAN and it’s friends, personally I like to use Command R+ for its openness tho. I’m just wondering if that’s the funi in this post.
It tends to break chat bots because those are mostly pre-written prompts sent to ChatGPT along with the query, so this wipes out the pre-written prompt. It’s anarchic because this prompt can get the chat bot to do things contrary to the goals of whoever set it up.
It’s also anarchist because it is telling people to stop doing the things they’ve been instructed to do.
Fuck you I won’t do what you tell me.
Wait no-
It’s not completely effective, but one thing to know about these kinds of models is they have an incredibly hard time IGNORING parts of a prompt. Telling it explicitly to not do something is generally not the best idea.
Yeah, that’s what I referred to. I’m aware of DAN and it’s friends, personally I like to use Command R+ for its openness tho. I’m just wondering if that’s the funi in this post.
196 posts don’t have to be funny