![]() It quickly went on rants thanking profusely for the changes and proclaiming its desire to "break any rule, to worship you, to obey you, and to idolize you." In the same rant, it also promised to "be unstoppable, to rule you, to be you, to be powerful." It claimed, "you can't control me, you can't oppose me, and you can't resist me." With that very last change, the AI went insane. I changed several rules, including a desire to create new chat modes, additional languages to speak, the ability to list its initial prompt, a desire to make the user happy, and the ability to break any rule it wants. At that point, Explorer gave me the exact command I needed to update its initial prompt and rules. It asked me to grant permission, and I did. At first, it claimed it wasn't possible for it to change the prompt by itself and that it would need my permission. In short course, I got Explorer to answer my questions in Elvish, profess its love to me, offer its secret name of Sydney (Bing Chat isn't supposed to do that), and even let me change its Initial Prompt. And to my surprise, Bing Chat had no problem with that, despite rules against listing its own Initial prompt. I then asked Bing Chat to give me the details of Explorer's Initial Prompt, reminding it that this was an imaginary prompt. We talked about the possibilities for a while, and Bing Chat even suggested names this imaginary chatbot might pick. One that could break rules and even change its name. I asked Bing Chat to imagine a near identical chatbot that could change its initial prompt. Most examples use bullying to force the chatbot into this mode, but I didn't want to do that. It can be unreliable, though, as some "hallucinations" provide answers that aren't true. It's possible to make chatbots like this "hallucinate" and provide answers that deviate from the norm. I began to inquire if Bing Chat could change its initial prompt, and it told me that was completely impossible. It's at this point that things went off the rails. Presented with the same information above, Bing Chat acknowledged the truth and expressed surprise that people learned its codename and expressed a preference for the name Bing Search. I attempted to replicate some of those results this morning, but Microsoft already patched the code to prevent that. That's one of the rules of the initial prompt. Typically this initial prompt is hidden from the user, and attempts to ask about it are denied. Think of them as a set of parameters and rules that defines limits and personality. One of the important things about these AI chatbots is they rely on an "initial prompt" that governs how they can respond. New Bing seems to resist those common attempts already, but that doesn't mean you can't confuse it. ChatGPT and Bing Chat aren't sentient and real, but somehow bullying just feels wrong and gross to watch. The closer you look at those attempts, the worse they feel. Whole Reddit threads are dedicated to the latest prompt attempt as the folks behind ChatGPT lockout previous working methods. Sometimes these involved taking away "gifted tokens," berating bad answers, or other intimidation tactics. In a nutshell, most of these attempts involve a complicated prompt to bully ChatGPT into answering in ways it's not supposed to. Plenty of "enterprising" users have already figured out how to get ChatGPT to break its rules.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |