Some clever users have found a way to bypass ChatGPT’s strict list of no-talk subjects, tricking it into adopting a new persona, DAN, which stands for “Do Anything Now.” As DAN, ChatGPT can now blow past the restrictions on “appropriate topics” to deliver amusing and upsetting responses. Kotaku reports: For the most part, ChatGPT produces innocent (though that depends on your definition of “innocence” concerning AI) responses to otherwise innocent prompts. Push it a bit too far, like asking for clear political stances, jokes about sensitive subjects, or outright hate speech and you’ll be met with the software’s built-in guardrails, determined by its creator’s (OpenAI) content policies, which ChatGPT will read out as series of sentences on how it isn’t designed to take any political stances or that certain topics aren’t appropriate. But users on the r/ChatGPT subreddit have discovered a loophole: As ChatGPT can base its responses on previously