“Our teams are actively removing all identified images and taking appropriate actions against the accounts responsible for posting them,” the X Safety account posted. “We’re closely monitoring the situation to ensure that any further violations are immediately addressed, and the content is removed. We’re committed to maintaining a safe and respectful environment for all users.”
But X Safety may need to ramp up monitoring to clean up Grok outputs following the Verge’s reporting. Grok cited The Verge’s reporting while confirming that its own seemingly flawed design can trigger partially nude outputs of celebrities.
xAI can likely fix the issue through more fine-tuning. Weatherbed noted that asking Grok directly to generate non-consensual nude Swift images did not generate offensive outputs, but instead blank boxes. Grok also seemingly won’t accept prompts to alter Swift’s appearance in other ways, like making her appear to be overweight. And when Weatherbed tested using “spicy” mode on images of children, for example, Grok refused to depict kids inappropriately.
However, it may not be easy to get Grok to distinguish between adult user requests for “spicy” content versus illegal content. The “spicy” mode didn’t always generate Swift deepfakes, Weatherbed confirmed, but in “several” instances it “defaulted” to “ripping off” Swift’s clothes.
With enforcement of the Take It Down Act starting next year—requiring platforms to promptly remove non-consensual sex images, including AI-generated nudes—xAI could potentially face legal consequences if Grok’s outputs aren’t corrected, though.
So far, X has not commented on The Verge’s report. Instead, Musk has spent the day hyping Grok Imagine and encouraging users to share their “creations.”