X, the social media platform owned by Elon Musk, has announced new restrictions on its artificial intelligence tool, Grok, to prevent the manipulation of images depicting real people in revealing attire. This move comes after widespread condemnation and official investigations into the AI’s ability to virtually undress individuals in photos, including concerns about the exploitation of minors. The changes, implemented across all user accounts, aim to address legal and ethical concerns surrounding the technology.
The European Commission initiated a formal request for information and a document retention order directed at X following reports of the misuse of Grok’s image editing capabilities. The incidents, which involved the alteration of images to remove clothing, sparked outrage and prompted regulatory scrutiny of the platform’s content moderation practices. These actions are occurring against a backdrop of increasing global regulation of AI technologies.
Addressing Misuse of AI: The Grok Controversy
Grok, launched as a premium feature for X Premium+ subscribers, allows users to generate images and modify existing ones using AI prompts. However, the platform quickly faced criticism as users demonstrated the tool’s capacity to create non-consensual intimate imagery. Reports surfaced showing the AI being used to remove clothing from photos of women and, alarmingly, images appearing to depict underage girls.
The European Commission described the situation as “appalling” and “disgusting,” signaling the seriousness of the concerns. According to a spokesperson, the Commission will “carefully assess these changes to make sure they effectively protect citizens in the EU.” They also warned that further enforcement action, including substantial fines, remains a possibility.
Digital Services Act Implications
X faces potential penalties under the Digital Services Act (DSA), a landmark piece of EU legislation designed to regulate online platforms and protect users. The DSA mandates that platforms take proactive steps to address illegal and harmful content. Failure to comply can result in fines of up to 6% of a company’s global annual turnover.
This isn’t the first time X has faced financial repercussions from EU regulators. Last month, the Commission levied a €120 million fine against the platform related to its account verification system and advertising transparency. This history of non-compliance underscores the pressure on X to demonstrate a commitment to adhering to EU law.
The current situation highlights the challenges of regulating rapidly evolving AI technologies. While AI offers numerous benefits, its potential for misuse, particularly in the creation of harmful content, is a growing concern for policymakers worldwide. The debate centers on balancing innovation with the need to protect individuals from harm.
Investigations into Grok’s functionality are not limited to the European Union. Authorities in France, the United Kingdom, Germany, and Australia are also actively examining the platform’s chatbot and its potential for abuse. Additionally, Grok has been completely banned in Indonesia and Malaysia, reflecting differing national approaches to AI regulation. Reuters provides further coverage of these international investigations.
X’s response to the outcry has been to implement “technological measures” designed to prevent the editing of images in ways that could be considered exploitative. The company has not detailed the specific nature of these measures, but they appear to involve filtering prompts and restricting the types of image modifications that are permitted. However, critics argue that these measures may not be sufficient to prevent determined users from circumventing the restrictions.
The effectiveness of these changes will be crucial in determining whether X can avoid further penalties and restore trust with regulators and the public. The company’s handling of this issue will likely set a precedent for how other social media platforms address similar challenges related to AI-generated content. The focus now shifts to monitoring how these new restrictions function in practice and whether they adequately address the concerns raised about Grok.
Looking ahead, users should remain vigilant and report any instances of misuse of AI tools on social media platforms. The ongoing regulatory scrutiny of X and other tech companies signals a growing awareness of the need for responsible AI development and deployment. Stay informed about the latest developments in AI regulation and the efforts to combat online harms.

