X Addresses Obscene Content Issues Related to Grok AI and Takes Action

Grok AI Obscene Content Issue: X Accepts Mistake, Removes 600 Accounts; Assures Compliance

On January 11, 2026, the microblogging platform X acknowledged its errors in managing obscene content associated with the AI chatbot Grok, resulting in the removal of approximately 3,500 pieces of content and the deletion of over 600 accounts. This decisive action followed concerns raised by the Indian IT Ministry regarding inappropriate materials linked to the Elon Musk-backed AI.

Compliance with Indian Laws

In light of the government’s concerns, X has assured compliance with Indian laws and has taken significant steps to rectify the situation. The platform has committed to preventing obscene imagery in the future, emphasizing its dedication to upholding legal standards.

Global Pressure and Content Moderation

The scrutiny of Grok is not limited to India; X is facing increasing pressure from governments worldwide as regulators intensify their oversight of generative AI technologies. Concerns have escalated regarding the moderation of content, data safety, and the proliferation of non-consensual sexually explicit images on the platform.

According to sources, X’s ‘Safety’ handle has reiterated its commitment to combating illegal content, including Child Sexual Abuse Material (CSAM). The platform has stated that users who leverage Grok to create illegal content will face the same consequences as those who upload such content directly.

Government Directives and Accountability

On January 2, the IT Ministry demanded immediate action from X to remove all vulgar, obscene, and unlawful content generated by Grok. The ministry issued a directive requiring X to submit a detailed action taken report (ATR) within 72 hours, outlining the measures adopted in relation to the Grok application.

The directive highlighted the misuse of Grok AI, noting that users were creating fake accounts to generate and share derogatory or vulgar images of women. The Ministry underscored the seriousness of this issue, indicating a significant failure of platform-level safeguards.

Consequences of Non-Compliance

The Indian government clarified that adherence to the IT Act and its rules is mandatory. X was warned that any failure to comply could lead to the loss of statutory exemptions under section 79 of the IT Act, exposing the platform to liability under various laws, including the IT Act and the Bharatiya Nyaya Sanhita.

Future Measures and Industry Reactions

In response to the government’s concerns, X has agreed to enforce its user terms and implement AI usage restrictions more rigorously. This includes strong deterrent measures such as account suspension and termination for violating users. The platform has also been instructed to remove or disable access to any content that has already been disseminated in violation of applicable laws.

Despite X’s extensive response to the initial notice, the IT Ministry deemed it lacking in critical information regarding specific actions taken against obscene content linked to Grok. Consequently, the government has intensified its demands for more detailed compliance measures.

As X navigates this complex landscape, it faces not only scrutiny from the Indian government but also pressure from the UK and EU over similar issues concerning Grok. The global conversation continues, with public and governmental calls for stricter regulations on the platform’s content moderation practices.

More Insights

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Embracing Responsible AI to Mitigate Legal Risks

Businesses must prioritize responsible AI as a frontline defense against legal, financial, and reputational risks, particularly in understanding data lineage. Ignoring these responsibilities could...

AI Governance: Addressing the Shadow IT Challenge

AI tools are rapidly transforming workplace operations, but much of their adoption is happening without proper oversight, leading to the rise of shadow AI as a security concern. Organizations need to...

EU Delays AI Act Implementation to 2027 Amid Industry Pressure

The EU plans to delay the enforcement of high-risk duties in the AI Act until late 2027, allowing companies more time to comply with the regulations. However, this move has drawn criticism from rights...

White House Challenges GAIN AI Act Amid Nvidia Export Controversy

The White House is pushing back against the bipartisan GAIN AI Act, which aims to prioritize U.S. companies in acquiring advanced AI chips. This resistance reflects a strategic decision to maintain...

Experts Warn of EU AI Act’s Impact on Medtech Innovation

Experts at the 2025 European Digital Technology and Software conference expressed concerns that the EU AI Act could hinder the launch of new medtech products in the European market. They emphasized...

Ethical AI: Transforming Compliance into Innovation

Enterprises are racing to innovate with artificial intelligence, often without the proper compliance measures in place. By embedding privacy and ethics into the development lifecycle, organizations...

AI Hiring Compliance Risks Uncovered

Artificial intelligence is reshaping recruitment, with the percentage of HR leaders using generative AI increasing from 19% to 61% between 2023 and 2025. However, this efficiency comes with legal...