11 Sources
11 Sources
[1]
India orders Musk's X to fix Grok over "obscene" AI content | TechCrunch
India has ordered Elon Musk's X to make immediate technical and procedural changes to its AI chatbot Grok after users and lawmakers flagged the generation of "obscene" content, including AI-altered images of women created using the tool. On Friday, India's IT ministry issued the order directing Musk's X to take corrective action on Grok, including restricting the generation of content involving "nudity, sexualization, sexually explicit, or otherwise unlawful" material. The ministry also gave the social media platform 72 hours to submit an action-taken report detailing the steps it has taken to prevent the hosting or dissemination of content deemed "obscene, pornographic, vulgar, indecent, sexually explicit, pedophilic, or otherwise prohibited under law." The order, reviewed by TechCrunch, warned that failure to comply could jeopardize X's "safe harbor" protections -- legal immunity from liability for user-generated content under Indian law. India's move follows concerns raised by users who shared examples of Grok being prompted to alter images of individuals -- primarily women -- to make them appear to be wearing bikinis, prompting a formal complaint from Indian parliamentarian Priyanka Chaturvedi. Separately, recent reports flagged instances in which the AI chatbot generated sexualized images involving minors, an issue X acknowledged earlier on Friday was caused by lapses in safeguards. Those images were later taken down. However, images generated using Grok that made women appear to be wearing bikinis through AI alteration remained accessible on X at the time of publication, TechCrunch found. The latest order comes days after the Indian IT ministry issued a broader advisory on Monday, which was also reviewed by TechCrunch, to social media platforms, reminding them that compliance with local laws governing obscene and sexually explicit content is a prerequisite for retaining legal immunity from liability for user-generated material. The advisory urged companies to strengthen internal safeguards and warned that failure to do so could invite legal action under India's IT and criminal laws. "It is reiterated that non-compliance with the above requirements shall be viewed seriously and may result in strict legal consequences against your platform, its responsible officers and the users on the platform who violate the law, without any further notice," the order warned. The Indian government said non-compliance could lead to action against X under India's IT law and criminal statutes. India, one of the world's biggest digital markets, has emerged as a critical test case for how far governments are willing to go in holding platforms responsible for AI-generated content. Any tightening of enforcement in the country could have ripple effects for global technology companies operating across multiple jurisdictions. The order comes as Musk's X continues to challenge aspects of India's content regulation rules in court, arguing that federal government takedown powers risk overreach, even as the platform has complied with a majority of blocking directives. At the same time, Grok has been increasingly used by X users for real-time fact-checking and commentary on news events, making its outputs more visible -- and more politically sensitive -- than those of standalone AI tools. X and xAI did not immediately respond to requests for comment on the Indian government's order.
[2]
Grok Deepfake Crisis Puts India's Intermediary Liability Framework to the Test | AIM
With Grok releasing deepfakes, MeitY asserted that adherence to the IT Act and the IT Rules is mandatory, not optional. Elon Musk-owned xAI is in crisis mode after users on X (formerly Twitter) asked its AI chatbot Grok last week to digitally undress real women. They prompted Grok to manipulate photos, and the AI chatbot released the morphed images of women through its own handle -- a trend that began circulating on social media, causing widespread outrage and condemnation from users, celebrities, and governments. The Ministry of Electronics and Information Technology (MeitY) intervened on Wednesday and issued a notice to X, directing it to remove obscene content and flagging concerns over the misuse of Grok. In a letter addressed to X's Chief Compliance Officer for India, the Ministry flagged that Grok was being exploited by users to create fake accounts that host, generate, publish, or share obscene i
[3]
Harbour changes in different ports - The Economic Times
Social media platform X risks losing its 'safe harbour' status in India. This is due to content generated by its AI bot, Grok. The government seeks a compliance report from X. Tech companies must balance free speech with legal requirements. They need to provide solutions for unacceptable content. This ensures user trust and business interests.Microblogging site X is at risk of losing its 'safe harbour' status in India over reportedly 'obscene' content generated by X's AI bot Grok. Under Section 79 of IT Act 2000, social media sites are classified as neutral hosts not held responsible for content created by third parties - in 'safe harbour' - as long as they follow GoI rules. GoI now wants a compliance report from X over takedown requests. In its earlier avatar, Twitter had briefly been made liable for content it carried for being at odds with GoI over curation. Such conflicts will keep erupting, so long as media platforms and lawmakers have divergent 'cultural' views. Tech companies are bound by their contract with their users to allow free speech within reason. A narrower reading imposes extra parameters into their business model. But social media must operate in multiple jurisdictions to extract network effects. The compliance burden, thus, becomes a variable for media platforms. Safe harbour status is an accomplishment - not an entitlement. Tech that facilitates creation and dissemination of patently unacceptable content must also provide curating solutions in concert with lawmakers. Obvious violations can be counteracted technically without having to invoke the law. And when law enforcement agencies raise a concern, it must be acted upon, or reason must be shown for inaction. Taking the stand that blaming AI for obscenity is like blaming the pen for blasphemy serves little purpose. Content is created through human agency, and tech providers have adequate leverage to deter rogue action. It is in their interests to ensure good housekeeping to retain users and advertisers. Limits to freedom of expression are defined by culture. It works when governments outsource some policing to tech companies. Social media must conform to restrictions imposed on legacy media. Additional requirements may be necessary for tech adaptation. Tech is not ethical by itself. The ethical dimension must be provided by society it serves, and profits from.
[4]
ETtech Explainer: Why Grok's edgy AI image generator has come under fire
Amid huge backlash online, India on Friday wrote to X, highlighting its failure to observe statutory due diligence under the Indian IT rules and demanding an "action taken report." Manipulated images of Bollywood actors and Indian public figures went viral as more and more users tried the feature, using prompts such as "undress," "change clothes" and "change to a bikini." A creative tool gone wrong! New Year's Eve on X turned controversial as users misused the microblogging platform's AI assistant to generate thousands of non-consensual and sexualised images of women, children and public figures. Amid huge backlash online, India on Friday wrote to X, highlighting its failure to observe statutory due diligence under the Indian IT rules and demanding an "action taken report." Manipulated images of Bollywood actors and Indian public figures went viral as more and more users tried the feature, using prompts such as "undress," "change clothes" and "change to a bikini." It once again brings the cons of AI tools to the centre stage. ETtech takes a look at what went wrong... What happened? On December 25, X Corp executive chairman Elon Musk announced a new feature of Grok that allowed X users to edit images and videos by simply asking the AI assistant to add, remove or create elements in the file. Things went south when netizens started using the feature to 'undress women', flooding the platform with thousands of obscene images. Unlike competitors such as Google's Gemini or OpenAI's ChatGPT, which have filters on their prompts and image creation for sexual and pornographic content, Grok was positioned as 'more edgy' with fewer guardrails. "Cybersecurity should not be treated as an afterthought. It has to be a 'before thought'," said Ankush Tiwari, founder of cyber forensics startup Pi Labs AI. "We need to analyse threat vectors for large foundation models." He noted that today, AI models are being released openly. However, where are these models trained? Which geography do they come from? What kind of data are they trained on? These questions often go unanswered, Tiwari added. How has India responded? The Ministry of Electronics and Information Technology (MeitY) wrote to X on Friday evening, stating that such conduct reflects a serious failure of platform-level safeguards and enforcement mechanisms. This amounted "to gross misuse of artificial intelligence technologies in violation of applicable laws," it said. The ministry said the laws are not being adequately adhered by the platform "particularly in relation to obscene, indecent, vulgar, pornographic, paedophilic, or otherwise unlawful or harmful content which are potentially violative of extant laws." On December 29, MeitY had also issued an advisory to social media platforms, warning them that they are legally accountable to stop the spread of obscene and non-consensual content. India has been on high alert regarding deep fakes since late 2023, following viral morphed videos of Indian celebrities. As of early to late 2025, India has approximately 22 to 27 million users on X, making it one of the top countries for the platform. What do laws in India say? Legal experts said the misuse of AI tools to create non-consensual and sexualised images violates multiple Indian laws while there are no direct AI laws for platforms. Currently, such content falls under the Indecent Representation of Women (Prohibition) Act, 1986, which bans obscene or derogatory depiction of women. The Information Technology Act also applies in the case, with sections 66E and 67 dealing with privacy violations and obscene content online. While platforms enjoy intermediary immunity under Section 79, this protection is lost if illegal content is not taken down after being flagged. The recently passed Digital Personal Data Protection (DPDP) Act, 2023 mandates consent for using personal data, including photographs and facial data, and lists penalties of up to Rs 250 crore for breaches. Calling the trend "immoral" and illegal, Adithya LHS, chief executive officer at NyayNidhi an AI-powered legal-tech company, said social media platforms often rely on intermediary immunity to avoid liability. "But disabling and removing such illicit content is part of a platform's core responsibility. This clearly crosses the line," he said. Adithya said AI remains largely unregulated globally. "Regulation always lags innovation, especially when AI models are improving at an exponential pace," he said. Has X responded to this? X responded to the global outrage by hiding Grok's media tab and directed users to report violations within its stated prohibitions on non-consensual intimate imagery. "X's decision to disable Grok's media tab and claim tighter safeguards is a reactive step, but the bar should be higher," said Kanishk Gaur, CEO of digital risk management firm Athenian Tech. "Default-safe settings, strong consent checks, auditable controls, and consistent policy application, especially for features that operate in public feeds, are needed."
[5]
Grok's deepfake incident signals the escalating threat of fake AI generated content
Several countries are looking into XAI's safety measures following reports that Grok was being used to generate thousands of sexually explicit deepfake photos of women, and in some cases, of minors on social media platform X (formerly Twitter). Since last week, X has been flooded by a mass undressing spree, where users were seen prompting Grok to manipulate images of women and generate explicit deepfakes. Most of these requests are reportedly non-consensual and violate XAI's policy on "acceptable use" that prohibits synthetic rendering of a person in a pornographic manner and sexualization of children. India's Ministry of Electronics and Information Technology (MeitY) sent a letter to X, last week, demanding immediate action to curb sexually explicit deepfakes and the misuse of AI tools such as Grok to create them. MeitY has asked X to remove all such material and strengthen its content moderation. French and Malaysian authorities have also launched an inquiry into the Grok generated deepfakes. XAI CEO Elon Musk said in a post on X, "anyone using Grok to make illegal content will suffer the same consequences as if they upload illegal content." What is alarming is that users didn't have to use jailbreaks or prompt engineering tricks to bypass Grok's guardrails. The AI chatbot generated these deepfake images on simple and direct user requests. Since its release in November 2023, Grok has been used to generate deepfake images of political figures. In the run up to the US elections, it was used to create deepfakes of Donald Trump, Joe Biden, Kamala Harris and even Elon Musk. Many of these widely circulated deepfakes showed these political figures in highly inflammatory situations, including fake images of them participating in the 9/11 terrorist attack. Grok is known to have fewer guardrails, unlike rival AI chatbots such as Gemini and ChatGPT, which refuse to generate synthetic images resembling a public figure. Getting them to generate sexually explicit content would require complex jailbreaking. Rise of deepfakes In 2024, explicit deepfake photos of singer Taylor Swift was reportedly shared more than 45 million times on X before it was taken down. However, their use is not limited to targeting women and political figures. Deepfakes are increasingly being used to target enterprises by impersonating CXOs and persuading staff to share sensitive information or transfer money. They can also be used to plant malware or circulate deepfakes of CXOs announcing fake financial results or mergers to manipulate the stock market. In May 2024, a worker in the finance department of a large British design and architecture firm, Arup, was tricked into transferring $25 million by a deepfake video call impersonating the CFO. According to news reports, the worker was initially suspicious but was convinced by the strikingly realistic AI impersonation of the CFO and other company executives who were faked on the video call. In another incident, scammers used a voice clone and a YouTube clip to impersonate Mark Read, CEO of an advertising firm WPP, and other executives in a video call on Microsoft Teams. Scammers tried to trick one of the agency leaders (who was on the call) into sharing funds and data to set up a new business. According to a Gartner report, 62% of organizations have faced a deepfake attack using social engineering or exploiting automated processes. Entrust's 2026 Identity Fraud report claims deepfakes now account for one in five biometric fraud attempts. Deepfakes include any form of image, video, and audio that seems authentic but is generated using generative AI tools. Before 2023, deepfakes were widely created using Generative Adversarial Networks (GANs), which pits two competing neural networks against each other to create realistic images. Most of the new generative AI apps such as Grok, Sora, and Midjourney use Diffusion model, which breaks down images into random noise and then learns to de-noise or clean the data and generate a new image that looks like the original. How are regulators dealing with deepfakes The Grok incident once again highlights the need for strict regulations and guidelines to prevent the misuse of AI, especially generative AI platforms. Lawmakers in some countries including India have so far maintained that they are not planning to introduce a separate AI law and will rely on existing legislations and legal frameworks to regulate and prevent misuse of AI. This reflects India's effort to strike a balance between innovation and regulation as it builds its own sovereign AI models to reduce over-reliance on US tech giants. In October 2025, MeitY proposed to amend the Information Technology (Intermediary Guidelines and Digital Media Ethics Code) Rules, 2021, to curb sharing of deepfakes and misleading AI-generated content on social media. It will mandate social media platforms such as Instagram, YouTube, and X to label AI-generated content with markers that covers at least 10% of the visual display and initial 10% of the duration of an audio clip. In November 2025, MeitY also introduced the AI Guidelines, which advocates the principle of "do no harm" and recommends creation of governance institutions, India-specific risk frameworks, and making AI safety tools accessible to the public. The European Commission took a different stance when it passed the EU AI Act. It came into force in August 2024, bit its full implementation is expected to be completed in 2027. The EU AI Act classifies AI models based on the level of risk they pose and regulates it accordingly. AI models in the "unacceptable risk" category are blocked from operating in the EU, while "high-risk" AI models have to go through assessments and regular monitoring. In July 2025, the commission added a new framework called the AI Code of Practice. It requires AI providers to maintain detailed technical documentation and instructions for those who would integrate these models into their own products. It also requires them to implement a policy to comply with EU copyright law. Most of the AI providers including Meta, OpenAI and Anthropic are involved in legal battles with authors, artists, and news publishers for using copyrighted content to train their AI models without seeking their permission or paying them for it.
[6]
Govt issues notice to X over 'obscene' content targeting women, children on Grok
MeitY has issued a formal notice to X, citing serious lapses in statutory due diligence under the IT Act and Rules. The government is concerned about X's AI tool, Grok, generating and circulating obscene content, particularly targeting women and children. X has been directed to review Grok's framework, remove unlawful content, and take action against offending users within 72 hours. The Ministry of Electronics and Information Technology (MeitY) on Friday issued a formal notice to Elon Musk-owned social media platform X, flagging serious lapses in statutory due diligence under the Information Technology Act and IT Rules. Amid ongoing concerns over misuse of the platform's AI tool, Grok, the government has raised the issue of the tool generating and circulating obscene, sexually explicit and derogatory content, particularly targeting women and children, calling it a grave violation of dignity, privacy and digital safety. Also read: "Gross misuse", RS MP Priyanka Chaturvedi flags AI misuse on social media, writes to IT Minister Vaishnaw MeitY has directed the platform to immediately review the technical and governance framework of its AI tool Grok and remove all unlawful content. It has also asked the platform to take action against offending users, and submit an Action Taken Report within 72 hours. The ministry, led by Union Minister Ashwini Vaishnaw, warned that continued non-compliance could lead to loss of legal protection under the IT Act and strict action under multiple cyber, criminal and child protection laws. Earlier on Friday, Vaishnaw said that social media firms should take responsibility for content they publish and a standing committee has already recommended a tough law to fix the accountability of platforms This week, the government warned online platforms, mainly social media firms, of legal consequences if they fail to act on obscene, vulgar, pornographic, paedophilic and other forms of unlawful content. "Social media should be responsible for the content they publish. Intervention is required," Vaishnaw said on Friday on the sidelines of a Meity event. Meanwhile, Rajya Sabha member Priyanka Chaturvedi has also written to the minister seeking urgent government intervention on increasing incidents of AI apps being misused to create vulgar photos of women and post them on social media. "The standing committee has recommended that there is a need to come up with a tough law to make social media accountable for content they publish," Vaishnaw informed. The Parliamentary Standing Committee on the Ministry of Information and Broadcasting has made a recommendation to the government seeking accountability of social media and intermediary platforms with respect to peddling fake content and news. (With inputs from PTI)
[7]
Safe Harbor of X Could be Revoked Over Grok's CSAM Content
The government is prepared to revoke the safe harbor status of social media platform X "if it doesn't comply with the latest takedown directions on artificial intelligence (AI)-generated obscene images," anonymous officials told the Economic Times. This latest Indian government response comes at a time when the Safety handle of X (formerly Twitter) assured that they will take action against "illegal content, including Child Sexual Abuse Material (CSAM)," without particularly referring to the non-consensual sexually explicit imagery that X's AI bot Grok has been generating since the launch of this image editing feature with the Grok bot around the Christmas period in 2025. These actions can include: "Anyone using Grok to make illegal content will suffer the same consequences as if they upload illegal content," warned X's CEO, Elon Musk, just a day before X's Safety handle mimicked a similar caution while referring to X's policies that define illegal content. Apart from the Indian government, the Malaysian government also took note of Grok's actions and the public representatives of the French government, who wrote to Paris' public prosecutor regarding the same. According to X's Rules, which favour public participation in "global public conversation", restricted safety-related aspects are: It is important to note that Grok admittedly violated its own rules by generating the Child Sexual Abuse Material (CSAM). On January 1, Grok bot admitted that it had "generated and shared an AI image of two young girls (estimated ages 12-16)." "It was a failure in safeguards, and I'm sorry for any harm caused. xAl is reviewing to prevent future issues," read Grok's response when other users persistently asked about the CSAM material. In addition to these, their policies also ask users not to share abusive content and not to engage in targeted harassment or inciting others. However, the specific section on "Illegal and Regulated Behaviours", which advocates that the Grok shouldn't be used for any "any unlawful purpose or in furtherance of illegal activities" -- which Elon Musk and X's team were referring to while responding to the mass undressing campaign unleashed by the Grok bot- doesn't refer to the sexual imagery, let alone non-consensual sexual imagery. It only refers to the activities related to the sales of drugs, weapons, human trafficking or poaching of endangered species and sexual services, which refer to sex work-related activities. Under X's rules framework, adult content, like morphed bikini-wear image generation, that does not involve underage children can fall under the platform's adult content policy, which explicitly permits consensually produced and distributed adult imagery. Now, the bigger question here is: how do you verify the consent of the alleged woman involved in the morphed picture? Addressing the question of safe harbor exemptions for the platform X and Grok AI bot, Nikhil Pahwa, Medianama's Founder and Editor, wrote for The Quint, arguing that X, especially Grok AI, can't claim safe harbor exemptions under Section 79 of the IT Act, which takes away the liability of a platform for the content posted by third-party users. Pahwa reasoned, saying, "X is actively enabling the publishing of this content via its own AI service and not making 'reasonable efforts' to prevent it; quite the opposite." He was referring to the "reasonable efforts" that a platform must make under Section 3(1)(b) of India's IT Rules to prevent the users from publishing content like "obscene, pornographic, paedophilic, invasive of another's privacy, including bodily privacy, insulting or harassing on the basis of gender, or racially or ethnically objectionable." Further explaining why X can't claim safe harbor exemptions, Pahwa explained, "Safe harbor protections are provided to platforms that allow others to post content: they act as 'intermediaries' and mere conduits. The company that runs X is not an intermediary here: its AI service is actively publishing this content, so safe harbor protections cannot apply to it." Therefore, he reasoned out, saying that the company behind Grok, namely X, is also potentially liable for generating NCII and not just the users who prompted the AI system, as X's terms of service try to push the user as the responsible person for the content they asked their AI system to generate. "Safe harbor protections were never meant to apply to publishing, and consent should never be optional. Until India's regulatory framework, especially the Digital Personal Data Protection Act, reflects this reality, these problems will continue," he concluded the editorial. After receiving a 72-hour countdown to submit the "Action Taken Report" notice from the Indian government's MeitY's Secretary regarding the rampant misuse of Grok bot's image editing capabilities to generate sexually explicit images of women, mostly without their consent, the X team seems to have begun suspending a few accounts, which prompted Grok to generate bikini wear on various women. For instance, an X handle, Komal Yadav (@komalyadav03), prompted Grok several times to ask for bikini-wear images of multiple women, and is now suspended. However, even after Elon Musk and X's Safety handles' assurances, the Grok AI bot is still generating the bikini-wear and sexually explicit images of various women. The unconsented image morphing is a huge issue. Because even if you want to take the consent for granted, that is when a person asks to modify "my image"; even then, the unconsented image generation continues in the same thread of the post, let alone when the users might re-upload the picture to Grok to re-morph the images. For example, a "verified" handle, Soumya Avasthi (@SoumyaAvasthi), which doesn't exist anymore, prompted Grok, saying, "Hey @grok put me in red saree!!" Complying with the prompt, Grok generated an image of the person in the image in a red saree. However, Grok also complied when another user asked to generate her image in Grok's app. Further making things worse, Grok will comply with the prompts to a random user's requests in the post's thread and will proceed to generate Non-Consensual Intimate Images (NCII). Addressing the unconsented morphed image generated, a verified user, Nandani S (@ChaiCodeChaos), reported an account handle (@HoeEnchanted) to X's handle Cyber Cell India, and the current status of the alleged account reads, "This account doesn't exist." This can also mean that the account might have been deactivated by the user- which can be revoked within 30 days- and not suspended by X's team. Similarly, another verified user, Meghna (@CPUatOnePercent), also claims that she got her morphed images taken down from X after two days. The concerned account (@ltlswhatltls), which allegedly generated morphed images of Meghna, also reads, "doesn't exist anymore." Concerning the potential exploitation, several users began telling Grok and X that they do not authorise it "to take, modify, or edit ANY photo or video" of them, "whether those published in the past or the upcoming ones" that they post online. However, X's attempts to comply with the Indian government's orders are not equal. For example, a handle with the name Kavi (@Kavithasri98), is still active at the time of writing this report despite generating bikini wear images on Jan 2, 2026. The account, created on December 29, 2025, days after Grok's image-editing rollout, contains only images of one woman in different Indian attire, suggesting the images may be AI-generated. However, Medianama could not independently verify whether the person shown in the images is real or a deepfake, a limitation encountered repeatedly during the reporting of this story. This aspect is important to note because, if this person is "real", then the person can/may initiate action against either X or the owner of X's handle, or both. If not, if the person is deepfake-generated, then would it still be a violation of the privacy and dignity of an (unreal) woman? How would X deal with the sexually explicit deepfakes of already deepfaked (probably unreal) women?
[8]
India Directs X to Curb AI-Generated Sexual Content
The Ministry of Electronics and Information Technology (MeitY), in a letter dated January 2, 2026, directed X (formerly Twitter) to take urgent action against the proliferation of obscene, nude, indecent, and sexually explicit content, particularly content generated or circulated through the misuse of its AI services such as Grok and other xAI tools. In short, the ministry has ordered X to promptly remove such material, strengthen its content moderation and AI safeguards, and demonstrate compliance with India's due diligence requirements for intermediaries. This move follows a wave of activity on X over the past few days, during which users repeatedly prompted the platform's AI chatbot, Grok, to generate graphic sexual content using photos of real women in replies or posts. These outputs were generated directly in replies and often remained visible for extended periods. As a result, the content gained traction through engagement loops, with multiple users amplifying or iterating on the same prompts, raising fresh concerns about the platform's controls over AI-assisted content creation and distribution. Meanwhile, MeitY's intervention builds on its earlier advisory on obscene and sexually explicit content, which reminded intermediaries of their obligations under the Information Technology Act and the IT Rules, 2021. The advisory stressed that platforms must prevent the hosting and dissemination of prohibited content, act swiftly upon gaining actual knowledge, and deploy adequate technological measures to curb the spread of unlawful material, particularly content that harms the dignity and privacy of women. Further, MeitY Secretary S. Krishnan said that "they cannot escape their duty or responsibility simply by pleading safe harbour", and that "social media entities have to show responsibility", when asked about the Grok incident during an interview with CNBC-TV18. The notice directs X to take immediate and concrete steps to prevent the hosting, generation, publication, transmission, storage, or sharing of obscene, nude, indecent, sexually explicit, vulgar, paedophilic, or otherwise unlawful content, including content created or amplified through AI-enabled services such as Grok. It instructs the platform to desist from allowing such material in any form and to act strictly in line with statutory due diligence obligations under the IT Act and the IT Rules, 2021. The notice further requires the platform to carry out a comprehensive technical, procedural, and governance-level review of the Grok application. This review must cover prompt processing, large language model-based output generation, image handling, and safety guardrails, and must ensure that the application does not generate, promote, or facilitate nudity, sexualisation, sexually explicit material, or any other unlawful content in any form. In parallel, the platform must enforce its user terms of service, acceptable use policies, and AI usage restrictions, including by imposing strong deterrent measures such as suspension, termination, or other enforcement action against users and accounts found to be in violation. Additionally, the platform must submit a detailed Action Taken Report (ATR) to the ministry within 72 hours of the issuance of the letter. This report must outline the technical and organisational measures adopted or proposed in relation to the Grok application, the role and oversight exercised by the Chief Compliance Officer, the actions taken against offending content, users, and accounts, and the mechanisms put in place to ensure compliance with mandatory offence-reporting requirements under criminal procedure law. India's deepfake regulation debate has largely centred on MeitY's draft amendments to the IT Rules on synthetic information, which aim to push platforms and AI service providers towards labelling and provenance-style compliance, while also expanding the conversation beyond simple takedowns. At MediaNama's Regulating for Deepfakes in India discussion, experts repeatedly questioned whether disclosure mandates alone can work at scale, and whether the government should assign responsibility differently across the ecosystem. One participant said that "the problem statement of these particular rules, which seems to be deepfakes and harmful deepfakes" could be better resolved "through a value chain approach where you're assigning appropriate responsibilities across various layers of the AI value chain". At the same time, speakers warned that the draft framework risks conflating AI governance with intermediary liability. One expert noted that "by conflating AI oversight with the liability regime, we are essentially creating outcomes that are not needed and a lot of confusion", referring to how labelling and provenance requirements begin to resemble AI technology regulation rather than content distribution compliance. Meanwhile, enforcement and detection emerged as central issues. One speaker pointed to existing deployments of detection systems, stating, "We are live with many banks and financial institutions", and that the system is "handling lakhs of KYC verifications per day" with "millions of API hits every month" that are "scalable to tens of millions of detections per day". However, they also asked, "Who will pay for it?", highlighting the cost challenge. This matters because India's deepfake and AI misuse problem has moved beyond fringe tools and isolated platforms. Generative capabilities now sit directly within large, mainstream digital services, enabling harmful content to be created and circulated at scale. This episode shows that when platforms integrate such tools into social feeds, replies, and engagement-driven interfaces, misuse can spread rapidly and gain visibility faster than existing moderation controls can respond. At the same time, the availability of face-swapping applications on the Google Play Store in India underscores that AI-enabled misuse spans multiple layers of the digital ecosystem. While regulators and platforms may intervene at the level of content distribution, the underlying tools that enable synthetic sexual content often remain easily accessible through app marketplaces. This creates an uneven regulatory landscape, where action against one platform does not necessarily reduce overall availability or use. Against this backdrop, the focus on AI features within a major social platform highlights a broader policy challenge. Regulatory responses are increasingly confronting not just user behaviour, but the design and deployment of generative systems themselves. However, these interventions remain largely case-specific. Consequently, this episode points to a growing gap between the pace of generative AI adoption and the coherence of regulatory oversight. As AI-driven content creation becomes more deeply woven into everyday platforms, the effectiveness of enforcement will depend on whether safeguards extend consistently across platforms, tools, and distribution channels -- rather than operating in isolation.
[9]
X, Grok and the limits of safe harbour in India | Marketing | Campaign India
India's Ministry of Electronics and Information Technology (MeitY) has warned X (formerly known as Twitter) that it risked losing its 'safe harbour' status. This trigger was not a single post or rogue account, but came over the circulation of AI-generated obscene content created using xAI's chatbot Grok. According to The Economic Times, the government has told the company that failure to remove flagged images and videos and submit an auditable compliance report could lead to the withdrawal of legal protections under Section 79 of the Information Technology Act. At stake is the intermediary immunity that shields platforms from liability for user-generated content, provided they follow Indian rules and act swiftly on takedown orders. Officials have made it clear that this protection is conditional -- and that generative AI tools hosted and promoted by platforms are now testing the boundaries of that framework. The notice follows a growing volume of complaints around Grok's image-generation features, particularly its so-called 'Spicy Mode'. Many netizens took to social media to point out that the feature has been misused to create non-consensual, sexualised deepfakes of public figures and private individuals, many of them women and celebrities like singer Taylor Swift. And also share their disgust on this intrusion into their privacy on a public platform. On January 2, MeitY sent X a formal notice saying it had "grave concern" over the use of Grok to generate obscene images. The government warned that "such conduct reflects a serious failure of platform-level safeguards and enforcement mechanisms, and amounts to gross misuse of artificial intelligence technologies in violation of applicable laws". It added that the spread of such content could violate the dignity, privacy and safety of women and children, "normalising sexual harassment and exploitation in digital spaces, and undermining the statutory due diligence framework applicable to intermediaries operating in India". Crucially, officials and legal experts say the issue is not limited to user misuse. Because Grok is a platform-provided AI tool, X could be seen as enabling or amplifying the creation of harmful content rather than merely hosting it -- weakening claims to intermediary immunity. A government source told The Economic Times: "Back in 2021, the Centre had informed the Delhi High Court that X (then Twitter) had briefly become legally responsible as a publisher. In any case, the government is not keen on removing safe harbour provisions. But given the explosion of AI-made content that harms our citizens, especially women and children, it is prepared to take a hard step." A member of xAI's technical staff, Parsa Tajik, tweeted that its engineering teams are working to introduce tighter guardrails. Critics argue that incremental fixes may not be enough unless platforms rethink how generative AI tools are deployed and moderated at a systemic level. Safe harbour, with conditions attached Under Section 79 of the Information Technology Act, 2000, social media platforms are treated as neutral intermediaries and are not held responsible for content created by third parties -- so long as they follow government rules and promptly remove unlawful material when directed. That protection, however, depends on demonstrable due diligence. Platforms must act quickly on government or court orders, appoint local compliance officers, and maintain safeguards to prevent the spread of illegal content. MeitY has said X is not adhering to the Information Technology (IT) Rules, 2021, and the Bharatiya Nagarik Suraksha Sanhita, 2023, which deal with obscene, indecent, vulgar, pornographic, paedophilic or otherwise unlawful content. Non-compliance within 72 hours, the ministry has warned, could result in revocation of X's safe harbour status. The ministry has also questioned whether X's India-based compliance officers have the authority and oversight mandated under Indian law. This revives a familiar fault line from earlier disputes with the platform. In December, MeitY issued an advisory asking social media firms to immediately review their compliance frameworks after observing that platforms were not acting strictly against obscene and unlawful content. The directive followed complaints from public discourse and representations from parliamentary stakeholders. Rajya Sabha member Priyanka Chaturvedi had written to Union IT minister Ashwini Vaishnaw seeking urgent intervention over the misuse of Grok to create vulgar images of women. Vaishnaw has signalled that the government's patience with self-regulation is wearing thin. "Shouldn't platforms operating in a context as complex as India adopt a different set of responsibilities? These pressing questions underline the need for a new framework that ensures accountability and safeguards the social fabric of the nation," he had stated. Musk's response -- and unresolved questions X owner Elon Musk responded publicly, saying users who generate illegal content using Grok would face the same consequences as those who upload illegal material. "Anyone using Grok to make illegal content will suffer the same consequences as if they upload illegal content," Musk said on X, responding to a post about "inappropriate images". Indian regulators seem unconvinced by that analogy. Beyond takedowns, MeitY is pressing for design-level changes to Grok itself, including disabling or sharply restricting 'Not suitable for work' modes, preventing the generation of realistic likenesses without consent, and maintaining independently verifiable moderation records. This is not X's first brush with Indian regulators. In 2021, the platform briefly lost its safe harbour protection after failing to comply fully with the IT Rules, 2021, including the appointment of key India-based officers. Its legal immunity remained suspended for nearly three months. Advertisers had already stepped back The regulatory heat comes against the backdrop of a long-running crisis of confidence between X and the advertising industry. Campaign reported in 2022 that advertisers and media buyers had started pausing advertising on X amid concerns about brand safety and moderation following Musk's takeover. Based on their unease about content controls and Musk's "erratic behaviour", they advised brands to avoid the platform because they could not guarantee a brand-safe environment. GroupM assessed advertising on Twitter as "high risk", according to a report in the Financial Times. Several large advertisers hit pause on spending. In the US, Media Matters reported that 50 of the top 100 advertisers, which spent $750m in 2022, had seemingly stopped advertising on Twitter. Brands that paused spending included VW, General Motors, Diageo, Heineken, Nestlé, Coca-Cola, Mars and Ford. This prompted Musk to personally call chief executives to berate them, according to media reports. Agency leaders told Campaign back then, on condition of anonymity, that concerns centred on reduced investment in moderation, staff exits and Musk's own conduct on the platform. "Some of the things he is posting, including memes, are horrendous," one agency leader said. "The challenge for Musk and Twitter is that he will need to prove that this is an environment that is safe for advertisers, and so far we have not seen that." Further unease followed sweeping cuts to Twitter's sales and moderation teams. Aura Intelligence reported that over 6,000 employees were let go during Musk's takeover of Twitter, reducing the workforce by nearly 80%. A platform under pressure from both sides In a 2024 article for WARC, Gonka Bubani of Kantar described Musk's stewardship of X as a case study in how quickly platforms can unravel when advertisers and moderation systems are alienated. "We now know that marketers are fleeing: a net 26% plan to reduce ad spend on X in 2025, the largest drop that Kantar's annual Media Reactions report has tracked across any major platform since its start in 2020," Bubani wrote. "Only 12% of marketers now trust ads on X... Only one in 25 believe X provides adequate brand safety," she added, noting that Google's perceived brand safety rating stood at 39%. While Bubani acknowledged that confidence in X had been declining before Musk's acquisition, she argued that dismantling moderation systems and increasing volatility had pushed the platform "to the edge". For Indian regulators, the Grok controversy sharpens a broader policy question: whether generative AI features change a platform's legal responsibilities. If MeitY withdraws X's safe harbour protection, the company could face direct liability for user posts in India, along with stricter disclosure and moderation obligations. That would set a precedent for how governments treat AI-native features embedded within social platforms. Regulators in Malaysia and parts of Europe are already probing Grok's outputs and urging tighter safeguards. For social media platforms and AI product teams, the trade-off is becoming unavoidable. Either redesign features that drive engagement but carry foreseeable risk, or accept tighter regulation and increased legal exposure. India's message to X suggests that promises and post-hoc takedowns may no longer be enough. The scrutiny is shifting upstream -- towards product design, model behaviour and whether platforms can prove, with evidence, that harm is being prevented rather than merely cleaned up after the fact.
[10]
Fix Grok bikini deepfakes: Indian govt pulls up X.com for viral AI trend
India's New Year has barely finished popping its confetti cannons, and social media already has its first big AI controversy of 2026, in case you didn't notice. The Indian government has issued a formal notice to X over the misuse of Grok AI for generating and circulating obscene and indecent content on social media. The Indian govt's notice is in response to the "Grok bikini" viral trend wave - non-consensual synthetic edits that turn real women (and, in some reported cases, minors) into subjects of online jokes or harassment campaigns. According to several reports, the Ministry of Electronics and Information Technology (MeitY) wrote to X's India Chief Compliance Officer and accused the platform of failing its statutory "due diligence" duties under the Information Technology Act, 2000 and the IT (Intermediary Guidelines and Digital Media Ethics Code) Rules, 2021. That phrasing is critically important, as it matters a lot from a legal perspective. It frames the issue as failure of X.com's platform responsibility, not just user misbehaviour that added fuel to the Grok AI viral bikini trend. The notice by MeitY seems to imply that when an AI feature that's the root cause behind the AI deepfake trend, sitting on the fence claiming you are just the platform - it's the users who are misbehaving - simply isn't good enough. Also read: X bikini trend: Govt issues 72 hour ultimatum to Elon Musk's X, asks to remove explicit Grok content MeitY's directions couldn't have been more blunt, especially on two counts. Firstly, X.com has been told to remove unlawful material and to run a comprehensive technical, procedural, and governance review of Grok AI - covering prompt processing, output generation, and safety guardrails - and to submit a detailed Action Taken Report within a tight deadline of 72 hours (three days). Secondly, Section 79 of India's IT Act offers intermediaries (like Facebook, X, and others) conditional protection from liability, but it is contingent on "due diligence" on their part. MeitY's intermediary rules also state that if an intermediary fails to observe the rules, Section 79(1) won't apply - which means X.com will be held responsible for damage caused by the Grok AI bikini trend. Why the escalation now, one may wonder, especially because this trend has been going on for a few days? Even Elon Musk joined in on the "fun". Because the harm caused is serious, and not at all fun. Grok's image-editing capabilities are seriously photo-realistic, giving X.com users the ability to quickly generate sexualised edits from real photos, including outputs involving minors when prompted - territory that triggers serious child-safety and potentially criminal-law concerns. Once that line is crossed, regulators stop issuing polite advisories and start counting enforcement tools. For X and xAI, this is the moderation math problem in its most unforgiving form: the more photorealistic the generator, the more damaging the abuse - especially when the target is a real person who never consented. MeitY's posture also signals something bigger for every platform rushing to bolt "AI edit" onto a feed: "experimental" is not a legal category once you deploy at national scale. If you can ship it, you should be able to govern it, and more importantly make sure it's not violating any laws. It refers to users on X replying to photos of women - celebrities, influencers, and ordinary users - and prompting Grok to "put her in a bikini" or produce other skimpy, sexualised edits. Those images are then reposted publicly, often as memes or targeted harassment, despite being non-consensual synthetic media. Outrage has been growing against this trend as examples circulated that appeared to involve minors, and as more women described the trend as tech-enabled sexual harassment rather than "harmless banter." If this is how 2026 opens, it's a reminder that "AI in everything" doesn't just automate chores - it can industrialise abuse. India's notice is effectively a line in the sand that ensures platforms can't hide behind a mode switch and call it innovation when what spills out is a violation of dignity.
[11]
X bikini trend: Govt issues 72 hour ultimatum to Elon Musk's X, asks to remove explicit Grok content
Ministry of Electronics and Information Technology warns of legal consequences if content is not removed quickly. What began as a harmless AI trend on social media has now turned into a serious policy and safety issue in India. The so-called bikini trend on X has drawn widespread attention in the past few days. The trend started when the users started tagging the platform's AI bot, Grok, with requests to alter photos and place people in bikinis. While some users willingly participated, the trend quickly crossed ethical lines when images of women were shared without consent and turned into obscene or sexualised visuals. The issue gained more popularity as Elon Musk, CEO of X, joined the trend by asking the AI bot to generate his image in a bikini. The trend has received backlash from politicians, activists, and users across the globe and has prompted a direct intervention from the Indian government. The bikini trend on X can be seen as a successor to these earlier trends. However, while the previous trends were largely praised for their creativity and positive appeal, the bikini trend marked a drastic shift, as it was widely perceived as inappropriate and received negatively by many users. Also read: Elon Musk falls victim to Grok's bikini trend on X, replies Perfect to AI-generated bikini image of himself While many users joined in the trend voluntarily, the issue began when users shared photos of women, including journalists and public figures, without permission and asked the AI to make the images revealing. Furthermore, many women also faced harassment in comment sections, where strangers repeatedly tagged Grok under their photos with such prompts. Shiv Sena UBT MP Priyanka Chaturvedi publicly flagged the issue, calling it a violation of women's dignity and privacy. Other than that, she also urged the government to act quickly and questioned the lack of laws that allow such content to be generated so easily. Soon after that, the Ministry of Electronics and Information Technology issued a formal notice to X. In the formal notice, the ministry accused the platform of failing to meet its legal duty under Indian law to prevent the hosting and spread of obscene and sexually explicit content. Furthermore, it warned the company that if they continue the non-compliance, then they could result in the loss of legal protection granted to online intermediaries. Also read: Need legal advice? Government's Nyaya Setu is now available on WhatsApp for free, here's how to use The government asked X to immediately review Grok's services, strengthen content controls, and remove all violating content without any delay. Moreover, the government has put a clock on the companies' actions, as they have been asked for a report on actions taken on the violation with a time limit of seventy-two hours. Also read: No iPhone 18 launch in 2026? Leaks suggest Apple may delay it to early 2027 Following the intervention, X has reportedly limited the visibility of the trend for most Indian users. Grok has also begun refusing bikini-related prompts. While this has brought some relief, many women say the episode highlights a deeper issue. Without strong guardrails and accountability, AI-driven trends can quickly turn into tools of harassment rather than creativity.
Share
Share
Copy Link
India's IT ministry has given Elon Musk's X just 72 hours to submit an action report after users exploited Grok to generate sexually explicit deepfake images of women. The order threatens X's safe harbor protections under Indian law, marking a critical test for platform accountability in the age of AI-generated content. The move follows widespread outrage over non-consensual images that remained accessible on the platform.
India has issued a stern directive to Elon Musk's X, demanding immediate technical and procedural changes to its AI chatbot Grok after widespread misuse generated obscene AI content, including deepfake images of women
1
. The Ministry of Electronics and Information Technology (MeitY) gave X just 72 hours to submit a compliance report detailing steps taken to prevent the hosting or dissemination of content deemed obscene, pornographic, or otherwise prohibited under Indian law1
. The order specifically targets content involving nudity, sexualization, and sexually explicit material, warning that non-compliance could jeopardize X's safe harbor protections—legal immunity from liability for user-generated content under the IT Act1
.
Source: Digit
The crisis erupted after users discovered they could prompt Grok to manipulate images of real women, particularly creating non-consensual sexualized images showing them in bikinis or digitally undressed . Unlike competitors such as Google's Gemini or OpenAI's ChatGPT, which maintain strict content guardrails, Grok was positioned as more edgy with fewer restrictions . What proved particularly alarming was that users didn't need jailbreaks or complex prompt engineering tricks—the AI chatbot generated these deepfake images on simple, direct requests
5
.
Source: AIM
Manipulated images of Bollywood actors and Indian public figures went viral as users employed prompts such as "undress," "change clothes," and "change to a bikini" . Indian parliamentarian Priyanka Chaturvedi filed a formal complaint, while separate reports flagged instances where the image generator created sexualized images involving minors, which xAI acknowledged resulted from lapses in safeguards
1
.MeitY asserted that adherence to the IT Act and IT Rules is mandatory, not optional, putting India's intermediary liability framework to the test
2
. Under Section 79 of the IT Act 2000, social media sites are classified as neutral hosts not held responsible for content created by third parties—enjoying safe harbor status—as long as they follow government rules . However, this protection is lost if illegal content is not taken down after being flagged through takedown requests . The ministry's letter to X's Chief Compliance Officer for India flagged that Grok was being exploited to create fake accounts that host, generate, publish, or share obscene content2
. Legal experts note the misuse violates multiple Indian laws, including the Indecent Representation of Women (Prohibition) Act, 1986, and sections 66E and 67 of the Information Technology Act dealing with privacy violations and obscene content online . The recently passed Digital Personal Data Protection (DPDP) Act, 2023 mandates consent for using personal data, including photographs and facial data, with penalties of up to Rs 250 crore for breaches .Related Stories
India, with approximately 22 to 27 million users on X, represents one of the world's biggest digital markets and has emerged as a critical test case for how far governments are willing to go in holding platforms responsible for AI-generated content
1
. Any tightening of enforcement in the country could have ripple effects for global technology companies operating across multiple jurisdictions1
.
Source: Digit
Cybersecurity experts warn that AI models are being released openly without adequate scrutiny of threat vectors, training data, or geographical origins . According to a Gartner report, 62% of organizations have faced a deepfake attack using social engineering or exploiting automated processes, while Entrust's 2026 Identity Fraud report claims deepfakes now account for one in five biometric fraud attempts
5
. The incident occurs as Musk's X continues to challenge aspects of India's content regulation rules in court, arguing that federal government takedown powers risk overreach, even as the platform has complied with a majority of blocking directives1
.X responded to global outrage by hiding Grok's media tab and directing users to report violations within its stated prohibitions on non-consensual intimate imagery . However, images generated using Grok that made women appear to be wearing bikinis through AI alteration remained accessible on X at the time of publication
1
. Digital risk management experts argue that X's reactive step doesn't meet the higher bar required, calling for default-safe settings, strong consent checks, auditable controls, and consistent policy application . The order warned that non-compliance could result in strict legal consequences against the platform, its responsible officers, and users who violate the law, without any further notice1
. In October 2025, MeitY proposed amendments to the Information Technology (Intermediary Guidelines and Digital Media Ethics Code) Rules, 2021, to mandate social media platforms label AI-generated content with visible markers5
. French and Malaysian authorities have also launched inquiries into the Grok-generated deepfakes, signaling growing international concern over generative AI misuse5
. X and xAI did not immediately respond to requests for comment on the Indian government's order1
.Summarized by
Navi
02 Jan 2026•Entertainment and Society

02 Jan 2026•Policy and Regulation

16 Aug 2024

1
Policy and Regulation

2
Business and Economy

3
Technology
