Updates
GS Paper IIIEnvironmental Ecology

Ministry tells Social Media Platforms to Block Obscene Content

LearnPro Editorial
31 Dec 2025
Updated 3 Mar 2026
9 min read
Share

24-Hour Takedowns: A Policy Directive or a Pipeless Echo?

On December 31, 2025, the Ministry of Electronics and Information Technology (MeitY) issued an advisory to social media platforms mandating the removal or disabling of obscene and illegal content within 24 hours upon receiving a complaint. The guidelines, issued under the Information Technology Rules, 2021, emphasize stricter enforcement mechanisms, including proactive use of automated moderation and expanded access to grievance redressal systems for users. Non-compliance may result in criminal prosecution under provisions such as Section 69A of the IT Act, 2000. Yet, beneath this headline lies a tangle of legal ambiguities, institutional bottlenecks, and questions about whether execution will match intent.

Why This Directive Marks a Departure

Previous attempts at regulating digital content have relied heavily on intermediary self-regulation or sporadic government interventions. For instance, the IT Rules, 2021 already mandated “reasonable efforts” for content moderation, but enforcement has been tepid. What stands out in the current advisory is the explicit 24-hour compliance rule for prima facie sexual or obscene content. This accelerates timelines and shifts the burden squarely onto platforms, marking a departure from erstwhile leniency.

Equally striking is the insistence on proactive mechanisms. Automated moderation technologies, though widespread elsewhere, have faced legal pushback in India. Even global giants like Meta and Google resisted full automation in content moderation, citing algorithmic biases and language challenges in India’s multilingual ecosystem. The advisory pushes these platforms to embrace a practice they’ve largely avoided, despite its fraught history.

Finally, the scope of "illegal content" in the directive breaks past precedent. It incorporates not only obscenity but impersonation, a largely under-regulated category in prior advisories. This implicitly ties social media accountability to the prevention of identity fraud, expanding their legal responsibilities without an obvious architecture to enforce compliance effectively.

The Machinery Behind the Mandate

The advisory derives its authority from the Information Technology Act, 2000, particularly Section 69A, which empowers the government to block digital content for reasons of security, decency, and public order. In conjunction, the Intermediary Guidelines and Digital Media Ethics Code, 2021 set the procedural backbone for compliance, laying out obligations for “significant social media intermediaries”, those with over 5 million registered users.

However, institutional enforcement remains fragmented. The Grievance Appellate Committee, formed earlier in 2023 to expedite complaints, has been sluggish and insufficiently staffed—handling barely 3,000 cases in its first year against an annual user complaint volume exceeding 50,000 for major platforms like WhatsApp and Instagram. Moreover, while platforms are now required to use automated tools for quicker takedowns, the advisory conspicuously lacks detail: Should these algorithms align with government-mandated standards? How will multilingual moderation gaps be mitigated?

Finally, the absence of penalties tied directly to the 24-hour compliance rule undercuts its bite. While Section 69A allows blocking mechanisms and criminal prosecution, it is unclear whether delays in takedown trigger measurable fines or judicial review for intermediaries who claim “good faith” moderation efforts. This legal vagueness is likely to lead to prolonged litigation rather than streamlined enforcement.

Behind the Data: Enforcement vs Reality

The gap between regulatory intent and ground-level outcomes is stark. MeitY claims that strict enforcement will curb digital obscenity. But independent audits, including those from Internet Freedom Foundation, suggest slow moderation rates. For instance, studies found that hate speech removal on Twitter takes an average of 4 days — far above the mandated 24-hour window — even for flagged posts under current regulatory pressure.

Further, platform transparency remains questionable. Despite provisions in the IT Rules requiring periodic reports on takedowns, only 14% of flagged content was disclosed by intermediaries in 2024 (MeitY report), creating a black box over moderation criteria. This lack of visibility raises uncomfortable questions about arbitrary removals, over-censorship, or selective enforcement triggered by ideological leanings.

Additionally, claims of automated moderation solving India’s problems are, at best, optimistic. Algorithms like variant AI tools from platforms such as TikTok flagged innocent content in over 25% of takedown attempts globally, particularly in multi-context cultural systems. In a country with 19,500 dialects (Census 2011), one must strongly doubt whether even advanced tools possess the capacity for fair linguistic moderation.

The Uncomfortable Questions Nobody Asks

The government's policy directive raises deeper concerns about institutional capacity. Enforcement largely falls on private platforms, but is MeitY prepared to monitor compliance at scale? Platforms such as Instagram reportedly host over 95 million photo uploads per day. From complaint registration to 24-hour takedowns, the regulatory machinery fails to address how oversight can realistically function for intermediaries navigating vast, multilingual user bases with limited human moderation teams.

Second, there’s the accountability paradox. Platforms may opt for excessive removals under vague definitions of “obscene” or “illegal,” fearing litigation more than user backlash. This risks overreach—think art censorship, LGBTQ+ content takedowns masquerading as decency regulation, or feminist advocacy campaigns flagged for nudity violations. Without safeguards against capricious moderation, India may find itself in a judicial quagmire worse than the problem it seeks to fix.

Third—and most politically fraught—is the impact on free speech. The directive’s framing ignores the thin line between decency regulation and viewpoint suppression. Who defines obscenity? A platform’s algorithms? The State censor? Without independent appellate adjudication mechanisms, the advisory reinforces concerns about power concentration, particularly ahead of the 2026 general elections.

Lessons from Australia’s "eSafety Commissioner"

India’s pivot towards stricter oversight may have drawn inspiration from global efforts, but its execution lacks nuance. Consider Australia’s eSafety Commissioner, which enforces 48-hour compliance for flagged harmful content, including cyberbullying and online child exploitation. Unlike India’s system, Australia integrates statutory penalties, independent investigation powers, and transparency benchmarks, ensuring platforms are both supported and held accountable. Importantly, its classification system distinguishes between harmful and artistic nudity—a nuance Indian regulators often bypass in favor of blanket categorization.

Despite its relative success, even Australia grapples with jurisdictional non-compliance for platforms headquartered abroad. India, which hosts few major intermediaries locally, should anticipate similar challenges in enforcing takedown directives.

📝 Prelims Practice
  • Question 1: Under which section of the IT Act, 2000 does the Indian government retain the power to block online content for reasons of public order?
    • a) Section 61
    • b) Section 69A
    • c) Section 54
    • d) Section 70A

    Answer: b) Section 69A

  • Question 2: The Intermediary Guidelines and Digital Media Ethics Code apply to which category of social media platforms?
    • a) Platforms registered within India only
    • b) Platforms with financial turnover exceeding ₹100 crore
    • c) Significant Social Media Intermediaries
    • d) All platforms where users exceed 1 million

    Answer: c) Significant Social Media Intermediaries

✍ Mains Practice Question
How far has the Information Technology Rules, 2021 succeeded in balancing free speech with content regulation? Critically evaluate the structural limitations of this legal framework in addressing obscene and illegal material online.
250 Words15 Marks

Practice Questions for UPSC

Prelims Practice Questions

📝 Prelims Practice
Consider the following statements about the 24-hour takedown advisory and the existing legal framework:
  1. By prescribing a 24-hour action window upon complaint, the advisory makes platform timelines more determinate compared to earlier expectations of “reasonable efforts”.
  2. Section 69A of the IT Act, 2000 is presented as enabling government action such as blocking and potential criminal prosecution in cases of non-compliance.
  3. The advisory clearly specifies monetary penalties that automatically apply if an intermediary misses the 24-hour takedown deadline.

Which of the above statements is/are correct?

  • a1 and 2 only
  • b2 and 3 only
  • c1 and 3 only
  • d1, 2 and 3
Answer: (a)
📝 Prelims Practice
Consider the following statements about automated moderation and enforcement capacity highlighted in the article:
  1. The advisory promotes proactive automated tools even though platforms have raised concerns about algorithmic bias and language challenges in a multilingual environment.
  2. The grievance redressal ecosystem is portrayed as fully capacity-adequate, with the appellate mechanism keeping pace with complaint volumes.
  3. Transparency in takedown reporting is depicted as incomplete, creating uncertainty about moderation criteria and risks of arbitrary or selective enforcement.

Which of the above statements is/are correct?

  • a1 and 3 only
  • b1 and 2 only
  • c2 and 3 only
  • d1, 2 and 3
Answer: (a)
✍ Mains Practice Question
Critically examine the 24-hour takedown advisory under the IT Rules, 2021 in terms of (i) legal clarity and enforceability, (ii) institutional capacity of grievance redressal, and (iii) risks from automated moderation in a multilingual society. Suggest measures to balance user safety, transparency, and due process. (250 words)
250 Words15 Marks

Frequently Asked Questions

How does the 24-hour takedown advisory change the compliance burden on social media intermediaries?

The advisory converts a broad “reasonable efforts” expectation into an explicit 24-hour timeline for prima facie sexual/obscene content once a complaint is received. This shifts compliance risk squarely onto platforms by compressing decision time and raising the stakes of delayed action. It also nudges intermediaries from reactive removals toward faster, system-driven workflows.

What is the legal basis cited for blocking or prosecuting non-compliance, and what ambiguity remains around penalties?

The advisory draws authority from the IT Act, 2000—especially Section 69A—alongside the Intermediary Guidelines and Digital Media Ethics Code, 2021. While Section 69A enables blocking and can support criminal prosecution, the advisory is unclear on whether missing the 24-hour deadline triggers measurable fines or structured judicial review. This vagueness can incentivize “good faith” defenses and prolonged litigation rather than swift enforcement.

Why is the push for automated moderation contentious in the Indian context as per the article?

Automated moderation has faced resistance due to algorithmic bias risks and the difficulty of accurately moderating across India’s multilingual ecosystem. The article notes that even large global firms resisted full automation, citing language challenges and bias concerns. Without clear standards for algorithms and safeguards for context, automation may amplify erroneous takedowns or inconsistent enforcement.

What institutional bottlenecks could undermine the advisory’s grievance-redressal and enforcement objectives?

The Grievance Appellate Committee formed in 2023 is described as sluggish and insufficiently staffed, handling about 3,000 cases in its first year against a much larger annual complaint load for major platforms. Such capacity gaps can create backlogs that dilute the intent of fast takedowns and effective remedies. Fragmented enforcement architecture further complicates consistent application across intermediaries.

What transparency and accountability concerns arise from the current takedown reporting and moderation practices?

Although the IT Rules require periodic reporting on takedowns, only a small fraction of flagged content was disclosed by intermediaries in 2024 as per a MeitY report, creating opacity over moderation criteria. This “black box” environment raises concerns of arbitrary removals, over-censorship, or selective enforcement influenced by ideological leanings. Limited disclosure also weakens public auditability of whether the 24-hour mandate improves outcomes.

Source: LearnPro Editorial | Environmental Ecology | Published: 31 December 2025 | Last updated: 3 March 2026

Share
About LearnPro Editorial Standards

LearnPro editorial content is researched and reviewed by subject matter experts with backgrounds in civil services preparation. Our articles draw from official government sources, NCERT textbooks, standard reference materials, and reputed publications including The Hindu, Indian Express, and PIB.

Content is regularly updated to reflect the latest syllabus changes, exam patterns, and current developments. For corrections or feedback, contact us at admin@learnpro.in.

Related Posts

Science and Technology

Missile Defence Systems

Context The renewed hostilities between the United States-led coalition (including Israel and United Arab Emirates) and Iran have tested a newly integrated regional air and missile defence network in West Asia. What is a missile defence system? Missile defence refers to an integrated military system designed to detect, track, intercept, and destroy incoming missiles before they reach their intended targets, thereby protecting civilian populations, military installations, and critical infrastruct

2 Mar 2026Read More
International Relations

US-Israel-Iran War

Syllabus: GS2/International Relations Context More About the News Background of the Current Escalation Global Implications Impact on India Way Forward for India About West Asia & Its Significance To Global Politics Source: IE

2 Mar 2026Read More
Polity

Securities and Exchange Board of India (SEBI) on Market Manipulators

Context The Securities and Exchange Board of India (SEBI) will enhance surveillance and enforcement on market manipulators and cyber fraudsters through technology and use Artificial Intelligence (AI). Securities and Exchange Board of India (SEBI) It is the regulatory authority for the securities and capital markets in India. It was established in 1988 and given statutory powers through the SEBI Act of 1992.

2 Mar 2026Read More
Polity

18 February 2026 as a Current Affairs Prompt: How to Convert a Date into UPSC Prelims-Grade Facts (Acts, Rules, Notifications, Institutions)

A bare date like “18-February-2026” is not a defensible current-affairs topic unless it is anchored to a primary instrument such as a Gazette notification, regulator circular, court judgment, or a Bill/Act. The exam-relevant task is to convert the date into verifiable identifiers—issuing authority, legal basis (Act/Rules/Sections), instrument number, effective date, and thresholds—because UPSC frames MCQs around precisely these hard edges. The central thesis: the difference between narrative awareness and Prelims accuracy is source hierarchy discipline.

2 Mar 2026Read More

Enhance Your UPSC Preparation

Study tools, daily current affairs analysis, and personalized study plans for Civil Services aspirants.

Try LearnPro AI Free

Our Courses

72+ Batches

Our Courses
Contact Us