Context:
- On December 10, Australia became the first country globally to enforce a blanket social media ban for users under 16 years of age.
- The move follows growing concerns over children’s mental health, cyberbullying, harmful content, and online predatory practices.
- The policy has attracted global attention, with several countries considering similar regulations.
Key Highlights:
The New Law
- Online Safety Amendment (Social Media Minimum Age) Bill, 2024, introduced in November 2024.
- Mandates a minimum age of 16 years to hold accounts on specified social media platforms.
- Parental consent is not allowed to bypass the age restriction.
Platforms Covered Under the Ban
- Major platforms blocked for children under 16 include:
- X (Twitter)
- TikTok
- Snapchat
- YouTube
- Twitch
- Kick (livestreaming platform)
Penalties for Non-Compliance
- Platforms failing to take “reasonable steps” to restrict underage users face fines of up to:
- A$49.5 million (≈ $33 million)
Government Rationale
- Social media identified as a fertile ground for:
- Cyberbullying
- Harmful and addictive content
- Online grooming and predatory behaviour
- Aim is to protect mental health and well-being of minors.
Relevant Prelims Points:
- Country: Australia (first mover globally).
- Law Name: Online Safety Amendment (Social Media Minimum Age) Bill, 2024.
- Minimum Age Prescribed: 16 years.
- Regulatory Authority: Australian Communications and Media Authority (ACMA).
- Penalty Provision: Up to A$49.5 million fine.
- Age Verification Methods Discussed:
- Government ID
- Facial recognition
- Video selfie (via third-party tools like Yoti)
- Behavioural age inference
- Concerns Identified:
- False rejection rates in facial age estimation
- Surveillance and privacy risks
Relevant Mains Points:
- Governance & Regulation of Big Tech:
- Marks a shift towards strong state intervention in digital platforms.
- Raises questions on platform accountability vs individual freedoms.
- Mental Health & Child Safety:
- Internal reports from Meta and TikTok revealed:
- Awareness of addictive nature of platforms
- Links with depression, anxiety, loneliness, and social comparison
- Admission that minors lack executive mental control over screen time
- Strengthens the case for precautionary regulation.
- Internal reports from Meta and TikTok revealed:
- Implementation Challenges:
- Accurate age verification without violating privacy.
- Risk of false positives excluding legitimate users aged 16–17.
- Potential digital exclusion and isolation of teenagers.
- Ethical & Rights Concerns:
- Surveillance of children through biometric verification.
- Balance between right to protection and right to information.
- Way Forward:
- Develop privacy-preserving age verification technologies.
- Standardise age checks at app-store level (Apple, Google Play).
- Continuous policy review and impact assessment.
- Global cooperation on child safety standards online.
UPSC Relevance (GS-wise):
- GS II: Governance, public policy, regulation of digital platforms
- GS III: Cyber security, technology and society
- GS IV: Ethics – child protection, corporate responsibility, digital morality
