Australia’s Social Media Under-16 Ban Faces Compliance Challenges, Says eSafety Regulator

Ryan Patel, Tech Industry Reporter
5 Min Read
⏱️ 4 min read

Australia’s initiative to prohibit social media access for users under the age of 16 is under scrutiny as the nation’s eSafety watchdog raises concerns about compliance among major platforms. Although the legislation took effect in December 2022, regulators have observed a lack of adequate enforcement by social media giants, including TikTok, Instagram, and YouTube. Campaigners argue that this measure is vital for shielding young users from harmful content and addictive algorithms, but the platforms contend that the implementation of age verification is fraught with difficulties.

Regulatory Concerns About Compliance

The eSafety Commissioner, Julie Inman Grant, has recently highlighted “significant concerns” regarding the adherence of key social media platforms to the new law. This regulation was designed to protect minors from the perils of online engagement, yet, according to the Commissioner’s first report since its inception, many platforms are falling short of expectations.

The report reveals troubling practices by platforms such as Meta and Snap, which continue to allow children who previously declared themselves as underage to manipulate age verification methods. Furthermore, the platforms are failing to effectively prevent new under-16 users from creating accounts or provide sufficient reporting mechanisms for parents and guardians.

Inman Grant stated, “While social media platforms have taken some initial action, I am concerned through our compliance monitoring that some may not be doing enough to comply with Australian law.” The report indicates that 4.7 million accounts were either restricted or removed within the first month of the ban, but the effectiveness of these measures remains questionable.

Industry Responses and Challenges

In response to the eSafety Commissioner’s findings, representatives from leading social media firms have defended their efforts. A spokesperson from Meta expressed the company’s commitment to adhering to the new regulations, emphasising that age determination is a challenge faced industry-wide. They argue that robust age verification systems coupled with parental approval at the app store level would be more effective in safeguarding young users.

Snap, the parent company of Snapchat, reported that they have locked 450,000 accounts deemed to belong to underage users and are actively continuing this process. Despite these efforts, the persistent access of many under-16s to social media raises questions about the overall effectiveness of the ban.

The Societal Context and Parental Support

The Australian policy has garnered considerable backing from parents who find themselves in disputes with their children over social media access. The government’s stance provides a powerful ally for parents attempting to limit their offspring’s screen time. However, critics of the ban argue that merely prohibiting access is not a viable long-term solution.

Experts in child welfare and technology advocate for a more educational approach, suggesting that children should be taught to navigate these platforms responsibly rather than being outright banned. The enforcement of the ban raises further concerns about its impact on underrepresented communities, including rural youths, disabled teenagers, and LGBTQ+ individuals, who often rely on these platforms for connection and support.

Inman Grant remarked on the broader implications of the reform, stating, “Durable, generational change takes time – but these platforms have the capability to comply today.” She emphasised the necessity of collaboration between parents and the platforms to foster a cultural shift in online engagement.

Why it Matters

The ongoing challenges surrounding Australia’s social media ban for under-16s highlight the complexities of regulating digital spaces in an era where social media is deeply embedded in daily life. As countries worldwide watch closely, the outcome of this initiative may set a precedent for future regulations, balancing the need for child protection with the rights of young users. The effectiveness of the ban will ultimately depend on the commitment of social media companies to implement robust measures, as well as on the willingness of parents and society to engage in meaningful conversations about safe online practices.

Share This Article
Ryan Patel reports on the technology industry with a focus on startups, venture capital, and tech business models. A former tech entrepreneur himself, he brings unique insights into the challenges facing digital companies. His coverage of tech layoffs, company culture, and industry trends has made him a trusted voice in the UK tech community.
Leave a Comment

Leave a Reply

Your email address will not be published. Required fields are marked *

© 2026 The Update Desk. All rights reserved.
Terms of Service Privacy Policy