URGENT UPDATE: Major tech companies including Snapchat, Facebook, Instagram, and TikTok have agreed to comply with Australia’s new social media ban for users under the age of 16, set to take effect on December 10, 2023. However, executives have voiced serious concerns, warning that this restriction may push minors into “darker corners” of the internet where safety protections are nonexistent.
During a heated parliamentary inquiry on December 5, tech leaders expressed their reservations about the effectiveness of the age ban. TikTok’s public policy lead, Ella Woods-Joyce, emphasized that the “blunt” restrictions would not adequately address the safety issues they aim to resolve. “We support evidence-based sensible legislation that improves safety standards for all internet users,” she stated. “A ban will push younger people into darker corners of the internet where rules, safety, tools and protections don’t exist.”
The new law mandates that companies must detect and remove approximately 1.5 million accounts belonging to underage users on platforms like Facebook, Instagram, YouTube, TikTok, Threads, and X. This sweeping compliance effort poses a significant challenge, particularly for Meta’s regional director of policy, Mia Garlick, who noted the complexities in distinguishing between ages 13 and 16, which is a “globally novel age boundary.”
In addition to the ban, platforms that fail to comply could face fines of up to $50 million. However, young users or their families will not face penalties for circumventing the new rules. The onus is clearly on tech companies to enforce these measures effectively.
Snap Inc. representative Jennifer Stout raised alarms about the potential negative impact on teens, asserting that restricting their access to social media could harm their mental well-being. “For teens, connection and communication are strong drivers of happiness and well-being,” she explained. “Taking that away does not necessarily make them safer and may instead push them towards other messaging services that lack Snapchat’s safety and privacy protections.”
The inquiry has intensified scrutiny on these companies, especially after Greens senator Sarah Hanson-Young previously threatened to compel executives from TikTok, Snapchat, and Meta to testify after they missed an earlier session. The law places significant responsibility on these platforms to ensure compliance, with experts warning that failure to adapt could lead to increased risks for minors.
In a related move, tech giants Apple and Google have removed the app OmeTV from their stores following reports of its use by predators to exploit Australian children. The app, which randomly connects users for video chats, was flagged by eSafety Commissioner Julie Inman Grant, who stated, “This app will no longer be able to reach Australians and they will no longer be able to make money off children’s misery.”
As the deadline approaches, the tech industry faces mounting pressure to prioritize the safety of young users. The implications of this ban extend beyond Australia, raising critical questions about global social media practices and child protection.
Stay tuned for more updates as this story develops.

































