NextFin News - Australia’s online safety regulator has issued a sharp rebuke to the world’s largest technology companies, accusing them of failing to implement basic safety measures to protect children from sexual exploitation and abuse. In a transparency report released on Thursday, February 5, 2026, the eSafety Commissioner, Julie Inman Grant, detailed systemic shortfalls across platforms operated by Meta Platforms Inc, Apple Inc, and Google, noting that the industry has largely ignored repeated calls to address known security gaps.
The report, the second in a series of four planned transparency assessments, highlights a stark disconnect between the technological capabilities of trillion-dollar companies and their actual deployment of safety tools. According to the eSafety Commissioner, key failings include the inadequate detection of live abuse during video calls and insufficient efforts to identify newly created material. Specifically, the regulator criticized the lack of language analysis tools designed to detect the sexual extortion of Australian children, even after the commission provided companies with specific online indicators to track such behavior.
The scale of the crisis is underscored by data from the Australian Centre to Counter Child Exploitation, which received nearly 83,000 reports of online child sexual abuse material (CSAM) in the 2024–25 financial year—a 41% increase from the previous period. Most of these reports originated from mainstream platforms. While some progress was noted—such as Snap Inc reducing its moderation response time from 90 minutes to 11 minutes and Microsoft expanding detection within Outlook—the regulator found that Meta and Google continue to leave video calling services like Messenger and Google Meet unmonitored for live-streamed abuse.
The eSafety report further revealed that Apple and Discord are failing to implement proactive detection, with Apple relying almost exclusively on user reports rather than automated safety technology. Furthermore, Apple, Discord, Google’s Chat and Meet, Microsoft Teams, and Snap are currently not utilizing available software to detect the sexual extortion of children. Inman Grant stated that it "beggars belief" that these tools have not been deployed, suggesting that the lack of progress represents a failure of corporate will rather than technical ability.
This regulatory friction occurs against the backdrop of Australia’s landmark social media ban for children under 16, which was inaugurated in December 2025. Since the ban took effect, government data indicates that over 4.7 million Australian social media accounts have been deactivated. However, the eSafety Commissioner argues that transparency and age-gating are insufficient. The regulator is now pushing for a legally mandated "Digital Duty of Care," a framework that would shift the burden of proof onto tech companies to demonstrate that their systems are "safe by design" before they are launched to the public.
From an analytical perspective, the persistent resistance of Big Tech to adopt comprehensive detection tools reflects a fundamental conflict between safety-by-design and the industry’s core growth metrics. For many of these firms, implementing intrusive automated scanning on encrypted or live-streamed services presents significant engineering challenges and potential privacy backlash. However, the eSafety Commissioner’s findings suggest that the industry is prioritizing user friction and operational costs over the prevention of high-impact harms. The fact that Meta uses detection technology for Facebook Live but not for Messenger video calls indicates a selective application of safety resources that leaves vulnerable gaps for predators to exploit.
The rise of AI-generated abuse material adds a new layer of complexity to this regulatory battle. As generative AI tools become more sophisticated, the volume of newly created, non-indexed abuse material is expected to surge. The eSafety Commissioner’s emphasis on language analysis and proactive detection is a direct response to this trend. Without these tools, platforms remain reactive, only removing material after it has already been reported by victims or third parties—a process that often occurs too late to prevent psychological trauma.
Looking forward, the Australian government’s aggressive stance is likely to serve as a global bellwether for tech regulation. If the proposed Digital Duty of Care is legislated, it will represent one of the most significant shifts in internet governance since the inception of the World Wide Web. This would move the industry away from the "notice and takedown" model toward a preventative regime where safety is a non-negotiable utility. For investors and industry analysts, this signals an era of increased compliance costs and potential liability for tech giants, as regulators move from requesting transparency to demanding systemic accountability.
The eSafety Commissioner’s dashboard, launched alongside the report, will continue to track metrics such as the size of trust and safety workforces and the efficacy of automated systems. As U.S. President Trump’s administration continues to navigate its own relationship with Big Tech, the Australian model of rigorous, data-driven transparency reports may provide a template for other jurisdictions seeking to curb the influence of algorithms that prey on the vulnerabilities of minors. The ultimate trend is clear: the era of self-regulation for Big Tech is rapidly coming to an end, replaced by a mandate to put corporate conscience on par with quarterly profits.
Explore more exclusive insights at nextfin.ai.
