Get AI summaries of any video or article — Sign up free
How the "Online Safety Act" might destroy the web thumbnail

How the "Online Safety Act" might destroy the web

Theo - t3․gg·
6 min read

Based on Theo - t3․gg's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

The Online Safety Act requires many UK-accessible services to implement “robust age checks,” often using sensitive identity data rather than a government-run verification system.

Briefing

The UK’s Online Safety Act is pushing websites to verify users’ ages—often through government-issued ID, face scans, or third-party “age assurance” services—at the same time it demands tighter controls on “harmful” content. The central worry is that forcing identity checks at scale will not reliably protect children online, while it will substantially increase privacy risks, data-leak exposure, censorship-by-proxy, and the operational burden that can reshape how the open web works.

Supporters of the law point to real harm experienced by young people online. Internet safety advocates cite survey findings that a significant share of children aged 9 to 17 have encountered harm online, and the act’s stated goal is to prevent minors from accessing content deemed harmful. But the critique is that the chosen mechanism—mandatory age checks across a wide range of services, including social media, search engines, music platforms, and adult content—turns a child-protection problem into a system-wide identity and moderation regime.

Under the act’s framework, platforms face steep consequences for failing to comply: fines up to 10% of global revenue and, in some cases, court-ordered blocking. The law defines harmful content in multiple categories, including pornography, content encouraging self-harm, abusive content tied to protected characteristics, incitement of hatred, instructions for serious violence, bullying, depictions of serious injury, “stunts and challenges” likely to cause harm, and content encouraging self-administration of harmful substances. It also includes broader “material risk” standards, requiring providers to judge whether content could cause significant harm to an appreciable number of children.

The practical implementation is where the backlash concentrates. The act requires platforms to run “robust age checks” rather than relying on a government-run verification system. That means companies must collect and process sensitive identity data—passport details, selfies or videos, or even bank-access permissions for age confirmation—often via third-party vendors. Critics argue this expands the number of organizations handling highly sensitive data, increasing the odds of breaches and identity theft. They also warn that face-based checks and identity pipelines can enable algorithmic discrimination and expose users to “opaque supply chains,” where people cannot easily see how their data is stored, shared, or secured.

The concern extends beyond privacy. Mandatory age verification can restrict free expression by forcing platforms to arbitrate what speech minors can access, and it can create a chilling effect for users who lack a personal device or ID. The rollout has already coincided with a surge in VPN downloads in the UK and a petition surpassing 400,000 signatures, which supporters interpret as evidence that users are seeking workarounds rather than embracing the new regime.

The argument also draws parallels to past internet-restricting efforts such as the Stop Online Piracy Act (SOAPA), which would have empowered ISPs and advertisers to cut off access based on copyright claims—an approach widely seen as damaging to the free web. In the same spirit, the Online Safety Act is framed as a policy that shifts responsibility down the chain—from parents to platforms to identity providers—so that when systems fail, the harm multiplies.

The proposed alternative is not abandoning child safety, but changing the method: fund education for parents and children, improve device-level controls, and avoid a universal identity-verification mandate that critics say will inevitably increase privacy risk, data-leak surface area, and circumvention—while failing to solve the underlying problem of children encountering harmful content.

Cornell Notes

The UK’s Online Safety Act aims to protect children online by requiring platforms to verify users’ ages before they can access large portions of the internet. The law threatens major penalties (including fines up to 10% of global revenue and potential blocking) for services that don’t implement “robust age checks” and content safeguards. Critics say the approach will not reliably make children safer because it forces companies to collect sensitive identity data—often via ID documents, face scans, or third-party age assurance—creating new privacy and security risks. They also argue it restricts free expression and shifts responsibility away from parents, increasing harm when systems fail. The backlash includes increased VPN use and calls for repeal, with similar age-check proposals discussed in other regions.

What mechanism does the Online Safety Act use to protect children, and why is it controversial?

It relies on mandatory age verification for access to services that could contain harmful content. Platforms must implement “robust age checks” (not a government-run system) and apply algorithmic and moderation measures to reduce minors’ exposure. The controversy centers on the fact that age checks often require collecting sensitive identity data—such as government-issued ID, passport details, or face selfies/videos—plus third-party processing. Critics argue this expands privacy risk and increases the chance of identity theft or data leaks while also restricting speech by forcing platforms to gate access.

How does the act define “harmful” content, and what kinds of services are affected?

Harmful content is grouped into categories including pornography, content encouraging self-harm, abusive content tied to race/religion/sex/sexual orientation, incitement of hatred, instructions for serious violence, bullying, serious injury depictions (real or fictional), and “stunts and challenges” likely to cause serious injury. It also includes content encouraging self-administration of harmful substances. The affected services are broad: social media, search engines, music sites, adult content providers, and other services accessible in the UK.

What are the compliance consequences, and how do they shape platform behavior?

Failure to comply can trigger fines up to 10% of global revenue and, in some circumstances, court blocking of the service. Critics argue these penalties incentivize companies to implement age-check systems quickly—often using third-party vendors—without having strong data security practices. That pressure can lead to rushed deployments, opaque data handling, and increased exposure if identity systems are breached.

Why do critics say age verification increases privacy and security risks?

Because age assurance creates more “surface area” for data leaks and identity attacks. Instead of only moderating content, platforms and identity providers must handle sensitive data (IDs, biometrics, and sometimes bank-linked confirmation). Critics also highlight that third-party supply chains can be opaque—users may not know where data is stored or how it’s protected—making it harder to assess risk. They point to historical breach patterns (e.g., “data leak” incidents) as evidence that scope can be limited in some cases, but could be much larger if a major identity provider is compromised.

How does the backlash show up in user behavior and public response?

After age checks took effect, VPN apps reportedly became the most downloaded on Apple’s App Store in the UK. A petition calling for repeal reportedly surpassed 400,000 signatures. Critics interpret this as evidence that users are seeking workarounds rather than accepting the new identity-gating model.

What alternative approach does the critique favor for protecting children online?

It argues for shifting from identity verification mandates to education and parental tools. The preferred model is funding resources that teach parents and children how to reduce exposure to harmful content, possibly improving device-level controls and screen-time enforcement. The critique also warns against offloading responsibility down the chain—parents to platforms to identity providers—because failures at any layer can produce broader harm than the original child-safety problem.

Review Questions

  1. What specific types of “harmful” content does the Online Safety Act target, and how does that breadth affect which services must comply?
  2. Why do critics argue that third-party age assurance increases the risk of data leaks and identity theft compared with content moderation alone?
  3. How does the critique use past internet-restricting efforts (like SOAPA) to frame the potential long-term effects of age verification mandates?

Key Points

  1. 1

    The Online Safety Act requires many UK-accessible services to implement “robust age checks,” often using sensitive identity data rather than a government-run verification system.

  2. 2

    Platforms face severe penalties for noncompliance, including fines up to 10% of global revenue and potential court blocking, which critics say encourages rushed and risky implementations.

  3. 3

    The law’s definition of harmful content spans pornography, self-harm, hate and abuse, serious violence, bullying, injury depictions, dangerous stunts, and self-administration of harmful substances.

  4. 4

    Mandatory age verification can restrict free expression and exclude users who lack a personal device or ID, while also increasing privacy and security exposure through identity pipelines.

  5. 5

    Critics argue that shifting responsibility from parents to platforms to third-party identity providers increases harm when systems fail, including the risk of large-scale identity theft.

  6. 6

    Workarounds have already gained traction, with VPN downloads rising in the UK and repeal petitions gathering hundreds of thousands of signatures.

  7. 7

    The favored alternative is education and better parental/device controls rather than universal identity verification mandates that critics say will be circumvented and will not solve the underlying safety problem.

Highlights

The act’s compliance model hinges on age verification, with penalties up to 10% of global revenue—pushing platforms toward collecting ID and biometric data at scale.
Critics argue that forcing identity checks increases the number of organizations handling sensitive information, raising the odds of breaches and identity theft.
The rollout has been met with user circumvention signals, including a spike in VPN downloads and a large repeal petition.
The critique frames the policy as a responsibility-shift chain (parents → platforms → identity providers) where failures can cause broader harm than the original child-safety goal.

Topics

  • Online Safety Act
  • Age Verification
  • Privacy Risks
  • Net Neutrality
  • Identity Theft

Mentioned

  • OSA
  • EFF
  • SOAPA