Discord is preventing some users from accessing sensitive content unless they allow the platform to verify their age by scanning their face or ID. The new age verification process is described as an “experiment” by Discord, and is being trialed in the United Kingdom and Australia in response to local laws that require online platforms to have guardrails in place for safeguarding children.
The new face and ID scanning requirements only apply to Discord users who haven’t previously verified their age on the platform. Discord says this is a “one-time process” that will be triggered in two situations: when users are exposed to nude or sexually explicit content that’s been flagged by Discord’s sensitive media filter, or when users attempt to change filter control…