Discord announced a global rollout of age verification and teen-by-default safety controls to restrict access to sensitive content. 

The update reflects mounting pressure on platforms to protect minors from harmful design and exposure.

Global age checks and teen-first defaults

Discord said it will require users worldwide to confirm their age to access adult content. 

Verification will occur through a facial scan or approved identification. 

The company plans to begin enforcement in early March.

The platform stated that all new and existing accounts will receive teen-appropriate defaults. 

These include tighter communication controls, content filters, and limits on age-gated spaces. 

Users must verify adulthood to view sensitive material or join restricted servers.

Discord already applies age checks in the United Kingdom and Australia. 

The expanded policy extends those measures across all regions. 

The company said the approach balances protection for teens with flexibility for verified adults.

The service reported more than 200 million monthly users. 

It allows people to form and join communities around shared interests. 

The update will also restrict direct messages from unknown users until age checks are completed.

Savannah Badalich, head of product policy, said teen safety remains a core priority. 

She said the global defaults build on existing protections while preserving privacy and meaningful connections.

Regulatory pressure intensifies across platforms

Discord will soon be expanding teen safety protections worldwide including teen-by-default settings and age assurance designed to create safer experiences for teens.We’re also launching recruitment for Discord's first Teen Council, creating a space for teen voices to help shape… pic.twitter.com/CW7G4sO38R

— Discord Support (@discord_support) February 9, 2026

Discord’s move follows rising global concern over social media design and youth well-being. 

Governments and regulators are examining whether platforms sufficiently protect minors.

The European Union recently accused TikTok of breaching digital rules. 

Regulators cited addictive features that may drive compulsive use among children. 

Officials pointed to autoplay and infinite scroll as risk factors.

EU investigators said TikTok failed to assess impacts on physical and emotional health. 

The European Commission urged changes to the service’s basic design. 

The findings followed a two-year probe.

Industry analysts said stricter standards are becoming unavoidable. 

Drew Benvie of Battenhall said safety measures support healthier online environments. 

He said clearer age controls can reduce exposure risks.

Major companies face landmark child harm trials

Legal scrutiny has also escalated in the United States. Several social media companies face trials over alleged harm to children. 

Opening arguments began in early February in Los Angeles County Superior Court.

Claims allege that YouTube and Instagram intentionally fostered addiction in minors. 

TikTok and Snap reached settlements for undisclosed amounts. 

The remaining defendants deny wrongdoing.

Attorney Mark Lanier argued that companies engineered addictive systems for young users. 

Prosecutors said algorithms kept minors online despite known risks. 

They cited exposure to exploitation and misleading safety claims.

The cases mark a pivotal moment for the industry. 

Courts may determine accountability for design choices affecting children. 

Outcomes could shape future platform standards.

Discord’s global changes arrive as scrutiny converges from regulators and courts. 

The company framed the update as a proactive step. 

The broader industry now faces sustained demands to prioritize youth safety.

The post Discord expands global age verification amid rising scrutiny on teen safety first appeared on Coinfea.