Facebook and Instagram face fresh EU digital scrutiny over child safety measures

Image

FILE - The Meta logo is seen at the Vivatech show in Paris, France, Wednesday, June 14, 2023. The European Union opened fresh investigations Thursday, May 16, 2024 into Facebook and Instagram over suspicions that they’re failing to protect children online, in violation of the bloc’s strict digital regulations for social media platforms. (AP Photo/Thibault Camus, File)

LONDON (AP) — The European Union opened fresh investigations Thursday into Facebook and Instagram over suspicions that they’re failing to protect children online, in violation of the bloc’s strict digital regulations for social media platforms.

It’s the latest round of scrutiny for parent company Meta Platforms under the 27-nation EU’s Digital Services Act, a sweeping set of regulations that took effect last year with the goal of cleaning up online platforms and protecting internet users.

The European Commission, the bloc’s executive arm, said it’s concerned that the algorithmic systems used by Facebook and Instagram to recommend content like videos and posts could “exploit the weaknesses and inexperience” of children and stimulate “addictive behaviour.” It’s worried that these systems could reinforce the so-called “rabbit hole” effect that leads users to increasingly disturbing content.

The commission is also looking into Meta’s use of age verification tools to prevent children from accessing Facebook or Instagram, or be shown inappropriate content. The platforms require users to be at least 13 years old to set up an account. It’s also looking into whether the company is complying with DSA rules requiring a high level of privacy, safety and security for minors.

“We want young people to have safe, age-appropriate experiences online and have spent a decade developing more than 50 tools and policies designed to protect them,” Meta said in a prepared statement. “This is a challenge the whole industry is facing, and we look forward to sharing details of our work with the European Commission.”

They’re the latest DSA cases to focus on child protection under the DSA, which requires platforms to put in place stringent measures to protect minors. The commission opened two separate investigations earlier this year into TikTok over concerns about risks to kids.

“We are not convinced that Meta has done enough to comply with the DSA obligations — to mitigate the risks of negative effects to the physical and mental health of young Europeans on its platforms Facebook and Instagram,” European Commissioner Thierry Breton said in a social media post.

The cases announced Thursday aren’t the first for Facebook and Instagram. They’re are already being investigated under the DSA over concerns they’re not doing enough to stop foreign disinformation ahead of EU elections next month.

Social media platform X and ecommerce site AliExpress are also being investigated over their compliance with the EU rules.

There’s no deadline for the investigations to wrap up. Violations could result in fines of up to 6% of a company’s annual worldwide revenue.