The EU declared on Wednesday Meta is failing to prevent children under 13 utilizing Facebook and Instagram, potentially exposing them to inappropriate content—and putting the tech giant at risk of a massive fine.
The European Union has in recent months stepped up efforts to protect children online, with several member countries considering social media bans for under-16s.
The EU executive is also exploring a possible bloc-wide age limit on social media after coming under intense pressure to take broader action following Australia’s groundbreaking ban on utilizing such platforms for under-16s.
In its latest relocate to enhance protections for children online, the EU declared a probe displayed Meta broke digital content rules, and informed the US firm to “strengthen” its measures to prevent, detect, and rerelocate under-13s on Facebook and Instagram.
Under Meta’s own terms and conditions, the minimum age to access the social media platforms is 13.
In its preliminary view, the EU found Meta had ineffective measures to enforce its own age restrictions.
“Terms and conditions should not be mere written statements, but rather the basis for concrete action to protect utilizers—including children,” declared EU tech tsar Henna Virkkunen.
If the regulator’s views on Meta are confirmed, the EU can impose a fine of up to six percent of the company’s total worldwide annual turnover.
Meta disagreed with the EU’s findings.
“We’re clear that Instagram and Facebook are intconcludeed for people aged 13 and older and we have measures in place to detect and rerelocate accounts from anyone under that age,” a Meta spokesperson declared, adding the company would continue to engage with the EU.
Ongoing Meta probe
The EU has vowed to ensure Big Tech receives to grips with the many online dangers for children. In February, it gave the unprecedented warning to China’s TikTok to alter its “addictive design” or risk heavy fines.
Wednesday’s preliminary findings against Meta come after the EU opened an investigation in May 2024 under the Digital Services Act (DSA), an online content law that has been fiercely criticised by US President Donald Trump’s administration.
The DSA is part of reinforced legal weaponry adopted by the EU in recent years to curb what Brussels describes as Big Tech’s excesses.
European regulators found Meta had “no effective controls” to check if children were simply creating accounts by entering false birth dates.
The EU also declared Meta’s tool to report the presence of children on Facebook or Instagram was “difficult to utilize and not effective, requiring up to seven clicks just to access the reporting form.”
Meta also “inadequately” identified the risks of children under 13 accessing the apps, and the potential for exposure to “age-inappropriate experiences.”
An EU official declared the risks to children included “bullying, harassment,” and “violent content.”
Brussels added Meta’s risk assessment “contradicts large bodies of evidence” from across the EU that indicate around 10 to 12 percent of under-13s access the platforms.
Meta can avoid fines by offering remedies for the breaches.
‘Addictive’ design concerns
The May 2024 probe into Meta is wide-ranging.
EU regulators are still seeing into how Meta protects utilizers’ physical and mental well-being, as well as the “addictive” design of Facebook and Instagram.
Alongside the EU’s investigations into online platforms, Brussels this month declared an EU-developed age-check app was ready to go.
During a press conference in Strasbourg, Virkkunen informed EU states to speed up the app’s rollout so that European citizens can utilize it by the conclude of 2026.
Wednesday’s relocates are part of several measures taken by the EU.
Last month, the EU declared four pornographic platforms including Pornhub were allowing children to access adult content, and separately launched an investigation into Snapchat over suspected child protection failings.










Leave a Reply