Explosion
Meta Uses AI Bone Structure Analysis to Catch Underage Users
Technology

Meta Uses AI Bone Structure Analysis to Catch Underage Users

Ava MitchellBy Ava Mitchell·

This week, Meta confirmed it’s using artificial intelligence to analyze users’ physical traits, such as height and bone structure, visible in photos and videos. This helps identify and remove underage users from its platforms.

How the System Works

The AI system scans visual signals in uploaded content to gauge if a user might be younger than the minimum age requirement of 13. Picture a bouncer who doesn’t check IDs but looks at your wrists and posture to estimate your age. Right now, Meta says this technology is available in select countries, with plans for a broader global rollout soon.

When we talk about bone structure analysis, it means the AI examines physical proportions seen in photos and videos. This includes factors like limb length in relation to body height, which typically vary between children and adults. It’s a similar concept to some medical imaging tools, but applied to social media content.

Meta is also expanding its Teen Account system, which adds restrictions for younger users, from Instagram to Facebook. Teen Accounts, which limit direct messages, content recommendations, and screen time for users under 18, were initially an Instagram-only feature launched in 2024.

Why Meta Is Doing This Now

Meta faces ongoing pressure from lawmakers, regulators, and parents regarding child safety on its platforms. Several U.S. states have passed or are working on legislation that demands stricter age verification for social media users. The EU’s Digital Services Act also requires large platforms to protect minors.

Traditional age verification methods, like having users enter their birthdate, can easily be bypassed. A 12-year-old can easily input a fake birth year. Meta’s approach aims to go further by using signals that users haven’t explicitly provided.

What This Means for Everyday Users

For most adult users, the impact should be minimal. The system targets accounts that seem to belong to kids, not every adult’s selfies. However, this technology raises important questions about how Meta analyzes the photos and videos you share, and what other conclusions they might draw from that visual data in the future.

If you’re a parent, it’s good to know about the Teen Account expansion to Facebook. This means if your teenager has a Facebook account, Meta’s age-detection and restriction tools apply there too, not just on Instagram.

There’s also a concern about false positives. Adults who appear younger could be flagged by the system. Meta hasn’t detailed how the appeal or review process will work if an adult’s account gets wrongly restricted.

Meta — Company Snapshot
Stock (META) $603.94 (-0.17%)
CEO Mark Zuckerberg
Founded 2004
Headquarters Menlo Park, CA
Sector Social Media
Teen Account expansion Instagram → Facebook
Rollout status Select countries, broader rollout planned

The Privacy Trade-Off

This is where things get tricky. Using AI to analyze users’ bodies, even indirectly through their posts, goes well beyond simply asking for a birthdate or requiring a government ID. Meta is essentially adapting its content-scanning systems into age-detection tools by interpreting biological signals from your photos.

Privacy advocates have long raised alarms about “biometric inference,” where a platform makes personal conclusions based on observed physical traits instead of information you willingly provided. Bone structure and height estimation fall into this category.

So far, Meta hasn’t released detailed information on accuracy rates, error margins, or the specific types of images that trigger this analysis.

Community Reaction

“So Meta is literally scanning your body in photos now to decide if you’re allowed to use the app. That’s insane. I get that protecting kids matters but this feels like a line crossed.”

— Reddit user, r/technology

“Honestly, if it keeps kids off these apps, I’m for it. Every other method has failed. The real question is what else they’re doing with that analysis.”

— YouTube comment on TechCrunch coverage

What To Watch

  • Global rollout timeline: Meta hasn’t provided a specific date for when the bone structure AI system will expand beyond its current test countries. Keep an eye out for announcements related to upcoming EU regulatory deadlines.
  • Legislative response: Ongoing U.S. congressional hearings on child safety and social media in 2026 will likely discuss this new system. Lawmakers will consider whether AI-based age detection is sufficient or if mandatory ID verification is necessary.
  • Privacy challenges: With Meta’s ongoing legal issues, including a separate 2026 copyright lawsuit from publishers regarding AI training data, privacy advocates and possibly regulators will closely examine the biometric implications of this system.
  • Accuracy data: If Meta shares technical details on how often the system wrongly flags adult users, that will be crucial to monitor.

Sources: TechCrunch | Engadget

Ava Mitchell

Ava Mitchell

Ava Mitchell is a digital culture journalist at Explosion.com covering social media platforms, streaming services, and the creator economy. With 4 years reporting on TikTok, Instagram, YouTube, and the apps that shape daily life, Ava specializes in explaining platform policy changes and their impact on everyday users. She previously managed social media strategy for a tech startup, giving her firsthand experience with the platforms she now covers.