Free SKILL.md scraped from GitHub. Clone the repo or copy the file directly into your Claude Code skills directory.
npx versuz@latest install onfire7777-universal-ai-skills-library-plugin-codex-skills-age-verification-methodsgit clone https://github.com/onfire7777/universal-ai-skills-library.gitcp universal-ai-skills-library/SKILL.MD ~/.claude/skills/onfire7777-universal-ai-skills-library-plugin-codex-skills-age-verification-methods/SKILL.md---
name: age-verification-methods
license: Apache-2.0
description: 'Evaluates and implements age estimation and verification technologies for online services. Covers facial age estimation, digital ID verification, self-declaration with risk assessment, AI-based age estimation, and the accuracy versus privacy tradeoff. Includes ICO guidance and euCONSENT framework. Keywords: age verification, age estimation, facial analysis, digital ID, children, online safety.'
metadata:
metadata:
author: mukul975
version: '1.0'
domain: privacy
subdomain: children-data-protection
tags: age-verification, age-estimation, facial-analysis, digital-id, children, online-safety
---
# Age Verification and Estimation Methods
## Overview
Age verification and age estimation are distinct but complementary approaches to determining whether a user is a child for the purpose of applying appropriate data protection safeguards. Age verification provides a definitive confirmation of age through documentary or transactional evidence. Age estimation provides a probabilistic assessment of age using technological methods such as facial analysis, behavioural analysis, or device signals. The selection of an appropriate method requires balancing accuracy, privacy impact, accessibility, and proportionality. This skill covers the full spectrum of available methods, their regulatory context under the GDPR, UK AADC, COPPA, and emerging legislation such as the EU Digital Services Act (DSA) and the UK Online Safety Act 2023, and provides implementation guidance based on ICO and CNIL recommendations.
## Regulatory Context
### GDPR Article 8(2)
"The controller shall make reasonable efforts to verify in such cases that consent is given or authorised by the holder of parental responsibility over the child, taking into consideration available technology."
The "reasonable efforts" standard is context-dependent. The EDPB has not prescribed specific technologies but expects controllers to adopt verification proportionate to the risk of the processing.
### UK AADC Standard 3 — Age-Appropriate Application
"Take a risk-based approach to recognising the age of individual users and ensure you effectively apply the standards in this code to child users." The ICO guidance states that the level of certainty required depends on the risks to children from the processing. Higher risks demand more robust age assurance methods.
### UK Online Safety Act 2023
Section 11(3) requires providers of regulated user-to-user services and search services to use "proportionate systems or processes" designed to prevent children from encountering primary priority content that is harmful to children. Ofcom's codes of practice specify age verification as a recommended measure for pornographic content and age estimation for broader content categories.
### EU Digital Services Act — Article 28
Providers of online platforms accessible to minors must put in place appropriate and proportionate measures to ensure a high level of privacy, safety, and security of minors on their service. This includes age verification for services with content restrictions.
### France — Loi SREN (2024)
France's law to regulate and secure the digital space requires age verification for access to pornographic websites, mandating technical solutions certified by CNIL that verify age without identifying the user. The CNIL-approved reference system requires a "double-blind" architecture where the identity verification provider and the content provider cannot link the user's identity to the content access.
## Age Verification Methods
### Method 1: Document-Based Verification
**Description**: User uploads or presents a government-issued identity document (passport, national ID card, driver's licence) which is verified against document security features and optionally against government databases.
**Technical Implementation**:
- Optical Character Recognition (OCR) extracts date of birth and document details
- Machine Readable Zone (MRZ) validation for passports and ID cards conforming to ICAO Doc 9303
- Document authenticity checks: hologram detection, microprint analysis, UV feature verification (for physical presentation)
- Optional: NFC chip reading for ePassports conforming to ICAO 9303 Part 10
- Optional: Liveness check to confirm the person presenting the document matches the photo
**Accuracy**: Very high (99%+ when combined with liveness detection)
**Privacy Considerations**:
- Collects highly sensitive identity data (ID number, full name, address, photo)
- Data minimisation: extract only date of birth, discard full document image immediately after verification
- Storage: do not retain the document image or full identity data; retain only a binary age-confirmed flag and a verification token
- DPIA required under Art. 35 due to large-scale processing of identity documents
**Accessibility**: Excludes individuals without government-issued ID (estimated 1.5 million UK adults lack photo ID per Electoral Commission 2021 data). Not appropriate as the sole method.
**Use Cases**: Age-restricted content (gambling, alcohol, adult content), high-risk services
### Method 2: Facial Age Estimation (AI-Based)
**Description**: Machine learning models estimate a user's age from a facial image captured by the device camera. The estimation provides an age range (e.g., "over 18" or "13-17") rather than a precise age.
**Technical Implementation**:
- Convolutional Neural Network (CNN) trained on large-scale age-labelled facial datasets
- Real-time processing on-device (edge computing) to avoid transmitting facial images to servers
- Liveness detection to prevent spoofing via photographs or video replay
- Age estimation outputs a confidence interval (e.g., estimated age 14 +/- 2 years with 95% confidence)
- The facial image is processed in volatile memory and not stored
**Accuracy**: Mean Absolute Error (MAE) of 1.5-3 years depending on the model and demographic. Accuracy varies by: age group (lower accuracy for children under 8 and adults over 65), ethnicity (documented bias in some commercial systems), lighting and image quality.
**Privacy Considerations**:
- Facial images constitute biometric data under GDPR Art. 4(14) and Art. 9(1) if used for unique identification
- When used solely for age estimation (not identification), the processing may not constitute "biometric data for the purpose of uniquely identifying" under Art. 9 — the ICO has confirmed this interpretation in its Children's Code guidance
- On-device processing with no server transmission significantly reduces privacy risk
- DPIA is recommended even when processing is on-device due to sensitivity of facial data
**Key Providers**: Yoti (Age Estimation), VerifyMyAge (EstimateMyAge), Privately SA
**ICO Position**: The ICO has stated that facial age estimation technology that processes images locally, does not store images, and does not identify the individual can be a proportionate method for age assurance. The ICO conducted a joint audit with the Australian Information Commissioner (OAIC) of Yoti's age estimation technology in 2022 and concluded it met data protection requirements when implemented with appropriate safeguards.
### Method 3: Digital Identity Verification
**Description**: User authenticates through a trusted digital identity provider (eID, digital wallet, Open Banking) that confirms age without disclosing full identity to the relying party (service provider).
**Technical Implementation**:
- OpenID Connect for Identity Assurance (OIDC4IDA) protocol for attribute-based verification
- The identity provider confirms a specific attribute (e.g., "is_over_18": true) via a signed assertion
- The relying party receives only the age attribute, not the user's name, address, or other identity data
- EU Digital Identity Wallet (eIDAS 2.0 Regulation, expected 2026 rollout) will enable selective attribute disclosure
- UK Digital Identity and Attributes Trust Framework (DIATF) provides a certification scheme for identity providers
**Accuracy**: Very high (dependent on the identity provider's verification of the underlying identity)
**Privacy Considerations**:
- Minimal data disclosure: only the specific age attribute is shared
- The identity provider knows the user's identity but not which service they are accessing (if double-blind architecture is used)
- The service provider knows the user is accessing their service but not their full identity
- Aligns with CNIL's recommended "double-blind" approach for age verification
**Use Cases**: EU/EEA services preparing for eIDAS 2.0 Digital Identity Wallet; UK services using DIATF-certified providers
### Method 4: Self-Declaration with Risk Mitigation
**Description**: User declares their age through a date-of-birth field or age-range selector. The declaration is treated as the baseline, supplemented by risk-based measures to detect false declarations.
**Technical Implementation**:
- Neutral age prompt: "What is your date of birth?" with a scrollable date picker (no calendar default to current date)
- No indication of the "correct" answer or the age threshold being applied
- Behavioural signals that may indicate false declaration: immediate re-entry with a different date, cookie evidence of prior declaration, typing speed patterns inconsistent with the declared age
- If false declaration is suspected, escalate to a higher-assurance verification method
**Accuracy**: Low as a standalone method. Children commonly misrepresent their age online. Ofcom's 2023 research found that 33% of UK 8-17 year olds have a social media profile despite being below the platform's minimum age.
**Privacy Considerations**: Minimal data collection (only declared date of birth). No biometric processing. No identity document collection.
**Use Cases**: Low-risk services as a first-line screening measure, always combined with additional safeguards for medium and high-risk services
### Method 5: Credit Card or Payment Verification
**Description**: User's age is inferred from possession of a credit card (typically issued only to adults 18+) through a monetary transaction.
**Technical Implementation**:
- Micro-transaction (USD/EUR/GBP 0.50) charged and refunded within 48 hours
- The card must be a credit card (not a debit card or prepaid card, which may be issued to minors)
- Transaction notification sent to the cardholder provides an audit trail
- Some implementations use 3D Secure (3DS2) authentication for additional identity assurance
**Accuracy**: Moderate. Establishes that the person has access to a credit card, which correlates with being over 18. Does not verify the specific age of the cardholder. Children may use a parent's card.
**Privacy Considerations**: Payment card data is subject to PCI DSS requirements. The service should not store full card details. Only the transaction confirmation and a binary "has credit card" flag should be retained.
### Method 6: Mobile Network Operator (MNO) Verification
**Description**: The mobile network operator confirms the user's age bracket based on the subscriber information associated with the SIM/eSIM, without disclosing the user's identity to the requesting service.
**Technical Implementation**:
- API call to MNO (via aggregator such as GBG, boku, or Sinch) with the user's mobile number
- MNO returns a binary response (e.g., "is_over_18": true/false) without disclosing identity
- Relies on the age data the MNO collected during subscriber registration (ID check at point of sale)
**Accuracy**: High for determining over/under 18, since MNO registration typically involves ID verification. Lower certainty for granular age (e.g., distinguishing 13 from 15) as MNOs may not record precise birth dates.
**Privacy Considerations**: The service learns only the age bracket. The MNO learns which service the user is accessing (unless intermediary architecture prevents this). DPIA recommended for the MNO's processing.
## Accuracy vs. Privacy Tradeoff Matrix
| Method | Accuracy | Privacy Impact | Proportionate For |
|--------|----------|---------------|-------------------|
| Document-Based | Very High | Very High (ID collection) | Age-restricted products (gambling, alcohol) |
| Facial Age Estimation | High (MAE 1.5-3y) | Medium (on-device) to High (server-side) | General online services, social media |
| Digital Identity | Very High | Low (attribute-only disclosure) | Any service; best privacy-accuracy balance |
| Self-Declaration | Low | Very Low | Initial screening; low-risk services only |
| Credit Card | Moderate | Medium (payment data) | Supplementary verification for parental consent |
| MNO Verification | High | Low-Medium | Mobile-first services; supplementary check |
## Risk-Based Selection Framework
### Step 1: Classify Service Risk Level
| Risk Level | Criteria | Examples |
|-----------|---------|---------|
| **High** | Direct messaging with strangers, user-generated content visible to strangers, age-restricted content, monetisation features targeting children | Social media, dating apps, gambling, online marketplaces |
| **Medium** | Content personalisation, in-app purchases, community features with moderation, educational services with profiling | EdTech platforms, gaming, streaming services |
| **Low** | Static content delivery, no social features, no data sharing, no profiling | Informational websites, single-player offline games |
### Step 2: Select Minimum Verification Level
| Risk Level | Minimum Verification | Recommended Approach |
|-----------|---------------------|---------------------|
| High | Document-Based OR Facial Estimation + Liveness | Document-based with digital identity as alternative |
| Medium | Facial Age Estimation (on-device) OR Self-Declaration + Risk Signals | Facial age estimation with escalation path |
| Low | Self-Declaration + Neutral Prompt | Self-declaration with cookie-based re-entry detection |
### Step 3: Apply Proportionality Test
For each selected method, document:
1. Why is this method necessary to protect children?
2. Could a less intrusive method achieve the same level of protection?
3. What safeguards minimise the privacy impact of the method (data minimisation, on-device processing, immediate deletion)?
4. How does this method account for accessibility (users without ID, users with disabilities)?
## BrightPath Learning Inc. — Age Verification Implementation
BrightPath Learning Inc. operates an educational platform classified as Medium risk (educational content with progress tracking and personalisation, no social features with strangers).
**Implemented Approach: Layered Verification**
1. **Layer 1 — Self-Declaration**: During registration, a neutral date-of-birth prompt with scrollable date picker. Determines initial routing (child vs. adult account).
2. **Layer 2 — Parental Account**: For users declaring an age below the applicable threshold, the parental consent flow is triggered. The parent creates an account and verifies via credit card micro-transaction.
3. **Layer 3 — Anomaly Detection**: Behavioural signals monitored for inconsistencies: IP address mismatch between child and parent accounts, identical email domains suggesting self-verification, multiple child accounts from the same parent in short succession.
4. **Layer 4 — Escalation**: If anomalies are detected, the parent is asked to complete a video verification call or upload a government-issued ID (with immediate deletion after verification).
**Data Retention for Age Verification**:
- Self-declared date of birth: retained for account lifecycle
- Credit card transaction record: retained for 90 days (reconciliation), then deleted; only "parent verified" flag retained
- Government ID image: processed in memory, not stored; verification outcome retained
- Facial images: not collected (BrightPath does not use facial age estimation)
## Emerging Standards and Legislation
### IEEE 2089.1-2024 — Age-Appropriate Digital Services Framework
Published standard providing a framework for implementing age-appropriate design in digital services, including guidance on age assurance methods and their application across different risk contexts.
### ISO/IEC 27566 — Age Assurance Systems (Under Development)
Working draft standard for age assurance systems covering both age verification and age estimation. Addresses accuracy, privacy, accessibility, interoperability, and governance requirements.
### EU Digital Identity Wallet (eIDAS 2.0)
The revised eIDAS Regulation mandates that EU Member States offer digital identity wallets to citizens by 2026. The wallet will support selective attribute disclosure, enabling users to prove they are over a specific age without revealing their full identity or date of birth. This will become the preferred age verification method for EU services.
## Integration Points
- **GDPR Parental Consent**: Age verification is a prerequisite for determining whether parental consent is required under Art. 8
- **UK AADC Implementation**: AADC Standard 3 requires risk-proportionate age assurance
- **Age-Gating Services**: Age verification results drive the age-gating decision (admit, redirect, or block)
- **COPPA Compliance**: COPPA requires age screening as part of the parental consent flow
- **Children's Privacy Notice**: The age verification result determines which version of the privacy notice to display (child-friendly vs. standard)