UCLA
datasetactive

FairFace: Face Attribute Dataset for Balanced Race, Gender, and Age

UCLA

View original resource

FairFace: Face Attribute Dataset for Balanced Race, Gender, and Age

Summary

FairFace addresses one of the most persistent challenges in computer vision: the lack of demographic balance in training and evaluation datasets. Created by UCLA researchers, this dataset provides 108,501 face images with balanced representation across seven racial groups, two genders, and nine age ranges. Unlike traditional face datasets that skew heavily toward certain demographics, FairFace was specifically engineered to enable rigorous bias testing in facial analysis systems. It's become a go-to resource for researchers and practitioners who need to validate that their face recognition, age estimation, or demographic classification models perform equitably across different population groups.

What makes this different

Most face datasets suffer from severe demographic imbalances - often 70-80% white faces, heavy male skew, and limited age diversity. FairFace flips this script by maintaining roughly equal representation across racial categories (White, Black, East Asian, Southeast Asian, Indian, Middle Eastern, Latino_Hispanic) and gender, with systematic coverage of age ranges from 0-2 years through 70+ years.

The images themselves are high-quality, front-facing photos sourced from the YFCC-100M Flickr dataset, then carefully filtered and annotated. Each image includes ground truth labels for race, gender, and age group, verified through multiple annotation rounds. The dataset also provides both training and validation splits, making it immediately usable for model development and evaluation.

Core bias evaluation capabilities

Demographic parity testing: Compare model accuracy, false positive rates, and false negative rates across different demographic groups to identify performance disparities.

Intersectional analysis: Evaluate how models perform on intersectional identities (e.g., young Black women vs. elderly white men) rather than just single demographic categories.

Age bias assessment: Test whether age estimation models show systematic errors for certain age groups or demographic combinations.

Representation learning: Train more balanced models by using FairFace as either primary training data or supplementary data to balance existing datasets.

Who this resource is for

Computer vision researchers developing or evaluating face analysis algorithms who need to demonstrate fairness across demographic groups. Essential for publishing in venues that now require bias evaluation.

AI ethics teams and bias auditors conducting fairness assessments of existing facial recognition or demographic classification systems in production.

Product teams at tech companies building consumer applications with facial analysis features who need to ensure equitable performance across their user base.

Regulatory compliance teams at organizations subject to AI fairness regulations who need standardized datasets for bias testing and documentation.

Academic institutions teaching AI ethics or computer vision courses who want hands-on datasets for demonstrating bias in machine learning systems.

Getting started with FairFace

Download the dataset from the GitHub repository, which includes both the image files and CSV annotation files. The total download is approximately 11GB. You'll find separate folders for training (86,744 images) and validation (21,757 images) sets.

The annotation format is straightforward - each row contains a filename and three labels (race, gender, age). Race categories use standardized labels, gender is binary (Male/Female), and age is grouped into ranges like "0-2", "3-9", "10-19", etc.

For bias evaluation, establish baseline performance metrics on the full dataset, then segment results by demographic groups to identify disparities. Many researchers use metrics like demographic parity difference and equalized odds to quantify fairness gaps.

Watch out for

The binary gender classification reflects the dataset's 2021 creation date and may not align with current understanding of gender diversity. Consider this limitation when using FairFace for applications serving diverse gender identities.

Race categories, while more comprehensive than most datasets, still represent broad groupings that may not capture the full spectrum of human diversity. The "Latino_Hispanic" category, for example, spans many distinct ethnicities.

Image quality and pose vary somewhat across demographic groups due to the source dataset characteristics, which could introduce confounding factors in bias evaluation. Always examine whether performance differences stem from demographic bias versus image quality differences.

The age groupings are broad ranges rather than specific ages, limiting its usefulness for precise age estimation evaluation compared to continuous age labels.

Tags

fairnessface recognitionbiasdataset

At a glance

Published

2021

Jurisdiction

Global

Category

Datasets and benchmarks

Access

Public access

Build your AI governance program

VerifyWise helps you implement AI governance frameworks, track compliance, and manage risk across your AI systems.

FairFace: Face Attribute Dataset for Balanced Race, Gender, and Age | AI Governance Library | VerifyWise