UCLA
Voir la ressource originaleFairFace addresses one of the most persistent challenges in computer vision: the lack of demographic balance in training and evaluation datasets. Created by UCLA researchers, this dataset provides 108,501 face images with balanced representation across seven racial groups, two genders, and nine age ranges. Unlike traditional face datasets that skew heavily toward certain demographics, FairFace was specifically engineered to enable rigorous bias testing in facial analysis systems. It's become a go-to resource for researchers and practitioners who need to validate that their face recognition, age estimation, or demographic classification models perform equitably across different population groups.
Most face datasets suffer from severe demographic imbalances - often 70-80% white faces, heavy male skew, and limited age diversity. FairFace flips this script by maintaining roughly equal representation across racial categories (White, Black, East Asian, Southeast Asian, Indian, Middle Eastern, Latino_Hispanic) and gender, with systematic coverage of age ranges from 0-2 years through 70+ years.
The images themselves are high-quality, front-facing photos sourced from the YFCC-100M Flickr dataset, then carefully filtered and annotated. Each image includes ground truth labels for race, gender, and age group, verified through multiple annotation rounds. The dataset also provides both training and validation splits, making it immediately usable for model development and evaluation.
Download the dataset from the GitHub repository, which includes both the image files and CSV annotation files. The total download is approximately 11GB. You'll find separate folders for training (86,744 images) and validation (21,757 images) sets.
The annotation format is straightforward - each row contains a filename and three labels (race, gender, age). Race categories use standardized labels, gender is binary (Male/Female), and age is grouped into ranges like "0-2", "3-9", "10-19", etc.
For bias evaluation, establish baseline performance metrics on the full dataset, then segment results by demographic groups to identify disparities. Many researchers use metrics like demographic parity difference and equalized odds to quantify fairness gaps.
The binary gender classification reflects the dataset's 2021 creation date and may not align with current understanding of gender diversity. Consider this limitation when using FairFace for applications serving diverse gender identities.
Race categories, while more comprehensive than most datasets, still represent broad groupings that may not capture the full spectrum of human diversity. The "Latino_Hispanic" category, for example, spans many distinct ethnicities.
Image quality and pose vary somewhat across demographic groups due to the source dataset characteristics, which could introduce confounding factors in bias evaluation. Always examine whether performance differences stem from demographic bias versus image quality differences.
The age groupings are broad ranges rather than specific ages, limiting its usefulness for precise age estimation evaluation compared to continuous age labels.
Publié
2021
Juridiction
Mondial
Catégorie
Datasets and benchmarks
Accès
Accès public
VerifyWise vous aide à implémenter des cadres de gouvernance de l'IA, à suivre la conformité et à gérer les risques dans vos systèmes d'IA.