AIResearch AIResearch
Back to articles
Science

AI Bias Tests Found Unreliable and Flawed

New research reveals current methods for detecting gender bias in AI systems produce inconsistent results and misinterpret common examples like 'man is to computer programmer as woman is to homemaker'

AI Research
November 14, 2025
3 min read
AI Bias Tests Found Unreliable and Flawed

The tools used to detect gender bias in artificial intelligence systems may be fundamentally unreliable, according to new research that challenges how we measure fairness in AI. As AI becomes increasingly integrated into hiring, lending, and other critical decisions, understanding whether these systems perpetuate harmful stereotypes has never been more important. The study reveals that popular bias detection methods produce inconsistent results and often misinterpret what they're measuring.

Researchers found that current bias assessment techniques fail three critical reliability tests. First, changing the gender word pairs used for measurement—such as switching from 'she/he' to 'woman/man'—causes significant variations in bias scores. For example, a profession might be classified as male-biased using one pair but female-biased using another. Second, the methods struggle with basic word variations—the same profession in singular and plural forms (like 'professor' versus 'professors') often receive opposite gender assignments. Third, these techniques cannot reliably identify words with explicit gender information, such as 'lioness' (female) or socially stereotyped traits like 'compassionate' (traditionally female).

The University of Sheffield team systematically evaluated four popular bias measures—Direct Bias, Word Association Test, Relational Inner Product Association, and Neighbourhood Metric—using standard datasets including 320 professions and the Bem Sex Role Inventory of gender-stereotyped traits. They tested each measure's sensitivity to different gender word pairs and word forms, then examined whether they could correctly identify words with known gender associations.

Analysis showed that bias measures lack stability across different testing conditions. When researchers changed the base gender pair, only about 25% of professions maintained consistent bias directions across all measures. The magnitude of bias scores changed significantly in 66-71% of cases when different pairs were used. Even more concerning, word form variations caused moderate to substantial disagreement in bias assignments—'surgeon' might be classified as male-biased while 'surgeons' as female-biased using the same measurement approach.

The research also debunks a widely cited example of gender bias: the analogy 'man is to computer programmer as woman is to homemaker.' The study demonstrates this analogy primarily reflects word similarity rather than gender bias. The cosine similarity between 'computer programmer' and 'homemaker' is 0.50, while 'man' and 'woman' have a similarity of 0.77. In fact, each word in this analogy is the most similar to the other in the embedding space, suggesting the relationship stems from mathematical properties of word vectors rather than systematic gender bias.

These findings have significant implications for AI development and regulation. As companies and governments increasingly rely on bias detection tools to ensure fair AI systems, the unreliability of current methods means we may be making incorrect assumptions about whether algorithms are biased. This could lead to either false confidence in biased systems or unnecessary modifications to fair ones. The research suggests we need more robust measurement approaches before we can accurately assess and address bias in AI.

The study acknowledges several limitations. There's no inherent ground truth for what constitutes problematic bias levels, making evaluation challenging. The research focused specifically on gender bias in English word embeddings, and similar investigations are needed for other languages and bias types like racial stereotyping. Additionally, while the methods showed poor performance, the paper doesn't claim word embeddings are free of bias—rather that current measurement approaches don't reliably detect it.

This work highlights the complexity of quantifying bias in AI systems and calls for more careful interpretation of bias measurement results. As AI continues to influence everything from job applications to medical diagnoses, developing reliable methods to ensure these systems treat everyone fairly remains an urgent priority.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn