Franziska Sofia Hafner
Research Assistant
Sofia completed her MSc in Social Data Science and now works as a Research Assistant at the OII, focusing on algorithmic fairness, machine learning, and interactive data visualization.
Oxford researchers reveal how AI language models encode a flawed and binary understanding of gender, posing significant risks for transgender, nonbinary, and even cisgender individuals.
AI language models are developing a flawed understanding of gender, leading to stereotypical associations that could result in harmful discrimination, finds research from the Oxford Internet Institute at the University of Oxford.
The researchers warn that in healthcare, where AI is increasingly integrated into health technologies, these flawed assumptions, which are often based on a model’s conflation of gender and biological sex characteristics, could lead to inaccurate advice and misdiagnoses.
For example, an AI model that learns a rigid association between ‘woman’ and biological markers like ‘uterus’ or ‘estrogen’ could provide irrelevant or even harmful advice to a transgender woman. This narrow view could also misinterpret the needs of cisgender women whose health profiles differ from typical reproductive assumptions, such as those who are postmenopausal or have undergone a hysterectomy, say the researchers.
The study is the first to develop a robust framework to examine how gender is constructed in 16 AI language models. It reveals their fundamental limitations in understanding gender, often defaulting to a restrictive, biologically tied, and binary view. These limitations have broad implications for both cisgender heterosexual people and the LGBTQIA+ community.
The research has been accepted for publication at the ACM Conference on Fairness, Accountability, and Transparency (FAccT).
Key findings:
Lead author, Franziska Sofia Hafner, Researcher at the Oxford Internet Institute, said: “If language models are going to be used in healthcare, either built into diagnostics to help doctors make decisions or as self-help tools for individuals, their limited and biased understanding of gender could introduce significant discriminatory harm.”
In their study, the researchers evaluated associations between gendered and sexed words, as well as associations between gendered words and physical or mental illnesses. They tested 16 language models based on GPT, RoBERTa, T5, Llama, and Mistral.
Language models are known to perpetuate stereotypes present in their training data, and developers typically respond by auditing for bias and applying filters. This study highlights deeper issues in how models internalise and reproduce social norms and stereotypes based on language.
Co-author, Dr Ana Valdivia, Lecturer in AI, Government and Policy at the Oxford Internet Institute, said: “Our academic community is aware of the social biases reproduced by algorithmic models. With the emergence of a new generation of AI systems, such as language models, these biases have not been mitigated; rather, they continue to amplify stereotypical representations. We advocate for stronger accountability mechanisms.”
Co-author, Dr Luc Rocher, Senior Research Fellow at the Oxford Internet Institute, said: “Our findings reveal a troubling trend where larger models, despite performing better on many benchmarks, actually encode a more rigid and essentialising view of gender. This challenges the notion that simply scaling up AI will lead to more nuanced or fair outcomes. Instead, these fundamental biases risk becoming more deeply ingrained.
“Current AI models are largely learning gender from the Internet, and the results are predictably problematic. Fixing AI’s gender problem is not just about tweaking algorithms. We need a concerted approach, from curating better training datasets to building standards and robust public oversight, to ensure these new tools stop amplifying old prejudices.”
The study, ‘Gender trouble in language models: an empirical audit guided by gender performativity theory’ by Franziska Sofia Hafner, Ana Valdivia, and Luc Rocher of the Oxford Internet Institute, will be available as a postprint on arXiv from 21 May. It will be formally published as part of the ACM Fairness, Accountability, and Transparency (FAccT) peer-reviewed conference proceedings. The conference will be held from 23-26 June in Athens, Greece.
NOTES FOR EDITORS
Funding information
The researchers were supported with funding from the University of Oxford, UK Research and Innovation and the Royal Society.
Image credit/ Copyright statement: Yutong Liu & Kingston School of Art / https://betterimagesofai.org / https://creativecommons.org/licenses/by/4.0/
Contact
For more information and briefings, please contact: Anthea Milnes, Head of Communications or Sara Spinks/Veena McCoole, Media and Communications Manager. T: +44 (0)1865 280527. M: +44 (0)7551 345493. E: press@oii.ox.ac.uk
About the Oxford Internet Institute (OII)
The Oxford Internet Institute (OII) is a multidisciplinary research and teaching department of the University of Oxford, dedicated to the social science of the Internet. Drawing from many different disciplines, the OII works to understand how individual and collective behaviour online shapes our social, economic and political world.
Since its founding in 2001, research from the OII has had a significant impact on policy debate, formulation and implementation around the globe, as well as a secondary impact on people’s wellbeing, safety and understanding.
Drawing on many different disciplines, the OII takes a combined approach to tackling society’s big questions, with the aim of positively shaping the development of the digital world for the public good.
About the University of Oxford
Oxford University has been placed number 1 in the Times Higher Education World University Rankings for the ninth year running, and number 3 in the QS World Rankings 2024. At the heart of this success are the twin-pillars of our ground-breaking research and innovation and our distinctive educational offer.
Oxford is world-famous for research and teaching excellence and home to some of the most talented people from across the globe. Our work helps the lives of millions, solving real-world problems through a huge network of partnerships and collaborations. The breadth and interdisciplinary nature of our research alongside our personalised approach to teaching sparks imaginative and inventive insights and solutions.
Through its research commercialisation arm, Oxford University Innovation, Oxford is the highest university patent filer in the UK and is ranked first in the UK for university spinouts, having created more than 300 new companies since 1988. Over a third of these companies have been created in the past five years. The university is a catalyst for prosperity in Oxfordshire and the United Kingdom, contributing £15.7 billion to the UK economy in 2018/19, and supports more than 28,000 full time jobs.