Hannah is a 3rd year DPhil student in Social Data Science at the OII. Hannah’s research centres on the role of granular and diverse human feedback for aligning large language models. At the core of all her research into AI, she is motivated by the prompt “who decides how AI systems behave”, and appreciates both technical and AI governance answers to this thorny question. Her body of published work spans computational linguistics, computer vision, ethics and sociology, addressing a broad range of issues such as AI alignment, bias, fairness, and hate speech from a multidisciplinary perspective.
Hannah holds degrees from the University of Oxford (MSc), the University of Cambridge (BA) and Peking University (MA). She is currently a visiting academic at New York University.
Alongside academia, she collaborates often with industry projects at Google, OpenAI and MetaAI, and previously worked as a Data Scientist in the online safety team at The Alan Turing Institute.
Artificial Intelligence; Machine Learning; NLP; Active Learning; Adversarial Learning; Online Harms; Hate Speech
6 December 2024
Several researchers and DPhil students from the Oxford Internet Institute, University of Oxford, will head to Vancouver for the Thirty-Eighth annual Conference on Neural Information Processing Systems (NeurIPS) from 10-15 December 2024.
23 April 2024
Personalisation has the potential to democratise who decides how LLMs behave, but comes with risks for individuals and society, say Oxford researchers.
12 March 2024
Oxford Internet Institute researchers are studying how to make Large Language Models (LLMs) more diverse by broadening the human feedback that informs them.
27 October 2023
The OII is leading the debate on Artificial Intelligence. Generative AI has been a key area of interest for faculty, researchers and students for many years. This article brings together some of this work to date and highlights forthcoming work.
Sky News, 03 December 2021
Spotify removes nearly 150 hours of content that it said violated its hate policy after Sky News reported it.
Sky News, 13 August 2021
Harmful posts can end up being missed altogether while acceptable posts are mislabelled as offensive, according to the Oxford Internet Institute. frontpage-pressSky NewsHannah Rose Kirk|Dr Scott A. Hale|Bertram Vidgen|Paul Röttger