Programme Launch Offer: Save 20% - Book Now

Track Talk, W8

Can We Measure Local Bias in AI Systems?

Wessel Braakman

Teresa Herland

12:00 - 12:45 CEST, Wednesday 17th June

AI bias is a well-known problem, but most studies have been conducted by institutions in the United States and focus on the English language. But what if we use such systems to make (partial) decisions that affect people in Norway, where we live? Can we investigate the same issues in Norwegian and in the context of Norwegian society?

We are three colleagues who worked on this research besides our day-job, because we feel like this is an important part of AI development and quality assurance. We have, since we started, set up our very own non-profit association in which we can do this research company-independent.

In this deep dive, we would like to share our findings from testing AI models for bias in the Norwegian language and society. We will share what worked, what didn’t, and what we learned along the way. The research we do is aimed at the Norwegian language and society, but can of course be mapped to other countries and cultures. What we provide is a clear guide on how to achieve this, including all we learned.

As a foundation, we use research from BBQ: A Hand-Built Bias Benchmark for Question Answering, which we adapted for the Norwegian context. We will explain about how we prepared our data, how we got responses from AI systems, and how we then scored them to find bias. We will also talk about automation, what worked and what didn’t.

We will walk through our method, the challenges we faced, and why this is important for AI development in Norway (and in other countries). You as a listener will gain concrete insights into how to evaluate and reduce bias in AI models so that they work fairly across languages and cultures.

This session will give you the tools to think critically about AI fairness, beyond the usual English-centered approach.