Algorithmic Injustice

Algorithms play an increasingly important role in our daily life but come with serious societal risks. In recent years we have seen many cases of algorithms that show unfair biased behavior towards particular groups or individuals, for instance the Dutch Toeslagenaffaire. This leads to growing concerns about harmful discrimination and reproduction of structural inequalities once these technologies become institutionalized in society. During this evening on algorithmic injustice we explore and discuss both the philosophical and technical aspects as well as the lived experiences of people who suffered from unfair algorithms.

Despite growing concerns about algorithmic injustice, in AI research and policy, the remedies against algorithmic discrimination are often narrowly framed as design challenges, rather than complex, structural, social-political problems. But is the solution always technological? Do we address harmful consequences of algorithms by fixing the data? And should engineers determine what is fair?

At this event, we bring together researchers from various disciplines. Su Lin Blodgett has been working on AI and fairness, Erin Beeghly on the wrong of stereotypes, Naomi Appelman on the unfairness of online proctoring, and documentary maker Nirit Peled documented firsthand stories of people who suffered the consequences of unfair police algorithms. Together, they will explore pressing matters around algorithmic injustice.

The event is organised by Dr. Marjolein Lanzing and Dr. Katrin Schulz as part of their project The politics of bias in AI: challenging the technocentric approach. The project is funded by the RPA Human(e) AI of the University of Amsterdam.

About the speakers

Su Lin Blodgett is senior researcher in the Fairness, Accountability, Transparency, and Ethics in AI (FATE) group at Microsoft Research Montréal. Blodgett is interested in examining the social and ethical implications of natural language processing technologies; she develops approaches for anticipating, measuring, and mitigating harms arising from language technologies, focusing on the complexities of language and language technologies in their social contexts, and on supporting NLP practitioners in their ethical work. She has also worked on using NLP approaches to examine language variation and change.

Erin Beeghly is Associate Professor of Philosophy at the University of Utah. Her research interests lie at the intersection of ethics, social epistemology, feminist philosophy, and moral psychology. Her current book project, What’s Wrong With Stereotyping? (under contract with OUP), examines the conditions under which judging people by group membership is wrong. She and Alex Madva are co-editors of the first philosophical introduction to implicit bias: An Introduction to Implicit Bias: Knowledge, Justice, and the Social Mind (Routledge 2020). Beeghly also writes and teaches about topics within legal theory, including discrimination law.

Nirit Peled is an independent filmmaker and writer based in the Netherlands. Drawing on techniques from journalism and documentary, she investigates the social impact of new technologies, structures of legality, systemic abuses of power and the nature of violence. Her latest documentary, MOTHERS tells the story of four women whose lives were forever changed when their adolescent sons entered a youth crime prevention program. TV archive and government documents reveal how their lives were impacted by an algorithmic reality that aims to assess the risks of their sons turning to crime. But can anyone’s life really be captured by data? Can they challenge the statistics that mark them as dangerous?

Naomi Appelman is a PhD-candidate in law and philosophy at the Institute for Information Law (IViR) interested in the role of law in online exclusion, speech governance, and platform power. Her research asks how European law should facilitate contestation of the content moderation systems governing online speech. The aim of facilitating this contestation is to minimise undue exclusion, often of already marginalised groups, from online spaces and democratise the power over how online speech is governed. Appelman is one of the founders of the Racism and Technology Center and together with bioinformatic Robin Pocornie filed a complaint at the Dutch Human Rights Institute for using online proctoring that discriminates against people of color.

Gerelateerde programma’s
23 05 24
Resisting Data Colonialism: A Practical Intervention

Some people think that colonialism is long over, while others are sure it has never stopped. This event marks the launch of a book exploring a third possibility: not only is colonialism still continuing, but right now it is morphing into possibly its most powerful version yet. This is data colonialism.

Datum
Donderdag 23 mei 2024 17:00 uur
Locatie
SPUI25
29 05 24
Challenging discriminatory algorithms through legal means
The Meta Case

Is Facebook potentially violating Dutch equality laws? In this event, leading up to a hearing of the case before the Netherlands Institute for Human Rights against digital giant Meta, we delve into the origins of this Meta Case and engage in a broader conversation about how technological developments can reflect, identify, and address societal issues such as sexism, racism, and other types of discrimination.

Datum
Woensdag 29 mei 2024 17:00 uur
Locatie
SPUI25
04 04 24
Blinde vlekken: hoe impliciete vooroordelen je (mis)leiden

We kunnen mensen onrechtvaardig behandelen zonder het te willen, en zelfs zonder het op te merken. Onbewust en ongewild maken we onderscheid tussen personen op basis van hun gender, etniciteit, leeftijd, lichaamsgewicht of seksuele oriëntatie. In haar nieuwe boek onderzoekt Lieke Asma die blinde vlekken. Om deze werkelijk te begrijpen, betoogt zij, is het zaak de blik niet naar binnen te keren, maar onze aandacht te richten naar de wereld buiten ons.

Datum
Donderdag 4 apr 2024 20:00 uur
Locatie
SPUI25