Skip to main content

Algorithmic Impact Assessment: Gianclaudio Malgieri receives 2019 Future of Privacy Forum Award

  • December 6, 2019

[Award] LSTS Researcher Gianclaudio Malgieri, along with his co-author Margot E. Kaminski, were among the five winners of the 10th Annual Privacy Papers for Policymakers (PPPM) Award held by the Future of Privacy Forum for their paper on Algorithmic Impact Assessment. Malgieri is the only winning author, representing a European university.

The Award recognises leading privacy scholarship that is a "must-read" for policymakers worldwide. Malgieri and Kaminski have been invited to present their research at the U.S. Senate on 6 February 2020, in the presence of policymakers, academics, and industry privacy professionals.

You can read the abstract of the winning paper below:

Policy-makers, scholars, and commentators are increasingly concerned with the risks of using profiling algorithms and automated decision-making. The EU’s General Data Protection Regulation (GDPR) has tried to address these concerns through an array of regulatory tools. As one of us has argued, the GDPR combines individual rights with systemic governance, towards algorithmic accountability. The individual tools are largely geared towards individual “legibility”: making the decision-making system understandable to an individual invoking her rights. The systemic governance tools, instead, focus on bringing expertise and oversight into the system as a whole, and rely on the tactics of “collaborative governance,” that is, use public-private partnerships towards these goals. How these two approaches to transparency and accountability interact remains a largely unexplored question, with much of the legal literature focusing instead on whether there is an individual right to explanation.
The GDPR contains an array of systemic accountability tools. Of these tools, impact assessments (Art. 35) have recently received particular attention on both sides of the Atlantic, as a means of implementing algorithmic accountability at early stages of design, development, and training. The aim of this paper is to address how a Data Protection Impact Assessment (DPIA) links the two faces of the GDPR’s approach to algorithmic accountability: individual rights and systemic collaborative governance. We address the relationship between DPIAs and individual transparency rights. We propose, too, that impact assessments link the GDPR’s two methods of governing algorithmic decision-making by both providing systemic governance and serving as an important “suitable safeguard” (Art. 22) of individual rights.
After noting the potential shortcomings of DPIAs, this paper closes with a call — and some suggestions — for a Model Algorithmic Impact Assessment in the context of the GDPR. Our examination of DPIAs suggests that the current focus on the right to explanation is too narrow. We call, instead, for data controllers to consciously use the required DPIA process to produce what we call “multi-layered explanations” of algorithmic systems. This concept of multi-layered explanations not only more accurately describes what the GDPR is attempting to do, but also normatively better fills potential gaps between the GDPR’s two approaches to algorithmic accountability.

 

News contact

For suggestions/feedback concerning the LSTS website news section, please contact us.