Quarterly Theory Workshop: Algorithms and their Social Impact
About this series
The Quarterly Theory Workshop brings in theoretical computer science experts to present their perspective and research on a common theme. The technical program starts in the morning and includes time for coffee, lunch and discussions between the speakers and participants.
Synopsis
The focus of this workshop will be on the societal impacts of algorithms. From designing self-driving cars to selecting the order of news posts on Facebook to automating credit checks, the use of algorithms for decision making is now commonplace. Hence it is more important than ever to consider fairness as a key aspect while designing these algorithms to prevent unwanted bias and prejudice. The speakers for this workshop are Rakesh Vohra, Michael Kearns, Samira Samadi, Steven Wu, and Suresh Venkatasubramanian . This is co-organized by IDEAL as part of the Special Quarter on Data Science and Law.
Logistics
- Location: Virtual (on Gather.Town and Zoom). watch the full event here
Schedule
- 10:30 – 11:00: Michael Kearns. watch the talk here
- 11:00 – 11:30: Samira Samadi. watch the talk here
- 11:30 – 11:40: break
- 11:40 – 12:10: Steven Wu. watch the talk here
- 12:10 – 1:00: lunch break
- 1:00 – 1:30: Suresh Venkatasubramanian. watch the talk here
- 1:30 – 2:00: Rakesh Vohra. watch the talk here
- 2:00 – 2:30: Discussion time
Abstracts
Speaker: Michael Kearns (University of Pennsylvania)
Title: Socially Fair k-Means Clustering
Speaker: Samira Samadi (MPI)
We show that the popular k-means clustering algorithm (Lloyd’s heuristic), can result in outcomes that are unfavorable to subgroups of data (e.g., demographic groups). Such biased clusterings can have deleterious implications for human-centric applications such as resource allocation. We present a fair 𝑘-means objective and algorithm to choose cluster centers that provide equitable costs for different groups. The algorithm, Fair-Lloyd, is a modification of Lloyd’s heuristic for 𝑘-means, inheriting its simplicity, efficiency, and stability. In comparison with standard Lloyd’s, we find that on benchmark datasets, Fair-Lloyd exhibits unbiased performance by ensuring that all groups have equal costs in the output 𝑘-clustering, while incurring a negligible increase in running time, thus making it a viable fair option wherever 𝑘-means is currently used.
Speaker: Steven Wu (CMU)
Recent work in fair machine learning has proposed dozens of technical definitions of algorithmic fairness and methods for enforcing these definitions. However, we still lack a comprehensive understanding of how to develop machine learning systems with fairness criteria that reflect relevant stakeholders’ nuanced viewpoints in real-world contexts. This talk will cover our recent work that aims to address this gap. We will first discuss an algorithmic framework that enforces the individual fairness criterion through interactions with a human auditor, who can identify fairness violations without enunciating a fairness (similarity) measure. We then discuss an empirical study on how to elicit stakeholders’ fairness notions in the context of a child maltreatment predictive system.
One of the more pressing concerns around the deployment of ML systems is explainability: can we understand why an ML system made the decision that it did. This question can be unpacked in a variety of ways, and one approach that has become popular is the idea of feature influence: that we can assign a score to features that represents their (relative) influence in an outcome (either locally for particular input, or globally).
One of the most influential of such approaches has been one based on cooperative game theory, where features are modeled as “players” and feature influence is captured as “player contribution” via the Shapley value of a game. The argument is that the axiomatic framework provided by Shapley values is well-aligned with the needs of an explanation system.
But is it? I’ll talk about two pieces of work that nail down mathematical deficiencies of Shapley values as a way of estimating feature influence and quantify the limits of Shapley values via a fascinating geometric interpretation that comes with interesting algorithmic challenges.
Speaker: Rakesh Vohra (University of Pennsylvania)
There is increasing regulatory interest in whether machine learning algorithms deployed in consequential domains (e.g. in criminal justice) treat different demographic groups “fairly.” Several proposed notions of fairness, typically mutually incompatible, have been examined in settings where the behavior being predicted is treated as exogenous.
Recent Comments