BEGIN:VCALENDAR METHOD:PUBLISH PRODID:-//Apple Computer\, Inc//iCal 1.0//EN X-WR-CALNAME;VALUE=TEXT:USC VERSION:2.0 BEGIN:VEVENT DESCRIPTION:Speaker: Hoda Heidari, Cornell Universtiy Talk Title: Distributive Justice for Machine Learning: An Interdisciplinary Perspective on Defining, Measuring, and Mitigating Algorithmic Unfairness Series: CS Colloquium Abstract: Automated decision-making tools are increasingly in charge of making high-stakes decisions for people-”in areas such as education, credit lending, criminal justice, and beyond. These tools can exhibit and exacerbate certain undesirable biases and disparately harm already disadvantaged and marginalized groups and individuals. In this talk, I will illustrate how we can bring together tools and methods from computer science, economics, and political philosophy to define, measure, and mitigate algorithmic unfairness in a principled manner. In particular, I will address two key questions:\n \n - Given the appropriate notion of harm/benefit, how should we measure and bound unfairness? Existing notions of fairness focus on defining conditions of fairness, but they do not offer a proper measure of unfairness. In practice, however, designers often need to select the least unfair model among a feasible set of unfair alternatives. I present (income) inequality indices from economics as a unifying framework for measuring unfairness--both at the individual- and group-level. I propose the use of cardinal social welfare functions as an alternative measure of fairness behind a veil of ignorance and a computationally tractable method for bounding inequality.\n \n - Given a specific decision-making context, how should we define fairness as the equality of some notion of harm/benefit across socially salient groups? First, I will offer a framework to think about this question normatively. I map the recently proposed notions of group-fairness to models of equality of opportunity. This mapping provides a unifying framework for understanding these notions, and importantly, allows us to spell out the moral assumptions underlying each one of them. Second, I give a descriptive answer to the question of "fairness as equality of what?". I mention a series of adaptive human-subject experiments we recently conducted to understand which existing notion best captures laypeople's perception of fairness.\n \n This lecture satisfies requirements for CSCI 591: Research Colloquium Biography: Hoda Heidari is currently a Postdoctoral Associate at the Department of Computer Science at Cornell University, where she collaborates with Professors Jon Kleinberg, Karen Levy, and Solon Barocas through the AIPP (Artificial Intelligence, Policy, and Practice) initiative. Hoda's research is broadly concerned with the societal aspects of Artificial Intelligence, and in particular, the issues of unfairness and discrimination for Machine Learning. She utilizes tools and methods from Computer Science (Algorithms, AI, and ML) and Social Sciences (Economics and Political Philosophy) to quantify and mitigate the inequalities that arise when socially consequential decisions are automated. Host: Aleksandra Korolova and Bistra Dilkina SEQUENCE:5 DTSTART:20200416T110000 LOCATION: Seminar will be exclusively online (no in-room presentation) DTSTAMP:20200416T110000 SUMMARY:Seminar will be exclusively online (no in-room presentation) - CS Colloquium: Hoda Heidari (Cornell University) - Distributive Justice for Machine Learning: An Interdisciplinary Perspective on Defining, Measuring, and Mitigating Algorithmic Unfairness UID:EC9439B1-FF65-11D6-9973-003065F99D04 DTEND:20200416T120000 END:VEVENT END:VCALENDAR