Overview and details of the sessions of this conference. Please select a date or location to show only sessions at that day or location. Please select a single session for detailed view (with abstracts and downloads if available).
Typically, fairness properties of algorithmic decisions are analyzed as if the machine predictions were implemented directly. However, many machine predictions are instead deployed to assist a human decision-maker who retains the ultimate decision authority. In this article, we therefore consider how properties of machine predictions affect the resulting human decisions through both a formal model and lab experiment.
Towards understanding the causes of developing biased algorithms by programmers
Mohammadreza Shahsahebi, Osman Alp, Justin Weinhardt, Alireza Sabouri
University of Calgary, Haskayne School of Business, Canada
The use of algorithms may be preferred over human judgment; however, algorithms are prone to bias. In this study, we designed a behavioral lab experiment environment to imitate the processes instated by the companies when asking their programmers to develop an ML algorithm for any given task. We investigate the causes of programmers developing biased ML algorithms and study how these causes can be mitigated through policies to encourage the development of fair and less-biased ML algorithms.
How good are privacy guarantees? Data sharing, privacy preservation, and platform behavior
Alireza Fallah1, Daron Acemoglu1, Ali Makhdoumi2, Azarakhsh Malekian3, Dmitry Mitrofanov4
1MIT; 2Duke University,; 3University of Toronto; 4Boston College