- Home
- Search Results
- Page 1 of 1
Search for: All records
-
Total Resources2
- Resource Type
-
20
- Availability
-
11
- Author / Contributor
- Filter by Author / Creator
-
-
Brun, Yuriy (2)
-
Giguere, Stephen (2)
-
Metevier, Blossom (2)
-
Brockman, Sarah (1)
-
Brunskill, Emma (1)
-
Kobren, Ari (1)
-
Niekum, Scott (1)
-
Thomas, Philip (1)
-
Thomas, Philip S. (1)
-
da Silva, Bruno Castro (1)
-
#Tyler Phillips, Kenneth E. (0)
-
& *Soto, E. (0)
-
& Ahmed, Khadija. (0)
-
& Akcil-Okan, O. (0)
-
& Akuom, D. (0)
-
& Andrews-Larson, C. (0)
-
& Archibald, J. (0)
-
& Attari, S. Z. (0)
-
& Ayala, O. (0)
-
& Babbitt, W. (0)
-
- Filter by Editor
-
-
& Ahn, J. (0)
-
& Bateiha, S. (0)
-
& Ruiz-Arias, P.M. (0)
-
& Spitzer, S. (0)
-
& Spitzer, S.M. (0)
-
A. Beygelzimer (0)
-
A. Ghate, K. Krishnaiyer (0)
-
A. I. Sacristán, J. C. (0)
-
A. Weinberg, D. Moore-Russo (0)
-
A.I. Sacristán, J.C. Cortés-Zavala (0)
-
ACS (0)
-
AIAA (0)
-
AIAA Propulsion and Energy 2021 (0)
-
AIAA SciTech (0)
-
ASEE Manufacturing Division (0)
-
ASME (0)
-
ASME ICEF (0)
-
ASSOCIATE EDITORS: Bahar, Ivet (Department (0)
-
Abdelaziz, Mohamed (0)
-
Abernethy, Jacob (0)
-
-
Have feedback or suggestions for a way to improve these results?
!
Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Recent studies found that using machine learning for social applications can lead to injustice in the form of racist, sexist, and otherwise unfair and discriminatory outcomes. To address this challenge, recent machine learning algorithms have been designed to limit the likelihood such unfair behavior occurs. However, these approaches typically assume the data used for training is representative of what will be encountered in deployment, which is often untrue. In particular, if certain subgroups of the population become more or less probable in deployment (a phenomenon we call demographic shift), prior work's fairness assurances are often invalid. In this paper, wemore »Free, publicly-accessible full text available April 25, 2023
-
Metevier, Blossom ; Giguere, Stephen ; Brockman, Sarah ; Kobren, Ari ; Brun, Yuriy ; Brunskill, Emma ; Thomas, Philip ( , Advances in neural information processing systems)We present RobinHood, an offline contextual bandit algorithm designed to satisfy a broad family of fairness constraints. Our algorithm accepts multiple fairness definitions and allows users to construct their own unique fairness definitions for the problem at hand. We provide a theoretical analysis of RobinHood, which includes a proof that it will not return an unfair solution with probability greater than a user-specified threshold. We validate our algorithm on three applications: a tutoring system in which we conduct a user study and consider multiple unique fairness definitions; a loan approval setting (using the Statlog German credit data set) in whichmore »well-known fairness definitions are applied; and criminal recidivism (using data released by ProPublica). In each setting, our algorithm is able to produce fair policies that achieve performance competitive with other offline and online contextual bandit algorithms.« less