Behavior Policy Search for Risk Estimators in RL
Elita Lobo, Yash Chandak, et al.
NeurIPS 2021
Post-processing in algorithmic fairness is a versatile approach for correcting bias in ML systems that are already used in production. The main appeal of post-processing is that it avoids expensive retraining. In this work, we propose a suite of general post-processing algorithms for individual fairness (IF). We consider a setting where the learner only has access to the predictions of the original model and a similarity graph between individuals guiding the desired fairness constraints. We cast the IF post-processing problem as a graph smoothing problem corresponding to graph Laplacian regularization that preserves the desired treat similar individuals similarly interpretation. Our theoretical results demonstrate the connection of the new objective function to a local relaxation of the original individual fairness. Empirically, our post-processing algorithms correct individual biases in large scale NLP models, e.g., BERT, while preserving accuracy.
Elita Lobo, Yash Chandak, et al.
NeurIPS 2021
Giovanni De Felice, Arianna Casanova Flores, et al.
NeurIPS 2025
Brianna Richardson, Kush R. Varshney
INFORMS 2021
Erik Miehling, Rahul Nair, et al.
NeurIPS 2023