Workshop
Algorithmic Fairness through the lens of Metrics and Evaluation
Awa Dieng · Miriam Rateike · Jamelle Watson-Daniels · Golnoosh Farnadi · Nando Fioretto
West Meeting Room 111, 112
Sat 14 Dec, 9 a.m. PST
We are proposing the Algorithmic Fairness through the lens of Metrics and Evaluation (AFME)workshop, which is the fifth edition of this workshop series on algorithmic fairness. While previouseditions have explored foundational work on causal approaches to fairness and the intersection offairness with other fields of trustworthy machine learning namely interpretability, robustness, privacyand temporal aspects, this year’s workshop aims to timely reflect on fairness metrics definitions andevaluation methods.Indeed, with rapid advances in large generative models and international regulatory efforts as well aspertinent calls to understand fairness in context, it is crucial to revisit the suitability of existing fairnessmetrics and explore new bias evaluation frameworks. Our workshop aims to provide a venue to haverigorous interdisciplinary discussions around these critical topics and foster reflections on the necessityand challenges in defining adaptable fairness metrics and designing reliable evaluation techniques.## TopicThe discussion on defining and measuring algorithmic (un)fairness has predominantly been afocus in the early stages of algorithmic fairness research [Dwork et al., 2012, Zemel et al., 2013, Hardtet al., 2016, Zafar et al., 2017, Agarwal et al., 2018] resulting in four main fairness denominations:individual or group [Binns, 2020], statistical or causal [Makhlouf et al., 2023], equalizing or non-equalizing [Diana et al., 2021], and temporal or non-temporal fairness [Rateike, 2024]. Since, muchwork in the field had been dedicated to providing methodological advances within each denominationand understanding various trade-offs between fairness metrics [Binns, 2020, Heidari et al., 2019,Kleinberg et al., 2017]. However, given the changing machine learning landscape, with both increasingglobal applications and the emergence of large generative models, the question of understanding anddefining what constitutes “fairness” in these systems has become paramount again.On one hand, definitions of algorithmic fairness are being critically examined regarding the historicaland cultural values they encode [Asiedu et al., 2024, Arora et al., 2023, Bhatt et al., 2022]. Themathematical conceptualization of these definitions and their operationalization through satisfyingstatistical parities has also raised criticism of not taking into account the context within which thesesystems are deployed [Weinberg, 2022, Green and Hu, 2018].On another hand, it is still unclear how to reconcile standard fairness metrics and evaluationsdeveloped mainly for prediction and classification tasks with large generative models. While someworks proposed adapting existing fairness metrics, e.g., to large language models [Li et al., 2023,Zhang et al., 2023, Gallegos et al., 2023], questions remain on how to systematically measure fairnessfor textual outputs, or even multi-modal generative models [Schmitz et al., 2022, Chen et al., 2023,Lum et al., 2024]. Large generative models also pose new challenges to fairness evaluation withrecent work showcasing how biases towards specific tokens in large language models can influencefairness assessments during evaluation [Ding et al., 2024]. Finally, regulatory requirements introducenew challenges in defining, selecting, and assessing algorithmic fairness [Deck et al., 2024, Lauxet al., 2024, Hellman, 2020].Given these critical and timely considerations, this workshop aims to investigate how to defineand evaluate (un)fairness in today’s machine learning landscape. We are particularly interested inaddressing open questions in the field, such as:• Through a retrospective lens, what are the strengths and limitations of existing fairnessmetrics?• How to operationalize contextual definitions of fairness in diverse deployment domains?• Given the plethora of use-cases, how to systematically evaluate fairness and bias in largegenerative models?• How do recent regulatory efforts demand the utilization of fairness metrics and evaluationtechniques, and do existing ones comply with regulations?
Schedule
Sat 9:00 a.m. - 9:05 a.m.
|
Opening Remarks
(
Opening Remarks
)
>
SlidesLive Video |
🔗 |
Sat 9:05 a.m. - 9:40 a.m.
|
Invited Talk: Reflections on Fairness Measurement: From Predictive to Generative AI
(
Invited Talk, Q&A
)
>
SlidesLive Video |
Hoda Heidari 🔗 |
Sat 9:40 a.m. - 9:45 a.m.
|
Short Break
|
🔗 |
Sat 9:45 a.m. - 10:20 a.m.
|
Invited Talk: Harm Detectors and Guardian Models for LLMs: Implementations, Uses, and Limitations
(
Invited Talk, Q&A
)
>
SlidesLive Video |
Kush Varshney 🔗 |
Sat 10:20 a.m. - 10:25 a.m.
|
Short Break
|
🔗 |
Sat 10:25 a.m. - 10:30 a.m.
|
Contributed talk: Evaluating and Mitigating Discrimination in Language Model Decisions
(
Oral
)
>
SlidesLive Video |
Alex Tamkin 🔗 |
Sat 10:30 a.m. - 10:35 a.m.
|
Contributed talk: Towards Fair RAG: On the Impact of Fair Ranking in Retrieval-Augmented Generation
(
Oral
)
>
SlidesLive Video |
To Eun Kim 🔗 |
Sat 10:35 a.m. - 11:20 a.m.
|
Poster Session 1
(
Poster Session
)
>
|
🔗 |
Sat 11:25 a.m. - 12:00 p.m.
|
Invited Talk: Evaluating the Ethical Competence of LLMs
(
Invited Talk, Q&A
)
>
SlidesLive Video |
Seth Lazar 🔗 |
Sat 12:00 p.m. - 12:10 p.m.
|
Short Break
|
🔗 |
Sat 12:10 p.m. - 1:00 p.m.
|
Roundtable Discussions
(
Discussions
)
>
|
🔗 |
Sat 1:00 p.m. - 2:00 p.m.
|
Lunch
|
🔗 |
Sat 2:00 p.m. - 2:35 p.m.
|
Invited Talk: Fairness through Difference Awareness: Measuring Desired Group Discrimination in LLMs
(
Invited Talk, Q&A
)
>
SlidesLive Video |
Sanmi Koyejo · Angelina Wang 🔗 |
Sat 2:35 p.m. - 2:40 p.m.
|
Short Break
|
🔗 |
Sat 2:40 p.m. - 2:45 p.m.
|
Contributed talk: Different Horses for Different Courses: Comparing Bias Mitigation Algorithms in ML
(
Oral
)
>
SlidesLive Video |
Prakhar Ganesh 🔗 |
Sat 2:45 p.m. - 2:50 p.m.
|
Contributed talk: Evaluating Gender Bias Transfer between Pre-trained and Prompt Adapted Language Models
(
Oral
)
>
SlidesLive Video |
Natalie Mackraz 🔗 |
Sat 2:50 p.m. - 2:55 p.m.
|
Contributed talk: The Search for Less Discriminatory Algorithms: Limits and Opportunities
(
Oral
)
>
SlidesLive Video |
Benjamin Laufer · Manish Raghavan · Solon Barocas 🔗 |
Sat 2:55 p.m. - 3:55 p.m.
|
Poster Session 2
(
Poster Session
)
>
|
🔗 |
Sat 3:55 p.m. - 4:15 p.m.
|
Short Break
|
🔗 |
Sat 4:15 p.m. - 4:55 p.m.
|
Panel Discussion: Rethinking fairness in the era of large language models
(
Panel
)
>
SlidesLive Video |
Hoda Heidari · Sanmi Koyejo · Jessica Schrouff · Alexander D'Amour · Seth Lazar 🔗 |
Sat 4:55 p.m. - 5:00 p.m.
|
Short Break
|
🔗 |
Sat 5:00 p.m. - 5:10 p.m.
|
Closing Remarks
(
Closing Remarks
)
>
SlidesLive Video |
🔗 |
Sat 5:10 p.m. - 5:30 p.m.
|
Poster Session 3
(
Poster Session
)
>
|
🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
M²FGB: A Min-Max Gradient Boosting Framework for Subgroup Fairness
(
Spotlight
)
>
SlidesLive Video |
Jansen Pereira · Giovani Valdrighi · Marcos M. Raimundo 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Toward Large Language Models that Benefit for All: Benchmarking Group Fairness in Reward Models
(
Spotlight
)
>
SlidesLive Video |
Kefan Song · Jin Yao · Shangtong Zhang 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Efficient Fairness-Performance Pareto Front Computation
(
Spotlight
)
>
|
Mark Kozdoba · Benny Perets · Shie Mannor 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Measuring Representational Harms in Image Generation with a Multi-Group Proportional Metric
(
Spotlight
)
>
|
Sangwon Jung · Claudio Mayrink Verdun · Alex Oesterling · Sajani Vithana · Taesup Moon · Flavio du Pin Calmon 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Mitigating Bias in Facial Recognition Systems: Centroid Fairness Loss Optimization
(
Spotlight
)
>
SlidesLive Video |
Jean-Rémy Conti · Stephan Clémençon 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Benchmark to Audit LLM Generated Clinical Notes for Disparities Arising from Biases and Stereotypes
(
Spotlight
)
>
|
Hongyu Cai · Swetasudha Panda · Naveen Jafer Nizar · Qinlan Shen · Daeja Oxendine · Sumana Srivatsa · Krishnaram Kenthapadi 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Beyond Internal Data: Constructing Complete Datasets for Fairness Testing
(
Spotlight
)
>
SlidesLive Video |
Varsha Ramineni · Hossein A. Rahmani · Emine Yilmaz · David Barber 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Beyond the Safety Bundle: Auditing the Helpful and Harmless Dataset
(
Spotlight
)
>
SlidesLive Video |
Khaoula Chehbouni · Jonathan Colaço Carr · Yash More · Jackie CK Cheung · Golnoosh Farnadi 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Optimal Selection Using Algorithmic Rankings with Side Information
(
Spotlight
)
>
|
Kate Donahue · Nicole Immorlica · Brendan Lucier 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Verifiable evaluations of machine learning models using zkSNARKs
(
Spotlight
)
>
|
Tobin South · Alexander Camuto · Shrey Jain · Robert Mahari · Christian Paquin · Jason Morton · Alex `Sandy' Pentland 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Measuring the Impact of Equal Treatment as Blindness via Explanations Disparity
(
Spotlight
)
>
|
Carlos Mougan · Salvatore Ruggieri · Laura State · Antonio Ferrara · Steffen Staab 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
From Models to Systems: A Comprehensive Fairness Framework for Compositional Recommender Systems
(
Spotlight
)
>
SlidesLive Video |
Brian Hsu · Cyrus DiCiccio · Natesh Pillai · Hongseok Namkoong 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Understanding The Effect Of Temperature On Alignment With Human Opinions
(
Spotlight
)
>
SlidesLive Video |
Maja Pavlovic · Massimo Poesio 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Towards Reliable Fairness Assessments of Multi-Label Image Classifiers
(
Spotlight
)
>
SlidesLive Video |
Melissa Hall · Bobbie Chern · Laura Gustafson · Denisse Ventura · Harshad Kulkarni · Candace Ross · Nicolas Usunier 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Better Bias Benchmarking of Language Models via Multi-factor Analysis
(
Spotlight
)
>
SlidesLive Video |
Hannah Powers · Ioana Baldini · Dennis Wei · Kristin P Bennett 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Adaptive Group Robust Ensemble Knowledge Distillation
(
Spotlight
)
>
SlidesLive Video |
Patrik Joslin Kenfack · Ulrich Aïvodji · Samira Ebrahimi Kahou 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
The Intersectionality Problem for Algorithmic Fairness
(
Spotlight
)
>
SlidesLive Video |
Johannes Himmelreich · Arbie Hsu · Kristian Lum · Ellen Veomett 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Counterpart Fairness – Addressing Systematic Between-Group Differences in Fairness Evaluation
(
Spotlight
)
>
SlidesLive Video |
Yifei Wang · Zhengyang Zhou · Liqin Wang · John Laurentiev · Peter Hou · Li Zhou · Pengyu Hong 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Multi-Output Distributional Fairness via Post-Processing
(
Spotlight
)
>
SlidesLive Video |
Gang Li · Qihang Lin · Ayush Ghosh · Tianbao Yang 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Fair Summarization: Bridging Quality and Diversity in Extractive Summaries
(
Spotlight
)
>
SlidesLive Video |
Sina Bagheri Nezhad · Sayan Bandyapadhyay · Ameeta Agrawal 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Improving Bias Metrics in Vision-Language Models by Addressing Inherent Model Disabilities
(
Spotlight
)
>
SlidesLive Video |
Lakshmipathi Balaji Darur · Shanmukha Sai Keerthi Gouravarapu · Shashwat Goel · Ponnurangam Kumaraguru 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Imitation Guided Automated Red Teaming
(
Spotlight
)
>
|
Sajad Mousavi · Desik Rengarajan · Ashwin Ramesh Babu · Vineet Gundecha · Soumyendu Sarkar 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Multilingual Hallucination Gaps in Large Language Models
(
Spotlight
)
>
SlidesLive Video |
Cléa Chataigner · Afaf Taik · Golnoosh Farnadi 🔗 |
Sat 5:27 p.m. - 5:30 p.m.
|
Fairness-Enhancing Data Augmentation Methods for Worst-Group Accuracy
(
Spotlight
)
>
SlidesLive Video |
Monica Welfert · Nathan Stromberg · Lalitha Sankar 🔗 |
-
|
The Search for Less Discriminatory Algorithms: Limits and Opportunities
(
Poster
)
>
|
Benjamin Laufer · Manish Raghavan · Solon Barocas 🔗 |
-
|
From Models to Systems: A Comprehensive Fairness Framework for Compositional Recommender Systems
(
Poster
)
>
|
Brian Hsu · Cyrus DiCiccio · Natesh Pillai · Hongseok Namkoong 🔗 |
-
|
M²FGB: A Min-Max Gradient Boosting Framework for Subgroup Fairness
(
Poster
)
>
|
Jansen Pereira · Giovani Valdrighi · Marcos M. Raimundo 🔗 |
-
|
Understanding The Effect Of Temperature On Alignment With Human Opinions
(
Poster
)
>
|
Maja Pavlovic · Massimo Poesio 🔗 |
-
|
Towards Reliable Fairness Assessments of Multi-Label Image Classifiers
(
Poster
)
>
|
Melissa Hall · Bobbie Chern · Laura Gustafson · Denisse Ventura · Harshad Kulkarni · Candace Ross · Nicolas Usunier 🔗 |
-
|
Toward Large Language Models that Benefit for All: Benchmarking Group Fairness in Reward Models
(
Poster
)
>
|
Kefan Song · Jin Yao · Shangtong Zhang 🔗 |
-
|
Efficient Fairness-Performance Pareto Front Computation
(
Poster
)
>
|
Mark Kozdoba · Benny Perets · Shie Mannor 🔗 |
-
|
Better Bias Benchmarking of Language Models via Multi-factor Analysis
(
Poster
)
>
|
Hannah Powers · Ioana Baldini · Dennis Wei · Kristin P Bennett 🔗 |
-
|
Adaptive Group Robust Ensemble Knowledge Distillation
(
Poster
)
>
|
Patrik Joslin Kenfack · Ulrich Aïvodji · Samira Ebrahimi Kahou 🔗 |
-
|
Evaluating Gender Bias Transfer between Pre-trained and Prompt Adapted Language Models
(
Poster
)
>
|
Nivedha Sivakumar · Natalie Mackraz · Samira Khorshidi · Krishna Patel · Barry-John Theobald · Luca Zappella · Nicholas Apostoloff 🔗 |
-
|
The Intersectionality Problem for Algorithmic Fairness
(
Poster
)
>
|
Johannes Himmelreich · Arbie Hsu · Kristian Lum · Ellen Veomett 🔗 |
-
|
Counterpart Fairness – Addressing Systematic Between-Group Differences in Fairness Evaluation
(
Poster
)
>
|
Yifei Wang · Zhengyang Zhou · Liqin Wang · John Laurentiev · Peter Hou · Li Zhou · Pengyu Hong 🔗 |
-
|
Measuring Representational Harms in Image Generation with a Multi-Group Proportional Metric
(
Poster
)
>
|
Sangwon Jung · Claudio Mayrink Verdun · Alex Oesterling · Sajani Vithana · Taesup Moon · Flavio du Pin Calmon 🔗 |
-
|
Mitigating Bias in Facial Recognition Systems: Centroid Fairness Loss Optimization
(
Poster
)
>
|
Jean-Rémy Conti · Stephan Clémençon 🔗 |
-
|
Towards Fair RAG: On the Impact of Fair Ranking in Retrieval-Augmented Generation
(
Poster
)
>
|
To Eun Kim · Fernando Diaz 🔗 |
-
|
Benchmark to Audit LLM Generated Clinical Notes for Disparities Arising from Biases and Stereotypes
(
Poster
)
>
|
Hongyu Cai · Swetasudha Panda · Naveen Jafer Nizar · Qinlan Shen · Daeja Oxendine · Sumana Srivatsa · Krishnaram Kenthapadi 🔗 |
-
|
Multi-Output Distributional Fairness via Post-Processing
(
Poster
)
>
|
Gang Li · Qihang Lin · Ayush Ghosh · Tianbao Yang 🔗 |
-
|
Fair Summarization: Bridging Quality and Diversity in Extractive Summaries
(
Poster
)
>
|
Sina Bagheri Nezhad · Sayan Bandyapadhyay · Ameeta Agrawal 🔗 |
-
|
Different Horses for Different Courses: Comparing Bias Mitigation Algorithms in ML
(
Poster
)
>
|
Prakhar Ganesh · Usman Gohar · Lu Cheng · Golnoosh Farnadi 🔗 |
-
|
Improving Bias Metrics in Vision-Language Models by Addressing Inherent Model Disabilities
(
Poster
)
>
|
Lakshmipathi Balaji Darur · Shanmukha Sai Keerthi Gouravarapu · Shashwat Goel · Ponnurangam Kumaraguru 🔗 |
-
|
Beyond Internal Data: Constructing Complete Datasets for Fairness Testing
(
Poster
)
>
|
Varsha Ramineni · Hossein A. Rahmani · Emine Yilmaz · David Barber 🔗 |
-
|
Beyond the Safety Bundle: Auditing the Helpful and Harmless Dataset
(
Poster
)
>
|
Khaoula Chehbouni · Jonathan Colaço Carr · Yash More · Jackie CK Cheung · Golnoosh Farnadi 🔗 |
-
|
What's in a Query: Examining Distribution-based Amortized Fair Ranking
(
Poster
)
>
|
Aparna Balagopalan · Kai Wang · Asia Biega · Marzyeh Ghassemi 🔗 |
-
|
Optimal Selection Using Algorithmic Rankings with Side Information
(
Poster
)
>
|
Kate Donahue · Nicole Immorlica · Brendan Lucier 🔗 |
-
|
Verifiable evaluations of machine learning models using zkSNARKs
(
Poster
)
>
|
Tobin South · Alexander Camuto · Shrey Jain · Robert Mahari · Christian Paquin · Jason Morton · Alex `Sandy' Pentland 🔗 |
-
|
Imitation Guided Automated Red Teaming
(
Poster
)
>
|
Sajad Mousavi · Desik Rengarajan · Ashwin Ramesh Babu · Vineet Gundecha · Soumyendu Sarkar 🔗 |
-
|
Measuring the Impact of Equal Treatment as Blindness via Explanations Disparity
(
Poster
)
>
|
Carlos Mougan · Salvatore Ruggieri · Laura State · Antonio Ferrara · Steffen Staab 🔗 |
-
|
Multilingual Hallucination Gaps in Large Language Models
(
Poster
)
>
|
Cléa Chataigner · Afaf Taik · Golnoosh Farnadi 🔗 |
-
|
Evaluating and Mitigating Discrimination in Language Model Decisions
(
Poster
)
>
|
Alex Tamkin · Amanda Askell · Liane Lovitt · Esin DURMUS · Nicholas Joseph · Shauna Kravec · Karina Nguyen · Jared Kaplan · Deep Ganguli 🔗 |
-
|
Fairness-Enhancing Data Augmentation Methods for Worst-Group Accuracy
(
Poster
)
>
|
Monica Welfert · Nathan Stromberg · Lalitha Sankar 🔗 |
-
|
Improving Fairness in Matching under Uncertainty
(
Poster
)
>
|
Piyushi Manupriya 🔗 |
-
|
LLMs Infer Protected Attributes Beyond Proxy Features
(
Poster
)
>
|
Dimitri Staufer 🔗 |
-
|
On Optimal Subgroups for Group Distributionally Robust Optimisation
(
Poster
)
>
|
Anissa Alloula · Daniel McGowan · Bartlomiej W. Papiez 🔗 |
-
|
Demographic (Mis)Alignment of LLMs' Perception of Offensiveness
(
Poster
)
>
|
Shayan Alipour · Indira Sen · Preetam Prabhu Srikar Dammu · Chris Choi · Mattia Samory · Tanu Mitra 🔗 |
-
|
Towards Better Fairness Metrics for Counter-Human Trafficking AI Initiatives
(
Poster
)
>
|
Vidya Sujaya · Pratheeksha Nair · Reihaneh Rabbany 🔗 |
-
|
Q-Morality: Quantum-Enhanced ActAdd-Guided Bias Reduction in LLMs
(
Poster
)
>
|
Shardul Kulkarni 🔗 |
-
|
Examining Distribution-based Amortized Fair Ranking
(
Poster
)
>
|
Aparna Balagopalan · Kai Wang · Asia Biega · Marzyeh Ghassemi 🔗 |
-
|
Exploring AUC-like metrics to propose threshold-independent fairness evaluation
(
Poster
)
>
|
Daniel Gratti · Thalita Veronese · Marcos M. Raimundo 🔗 |
-
|
Integrating Participatory Methods with Technical Fairness Solutions: Enhancing Bias Mitigation and Equity in AI Systems
(
Poster
)
>
|
Abdoul Jalil Djiberou Mahamadou · Lea Goetz 🔗 |