| |
| |
|
Towards Explainable Neural Network Fairness
|
|

|
ZHANG Mengdi
PhD Candidate
School of Computing and Information Systems
Singapore Management University
|
|
Research Area
Dissertation Committee
Research Advisor
Committee Members
|
|
|
|
Date
9 November 2022 (Wednesday)
|
|
Time
2:00pm - 3:00pm
|
|
Venue
Meeting room 5.1, Level 5,
School of Computing and Information Systems 1,
Singapore Management University,
80 Stamford Road Singapore 178902
|
|
We look forward to seeing you at this research seminar.

|
|
|
| |
|
About The Talk
Neural networks are achieving higher accuracy and exceptional performance and widely applied in solving many real-world problems. At the same time, they are shown to be vulnerable to attacks, difficult to debug and subject to fairness issues. To improve people's trust in the technology, it is necessary to provide some human-understandable explanation of neural networks' decisions, e.g., why is that my loan application is rejected whereas hers is approved? It becomes a fairness issue if the only explanation possible is based on certain sensitive features (such as race and gender). In this thesis, we develop multiple approaches towards improving people's trust on neural networks.
We propose measurements on the decision explainability of neural networks. Afterwards, we develop multiple algorithms that allow us to automatically evaluate and compare neural networks in terms of the measurements. We propose \sgde, an interpretable testing approach which systematically identifies and measures hidden (which we call `subtle') group discrimination of a neural network characterized by conditions over combinations of the sensitive features. We propose an approach which adaptively chooses the fairness improving method based on causality analysis.
While most existing fairness improvement methods have shown effectiveness on tabular data, we are still working on textual data and mitigating bias on NLP models such as BERT.
|
| |
|
Speaker Biography
Mengdi ZHANG is a Ph.D. candidate at SMU SCIS, supervised by Prof. Jun SUN. Her research interests are mainly on AI security, including machine learning interpretability, fairness testing and discrimination mitigation.
|
|