Fairness and bias in Machine Learning - QCon 2019 - Bias Fairness presentation

Page created by Troy Clarke
 
CONTINUE READING
Fairness and bias in Machine Learning - QCon 2019 - Bias Fairness presentation
QCon 2019

                Fairness and bias in
                 Machine Learning
                        A quick review on tools to detect biases in machine learning model

thierry.silbermann@nubank.com.br                 Thierry Silbermann, Tech Lead Data Science at Nubank
Fairness and bias in Machine Learning - QCon 2019 - Bias Fairness presentation
Data collection
•   Today’s applications collect and mine vast quantities of
    personal information.

•   The collection and use of such data raise two important
    challenges.

    •   First, massive data collection is perceived by many as a
        major threat to traditional notions of individual privacy.

    •   Second, the use of personal data for algorithmic decision-
        making can have unintended and harmful consequences,
        such as unfair or discriminatory treatment of users.
Fairness and bias in Machine Learning - QCon 2019 - Bias Fairness presentation
Data collection
•   Today’s applications collect and mine vast quantities of
    personal information.

•   The collection and use of such data raise two important
    challenges.

    •   First, massive data collection is perceived by many as a
        major threat to traditional notions of individual privacy.

    •   Second, the use of personal data for algorithmic decision-
        making can have unintended and harmful consequences,
        such as unfair or discriminatory treatment of users.
Fairness and bias in Machine Learning - QCon 2019 - Bias Fairness presentation
Fairness and bias in Machine Learning - QCon 2019 - Bias Fairness presentation
Fairness and bias in Machine Learning - QCon 2019 - Bias Fairness presentation
Fairness and bias in Machine Learning - QCon 2019 - Bias Fairness presentation
Fairness
•   Fairness is increasingly important concern as machine
    learning models are used to support decision making in
    high-stakes applications such as:

      •   Mortgage lending

      •   Hiring

      •   Prison sentencing

      •   (Approve customers, increase credit line)
Fairness and bias in Machine Learning - QCon 2019 - Bias Fairness presentation
Definitions of fairness

               http://fairware.cs.umass.edu/papers/Verma.pdf
Definitions of fairness
•   It is impossible to satisfy all definitions of fairness at the
    same time [Kleinberg et al., 2017]

•   Although fairness research is a very active field, clarity on
    which bias metrics and bias mitigation strategies are best
    is yet to be achieved [Friedler et al., 2018]

•   In addition to the multitude of fairness definitions,
    different bias handling algorithms address different parts
    of the model life-cycle, and understanding each research
    contribution, how, when and why to use it is challenging
    even for experts in algorithmic fairness.

        Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk
Example: Prison sentencing

     Did not           True Negative         False Positive
    recidivate

    Recidivate         False Negative         True Positive

                            Label                Label
                          low-risk             high-risk

  Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk
Example: Prison sentencing
Decision maker: Of those I’ve labeled
high-risk, how many will recidivate ?
          Predictive value
                                                  Did not           True Negative        False Positive
                                                 recidivate

                                                Recidivate         False Negative         True Positive

                                                                         Label                Label
                                                                       low-risk             high-risk

           Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk
Example: Prison sentencing
Decision maker: Of those I’ve labeled
high-risk, how many will recidivate ?
           Predictive value
                                                  Did not          True Negative         False Positive
Defendant: What’s the probability I’ll           recidivate
be incorrectly classifying high-risk ?
         False positive rate
                                                Recidivate         False Negative         True Positive

                                                                         Label                Label
                                                                       low-risk             high-risk

           Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk
Example: Prison sentencing
Decision maker: Of those I’ve labeled
high-risk, how many will recidivate ?
           Predictive value
                                                  Did not           True Negative        False Positive
Defendant: What’s the probability I’ll           recidivate
be incorrectly classifying high-risk ?
         False positive rate
                                                Recidivate         False Negative         True Positive
  Society [think hiring rather than
  criminal justice]: Is the selected
  set demographically balanced ?
                                                                         Label                Label
            Demography                                                 low-risk             high-risk

           Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk
18 scores/metrics

            https://en.wikipedia.org/wiki/Confusion_matrix
Terminology
•   Favorable label: a label whose value corresponds to an outcome that provides an advantage
    to the recipient.

         •   receiving a loan, being hired for a job, and not being arrested

• Protected attribute: attribute that partitions a population into groups that have parity in
    terms of benefit received

         • race, gender, religion
         • Protected attributes are not universal, but are application specific
• Privileged value of a protected attribute: group that has historically been at a systematic
    advantage

• Group fairness: the goal of groups defined by protected attributes receiving similar
    treatments or outcomes

• Individual fairness: the goal of similar individuals receiving similar treatments or outcomes
Terminology
• Bias: systematic error
    • In the context of fairness, we are concerned with
       unwanted bias that places privileged groups at a
       systematic advantage and unprivileged groups at a
       systematic disadvantage.

• Fairness metric: a quantification of unwanted bias in
  training data or models.

• Bias mitigation algorithm: a procedure for reducing
  unwanted bias in training data or models.
But wait !
•   I’m not using any feature that is discriminatory for my
    application !

    •   I’ve never used gender or even race !
But wait !

      https://demographics.virginia.edu/DotMap/index.html
But wait !
          Chicago Area, IL, USA

      https://demographics.virginia.edu/DotMap/index.html
Fairness metric
• Confusion matrix
    • TP, FP, TN, FN, TPR, FPR, TNR, FNR
    • Prevalence, accuracy, PPV, FDR, FOR, NPV
    • LR+, LR-, DOR, F1
Fairness metric
• Difference of Means
• Disparate Impact
• Statistical Parity
• Odd ratios
• Consistency
• Generalized Entropy Index
Statistical parity difference
•   Group fairness == statistical parity difference == equal acceptance rate
    == benchmarking

•   A classifier satisfies this definition if subjects in both protected and
    unprotected groups have equal probability of being assigned to the
    positive predicted class.

•   Example, this would imply equal probability for male and female
    applicants to have good predicted credit score:

           •   P(d = 1 | G = male) = P (d = 1 | G = female)

•   The main idea behind this definition is that applicants should have an
    equivalent opportunity to obtain a good credit score, regardless of their
    gender.
Disparate impact
The 80% test was originally framed by
a panel of 32 professionals assembled                       X=0   X=1
by the State of California Fair
Employment Practice Commission
(FEPC) in 1971

                                                    FALSE    A    B
                                        Predicted
                                        condition
                                                    TRUE    C     D
Disparate impact
                                     X=0   X=1

                             FALSE    A    B
                 Predicted
                 condition
                             TRUE    C     D

The 80% rule can then be quantified as:
Aequitas approach

https://dsapp.uchicago.edu/projects/aequitas/
How about some solutions?
Disparate impact remover

                               Relabelling

Learning Fair representation
Disparate impact remover

                                     Prejudice remover regulariser

      Reject Option Classification

                                                    Relabelling

      Adversarial Debiasing

                                             Optimised Preprocessing

Learning Fair representation
Disparate impact remover

                                         Prejudice remover regulariser

      Reject Option Classification
                                         Equalised Odds Post-processing

Meta-Algorithm for Fair Classification
                                                        Relabelling

      Adversarial Debiasing                   Reweighing

                                                 Optimised Preprocessing
  Additive counterfactually fair estimator

Learning Fair representation
                                 Calibrated Equalised Odds Post-processing
Tools
How about fixing
                predictions?

•   There are three main paths to the goal of making fair
    predictions:

      •   fair pre-processing,

      •   fair in-processing, and

      •   fair post-processing
AIF360, https://arxiv.org/abs/1810.01943
Pre-Processing
•   Reweighing generates weights for the training examples in each
    (group, label) combination differently to ensure fairness before
    classification.

•   Optimized preprocessing (Calmon et al., 2017) learns a probabilistic
    transformation that edits the features and labels in the data with group
    fairness, individual distortion, and data fidelity constraints and
    objectives.

•   Learning fair representations (Zemel et al., 2013) finds a latent
    representation that encodes the data well but obfuscates information
    about protected attributes.

•   Disparate impact remover (Feldman et al., 2015) edits feature values to
    increase group fairness while preserving rank-ordering within groups.
In-Processing
•   Adversarial debiasing (Zhang et al., 2018) learns a
    classifier to maximize prediction accuracy and
    simultaneously reduce an adversaries ability to determine
    the protected attribute from the predictions. This
    approach leads to a fair classifier as the predictions
    cannot carry any group discrimination information that the
    adversary can exploit.

•   Prejudice remover (Kamishima et al., 2012) adds a
    discrimination-aware regularization term to the learning
    objective
Post-Processing
•   Equalized odds postprocessing (Hardt et al., 2016) solves a
    linear program to find probabilities with which to change
    output labels to optimize equalized odds.

•   Calibrated equalized odds post-processing (Pleiss et al.,
    2017) optimizes over calibrated classifier score outputs to
    find probabilities with which to change output labels with an
    equalized odds objective.

•   Reject option classification (Kamiran et al., 2012) gives
    favorable outcomes to unprivileged groups and unfavorable
    outcomes to privileged groups in a confidence band around
    the decision boundary with the highest uncertainty.
Experiments
   Datasets              Adult Census Income, German Credit, COMPAS

                                         Disparate impact
                                    Statistical parity difference
    Metrics
                                     Average odds difference
                                   Equal opportunity difference
                  Logistic Regression (LR), Random Forest Classifier (RF), Neural
  Classifiers
                                           Network (NN)

                             Re-weighing (Kamiran & Calders, 2012)
Pre-processing           Optimized pre-processing (Calmon et al., 2017)
  Algorithms             Learning fair representations (Zemel et al., 2013)
                         Disparate impact remover (Feldman et al., 2015)

 In-processing               Adversarial debasing (Zhang et al., 2018)
   Algorithms               Prejudice remover (Kamishima et al., 2012)

                       Equalized odds post-processing (Hardt et al., 2016)
Post-processing
                     Calibrated eq. odds post-processing (Pleiss et al., 2017)
  Algorithms
                         Reject option classification (Kamiran et al., 2012)
                                                             AIF360, https://arxiv.org/abs/1810.01943
Results - Statistical Parity
    Difference (SPD)

          SPD Fair Value is 0
                                AIF360, https://arxiv.org/abs/1810.01943
Results - Disparate Impact
            (DI)

         DI Fair Value is 1
                              AIF360, https://arxiv.org/abs/1810.01943
Adult census dataset

                            Results
Protected attribute: race

                                      AIF360, https://arxiv.org/abs/1810.01943
Results

          AIF360, https://arxiv.org/abs/1810.01943
Thank you
References
•   Conference

    •   ACM Conference on Fairness, Accountability, and
        Transparency (ACM FAT*) https://fatconference.org/

    •   IJCAI 2017 Workshop on Explainable Artificial
        Intelligence (XAI) http://home.earthlink.net/~dwaha/
        research/meetings/ijcai17-xai/

    •   Interpretable ML Symposium - NIPS 2017 http://
        interpretable.ml/
References
•   Books

      •   https://fairmlbook.org/

•   Course materials

      •   Berkeley CS 294: Fairness in machine learning

      •   Cornell INFO 4270: Ethics and policy in data science

      •   Princeton COS 597E: Fairness in machine learning
References
•   Papers

      •   Fairness Definitions Explained: http://fairware.cs.umass.edu/
          papers/Verma.pdf

      •   AIF360: An Extensible Toolkit for Detecting, Understanding,
          and Mitigating Unwanted Algorithmic Bias https://arxiv.org/
          pdf/1810.01943.pdf

      •   Aequitas: A Bias and Fairness Audit Toolkit: https://arxiv.org/
          pdf/1811.05577.pdf

      •   FairTest: Discovering Unwarranted Associations in Data-
          Driven Applications: https://arxiv.org/pdf/1510.02377.pdf
References

•   Videos

      •   Tutorial: 21 fairness definitions and their politics
          https://www.youtube.com/watch?v=jIXIuYdnyyk

      •   AI Fairness 360 Tutorial at ACM FAT* 2019 https://
          www.youtube.com/watch?v=XCFDckvyC0M
You can also read