02/26/2020 – Subil Abraham – Explaining models

A big concern with the usage of current ML systems is the issue of fairness and bias when making their decisions. Bias can creep into ML decisions through either the design of the algorithm or through training datasets that are labeled in a way to bias against certain kinds of things. The example used in this paper is the bias against African Americans in an ML system used by judges to predict the probability of a person re-offending after committing a crime. Fairness is hard to judge when ML systems are black boxes so this paper proposes that if ML systems expose reasons behind the decisions (i.e. the idea of explainable AI), a better judgement of the fairness of the decision can be made by the user. To this end, this paper examines the effect of four of different kinds of explanations of the ML decisions on people’s judgements of the fairness of that decision.

I believe this is a very timely and necessary paper in these times, with ML systems being used more and more for sensitive and life changing decisions. It is probably impossible to stop people from adopting these systems so the next best thing is making explainability of the ML decisions mandatory, so people can see and judge if there was potentially bias in the ML system’s decisions. It is interesting that people were mostly able to perceive that there were fairness issues in the raw data. You would think that that would be hard but the generated explanations may have worked well enough to help with that (though I do wish they could’ve shown an example comparing a raw data point and a processed data point that showed how their pre-processing cleaned things). I did wonder why they didn’t show confidence levels to the users in the evaluation, but their explanation that it was something they could not control for makes sense. People could have different reactions to confidence levels, some thinking that anything less than a 100% is insufficient, while others thinking that 51% is good enough. So keeping it out is a limiting but is logical.

  1. What other kinds of generated explanations could be beneficial, outside of the ones used in the paper?
  2. Checking for racial bias is an important case for fair AI. In what other areas is fairness and bias correction in AI critical?
  3. What would be ways that you could mitigate any inherent racial bias of the users who are using explainable AI, when they are making their decisions?

One thought on “02/26/2020 – Subil Abraham – Explaining models

  1. I believe fairness and bias is important in all fields in AI.
    Or else we wouldn’t have a conference (ACM FAT*) suddenly forming as this field grows. It sheds light on the importance of fairness, accountability, and transparency in this emerging discipline.
    I’d say gender bias is another critical issue, if we have trained AI systems monitoring job applicants and those models are biased like language or word embeddings [1] then probably more men applications would be considered than women.

    1 Bolukbasi, T., Chang, K. W., Zou, J. Y., Saligrama, V., & Kalai, A. T. (2016). Man is to computer programmer as woman is to homemaker? debiasing word embeddings. In Advances in neural information processing systems (pp. 4349-4357).

Leave a Reply