Excellent summary on some of the current topics that currently permeates in the burgeoning ethics of AI field. As the authors point out in the first few pages, they realize the topic choices made here may not be timeless, and that in just a few years, the field will have moved on to tackle other topics, either because the current crop of of areas have been mastered (or it’s been shown to be intractable), and/or society has not focused its attention on some other topic that may become more timely at that time (e.g. the ethics of automatic warfare etc.). The purpose of this book is not be a “textbook” for the field, but to summarize current trends, and to quickly educate other practitioners in the broader ML/AI/DS space (what I’ll call the ‘machine intelligence fields’), on these topics.
Although the topic of “ethics” in AI is not new, and in some ways, almost all practitioners in the broader machine intelligence fields will have been exposed to a facet of the topic when they grapple with the precision/recall conundrum of classification in either their work or education, this book is possibly the first text in the non-specialist market that directly speak to the designers of the algorithms or machine learning pipeline. To be sure other books have been published in this domain, especially after 2016. Two that come to mind that I’ve read recently include “Automating Inequality” and “Weapons of Math Destruction”, yet again these were mostly aimed at either a general reader, or a policy specialist/thinker. Not that “The Ethical Algorithm” is overly technical and opaque to non-specialist, just that it really feels like it’s speaking to a more technically-oriented crowd.
5 topics are covered in the book, algorithmic privacy (or more commonly known in the field as differential privacy), the notion of fairness in machine intelligence, biases that occur from machine learning algorithms in a variety of domains (travel route-recommendation/optimization, dating, etc.), a chapter on ‘p-hacking’ and ‘hill-climbing’ in experimentation and model-space exploration, and a sort of generic “didn’t have time to write a full chapter off” last chapter on interpretation in ML and “future” topics like the singularity.
A few months ago I previously attended a brown-bag lunch where the census’s chief scientist discussed their application on differential privacy and I found the topic fascinating. In my opinion, this may be the most “low-key” impactful use-case of ethical AI or ML because it seeks to preserve a level of privacy (which is tunable) by “masking” samples that go out into the ether. These samples can then be used as input for ML algorithms or experimentation fodder, but the key here is that there is a provable way to ensure that reconstruction, or a mapping, of the critical ground-truth is close to impossible. Given how so many of the “first wave” of ethical dilemmas in “big tech” have centered on how personal data could be used to detriment individuals (e.g. healthcare providers increasing premiums because of health or behavioral data etc.), the implementation of this layer to all machine intelligence processes in the enterprise will be critical over the next few years.
The two most practical chapters in the book are probably the 2nd and 4th however, which deals with the algorithmic notions of ‘fairness’ and hill-climbing/p-hacking respectively. Parts of chapter 2 may be viewed as remedial for anyone but the most junior data scientist or ML practitioner, but it's a good review. I was disappointed a bit in the chapter as there was a real opportunity here to discuss the latest/greatest fairness-preserving algorithmic procedures being worked on currently, which ranges from re-weighting by stratifying protected classes, to something a bit more exotic like adversarial de-biasing. Much of those topics would be covered in a standard graduate school course in AI Ethics. I think this chapter would be a great “prelude” chapter in such a course before discussing the nitty-gritty of those technical procedures.
Likewise, the chapter on hill-climbing and p-hacking was also instructive and well-written. This chapter uses the guise of cheating in machine learning competitions to discuss the perils of naive model-exploration, often referred to as hill-climbing to practitioners. The problem here is that one wishes to build an algorithm that is generalizable, and there are well-known validation pipelines one can set up to measure/observe how generalizable your model may be as it’s being trained. However, in a context of a Kaggle competition, where the purpose is not to build the best performing algorithm to solve a business or other objective, but to increase your score, this process can be perverted (since it is known that there is a ‘max-score’ one can achieve).
When this occurs, naive feature-engineering techniques can be leveraged to “hill-climb” (or increase the out-of-sample performance score) iteratively. Yet, because the methodology is in some ways “backwards”, the resultant model could end up being a sort of “Frankenstein” algorithm that is poorly generalizable, but is efficacious in the narrow case of the competition. The chapter goes through topics relevant to both repeated experimentation and multiple linear regression, like the Bonferroni correction of the critical value for hypothesis tests and applies it to exploration in the model-space (or “the Garden of the Forking Path”). Pretty good read here.
Chapter 3 was interesting, but I suspect this will be the one that is expanded in the future as it deals mostly unforeseen consequences of algorithms to human-activity, like driving recommendation, online-dating, and things that in general can be characterized as matching problems. Alvin Roth is mentioned and discussed in this chapter, with his famous algorithm given some spotlight as well, but I felt it really was mostly a chapter on matching as opposed to a chapter on general ethical issues emanating from algorithmic processes. As algorithms penetrate other parts of human activity (like warfare), those topics will likely be appended to this chapter. This chapter also felt the closest to traditional social science to me, as many of the adversarial/gaming processes described within the matching domain leverages the traditional notion of utility as well as the traditional notion of characterizing multi-agent behavior as a non-cooperative game.
Overall, I am very satisfied with this book, it educates well as well as informs, and the narrative provokes thoughts on this topic that could lead to fruitful research or ideas in this field. I believe both the current practitioner, and even adjacent researchers in the AI/ML domains could benefit from this reading. Students of the subject could also benefit from the text both as ancillary reading of quasi-instruction to supplement a main-text or notes, but to give them advanced notice of issues that will soon become standard in all design in the machine intelligence fields shortly. Highly recommended.