Abstract
Ethics has powerful teeth, but these are barely being used in the ethics of AI today – it is no wonder the ethics of AI is then blamed for having no teeth. This article argues that ‘ethics’ in the current AI ethics field is largely ineffective, trapped in an ‘ethical principles’ approach and as such particularly prone to manipulation, especially by industry actors. Using ethics as a substitute for law risks its abuse and misuse. This significantly limits what ethics can achieve and is a great loss to the AI field and its impacts on individuals and society. This article discusses these risks and then highlights the teeth of ethics and the essential value they can – and should – bring to AI ethics now.
Ethics has great powerful teeth. Unfortunately, we are barely using them in AI ethics – it is no wonder then that the ethics of AI is called toothless. This article reflects on the various ethics mechanisms that have emerged over the past couple of years to respond to the massive deployment and use of AI in society and its associated risks. These mechanisms include lists of principles, ethics codes, recommendations and guidelines. Yet, as many have shown, these ethics developments, while promising, are also problematic: their effectiveness is still to be demonstrated and they are particularly prone to manipulation, especially by industry. This is a loss to the AI field and significantly limits what ethics can achieve for society and individuals. However, as this article shows, the issue is not that ethics is worthless (or toothless) in the face of current AI deployment; it is rather that ethics is used (or manipulated) in such a way that it is rendered ineffective for AI ethics.
The article begins by highlighting the nature of ethics in the AI ethics field today – AI ethics is primarily principled, i.e. it follows a ‘law’ conception of ethics. It then shows how this approach to ethics fails to achieve what it pretends to achieve. This article next builds on the real value of ethics – its ‘teeth’ – which we define as a constantly renewed ability to see the new as it emerges. We show how this capacity to avoid cognitive and perceptive inertia that make us passive in the face of new developments is highly critical for AI ethics today. Finally, while we recognise that the legalistic approach to ethics is not completely off the mark, we argue that it is the
AI ethics today
There are many ongoing discussions and initiatives on the ethics of AI in various stakeholder quarters (policy, academia, industry and even the media). We can certainly rejoice about this. In particular, policymakers (e.g. European Commission and the European Parliament) and industry are showing much concern about getting things right to ensure the ethical and responsible development and deployment of AI in society. 1 It is now well recognised that things could go really wrong if AI is implemented without due regard and consideration for its potentially harmful impacts on individuals, on specific communities and on society as a whole (including, for example, bias and discrimination, injustice, privacy infringements, increase in surveillance, loss of autonomy, overdependency on technology, etc.). 2 We can then observe a turn to ethics to ensure that AI is deployed in a manner that respects dearly held societal values and norms, and puts them at the heart of responsible technology development and deployment (Hagendorff, 2020; Jobin et al., 2019). The ‘Ethics guidelines for trustworthy AI’ drafted by the High-Level Expert Group on AI – a group set up by the European Commission in 2018 – is one example of such recent ethics initiatives (High-Level Expert Group on Artificial Intelligence, 2019).
However, the manner in which ‘ethics’ is currently being used in the field of AI ethics is problematic. Today’s AI ethics is dominated by what the British philosopher G.E.M. Anscombe calls a ‘
Misusing ethics as a replacement for regulation
First, it is problematic because of its potential for misuse as a replacement for regulation (whether through law, policies or standards). Many articles have argued the following point over the last couple of years: AI ethics is, by itself, deficient in regulating behaviours and practices for proper development and deployment of AI (Article 19, 2019; Greene et al., 2019; Hagendorff, 2020; Jobin et al., 2019; Klöver and Fanta, 2019; Mittelstadt, 2019; Wagner, 2018). Wagner points, for instance, to the case of a member of the Google DeepMind ethics team at the
Considering it has no means to ensure compliance, ethics does indeed fall short in this respect. As Hagendorff puts it, ethics ‘lacks mechanisms to reinforce its own normative claims’ (2020: 99). If ethics is about regulation, then indeed ethics has no teeth. This is the view of the human rights organisation Article 19 for which, although ethics initiatives ‘put forth admirable goals’, ‘their common lack of accountability and enforcement mechanisms’ makes these initiatives ineffective (Article 19, 2019: 18). Ultimately and unsurprisingly, ethics is then blamed for being toothless.
However, it is essential to be clear here: the issue is not that ethics is asked to do something for which it is too weak, or too soft. It is rather that it is asked to do something that it is
Some argue that this misuse of ethics is an intended one, driven by the desire to ensure that AI will not be regulated by law, i.e. more flexibility is possible and no hard lines will be put in place restricting industrial and commercial interests related to this technology (Klöver and Fanta, 2019). This critique has been addressed to the AI HLEG guidelines, for instance. Article 19 points to the fact that: during deliberations at the European High-Level Expert Group on Artificial Intelligence (EU-HLEG), industry was heavily represented, but academics and civil society did not enjoy the same luxury. And while some non-negotiable ethical principles were originally articulated in the document, these were omitted from the final document due to industry pressure. (Article 19, 2019: 18)
We are not using the teeth of ethics in AI ethics today
Second, because the AI ethics field tends to be dominated by this ‘law conception of ethics’, it does not actually truly use what ethics has to offer, its actual proper teeth, despite the great need for them. What are these ethics teeth and what can they bring to the field?
The real teeth of ethics consist of a
This constantly refreshed capacity to perceive the world is the one that helps us not to be boiled alive like the frog: it helps us notice small changes as they unfold. In the context of AI, the increasingly hot water is a combination of evolutions that include an expansion and deepening of surveillance by governments and private companies, an increasing dependency on technology and the deployment of biased systems that lead to discrimination towards women and minorities. The progressive evolutions these bring to society need to be closely studied and resisted when their negative impacts outweigh their benefits.
In that sense, ethics entertains a very close link with social sciences, as an effort to see what we do not otherwise see. Ethics helps us look concretely at how the world changes. It helps clean up the lens through which we see the world so that we can be more attentive to its transformations (and AI does bring many of these). It is essential for ethics to support us in this regard. It helps us to be less passive towards these changes and puts us in a better position to then steer them in ways that do not harm individuals and society, and that help us live better. In his piece on the ‘Ethics of AI ethics’, Hagendorff makes a similar proposition by questioning the dominant deontological approach to ethics in AI ethics (what we have called in this article a legalistic approach to ethics) whose purpose is primarily ‘to limit, control, or steer’ (2020: 112). He presses the need to turn to virtue ethics for the AI field, an ethics that aims at ‘broadening the scope of action, uncovering blind spots, promoting autonomy and freedom, and fostering self-responsibility’ (Hagendorff, 2020: 112). Other ethical theory frameworks that would bring significant value to the AI ethics debate today include in particular the Spinozist approach that focuses on increase or decrease of agency and capacity of action.
Ethical principles, norms and values are an ‘end’ of ethics, not ethics itself
So, are we simply getting it wrong in AI ethics, which today, as we have shown, is dominated by a ‘law-conception of ethics’? Is the current legalistic approach to ethics completely off the mark? Not exactly. The issue is rather that principles, norms, values – this law conception of ethics that is dominant today in AI ethics – are rather an
First (1), it is an end of ethics in the sense of its
However, it is essential to remember that ethics as ethical principles is also an end of ethics in another sense (2):
Ethics is about navigating murky and risky waters; to do so, it needs to be watchful. For instance, in the field of AI, ethical principles do not, by themselves, help practically explore challenging issues such as fairness in highly complex socio-technical systems. These need to be investigated in-depth to ensure we are not implementing systems that go against dearly held norms and values. Without a continuous process of questioning what is or may be obvious, of digging behind what seems to be settled, of keeping alive this interrogation, ethics is rendered ineffective. And thus, the settling of ethics into established norms and principles comes down to its termination.
Considering the radical, massive and widespread impact of AI on the society, it is critical to keep ethics agile and alive. AI ethics is in a deep and vital need of the continuously renewed process of questioning the world and the lenses through which we see it – consciously, constantly and iteratively.
Footnotes
Declaration of conflicting interests
The author(s) declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: This article reflects only the views of the authors and does not intend to reflect those of the European Commission. The European Commission is not responsible for any use that may be made of the information it contains.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This article was developed as part of the SIENNA project (Stakeholder-informed ethics for new technologies with high socio-economic and human rights impact) which has received funding under the European Union’s H2020 research and innovation programme under grant agreement No 741716.
