To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Many types of emergence exist. This chapter will discuss some of the most prominent, and broadly occurring, examples of emergent structure in space and time.
The law scholar Dustin Lewis explores the requirements of international law with regard to the employments of AI-related tools and techniques in armed conflict. The scope of this chapter is not limited to Lethal Autonomous Weapons (AWS) but also encompasses other AI-related tools and techniques related to warfighting, detention, and humanitarian services. After providing an overview of international law applicable to armed conflict, the author outlines some preconditions necessary to respect international law. According to Lewis, current international law essentially presupposes humans – and not artificial, non-humans – as legal agents. From that premise, the author argues that any employment of AI-related tools or techniques in an armed conflict needs to be susceptible to being administered, discerned, attributed, understood, and assessed by human agents.
In the past decade, Artificial Intelligence (AI) as a general-purpose tool has become a disruptive force globally. By leveraging the power of artificial neural networks, deep learning frameworks can now translate text from hundreds of languages, enable real-time navigation for everyone, recognise pathological medical images, as well as enable many other applications across all sectors in society. However, the enormous potential for innovation and technological advances and the chances that AI systems provide come with hazards and risks that are not yet fully explored, let alone fully understood. One can stress the opportunities of AI systems to improve healthcare, especially in times of a pandemic, provide automated mobility, support the protection of the environment, protect our security, and otherwise support human welfare. Nevertheless, we must not neglect that AI systems can pose risks to individuals and societies; for example by disseminating biases, by undermining political deliberation, or by the development of autonomous weapons. This means that there is an urgent need for responsible governance of AI systems. This Handbook shall be a basis to spell out in more detail what could become relevant features of Responsible AI and how we can achieve and implement them at the regional, national, and international level. Hence, the aim of this Handbook is to address some of the most pressing philosophical, ethical, legal, and societal challenges posed by AI.
In this chapter, the philosophers Oliver Mueller and Boris Essmann address AI-supported neurotechnology, especially Brain–Computer Interfaces (BCIs) that may in the future supplement and restore functioning in agency-limited individuals or even augment or enhance capacities for natural agency. The authors propose a normative framework for the evaluation of neurotechnological and AI-assisted agency based on ‘cyberbilities’. These are capabilities that emerge from human–machine interactions in which agency is distributed across human and artificial elements. The authors conclude by providing a list of cyberbilities that is meant to support the well-being of individuals.
The chapter by the philosopher Catrin Misselhorn provides an overview of the most central debates in artificial morality and machine ethics. Artificial moral agents are AI systems which are able to recognise the morally relevant aspects of a situation and take them into account in their decisions and actions. Misselhorn shows that artificial morality is not just a matter of Science Fiction scenarios but rather an issue that has to be considered today. She lays the conceptual foundations of artificial morality and discusses the ethical issues that arise. She addresses questions like: which morality should be part of an AI system? Can AI systems be aligned with human morality, or do they need a machine-specific morality? Are there decisions, which should never be transferred to machines? Could artificial morality have impacts on human morality if it becomes more pervasive? These and other questions relating to AI are discussed and answered.
In this chapter, the law scholar Ralf Poscher sets out to show how AI challenges the traditional understanding of the right to data protection and presents an outline of an alternative conception that better deals with emerging AI technologies. Firstly, Poscher explains how the traditional conceptualisation of data protection as an independent fundamental right on its own collides with AI’s technological development, given that AI systems do not provide the kind of transparency required by the traditional approach. Secondly, the author proposes an alternative model, a no-right thesis, which shifts the focus from data protection as an independent right to other existing fundamental rights, such as liberty and equality. He argues that this allows us to step back from the idea that each and every instance of personal data processing concerns a fundamental right. Instead, it is important to assess how an AI system ‘behaves’, what type of risks it generates, and which substantive fundamental rights are being affected.
In this chapter, the law scholar Jan von Hein analyses and evaluates the European Parliament’s proposal on a civil liability regime for artificial intelligence against the background of the already existing European regulatory framework on private international law, in particular the Rome I and II Regulations. The draft regulation (DR) proposed by the European Parliament is noteworthy from a private international law perspective because it introduces new conflicts rules for AI. In this regard, the proposed regulation distinguishes between a rule delineating the spatial scope of its autonomous rules on strict liability for high-risk AI systems (Article 2 DR) on the one hand, and a rule on the law applicable to fault-based liability for low-risk systems (Article 9 DR) on the other hand. The latter rule refers to the domestic laws of the Member State in which the harm or damage occurred. In sum, compared with Rome II, the conflicts approach of the draft regulation would be a regrettable step backwards in many ways.
The law scholars Weixing Shen and Yun Liu focus on China’s efforts in the field of AI regulation and spell out recent legislative actions. While there is no unified AI law today in China, many provisions from Chinese data protection law are in part applicable to AI systems. The authors particularly analyse the rights and obligations from the Chinese Data Security Law, the Chinese Civil Code, the E-Commerce Law, and the Personal Information Protection Law and explain the relevance of these regulations with regard to responsible AI and algorithm governance. The authors introduce as well the Draft Regulation in Internet Information Service Based on Algorithm Recommendation Technology. This adopts many AI specific principles such as transparency, fairness, and reasonableness. Regarding the widely discussed field of facial recognition by AI systems, they introduce a Draft Regulation, and a judicial Opinion by the Supreme People’s Court of China. Finally, Weixing Shen and Yun Liu refer to the AI Act proposed by the European Commission, which could also inspire future Chinese regulatory approaches.
The process of entities successively splitting into two or more is of great relevance: biological reproduction, infection spreading, rumour spreading, nuclear reactions and much more.
In this chapter, the law scholar Christine Wendehorst analyses the different potential risks posed by AI as part of two main categories, safety risks and fundamental rights risks. Based on this, the author considers why AI challenges existing liability regimes. She spells out the main solutions put forward so far and evaluates them. This chapter highlights the fact that liability for fundamental rights risks is largely unchartered while being AI-specific. Such risks are now being addressed at the level of AI safety law, by way of prohibiting certain AI practices and by imposing strict legal requirements concerning data governance, transparency, and human oversight. Wendehorst nevertheless argues that a number of changes have to be made for the emerging AI safety regime to be used as a ‘backbone’ for the future AI liability regime if this is going to help address liability for fundamental rights risks. As a result, she suggests that further negotiations about the AI Act proposed by the European Commission should be closely aligned with the preparatory work on a future AI liability regime.
Emergent phenomena require some interdependence between components. Networks of nodes and connecting links are therefore a very natural and powerful language for the analysis and characterisation of complex systems.
In this chapter, the philosopher Thomas Metzinger lists five main problem domains related to AI systems. For each problem field, he proposes several measures which should be taken. Firstly, there should be worldwide safety standards concerning the research and development of AI. If not, Metzinger fears a ‘race to the bottom’ in safety standards. Additionally, a possible AI arms race must be prevented as early as possible. Thirdly, he stresses that any creation of artificial consciousness should be avoided, as it is highly problematic from an ethical point of view. He argues that synthetic phenomenology could lead to non-biological forms of suffering and might lead to a vast increase of suffering in the universe, as AI can be copied rapidly. While AI might improve different kinds of governance, there is the risk of unknown risks, the ‘unknown unknowns’. Accordingly, as a fourth problem domain, the author proposes allocating resources to research and prepare for unexpected and long-term risks. Finally, Metzinger highlights the need for a concrete code of ethical conduct for anyone researching AI.