To send content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about sending content to .
To send content items to your Kindle, first ensure email@example.com
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about sending to your Kindle.
Note you can select to send to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be sent to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
1. Artificial intelligence is becoming increasingly important in our daily professional and social lives. Although the use of AI systems has many benefits for a variety of sectors, different legal challenges remain. Some of these challenges are extensively discussed in this book. In this chapter, we will focus on the application of liability for damage caused by AI systems. The importance of liability and AI systems has already been mentioned in several recent documents issued by the European Union (EU). The White Paper on Artificial Intelligence, for instance, stresses that the main risks related to the use of AI concern the application of rules designed to protect fundamental rights as well as safety and liability-related issues. Scholars have also concluded that ‘[l]iability certainly represents one of the most relevant and recurring themes’ when it comes to AI systems.
2. This emphasis on liability is not surprising considering that AI systems will increasingly cause damage. Reference can be made to recent accidents involving autonomous vehicles. The autopilot of a Tesla car, for instance, was not able to distinguish a white tractor-trailer crossing the road from the bright sky above, leading to a fatal crash. A self-driving Uber car recently hit a pedestrian in Arizona. The woman later died in the hospital. A robot also attacked and injured a man at a tech fair in China. A surgical robot at a hospital in Philadelphia malfunctioned during a prostate surgery, thereby severely injuring the patient. In February 2015, a South Korean woman was sleeping on the floor when her robot vacuum ‘ate’ her hair forcing her to call for emergency help.
These examples show that accidents may happen despite optimising national and supranational safety rules for AI. This is when questions of liability become important. Nevertheless, the application of liability regimes for damage caused by AI systems can be challenging. The characteristics of AI systems such as opaqueness, autonomy, connectivity, data dependency or self-learning abilities make it difficult to trace back potentially problematic decisions made with the involvement of such systems.
Email your librarian or administrator to recommend adding this to your organisation's collection.