Self-Driving Cars: Travelling Towards the Law
They’re not bothered by everyday problems and thoughts; they don’t get hungry or develop headaches. Self-driving cars don’t drink alcohol or drive under the influence of drugs. In short, human error, the number one cause of road traffic accidents, could be made a thing of the past in one fell swoop if manual driving was banned immediately. Is that right? It would be, if there hadn’t recently been reports of two deaths, one during the test drive for a self-driving car (Uber) and one while a semi-autonomous vehicle was driving on a motorway and using its lane assist system (Tesla), both of which regrettably occurred in the USA in March 2018. In Tesla’s case it seems that the semi-autonomous driving assistant was switched off at the moment of the accident.
Main cause: human error
Around the globe, people die every day due to careless driving, with around 90% of all accidents caused by human error, and just a small percentage due to a technical fault related to the vehicle. Despite human error, we have not banned driving on these grounds. Two accidents with fatal consequences involving autonomous vehicles being test-driven have attracted the full glare of the media spotlight, and call into question the technical development of a rapidly progressing industry. Are self-driving cars now just hype, or is this a trend that cannot be contained, despite every additional human life that is lost as a result of mistakes made by self-driving technology?
For many, the thought that fully autonomous vehicles (a self-driving car without a driver) might exist in the future is rather unsettling. The two recent deaths in the USA resulting from (semi-) autonomous cars may well have created fear of further accidents. From a legal perspective, it makes no difference whatsoever for the injured party whether the accident was caused by a careless human or by technology that was functioning inadequately. The reason for the line drawn between the two, despite this fact, is probably that while every human error represents a separate accident, the failure or malfunction of technology cannot be seen as a one-off: rather, understandably and probably correctly, it is viewed as a system error or series error caused by a certain technology available at a particular point in time.
Still little confidence in the self-driving car
From a legal angle, a technical defect generally also represents a design defect that affects the entire run of a particular vehicle range. Deaths caused by software malfunctions cause people to quickly lose trust in other vehicles equipped with the same faulty software. Conversely, if a drunk driver injures or kills another road user, it is not assumed that the majority of other drivers (or all of them) could potentially cause accidents due to the influence of alcohol.
The fundamental question for all the technological developments in this area is this: do people want self-driving cars?
When we talk of self-driving (or autonomous) vehicles, we mean machines guided by computers. On-board computers are common practice in aviation, with pilots not actually flying the planes themselves, and from a statistical point of view, airplanes are the safest mode of transport. Couldn’t cars become just as safe? However, a comparison between planes and cars cannot be justified, due to the different user groups, the number of cars driven every day, and the ever-present risk of a collision with other road users, including pedestrians.
While driver assistance systems, such as lane assist, park assist or adaptive cruise control, can be found in many widespread models and are principally permitted and allowed in Europe, current legislation in Europe and also Austria only permits (semi-) autonomous vehicles to be used for test purposes. Additionally, in Austria these test drives can, inter alia, only take place on motorways or with minibuses in an urban environment following specially marked routes (cf. the test drives with minibuses in the towns of Salzburg and Velden). Test drives have been carried out on Austria’s roads in line with particular legal requirements for a little more than a year, and it has been necessary to have a person in the vehicle at all times who must be able to intervene immediately if an accident is on the horizon, correct wrong steering by the computer or get the vehicle back under (human) control.
Indeed, under the legislation in those US states that permit test drives, there still (currently) have to be people inside the car (before the two accidents mentioned above, California had announced a law that would have made it no longer necessary to have a person in the vehicle). As a result, three questions arise regarding the Uber accident which occurred during a test drive in the US state of Arizona, resulting in a fatal collision with a cyclist:
- Could the person who was inside the vehicle to control it for safety reasons have activated the emergency brake, and thus averted the collision with the cyclist who suddenly crossed the road?
- Why did the sensors built into the car not recognize the cyclist in time?
- Why did the vehicle not stick to the legal speed limit?
Driving systems are currently being tested in Europe and the USA. In the USA, this can take place on national roads and, contrary to European legislation, also on urban streets. As long as we are still in the test phase we cannot talk of technically proven, let alone officially approved, driving systems. The technical development of self-driving cars, however, has already made it clear that legal responsibility is shifting away from the driver and towards the vehicle manufacturers and software developers.
Humanoid robot on four wheels
Whether, and when, self-driving cars could become an everyday phenomenon is greatly dependent on certain (future) questions: are we right to expect absolute safety from self-driving cars? What decisions should self-driving cars make in the event that one life can only be saved at the cost of another, and how should this dilemma be resolved?
If artificial intelligence (AI) and self-learning systems could also be included within the technology for self-driving cars, vehicles of this type might possibly one day become “humanoid robots on four wheels”, but they could not be compared to a human being with particular notions of value and morality. If every individual personally bears responsibility for their intuitive behavior in a specific accident situation, the limits of our legal system will be laid bare if algorithms using huge quantities of data make decisions in advance for a subsequent accident situation: these decisions can no longer be wholly ascribed to a particular person or software developer if a self-driving car is involved. It will be our task as lawyers to offer legal support to legislators as they attempt to meet these challenges.
About Dr. Andreas Eustacchio
Dr. Andreas Eustacchio LL.M. (London LSE), Attorney-at-Law in Austria; born in Zambia; Partner at EUSTACCHIO Attorneys-at-law, a Vienna-based law firm operating internationally with a focus on commercial law; he advises both Austrian and foreign companies in the field of digitalized and automated industry; since 2016 Associate Legal Partner of the ‘Virtual Vehicle’ Research Centre for Autonomous Driving in Graz; author and lecturer at several Austrian universities and visiting professor at the universities of Hanoi (Vietnam) and Sanya (China); Cavaliere (Order of Merit of the Italian Republic).