A recent decision by the National Transportation Safety Board (NTSB) on the March 2018 Uber crash that killed a pedestrian in Arizona split the blame among Uber, the company’s autonomous vehicle (AV), the safety driver in the vehicle, the victim, and the state of Arizona. With the advent of self-driving cars, the NTSB findings raise a number of questions about the uncertainty in today’s legal liability system. In an accident involving an AV and a human driver, who is liable? If both are liable, how should the accident loss be apportioned between them?
AVs remove people from the hands-on task of driving and thus pose a complex challenge to today’s accident tort law, which primarily punishes humans. Legal experts anticipate that, by programming driving algorithms, self-driving car manufacturers, including car designers, sensor vendors, software developers, car producers, and related parties who contribute to the design, manufacturing, and testing, will have a direct influence on traffic. While these algorithms make manufacturers indispensable actors, with their product’s liability potentially playing a critical role, policy makers have not yet devised a quantitative method to assign the loss between the self-driving car and the human driver.
To tackle this problem, researchers at Columbia Engineering and Columbia Law School have developed a joint fault-based liability rule that can be used to regulate both self-driving car manufacturers and human drivers. They propose a game-theoretic model that describes the strategic interactions among the law maker, the self-driving car manufacturer, the self-driving car, and human drivers, and examine how, as the market penetration of AVs increases, the liability rule should evolve.
Their findings are outlined in a new study to be presented on January 14 by Sharon Di, assistant professor of civil engineering and engineering mechanics, and Eric Talley, Isidor and Seville Sulzbacher Professor of Law, at the Transportation Research Board’s 99th Annual Meeting in Washington, D.C.
While most current studies have focused on designing AVs’ driving algorithms in various scenarios to ensure traffic efficiency and safety, they have not explored human drivers’ behavioral adaptation to AVs. Di and Talley wondered about the “moral hazard” effect on humans, whether with exposure to more and more traffic encounters with AVs, people might be less inclined to exercise “due care” when faced with AVs on the road and drive in a more risky fashion.
“Human drivers perceive AVs as intelligent agents with the ability to adapt to more aggressive and potentially dangerous human driving behavior,” says Di, who is a member of Columbia’s Data Science Institute. “We found that human drivers may take advantage of this technology by driving carelessly and taking more risks, because they know that self-driving cars would be designed to drive more conservatively.”
“The tragic fatality in Arizona involving a self-driving automobile elicited tremendous attention from the public and policy makers about how to draw the lines of legal liability when AVs interact with human drivers, cyclists, and pedestrians,” Talley adds. “The emergence of AVs introduces a particularly thorny type of uncertainty into the status quo, and one that feeds back onto AV manufacturing and design. Legal liability for accidents between automobiles and pedestrians typically involves a complex calculus of comparative fault assessments for each of the aforementioned groups. The introduction of an autonomous vehicle can complicate matters further by adding other parties to the mix, such as the manufacturers of hardware and programmers of software. And insurance coverage distorts matters further by including third party stakeholders. We hope our analytical tools will assist AV policy-makers with their regulatory decisions, and in doing so, will help mitigate uncertainty in the existing regulatory environment around AV technologies.”
Di and Talley are now looking at multiple AV manufacturers that target different global markets with different technological specifications, making the development of legal rules even more complex.
“We know that human drivers will take more risks and develop moral hazard if they think their road environment has become safer,” Di notes. “It’s clear that an optimal liability rule design is crucial to improve social welfare and road safety with advanced transportation technologies.”