Self-Driving Cars may Soon Be Able to Make Moral and Ethical Decisions

Can a self-driving vehicle be moral, act like humans do, or act like humans expect humans to? Contrary to previous thinking, a ground-breaking new study has found for the first time that human morality can be modelled meaning that machine based moral decisions are, in principle, possible. The research, Virtual Reality experiments investigating human behavior and moral assessments, from The Institute of Cognitive Science at the University of Osnabrück, and published in Frontiers in Behavioral Neuroscience, used immersive virtual reality to allow the authors to study human behavior in simulated road traffic scenarios.

The participants were asked to drive a car in a typical suburban neighborhood on a foggy day when they experienced unexpected unavoidable dilemma situations with inanimate objects, animals, and humans and had to decide which was to be spared. The results were conceptualized by statistical models leading to rules, with an associated degree of explanatory power to explain the observed behavior. The research showed that moral decisions in the con?ned scope of unavoidable traffic collisions can be explained well, and modeled, by a single value-of-life for every human, animal, or inanimate object.

Leon Sütfeld, first author of the study, says that until now it has been assumed that moral decisions are strongly context dependent and therefore cannot be modeled or described algorithmically, "But we found quite the opposite. Human behavior in dilemma situations can be modeled by a rather simple value-of-life-based model that is attributed by the participant to every human, animal, or inanimate object." This implies that human moral behavior can be well described by algorithms that could be used by machines as well.

The study's findings have major implications in the debate around the behavior of self-driving cars and other machines, like in unavoidable situations. For example, a leading new initiative from the German Federal Ministry of Transport and Digital Infrastructure (BMVI) has defined 20 ethical principles related to self-driving vehicles, for example, in relation to behavior in the case of unavoidable accidents, making the critical assumption that human moral behavior could not be modeled.

Prof. Gordon Pipa, a senior author of the study, says that since it now seems to be possible that machines can be programmed to make human like moral decisions it is crucial that society engages in an urgent and serious debate, "we need to ask whether autonomous systems should adopt moral judgements, if yes, should they imitate moral behavior by imitating human decisions, should they behave along ethical theories and if so, which ones and critically, if things go wrong who or what is at fault?"

As an example, within the new German ethical principles, a child running onto the road would be classified as significantly involved in creating the risk, thus less qualified to be saved in comparison to an adult standing on the footpath as a non-involved party. But is this a moral value held by most people and how large is the scope for interpretation?

"Now that we know how to implement human ethical decisions into machines we, as a society, are still left with a double dilemma," explains Prof. Peter König, a senior author of the paper. "Firstly, we have to decide whether moral values should be included in guidelines for machine behavior and secondly, if they are, should machines should act just like humans."

The study's authors say that autonomous cars are just the beginning as robots in hospitals and other artificial intelligence systems become more common place. They warn that we are now at the beginning of a new epoch with the need for clear rules otherwise machines will start marking decisions without us.

Leon R. Sütfeld, Richard Gast, Peter König, Gordon Pipa.
Using Virtual Reality to Assess Ethical Decisions in Road Traffic Scenarios: Applicability of Value-of-Life-Based Models and Influences of Time Pressure.
Front. Behav. Neurosci, 11:122, doi: 10.3389/fnbeh.2017.00122.

Most Popular Now

AI Tools Help Predict Severe Asthma Risk…

Mayo Clinic researchers have developed artificial intelligence (AI) tools that help identify which children with asthma face the highest risk of serious asthma exacerbation and acute respiratory infections. The study...

ChatGPT 4o Therapeutic Chatbot 'Ama…

One of the first randomized controlled trials assessing the effectiveness of a large language model (LLM) chatbot 'Amanda' for relationship support shows that a single session of chatbot therapy...

AI Distinguishes Glioblastoma from Look-…

A Harvard Medical School–led research team has developed an AI tool that can reliably tell apart two look-alike cancers found in the brain but with different origins, behaviors, and treatments. The...

AI Model Forecasts Disease Risk Decades …

Imagine a future where your medical history could help predict what health conditions you might face in the next two decades. Researchers have developed a generative AI model that uses...

Overcoming the AI Applicability Crisis a…

Opinion Article by Harry Lykostratis, Chief Executive, Open Medical. The government’s 10 Year Health Plan makes a lot of the potential of AI-software to support clinical decision making, improve productivity, and...

Smart Device Uses AI and Bioelectronics …

As a wound heals, it goes through several stages: clotting to stop bleeding, immune system response, scabbing, and scarring. A wearable device called "a-Heal," designed by engineers at the University...

Dartford and Gravesham Implements Clinis…

Dartford and Gravesham NHS Trust has taken a significant step towards a more digital future by rolling out electronic test ordering using Clinisys ICE. The trust deployed the order communications...