Driverless Cars at the Moral Crossroads

Sunday, December 8, 2019
First Aired:
Sunday, July 30, 2017

What Is It

Autonomous vehicles are quickly emerging as the next innovation that will change society in radical ways. Champions of this new technology say that driverless cars, which are programed to obey the law and avoid collisions, will be safer than human controlled vehicles. But how do we program these vehicles to act ethically? Should we trust computer programmers to determine the most ethical response to all possible scenarios the vehicle might encounter? And who should be held responsible for the bad − potentially lethaldecisions these cars make? Our hosts take the wheel with Harvard psychologist Joshua Greene, author of "Our Driverless Dilemma: When Should Your Car be Willing to Kill You?"

Recorded live at Cubberley Auditorium on the Stanford campus with support from theSymbolic Systems Programand theMcCoy Center for Ethics in Society.

Listening Notes

Live from Cubberley Auditorium at Stanford University, Ken and Laura Maguire, Philosophy Talk director of research, discuss a familiar topic: bad drivers. Between drinking, texting, and standard human error, driving is one of the most dangerous responsibilities that humans are entrusted with every day. But could the dawn of driverless cars controlled by computer algorithms change everything? Sure, computers may be safer drivers than humans on average, but can they care about human life the same way that people can?

Harvard psychology professor Joshua Greene joins Ken and Laura to discuss the moral dilemmas that come with the advent of driverless cars. Josh admits that it is difficult for people to accept handing their capacity for decision-making over to a computer but explains that computerized driving will ultimately lead to a much safer world. Still, there is justified caution about “mechanized morality” – can we trust computers to make morally fraught decisions? Josh explains that from the perspective of neuroscience, moral decisions are just like any others, meaning that they can be programmed into computer algorithms as easily as commanding the computer to turn left or right.

在最后一个环节,肯,劳拉和乔什回答了观众关于道德机械化的问题。一位律师指出,从保险到产品设计等各个领域的道德问题每天都必须被量化。一名学生指出,自动驾驶汽车可能会偏向自己的乘客,导致不同社会经济阶层的乘客之间不一致。其他听众关注的是自动驾驶汽车的特定问题,但Josh强调,除了挑剔的例外,任何一种无人驾驶技术都需要某种量化的、程序化的道德系统。现在的挑战只是决定该如何建立这个系统。

  • Roving Philosophical Report (Seek to 8:13):Liza Veale visits an autonomous driving research lab at Stanford University to see how they are dealing with the technological and ethical challenges that accompany self-driving cars. While it might be 50 years before all cars on the road are autonomous, cars already are becoming far more automated.
  • Sixty-Second Philosopher (Seek to 45:55):伊恩·肖尔斯质疑人们是否应该想要无人驾驶汽车。他指出,无人驾驶汽车的广泛应用可能会在交通、残疾、公共交通和司机就业市场等方面引发各种不可预见的后果。

Transcript

Comments(2)


Gerald Fnord's picture

Gerald Fnord

Sunday, July 30, 2017 -- 10:57 AM

Driverless cars: my will be done

As a moral being and, significantly, one who privileges ratiocination over passion and reflex, I should actively _prefer_ that a car with much faster reaction times and otherwise capable of driving much better than I can would implement my moral decisions than I. I would otherwise risk momentary weakness of the body or spirit's interference with the judgements I would make were I _not_ about to crash…and, preferably, when I'm not in a misanthropic mood.

Harold G. Neuman's picture

Harold G. Neuman

Thursday, December 5, 2019 -- 10:12 AM

In 2017 I was not too worried

2017年,我还不太担心无人驾驶汽车。现在想想,再看看其他的交通事故(人们忙着发短信或打电话,无暇顾及自己的安全;摩托车,本来就不该在路上行驶;骑自行车不遵守交通法规的;等等),我意识到我们正在接近交通熵:在这个饱和点上,在安全操作任何车辆(无论是无人驾驶还是有人驾驶)的同时,跟踪不断增加的危险几乎是不可能的。我正在考虑完全不开车了——尽管至少会很不方便。