Tag: self driving cars
Will Your Driverless Car Kill You So Others May Live?

Will Your Driverless Car Kill You So Others May Live?

By Eric Schwitzgebel, Los Angeles Times (TNS)

It’s 2025. You and your daughter are riding in a driverless car along Pacific Coast Highway. The autonomous vehicle rounds a corner and detects a crosswalk full of children. It brakes, but your lane is unexpectedly full of sand from a recent rock slide. It can’t get traction. Your car does some calculations: If it continues braking, there’s a 90 percent chance that it will kill at least three children. Should it save them by steering you and your daughter off the cliff?

This isn’t an idle thought experiment. Driverless cars will be programmed to avoid collisions with pedestrians and other vehicles. They will also be programmed to protect the safety of their passengers. What happens in an emergency when these two aims come into conflict?

The California Department of Motor Vehicles is now trying to draw up safety regulations for autonomous vehicles. These regulations might or might not specify when it is acceptable for collision-avoidance programs to expose passengers to risk to avoid harming others — for example, by crossing the double-yellow line or attempting an uncertain maneuver on ice.

Google, which operates most of the driverless cars being street-tested in California, prefers that the DMV not insist on specific functional safety standards. Instead, Google proposes that manufacturers “self-certify” the safety of their vehicles, with substantial freedom to develop collision-avoidance algorithms as they see fit.

That’s far too much responsibility for private companies. Because determining how a car will steer in a risky situation is a moral decision, programming the collision-avoiding software of an autonomous vehicle is an act of applied ethics. We should bring the programming choices into the open, for passengers and the public to see and assess.

Regulatory agencies will need to set some boundaries. For example, some rules should presumably be excluded as too selfish. Consider the over-simple rule of protecting the car’s occupants at all costs. This would imply that if the car calculates that the only way to avoid killing a pedestrian would involve sideswiping a parked truck, with a 5 percent chance of injury to the car’s passengers, then the car should instead kill the pedestrian.

Other possible rules might be too sacrificial of the passengers. The equally over-simple rule of maximizing lives saved without any special regard for the car’s occupants would unfairly disregard personal accountability. What if other drivers — human drivers — have knowingly put themselves in danger? Should your autonomous vehicle risk your safety, perhaps even your life, because a reckless motorcyclist chose to speed around a sharp curve?

A Mountain View lab must not be allowed to resolve these difficult questions on our behalf.

That said, a good regulatory framework ought to allow some manufacturer variation and consumer choice, within ethical limits. Manufacturers or fleet operators could offer passengers a range of options. “When your child is in the car, our onboard systems will detect it and prioritize the protection of rear-seat passengers!” Cars might have aggressive modes (maximum allowable speed and aggressiveness), safety modes, ethical utilitarian modes (perhaps visibly advertised so that others can admire your benevolence) and so forth.

Some consumer freedom seems ethically desirable. To require that all vehicles at all times employ the same set of collision-avoidance procedures would needlessly deprive people of the opportunity to choose algorithms that reflect their values. Some people might wish to prioritize the safety of their children over themselves. Others might want to prioritize all passengers equally. Some people might wish to choose algorithms more self-sacrificial on behalf of strangers than the government could legitimately require of its citizens.

There will also always be trade-offs between speed and safety, and different passengers might legitimately weigh them differently, as we now do in our manual driving choices.

Furthermore, although we might expect computers to have faster reaction times than people, our best computer programs still lag far behind normal human vision at detecting objects in novel, cluttered environments. Suppose your car happens upon a woman pushing a rack of coats in a windy swirl of leaves. Vehicle owners may insist on some sort of preemptive override, some way of telling their car not to employ its usual algorithm, lest it sacrifice them for a mirage.

There is something romantic about the hand upon the wheel — about the responsibility it implies. But future generations might be amazed that we allowed music-blasting 16-year-olds to pilot vehicles unsupervised at 65 mph, with a flick of the steering wheel the difference between life and death. A well-designed machine will probably do better in the long run.

That machine will never drive drunk, never look away from the road to change the radio station or yell at the kids in the back seat. It will, however, have power over life and death. We need to decide — publicly — how it will exert that power.

ABOUT THE WRITER

Eric Schwitzgebel is a professor of philosophy at UC Riverside and the author of “Perplexities of Consciousness.” He blogs at the Splintered Mind. He wrote this for the Los Angeles Times.

©2015 Los Angeles Times. Distributed by Tribune Content Agency, LLC.

Photo: Ali Eminov via Flickr

Are Insurers Prepared For Self-Driving Cars?

Are Insurers Prepared For Self-Driving Cars?

By Greg Gardner, Detroit Free Press (TNS)

DETROIT — Auto insurers are unprepared for the changes self-driving vehicles will create for their business and most don’t think the technology will have a significant impact for at least a decade, according to a study released Wednesday by consulting firm KPMG.

The survey asked executives of insurance companies with about $85 billion of auto insurance premiums a series of questions designed to measure their knowledge and opinions of autonomous transportation.

Among the findings: more than four out of five (84 percent) insurance companies don’t think driverless vehicles will have a significant impact on their business until 2025.

Google is testing a fleet of 100 autonomous cars around its northern California headquarters. Audi and Delphi Automotive have teamed up on a self-driving car that completed a cross-country trip earlier this year. Tesla Motors is rolling out a software update that will enable its Model S to drive autonomously from San Francisco to Seattle.

Cadillac is planning to introduce an advanced system it calls Super Cruise on a 2017 model that will take over the steering, acceleration and braking in highway driving. Toyota plans to offer crash-avoidance technology in all Toyota and Lexus models by 2017.

“The disruption of autonomous vehicles to the automotive ecosystem will be profound, and the change will happen faster than most in the insurance industry think,” said Jerry Albright, a leader of KPMG’s actuarial and insurance risk practice. “To remain relevant in the future, insurers must evaluate their exposure and make necessary adjustments to their business models, corporate strategy, and operations.”

How quickly consumers will embrace self-driving vehicles and how much they will cost remain unclear for the near future. But most of the technology that enables autonomy was created to improve safety, such as lane-departure alerts, cruise control and collision-avoidance sensors.

Most of the insurance executives did acknowledge they need to learn more about the technology and its impact on insurance underwriting. For example, one possible channel for self-driving cars could be through taxi-like fleets where consumers purchase the service by number of trips or by miles.

So individual consumers may be driving less in their own vehicles, but more in shared transportation networks.

Advocates of self-driving cars say safety will be their primary selling point as the software and sensors that control them are refined.

“The share of the personal auto insurance sector will likely continue to shrink as the potential liability of the software developer and manufacturer increases,” said Alex Bell, managing director of KPMG’s CIO advisory group. “At the same time, losses covered by product liability policies are likely to increase given that the sophisticated technology that underpins autonomous vehicles will also need to be insured.”

Earlier this year, the Insurance Information Institute released a study that acknowledged the likely impact of autonomous vehicles.

“There will still be a need for liability coverage, but over time the coverage could change, as manufacturers and suppliers and possibly even municipalities are called upon to take responsibility for what went wrong,” the Insurance Information Institute study stated.

During the transition to wholly autonomous driving, insurers may rely more on telematics devices, known as “black boxes,” that monitor driver activity. Some drivers may object to them based on concerns about privacy.

The institute also warned that autonomous vehicle may reduce the number of accidents, but the cost of replacing damaged parts “is likely to increase because of the complexity of the components.”

But KPMG says the combination of fewer cars on the road and consumers driving less in their own vehicles will raise expectations of lower premiums.

“Assuming consumers demand lower premiums to reflect fewer accidents, there is the possibility of frenzied competition,” said Joe Schneider, director of KPMG corporate finance practice.

Photo: Google is testing Tesla’s Model S as an autonomous vehicle. Maurizio Pesce via Flickr