By MARSHA MERCER
I learned as a rookie newspaper reporter that there are no accidents in life – none involving cars, trucks or motorcycles, anyway.
“Accident” implies an event that’s beyond our control. Because reporters cannot know what’s in a driver’s mind, we can’t say something is an accident.
Newspapers that shunned the A word reported on automotive mishaps, incidents and collisions. Blow-dried TV announcers and traffic reporters shouting over helicopter noise favored the more vivid noun crash.
Copy editors weren’t the only sticklers for the right word to describe highway misfortune. Years ago, the National Highway Traffic Safety Administration tried to expunge the concept of traffic accidents.
By eliminating the word in speeches, news releases and publications, it hoped in 1997 to re-educate people that car wrecks weren’t acts of God but predictable and preventable events.
The agency also launched a “Crashes are not accidents” campaign. From time to time, other groups have taken up the cause, but people hang onto the idea of tragic highway accidents. Now, though, while we can’t stop the lethal combination of speed and distance, we at least buckle our seat belts.
Technology is making cars safer. Backup cameras help you park and avoid pedestrians. Lasers or radar in adaptive cruise control keep your car a safe distance from the one ahead and warn if you veer out of your lane.
But what happens when technology – and not a human -- controls the car?
This summer, Google will test the prototype of its fully self-driving car on city streets in Mountain View, Calif. The cautious two-seater won’t exceed the grand speed of 25 mph. Other companies working on driverless cars include General Motors, Tesla, Mercedes-Benz and the Chinese Web company Baidu.
States are eager to get on board. Michigan, Florida, Nevada and Washington, D.C., have passed laws allowing automakers to test driverless cars. Virginia Gov. Terry McAuliffe in March proclaimed the commonwealth “open for business” for the testing and deployment of autonomous vehicles.
In Northern Virginia, 70 miles of busy highways – parts of Interstates 66, 95 and 495 and U.S. Routes 29 and 50 -- could be used to test self-driving cars within a year, the Richmond Times-Dispatch reported June 1.
The potential for safer highway travel has instant appeal. Autonomous vehicles don’t drive drunk or send texts or take selfies behind the wheel. But they do have to deal with humans in other cars.
Since Google began testing self-driving cars in 2009, the vehicles have logged nearly 2 million miles with a dozen minor accidents – but “Not once was the self-driving car the cause of the accident,” the company announced. Other motorists rear-ended or sideswiped the autonomous car, or the car was in manual mode and its human driver at fault.
The cars don’t have to be perfect; they just have to beat humans at driving, developers say. But that seems a low bar. Humans are terrible drivers.
In 2010, there were 33,000 traffic deaths, nearly 4 million injuries and 24 million vehicles damaged in motor vehicle crashes in the United States, the highway traffic safety administration reported last month. It still avoids the word accident.
Meanwhile, states are struggling with new rules of the road. California allows self-driving cars on public roads but requires a steering wheel, brake pedal and accelerator – and a trained safety driver who can take over in case there’s a problem.
That may lead to a false sense of security. Google in the fall of 2012 started letting its employees take home self-driving Lexus SUVs. The employees promised to be ever-alert and take the wheel as necessary. It didn’t work out.
“We discovered something we really hadn’t seen coming but was obvious in retrospect,” Astro Teller, the improbably named director of Google’s Project X, the division that explores innovative products, said last month in a South by Southwest talk.
“Once people trust the system, they trust it. Our success was itself a failure,” he said. “The assumption that people can be a reliable backup for the system was a total fallacy.”
The only way a driverless car will work, Google decided, is to make it totally driverless. No steering wheel, no brake pedal, no accelerator. The car must drive itself from Point A to Point B at the push of a button.
“That has been a lot more work than we thought, but it’s the right thing to do,” Teller said.
Google’s decision to adjust the technology to fit the human – rather than expecting the human to adjust to the technology -- is a good lesson for us all. Experiment, yes, and take the time for safety. Crashes are not accidents.
©2015 Marsha Mercer. All rights reserved.