‘It happened so fast’: Inside a fatal Tesla Autopilot crash

Gaps in Tesla’s driver-assistance systems and distractions can have tragic consequences


George Brian McGee, a finance executive in Florida, was driving home in a Tesla Model S operating on Autopilot, a system that can steer, brake and accelerate a car on its own, when he dropped his phone during a call and bent down to look for it. Neither he nor Autopilot noticed that the road was ending. The Model S drove past a stop sign and a flashing red light. The car smashed into a parked Chevrolet Tahoe, killing a 22-year-old college student, Naibel Benavides.

One of a growing number of fatal crashes involving Tesla cars operating on Autopilot, McGee’s case is unusual because he survived and told investigators what had happened: he got distracted and put his trust in a system that did not see and brake for a parked car in front of it. American Tesla drivers using Autopilot in other fatal crashes have often been killed, leaving investigators to piece together the details from data stored and videos recorded by the cars.

“I was driving and dropped my phone,” McGee told an officer who responded to the crash, according to a recording from a police body camera. “I looked down, and I ran the stop sign and hit the guy’s car.”

Distracted driving can be deadly in any car. But safety experts say Tesla's Autopilot may encourage distraction by lulling people into thinking that their cars are more capable than they are

McGee’s statements to investigators, the crash report and court filings paint a tragic picture of overreliance on technology. They also strongly suggest that Autopilot failed at a basic function – automatic emergency braking – that engineers developed years ago. Many newer cars, including models much more affordable and less sophisticated than Teslas, can slow or stop themselves when a crash seems likely.

READ MORE

Last week the US National Highway Traffic Safety Administration said it had opened a formal investigation into Autopilot. The agency said it was aware of 11 crashes since 2018 involving Teslas that hit police, fire and other emergency vehicles with flashing lights parked on roads and highways. In one of them, a Tesla ploughed into a fire truck in December 2019 in Indiana, killing a passenger in the car and seriously injuring the driver.

Distracted driving can be deadly in any car. But safety experts say Autopilot may encourage distraction by lulling people into thinking that their cars are more capable than they are. And the system does not include safeguards to make sure drivers are paying attention to the road and can retake control if something goes wrong.

Tesla, the world's most valuable automaker, and its chief executive, Elon Musk, describe Autopilot as a way to make driving easier and safer. Despite its name, Autopilot does not make Teslas autonomous. The auto industry classifies it, and similar systems offered by General Motors and other companies, as Level 2 self-driving. Cars that can operate autonomously at all times would be Level 5, a distinction that no vehicle on sale today is close to achieving.

Tesla’s critics contend that Autopilot has several weaknesses, including the ability for drivers like McGee to use it on local roads. With the help of GPS and software, GM, Ford and other automakers restrict their systems to divided highways where there are no stop signs, traffic lights or pedestrians.

Tesla owners’ manuals warn customers not to use Autopilot on city streets. “Failure to follow these instructions could cause damage, serious injury or death,” the manual for 2019 models says.

“The technology exists to limit where Autopilot can operate, but Tesla allows drivers to use it on roads it shouldn’t operate on,” says Jason Levine, executive director of the Centre for Auto Safety, a Washington, DC, nonprofit group. “They made a corporate decision to do that, and it’s resulted in preventable tragedies. That should be enraging.”

Musk and Tesla's associate general counsel, Ryan McCarthy, have not responded to emails seeking comment. Regulators are looking into other potential Autopilot flaws. The system, which includes cameras, radar and software, sometimes fails to recognise other vehicles and stationary objects. In July a Tesla ran into an SUV parked at the site of an earlier crash on a highway near San Diego, in California. The driver had Autopilot on, fell asleep and, later, failed a sobriety test, police say. This year a California couple sued Tesla in connection with a 2019 crash that killed their 15-year-old son.

The National Highway Traffic Safety Administration is investigating more than two dozen crashes that occurred when Autopilot was in use. The agency says it is aware of at least 10 deaths in those accidents.

McGee, who is 44, is a managing partner at a small private-equity firm, New Water Capital. He bought his Model S, a performance model, in 2019. On the night of the crash he left Boca Raton and headed south over major highways. South of Miami, he got on US Route 1, took a narrow toll bridge from the mainland to Key Largo and continued on the two-lane Card Sound Road, which ends at County Road 905. McGee had Autopilot on, and the speed was set at 70km/h, according to data that police retrieved from the car.

About the same time, Benavides was on a date with Dillon Angulo. He was driving his mother's black Tahoe, and pulled off to the wide shoulder of County Road 905 near Card Sound Road. Angulo stopped about 13m from the edge of the junction, parked on a gravel strip and stepped out. Benavides got out of the passenger seat and walked around to the driver's side, according to the investigation.

Data from the Tesla shows the Model S accelerated from 70km/h to 95km/h a few seconds before crashing into the Tahoe. It is unclear if Autopilot or McGee raised the speed. Vehicle data and skid marks indicated McGee jammed on the brakes less than a second before impact. He told the police that he could not tell how close he was to the intersection when he started looking for his phone.

Benavides' estate has sued Tesla in the circuit court for Miami-Dade County, claiming the company's cars are "defective and unsafe". Todd Poses, a Miami lawyer representing the estate, says McGee is expected to give a deposition in that case. A separate lawsuit that the estate filed against McGee was settled, Poses says, but he will not disclose the terms.

In court, Tesla has filed a brief response denying the estate’s claims without elaborating. In similar cases, the company has said any blame rests solely with the drivers of its cars.

As in other crashes involving Autopilot, the system appears not to have done much to make sure McGee was paying attention to the road. Tesla recently activated an in-car camera in certain newer models to monitor drivers, but it cannot see in the dark. Tesla owners have posted videos on YouTube showing that the camera sometimes fails to notice when drivers look away from the road and that it can be fooled if they cover the lens. When the camera notices a Tesla driver looking away from the road, it sounds a warning chime but does not turn Autopilot off.

GM and Ford systems use infrared cameras to monitor drivers’ eyes. If drivers look away for more than two or three seconds, warnings remind them to look straight ahead. If drivers fail to comply, the GM and Ford systems will shut off and tell drivers to take control of the car.

Musk has said monitoring drivers while cars operate on systems like Autopilot may not be needed because machines are much safer than people. “Having a human intervene will decrease safety,” he said in an April 2019 interview with a Massachusetts Institute of Technology researcher.

Benavides emigrated from Cuba in 2016 and lived with her mother in Miami. She worked at a Walgreens pharmacy and a clothing store while attending community college. An older sister, Neima, who is 34, and executor of the estate, says Naibel had been working to improve her English in the hope of getting a college degree.

“She was always laughing and making people laugh,” Neima Benavides says. “Her favourite thing was to go to the beach. She would go almost every day and hang out with friends or just sit by herself and read.” Neima Benavides says she hopes the lawsuit will prod Tesla into making Autopilot safer. “Maybe something can change so other people don’t have to go through this.”

Benavides had just started dating Angulo when they went fishing on Key Largo. That afternoon she sent her sister a text message indicating she was having a good time. At 9pm Benavides called her mother from Angulo’s phone to say she was on the way home. She had lost her phone that day.

On the 911 call, McGee reported that a man was on the ground, unconscious and bleeding from the mouth. Several times McGee said, “Oh, my God,” and shouted “Help!” When an emergency operator asked if the man was the only injured person, McGee replied, “Yes, he’s the only passenger.”

Angulo was airlifted to a hospital. He later told investigators that he had no recollection of the accident or why they had stopped at the junction. An emergency medical technician spotted a woman’s sandal under the Tahoe and called on others to start searching the area for another victim. “Please tell me no,” McGee can be heard saying in the police video. “Please tell me no.” Benavides’ body was found about 20m away. – New York Times