The US National Highway Traffic Safety Administration (NHTSA) launched a preliminary evaluation on Thursday into Tesla's Full Self-Driving (FSD) system's alleged failure to "detect and respond appropriately to reduced roadway visibility conditions."
It covers an estimated more than 2.4M Tesla vehicles that offer the optional FSD feature, including the 2016-2024 Model S and Model X, 2017-2024 Model 3, 2020-2023 Model Y, and 2023-2024 Cybertruck.
Recent crashes involving Tesla's FSD system sent a warning signal to auto regulators about serious safety concerns in the system. Now, they are finally taking action against a technology that should never have been allowed on the road in the first place, especially as data shows that this system is far more dangerous at driving than humans.
It's hard to fault a driver assistant system clearly labeled as "supervised" for accidents when those crashes are mostly — if not entirely — to blame either on human error or bad road conditions. In fact, issues with this technology apparently stem from the fact that it is safe because it can make drivers too complacent.