Tesla and Elon Musk are killing people for no reason and will never get close to producing a truly autonomous vehicle

First — I am aware of the words I have chosen in the title. That often people who use them are trying to hype or exaggerate them to gain attention. I chose to use them because I think minimizing what is occurring is enabling the problems to continue and even get worse.

Recently a driver was banned from driving in the UK for using Tesla’s Autopilot (AP) from the passenger seat. This demonstrates Tesla wants its drivers to let go of the steering wheel no matter what their legalese states. If Tesla actually wanted people to keep their hands on the wheel their system wouldn’t allow it to happen. Of course the reason for this is that they need for their customers to not only let go of the wheel but get in the very same accidents they will later scapegoat and abandon them for getting in to. Elon Musk needs to mislead, use and then abandon his paying customers so he can use them and their families as Guinea pigs to train the AI in the “Autopilot” system. The problem is this approach is futile and killing people for no reason.

Public shadow driving is the process where a driver alternates between hands on driving to train the AI and letting go of the wheel so it can be evaluated. Far too many other companies do this as well, Uber being another example. However Tesla, comma.ai and some others take it further by use their paying customers and their families, not paid drivers, as Guinea pigs.

The process is futile because it is impossible to drive and redrive, stumble and restumble on enough scenarios enough times to get anywhere close to L4. Toyota estimates the miles needed are one trillion. My very conservative calculation for cheap cars, sensors and drivers, no other costs, says that is $300B over ten years. To make matters worse though is how unnecessarily dangerous the process is. The process of the vehicle handing over control back to the driver cannot be made consistently safe by any monitoring and control system, especially in complex and dangerous scenarios. This because they cannot provide you the 6–45 seconds of time to gain the proper level of situation awareness to take the right action. Beyond this the AV makers will have to move from public shadow driving hyped benign or simple scenarios to those that are complex and dangerous. Eventually running thousands of accident scenarios thousands of times each. That will mean thousands of casualties.

Unfortunately Elon Musk doesn’t get any of this. He actually said the lives lost in the pursuit of autonomous vehicles are an ends to a means. That some lives need to be lost now to save many more later. The problem being this is not accurate regarding the use of public shadow driving. Musk has also said he only needs 6B miles to get to L4. I believe 4B have been driven so far. That has resulted in Tesla bragging it can now drive straight on a narrow bridge, killing Walter Huang by having his car follow pavement lines vs street marking directly in to a barrier and driving one of their cars into a stopped firetruck. (And that because these systems can’t see non-moving objects right in front of them well. Leaving them to ignore them and drive right through them). On the other side if this though Elon Musk would like to avoid accountability and liability. So Tesla built some legal CYA into their manuals and contracts telling the driver to not let go of the wheel. And if they do they will be to blame for the accidents they are involved in

Elon wants it both ways. He wants to convince his overly trusting customers to pay $5000 for the privilege to risk their lives as Guinea pigs to create a system that will supposedly save lives. Appealing to their desire to be part of a higher good. But he also wants to avoid blame and liability so he creates a legal protection system that abandons his customers when they get in to the accident Tesla needs them to get in. This so the AI can be properly tested. This is why you can see videos of him in his cars with his hands off the wheel. He sends mixed messages to fool people into being his beta testers. All while using a process that will never get close to creating the systems he has convinced his customers and the public will eventually save lives.

What is needed here is for the industry to switch from the use of public shadow driving to the use of aerospace/DoD level simulation. And for someone to step in and force Tesla and Elon Musk to do the right thing. Not unlike what NASA did in 2011 when it said SpaceX code failed safety testing. SpaceX was forced to regroup, hire aerospace engineers and do things right. NASA saved Elon from himself. NHTSA should put a stop to this practice or at least impose a moratorium while these issues are properly vetted. The problem there being that NHTSA is part of the problem. They stated in 2015 that handover can be made safe. The issue there being the study they used to support that position was poorly executed. They determined full control of the vehicle was attained, after being distracted, by simply grabbing the wheel and facing forward. They purposefully chose to ignore the quality of the actions taken after gaining “total control”, the proper situational awareness time needed to perform them properly and if that time could be provided by the monitoring and alert system.

As you can see we have a perfect storm here. The watchdogs are part of the problem. So what is the real solution? Short of more people dying, especially the first child or family needlessly? We need to people, especially those who make up the press and various levels of government, to stopped being overly impressed by the folks who make the apps of their devices or the games that they play, do their due diligence and police an industry that is clearly unable to police itself.

For more information on this please find my other articles

Impediments to Creating an Autonomous Vehicle

Autonomous Levels 4 and 5 will never be reached without Simulation vs Public Shadow Driving for AI

DoT, NHTSA and NTSB are Enabling Autonomous Vehicle Tragedies

The Dangers of Inferior Simulation for Autonomous Vehicles

Open Letter to Elon Musk — Please consider changing your approach in creating autonomous vehicles

Systems Engineer, Engineering/Program Management -- DoD/Aerospace/IT - Autonomous Systems Air & Ground, FAA Simulation, UAM, V2X, C4ISR, Cybersecurity