Seeing Ghosts: Self-driving Cars Aren't Immune From Hackers
Autonomous vehicles “feel” the road ahead with a variety of sensors, with data received sent through the vehicle’s brain to stimulate a response. Brake action, for example. It’s technology that’s far from perfected, yet self-driving trials continue on America’s streets, growing in number as companies chase that elusive driver-free buck.
In one tragic case, a tech company (that’s since had a come-to-Jesus moment regarding public safety) decided to dumb down its fleet’s responsiveness to cut down on “false positives” — perceived obstacles that would send the vehicle screeching to a stop, despite the obstacle only being a windblown plastic bag — with fatal implications. On the other side of the coin, Tesla drivers continue to plow into the backs and sides of large trucks that their Level 2 self-driving technology failed to register.
Because all things can be hacked, researchers now say there’s a way to trick autonomous vehicles into seeing what’s not there.
If manufacturing ghosts is your bag, read this piece in The Conversation. It details work performed the RobustNet Research Group at the University of Michigan, describing how an EV’s most sophisticated piece of tech, LiDAR, can be fooled into thinking it’s about to collide with a stationary object that doesn’t exist.
LiDAR sends out pulses of light, thousands per second, then measures how long it takes for those signals to bounce back to the sender, much like sonar or radar. This allows a vehicle to paint a picture of the world around it. Camera systems and ultrasonic sensors, which you’ll find on many new driver-assist-equipped models, complete the sensor suite.
From The Conversation:
The problem is these pulses can be spoofed. To fool the sensor, an attacker can shine his or her own light signal at the sensor. That’s all you need to get the sensor mixed up.
However, it’s more difficult to spoof the LiDAR sensor to “see” a “vehicle” that isn’t there. To succeed, the attacker needs to precisely time the signals shot at the victim LiDAR. This has to happen at the nanosecond level, since the signals travel at the speed of light. Small differences will stand out when the LiDAR is calculating the distance using the measured time-of-flight.
If an attacker successfully fools the LiDAR sensor, it then also has to trick the machine learning model. Work done at the OpenAI research lab shows that machine learning models are vulnerable to specially crafted signals or inputs – what are known as adversarial examples. For example, specially generated stickers on traffic signs can fool camera-based perception.
The research group claims that spoofed signals designed specifically to dupe this machine learning model are possible. “The LiDAR sensor will feed the hacker’s fake signals to the machine learning model, which will recognize them as an obstacle.”
Were this to happen, an autonomous vehicle would slam to a halt, with the potential for following vehicles to slam into it. On a fast-moving freeway, you can imagine the carnage resulting from a panic stop in the center lane.
The team tested two possible light pulse attack scenarios using a common autonomous drive system; one with a vehicle in motion, the other with a vehicle stopped at a red light. In the first setup, the vehicle braked, while the other remained immobile at the stoplight.
Needless fear mongering? Not in this case. With the advent of new technology, especially one that exists in a hazy regulatory environment, there will be people who seek to exploit the tech’s weaknesses. The team said it hopes “to trigger an alarm for teams building autonomous technologies.”
“Research into new types of security problems in the autonomous driving systems is just beginning, and we hope to uncover more possible problems before they can be exploited out on the road by bad actors,” the researchers wrote.
[Image: SAE, Ford]
Pwrwrench on Mar 07, 2020
Aside from anyone deliberately trying to mess with an 'autonomous' vehicle there's EMI to consider. With more and more underground cables EMI is pervasive. Sure you can sheild a system. IIRC in the early days of Electronic Fuel Injection, high powered 2-way radios could cause havoc and engine shut down. I installed a few anti-static kits with grounding cables from hoods and trunks to body. As late as the 80s some Bosch systems malfunctioned due to static discharges under very specific conditions. A fix was created with a filter in a cable leading to the ECM. As another post has it,"you need to test the crap out of it"
Latest Car ReviewsRead more
Latest Product ReviewsRead more
- Buickman what about EMFs from riding on a giant battery?is there a vax for that?
- ScarecrowRepair $1.2M at $1K per car is only 1200 cars, and if you spread that over 5 years, 240 cars per year, roughly one per work day and one more every weekend. Sell another every weekend for the interest. That seems plausible to me.
- FreedMike There are the guys charging $20000 over sticker for a F150 Lightning. They won’t go broke.
- CKNSLS Sierra SLT Some sales people I have met at Ford dealers don't appear smart enough to answer questions about EVs. They can't answer questions about payload, towing capacity and axle ratios when truck shopping.
- Dukeisduke The E23 here is wearing steelies - so is it because it arrived at the yard wearing winter shoes, or because someone swapped the factory alloys for steelies (either before or after it arrived)?Fun fact - the turbocharged 745i was called the 745i because it used the 3.2l six with a turbo, and at the time, F1 was using a 1.4 multiplication factor for turbocharged engines to arrive at a computed displacement. So, 3,210 times 1.4, divided by 1,000 equals 4.5 (rounded up from 4.494). The 745i was way cool, with its big 7" round low beam headlights paired with the 5-3/4" high beams.