{"id":4834,"date":"2016-07-22T08:16:15","date_gmt":"2016-07-22T00:16:15","guid":{"rendered":"http:\/\/people.utm.my\/haslinasarkan\/?p=4834"},"modified":"2016-07-22T08:20:27","modified_gmt":"2016-07-22T00:20:27","slug":"fatal-tesla-self-driving-car-crash-reminds-us-that-robots-arent-perfect","status":"publish","type":"post","link":"https:\/\/people.utm.my\/haslinasarkan\/fatal-tesla-self-driving-car-crash-reminds-us-that-robots-arent-perfect\/","title":{"rendered":"Fatal Tesla Self-Driving Car Crash Reminds Us That Robots Aren&#8217;t Perfect"},"content":{"rendered":"<p><a href=\"https:\/\/people.utm.my\/haslinasarkan\/files\/2016\/07\/Tesla-Model-S-Elon-Musk-2011.jpg\"><img fetchpriority=\"high\" decoding=\"async\" class=\"aligncenter size-medium wp-image-4836\" src=\"https:\/\/people.utm.my\/haslinasarkan\/files\/2016\/07\/Tesla-Model-S-Elon-Musk-2011-300x169.jpg\" alt=\"Tesla-Model-S-Elon-Musk-2011\" width=\"300\" height=\"169\" \/><\/a><\/p>\n<p>This article is obtained from\u00a0http:\/\/spectrum.ieee.org\/cars-that-think\/transportation\/self-driving\/fatal-tesla-autopilot-crash-reminds-us-that-robots-arent-perfect?utm_campaign=TechAlert_07-21-16&amp;utm_medium=Email&amp;utm_source=TechAlert&amp;bt_alias=eyJ1c2VySWQiOiAiOWUzYzU2NzUtMDNhYS00YzBjLWIxMTItMWUxMjlkMjZhNjE0In0%3D<\/p>\n<h1 class=\"article-title\"><\/h1>\n<div class=\"metadata\">\n<div class=\"byline\">\n<p>By <a href=\"http:\/\/spectrum.ieee.org\/author\/ackerman-evan-\">Evan Ackerman<\/a><\/p>\n<div>Posted <label>1 Jul 2016 | 20:00 GMT<\/label><\/div>\n<\/div>\n<div class=\"addthis_sharing_toolbox\">\n<div id=\"atstbx\" class=\"at-share-tbx-element addthis_32x32_style addthis-smartlayers addthis-animated at4-show\">\n<div id=\"atstbx-share-label\" class=\"at4-visually-hidden\">AddThis Sharing Buttons<\/div>\n<\/div>\n<\/div>\n<\/div>\n<figure class=\"xlrg\"><img decoding=\"async\" src=\"http:\/\/spectrum.ieee.org\/img\/TeslaBloombergGettyImages501583098-1467397345920.jpg\" alt=\"null\" \/><figcaption class=\"hi-cap\">Photo: Bloomberg\/Getty Images<\/figcaption><\/figure>\n<div class=\"entry-content\">\n<p><a href=\"https:\/\/www.teslamotors.com\/blog\/tragic-loss\">On 7 May, a Tesla Model S was involved in a fatal accident in Florida<\/a>. At the time of the accident, the vehicle was driving itself, using its Autopilot system. The system didn\u2019t stop for a tractor-trailer attempting to turn across a divided highway, and the Tesla collided with the trailer. In a <a href=\"http:\/\/www.teslamotors.com\/blog\/tragic-loss\">statement<\/a>, Tesla\u00a0Motors said this is the \u201cfirst known fatality in just over 130 million miles\u00a0[210 million km]\u00a0where Autopilot was activated\u201d and suggested\u00a0that\u00a0this ratio\u00a0makes the Autopilot safer than an average vehicle. Early this year, Tesla CEO\u00a0Elon Musk\u00a0told reporters\u00a0that\u00a0the Autopilot system\u00a0in the Model S was \u201cprobably better than a person right now.\u201d<\/p>\n<p>The U.S.\u00a0National Highway Transportation Safety Administration (NHTSA) has opened a preliminary evaluation into the performance of Autopilot, to determine whether the system worked as it was expected to. For now, we\u2019ll take a closer look at what happened in Florida, how the accident may could have been prevented, and what this could mean\u00a0for <a href=\"http:\/\/spectrum.ieee.org\/transportation\/self-driving\">self-driving cars<\/a>.<\/p>\n<p>According to an official report\u00a0of the accident, the crash occurred on a divided highway with a median strip. A tractor-trailer truck in the westbound lane made a left turn onto a side road, making a perpendicular crossing in front of oncoming traffic in the eastbound lane. The driver of the truck didn\u2019t see the Tesla, nor did\u00a0the self-driving Tesla and its human occupant\u00a0notice\u00a0the trailer.\u00a0\u00a0The Tesla collided with the truck without the human or the Autopilot system ever applying the brakes.\u00a0The Tesla passed under the center of the trailer at windshield height and came to rest at the side of the road after hitting a fence and a pole.<\/p>\n<figure class=\"xlrg\"><img decoding=\"async\" src=\"http:\/\/spectrum.ieee.org\/image\/Mjc3MjgxNw.jpeg\" alt=\"img\" \/><figcaption class=\"hi-cap\">Image: Florida Highway Patrol<\/figcaption><\/figure>\n<p><a href=\"https:\/\/www.teslamotors.com\/blog\/tragic-loss\">Tesla\u2019s statement<\/a>\u00a0and <a href=\"https:\/\/twitter.com\/elonmusk\/status\/748625979271045121\">a tweet\u00a0from Elon Musk<\/a> provide some insight as to why the Autopilot system failed to stop for the trailer. The autopilot relies on cameras and radar to detect and avoid obstacles, and the cameras weren\u2019t able to effectively differentiate \u201cthe white side of the tractor trailer against a brightly lit sky.\u201d The radar should not have had any problems detecting the trailer, but <a href=\"https:\/\/twitter.com\/elonmusk\/status\/748625979271045121\">according to Musk<\/a>, \u201cradar tunes out what looks like an overhead road sign to avoid false braking events.\u201d<\/p>\n<p>We don\u2019t know all the details of how the Tesla S\u2019s radar works, but the fact that the radar could likely see underneath the trailer (between its front and rear wheels), coupled with a position that was perpendicular to the road (and mostly stationary) could easily lead to a situation where a computer could reasonably assume that it was looking at an overhead road sign. And most of the time, the computer would be correct.<\/p>\n<p>Tesla\u2019s statement also emphasized that, despite being called \u201cAutopilot,\u201d the system is assistive only and is not intended to assume complete control over the vehicle:<\/p>\n<blockquote><p><em>It is important to note that Tesla disables Autopilot by default and requires explicit acknowledgement that the system is new technology and still in a public beta phase before it can be enabled. When drivers activate Autopilot, the acknowledgment box explains, among other things, that Autopilot \u201cis an assist feature that requires you to keep your hands on the steering wheel at all times,&#8221; and that &#8220;you need to maintain control and responsibility for your vehicle\u201d while using it. Additionally, every time that Autopilot is engaged, the car reminds the driver to \u201cAlways keep your hands on the wheel. Be prepared to take over at any time.\u201d The system also makes frequent checks to ensure that the driver\u2019s hands remain on the wheel and provides visual and audible alerts if hands-on is not detected. It then gradually slows down the car until hands-on is detected again.<\/em><\/p><\/blockquote>\n<p>I don\u2019t believe that it\u2019s Tesla\u2019s intention to blame the driver in this situation, but the issue (and this has been an issue from the beginning) is that it\u2019s not entirely clear whether drivers are supposed to feel like they can rely on the Autopilot or not. I would guess Tesla\u2019s position on this would be that most of the time, yes, you can rely on it, but because Tesla has no idea when you <em>won\u2019t<\/em>be able to rely on it, you can\u2019t really rely on it. In other words, the Autopilot works very well under ideal conditions. You shouldn\u2019t use it when conditions are not ideal, but the problem with driving is that conditions can very occasionally turn from ideal to not ideal almost instantly, and the Autopilot can\u2019t predict when this will happen. Again, this is a fundamental issue with any car that has an \u201cassistive\u201d autopilot that asks for a human to remain in the loop, and is why companies like Google have made their explicit goal to remove human drivers from the loop entirely.<\/p>\n<p>The fact that this kind of accident has happened once means that there is a reasonable chance that it, or something very much\u00a0like it, could happen again. Tesla will need to address this, of course, although this particular situation also suggests ways in which vehicle safety in general could be enhanced.<\/p>\n<p>Here are a few ways in which this accident scenario could be addressed, both by Tesla itself, and by lawmakers\u00a0more generally:<\/p>\n<p><strong>A Tesla Software Fix:<\/strong> It\u2019s possible that Tesla\u2019s Autopilot software could be changed to more reliably differentiate between trailers and overhead road signs, if it turns out that that was\u00a0the issue. There may be a bug in the software, or it could be calibrated too heavily in favor of minimizing false braking events.<\/p>\n<p><strong>A Tesla Hardware Fix:<\/strong> There are some common lighting conditions in which cameras do very poorly (wet roads, reflective surfaces, or low sun angles), and the resolution of radar is relatively low. Almost every other self-driving car with a goal of sophisticated autonomy uses LIDAR to fill this kind of sensor gap, since LIDAR provides high resolution data out to a distance of several hundred meters with much higher resiliency to ambient lighting effects. Elon Musk doesn\u2019t believe that LIDAR is necessary for autonomous cars, however:<\/p>\n<blockquote><p><em>For full autonomy you\u2019d really want to have a more comprehensive sensor suite and computer systems that are fail proof.<\/em><\/p>\n<p><em>That said, I don\u2019t think you need LIDAR. I think you can do this all with passive optical and then with maybe one forward RADAR\u2026 if you are driving fast into rain or snow or dust. I think that completely solves it without the use of LIDAR. I\u2019m not a big fan of LIDAR, I don\u2019t think it makes sense in this context.<\/em><\/p><\/blockquote>\n<p>Musk may be right, but again, almost every other self-driving car uses LIDAR. Virtually every other company trying to make autonomy work has agreed that the kind of data that LIDAR can provide is necessary and unique, and it does seem like it might have prevented this particular accident, and could prevent accidents like it.<\/p>\n<p><strong>Vehicle-to-Vehicle Communication:<\/strong> The NHTSA is currently studying vehicle-to-vehicle (V2V) communication technology, which would allow vehicles \u201cto communicate important safety and mobility information to one another that can help save lives, prevent injuries, ease traffic congestion, and improve the environment.\u201d If (or hopefully when) vehicles are able to tell all other vehicles around them exactly where they are and where they\u2019re going, accidents like these will become much less frequent.<\/p>\n<p><strong>Side Guards on Trailers:<\/strong> The U.S. has relatively weak safety regulations regarding trailer impact safety systems. Trailers are required to have rear underride guards, but compared with\u00a0other countries (like Canada), the strength requirements are low. The U.S. does not require side underride guards. Europe does, but they\u2019re designed to protect pedestrians and bicyclists, not passenger vehicles. An IIHS analysis of fatal crashes involving passenger cars and trucks found that \u201c88 percent involving the side of the large truck\u2026 produced underride,\u201d where the vehicle passes under the truck. This bypasses almost all front-impact safety systems on the passenger vehicle, and as Tesla points out, \u201chad the Model S impacted the front or rear of the trailer, even at high speed, its advanced crash safety system would likely have prevented serious injury as it has in numerous other similar incidents.\u201d<\/p>\n<hr \/>\n<p>If Tesla comes up with a software fix, which seems like the most likely scenario, all other Tesla Autopilot systems will immediately benefit from improved safety. This is one of the major advantages of autonomous cars in general: accidents are inevitable, but unlike with humans, each kind of accident only has to happen once. Once a software fix has been deployed, no Tesla autopilot will make this same mistake ever again. Similar mistakes are possible, but as Tesla says, \u201cas more real-world miles accumulate and the software logic accounts for increasingly rare events, the probability of injury will keep decreasing.\u201d<\/p>\n<p>The near infinite variability of driving on real-world roads full of unpredictable humans means that it\u2019s unrealistic to think that the probability of injury while driving, even if your car is fully autonomous, will ever reach zero. But the point is that autonomous cars, and cars with assistive autonomy, are already much safer than cars driven by humans without the aid of technology. This is Tesla\u2019s first Autopilot-related fatality in 130 million miles [210 million km]: humans in the U.S. experience a driving fatality on average every 90 million miles [145 million km], and in the rest of the world, it\u2019s every 60 million miles [100 million km]. It\u2019s already far safer to have these systems working for us, and they\u2019re only going to get better at what they do.<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>This article is obtained from\u00a0http:\/\/spectrum.ieee.org\/cars-that-think\/transportation\/self-driving\/fatal-tesla-autopilot-crash-reminds-us-that-robots-arent-perfect?utm_campaign=TechAlert_07-21-16&amp;utm_medium=Email&amp;utm_source=TechAlert&amp;bt_alias=eyJ1c2VySWQiOiAiOWUzYzU2NzUtMDNhYS00YzBjLWIxMTItMWUxMjlkMjZhNjE0In0%3D By Evan Ackerman Posted 1 Jul 2016 | 20:00 GMT AddThis Sharing Buttons Photo: Bloomberg\/Getty Images On 7 May, a Tesla Model S was involved in a fatal accident in Florida. At the time of the accident, the vehicle was driving itself, using its Autopilot system. The system didn\u2019t stop [&hellip;]<\/p>\n","protected":false},"author":6477,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[30],"tags":[],"class_list":["post-4834","post","type-post","status-publish","format-standard","hentry","category-research-blog"],"_links":{"self":[{"href":"https:\/\/people.utm.my\/haslinasarkan\/wp-json\/wp\/v2\/posts\/4834","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/people.utm.my\/haslinasarkan\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/people.utm.my\/haslinasarkan\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/people.utm.my\/haslinasarkan\/wp-json\/wp\/v2\/users\/6477"}],"replies":[{"embeddable":true,"href":"https:\/\/people.utm.my\/haslinasarkan\/wp-json\/wp\/v2\/comments?post=4834"}],"version-history":[{"count":0,"href":"https:\/\/people.utm.my\/haslinasarkan\/wp-json\/wp\/v2\/posts\/4834\/revisions"}],"wp:attachment":[{"href":"https:\/\/people.utm.my\/haslinasarkan\/wp-json\/wp\/v2\/media?parent=4834"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/people.utm.my\/haslinasarkan\/wp-json\/wp\/v2\/categories?post=4834"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/people.utm.my\/haslinasarkan\/wp-json\/wp\/v2\/tags?post=4834"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}