A self-driving Uber SUV struck and killed a pedestrian in Tempe, Arizona. It was in autonomous mode at the time of the collision, with a vehicle operator behind the wheel. Uber has suspended testing of its self-driving cars.
http://money.cnn.com/2018/03/19/technology/uber-autonomous-car-fatal-crash/index.html
https://www.nytimes.com/2018/03/19/technology/uber-driverless-fatality.html
https://finance.yahoo.com/news/self-driving-uber-kills-arizona-171055918.html
https://www.npr.org/sections/thetwo-way/2018/03/19/594950197/uber-suspends-self-driving-tests-after-pedestrian-is-killed-in-arizona
https://www.wsj.com/articles/uber-suspends-driverless-car-program-after-pedestrian-is-struck-and-killed-1521480386
https://www.theverge.com/2018/3/19/17139518/uber-self-driving-car-fatal-crash-tempe-arizona
https://techcrunch.com/2018/03/19/uber-self-driving-test-car-involved-in-accident-resulting-in-pedestrian-death/
I couldn't find any good analysis of the liability situation here.
(Score: 0) by Anonymous Coward on Tuesday March 20 2018, @12:07AM
The network needs more incidents like this as negative data points to train on. It literally becomes smarter after each additional accident, and the more deadly ones will be given ever larger weights as examples of what *not* to do under any circumstance. Then what happens if somehow all the weights are multiplied by -1 someday? Possibly a single cosmic ray could do it.