Stories
Slash Boxes
Comments

SoylentNews is people

Submission Preview

Link to Story

Protecting smart machines from smart attacks

Accepted submission by KritonK mailto:kyrimis@alumni.princeton.edu at 2019-11-27 08:16:24 from the Science dept.
Security

Machines' ability to learn by processing data gleaned from sensors underlies automated vehicles, medical devices and a host of other emerging technologies. But that learning ability leaves systems vulnerable to hackers in unexpected ways, researchers at Princeton University have found [princeton.edu].

In a series of recent papers, a research team has explored how adversarial tactics applied to artificial intelligence (AI) could, for instance, trick a traffic-efficiency system into causing gridlock or manipulate a health-related AI application to reveal patients’ private medical history. As an example of one such attack, the team altered a driving robot’s perception of a road sign from a speed limit to a "Stop" sign, which could cause the vehicle to dangerously slam the brakes at highway speeds; in other examples, they altered Stop signs to be perceived as a variety of other traffic instructions.

(Summary taken from the article.)


Original Submission