Stories
Slash Boxes
Comments

SoylentNews is people

SoylentNews is powered by your submissions, so send in your scoop. Only 18 submissions in the queue.
posted by on Monday April 24 2017, @03:49PM   Printer-friendly
from the Is-it-live-or-is-it... dept.

According to a story appearing in The Economist, several companies are developing software that can, with only a relatively short sample of a person's speech, produce a "clone" of their saying nearly anything. CandyVoice is a phone app developed by a new Parisian company that only needs 160 or so French or English phrases from which it can extract sufficient information to read plain text in that person's voice. Carnegie Mellon University has a similar program called Festvox. The Chinese internet giant, Baidu, claims it has software which needs only about 50 sentences. And now, Vivotext, a voice-cloning firm headed by Gershon Silbert in Tel Aviv, looks to expand on that as it licenses its software to Hasbro.

More troubling, any voice—including that of a stranger—can be cloned if decent recordings are available on YouTube or elsewhere. Researchers at the University of Alabama, Birmingham, led by Nitesh Saxena, were able to use Festvox to clone voices based on only five minutes of speech retrieved online. When tested against voice-biometrics software like that used by many banks to block unauthorised access to accounts, more than 80% of the fake voices tricked the computer. Alan Black, one of Festvox's developers, reckons systems that rely on voice-ID software are now "deeply, fundamentally insecure".

And, lest people get smug about the inferiority of machines, humans have proved only a little harder to fool than software is. Dr Saxena and his colleagues asked volunteers if a voice sample belonged to a person whose real speech they had just listened to for about 90 seconds. The volunteers recognised cloned speech as such only half the time (ie, no better than chance). The upshot, according to George Papcun, an expert witness paid to detect faked recordings produced as evidence in court, is the emergence of a technology with "enormous potential value for disinformation". Dr Papcun, who previously worked as a speech-synthesis scientist at Los Alamos National Laboratory, a weapons establishment in New Mexico, ponders on things like the ability to clone an enemy leader's voice in wartime.

Efforts are now afoot to develop voice cloning detection software to identify whether a voice sample is genuine or machine-created.

[...] Nuance Communications, a maker of voice-activated software, is working on algorithms that detect tiny skips in frequency at the points where slices of speech are stuck together. Adobe, best known as the maker of Photoshop, an image-editing software suite, says that it may encode digital watermarks into speech fabricated by a voice-cloning feature called VoCo it is developing. Such wizardry may help computers flag up suspicious speech. Even so, it is easy to imagine the mayhem that might be created in a world which makes it easy to put authentic-sounding words into the mouths of adversaries—be they colleagues or heads of state.

Ever since I read "The Moon is a Harsh Mistress" by Robert Heinlein, I've been aware of the risks of assuming that the voice I hear is actually that of the person that I think it is. That was further confirmed when, as a teenager, callers to our home phone would be unable to distinguish between my Dad, my brother, and myself by voice alone.

How soon will it be that audio, pictures, and videos will be able to be manipulated to such a degree that it becomes impossible to detect an original from a manipulated version? How will that affect evidence introduced in court? Where will things go from here?

Previously:
Adobe Voco 'Photoshop-for-Voice' Causes Concern


Original Submission

 
This discussion has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 1, Interesting) by Anonymous Coward on Monday April 24 2017, @04:10PM (2 children)

    by Anonymous Coward on Monday April 24 2017, @04:10PM (#498908)

    So what we're seeing is that for a brief moment, there were video cameras everywhere. People on the ground when something happened became instant journalists, recording events as they were, much to the chagrin of corporations and governments.

    Then, one day, ubiquitous video recording, which hadn't been available decades before, suddenly became untrustworthy. And we're back to where we started.

    Starting Score:    0  points
    Moderation   +1  
       Interesting=1, Total=1
    Extra 'Interesting' Modifier   0  

    Total Score:   1  
  • (Score: 0) by Anonymous Coward on Monday April 24 2017, @04:20PM

    by Anonymous Coward on Monday April 24 2017, @04:20PM (#498916)

    The Butlers were on to something I think. Technology is fun, but unless we're very responsible then it looks like we're gonna have some serious "fun" as we go through a round of adjustment. Said adjustment will of course involve nasty methods used to gain control of humanity and then we'll have the inevitable revolution. Personally I would like to skip the horror part and have the world listen to the smart people saying these things are a bad idea. The premise of "us vs. them" is so very wrong, we need to manage our planet's resources as a species and get over this dumb nation state crap.

  • (Score: 0) by Anonymous Coward on Monday April 24 2017, @07:10PM

    by Anonymous Coward on Monday April 24 2017, @07:10PM (#499004)

    Then, one day, ubiquitous video recording, which hadn't been available decades before, suddenly became untrustworthy.

    Authenticating video is theoretically easy enough - have the camera do a digital signature of the original footage. No signature, no trust.
    Take it one step further and have the camera automatically put something like an HMAC of the video into a public blockchain in near real-time to reduce the window of time available for tampering. Sure, nothing is 100% trustworthy. But film never was either.