Stories
Slash Boxes
Comments

SoylentNews is people

posted by martyb on Thursday August 29 2019, @11:43PM   Printer-friendly
from the speech-to-text-to-speech dept.

Researchers Create AI that Hides your Emotions from Other AI

Humans can communicate a range of nonverbal emotions, from terrified shrieks to exasperated groans. Voice inflections and cues can communicate subtle feelings, from ecstasy to agony, arousal and disgust. Even when simply speaking, the human voice is stuffed with meaning, and a lot of potential value if you're a company collecting personal data.

Now, researchers at the Imperial College London have used AI to mask the emotional cues in users' voices when they're speaking to internet-connected voice assistants. The idea is to put a "layer" between the user and the cloud their data is uploaded to by automatically converting emotional speech into "normal" speech. They recently published their paper "Emotionless: Privacy-Preserving Speech Analysis for Voice Assistants" on the arXiv preprint server.

Our voices can reveal our confidence and stress levels, physical condition, age, gender, and personal traits. This isn't lost on smart speaker makers, and companies such as Amazon are always working to improve the emotion-detecting abilities of AI.

[...] Their method for masking emotion involves collecting speech, analyzing it, and extracting emotional features from the raw signal. Next, an AI program trains on this signal and replaces the emotional indicators in speech, flattening them. Finally, a voice synthesizer re-generates the normalized speech using the AIs outputs, which gets sent to the cloud. The researchers say that this method reduced emotional identification by 96 percent in an experiment, although speech recognition accuracy decreased, with a word error rate of 35 percent.

[...] Alexa speech engineers unveiled research on using adversarial networks to discern emotion in Amazon's home voice assistants. Emotion, they wrote, "can aid in health monitoring; it can make conversational-AI systems more engaging; and it can provide implicit customer feedback that could help voice agents like Alexa learn from their mistakes."

[...] Amazon's 2017 patent for emotional speech recognition uses illness as an example: "physical conditions such as sore throats and coughs may be determined based at least in part on a voice input from the user, and emotional conditions such as an excited emotional state or a sad emotional state may be determined based at least in part on voice input from a user," the patent says. "A cough or sniffle, or crying, may indicate that the user has a specific physical or emotional abnormality."

A virtual assistant such as Alexa would then use these cues—combined with your browsing and purchase history, according to the patent—to offer you hyper-specific advertisements for medications or other products.


Original Submission

 
This discussion has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 5, Funny) by Snotnose on Thursday August 29 2019, @11:54PM (1 child)

    by Snotnose (1623) on Thursday August 29 2019, @11:54PM (#887558)

    What's wrong? Nothing. Really, what's wrong? Nothing. You're giving me the silent treatment, what did I do? If you don't know I'm not going to tell you.

    Yeah, this is a huge advance in AI.

    --
    Why shouldn't we judge a book by it's cover? It's got the author, title, and a summary of what the book's about.
    Starting Score:    1  point
    Moderation   +3  
       Funny=2, Touché=1, Total=3
    Extra 'Funny' Modifier   0  
    Karma-Bonus Modifier   +1  

    Total Score:   5  
  • (Score: 4, Funny) by PartTimeZombie on Friday August 30 2019, @12:33AM

    by PartTimeZombie (4827) on Friday August 30 2019, @12:33AM (#887582)

    Shit. I may have dated your ex.