Before a car crash in 2008 left her paralysed from the neck down, Nancy Smith enjoyed playing the piano. Years later, Smith started making music again, thanks to an implant that recorded and analysed her brain activity. When she imagined playing an on-screen keyboard, her brain–computer interface (BCI) translated her thoughts into keystrokes — and simple melodies, such as 'Twinkle, Twinkle, Little Star', rang out
But there was a twist. For Smith, it seemed as if the piano played itself. "It felt like the keys just automatically hit themselves without me thinking about it," she said at the time. "It just seemed like it knew the tune, and it just did it on its own."
Smith's BCI system, implanted as part of a clinical trial, trained on her brain signals as she imagined playing the keyboard. That learning enabled the system to detect her intention to play hundreds of milliseconds before she consciously attempted to do so, says trial leader Richard Andersen, a neuroscientist at the California Institute of Technology in Pasadena.
[...] Andersen's research also illustrates the potential of BCIs that access areas outside the motor cortex. "The surprise was that when we go into the posterior parietal, we can get signals that are mixed together from a large number of areas," says Andersen. "There's a wide variety of things that we can decode."
The ability of these devices to access aspects of a person's innermost life, including preconscious thought, raises the stakes on concerns about how to keep neural data private. It also poses ethical questions about how neurotechnologies might shape people's thoughts and actions — especially when paired with artificial intelligence.
Meanwhile, AI is enhancing the capabilities of wearable consumer products that record signals from outside the brain. Ethicists worry that, left unregulated, these devices could give technology companies access to new and more precise data about people's internal reactions to online and other content.
Ethicists and BCI developers are now asking how previously inaccessible information should be handled and used. "Whole-brain interfacing is going to be the future," says Tom Oxley, chief executive of Synchron, a BCI company in New York City. He predicts that the desire to treat psychiatric conditions and other brain disorders will lead to more brain regions being explored. Along the way, he says, AI will continue to improve decoding capabilities and change how these systems serve their users. "It leads you to the final question: how do we make that safe?"
[...] Although accurate user numbers are hard to gather, many thousands of enthusiasts are already using neurotech headsets. And ethicists say that a big tech company could suddenly catapult the devices to widespread use. Apple, for example, patented a design for EEG sensors for future use in its Airpods wireless earphones in 2023.
Yet unlike BCIs aimed at the clinic, which are governed by medical regulations and privacy protections, the consumer BCI space has little legal oversight, says David Lyreskog, an ethicist at the University of Oxford, UK. "There's a wild west when it comes to the regulatory standards," he says.
In 2018, Ienca and his colleagues found that most consumer BCIs don't use secure data-sharing channels or implement state-of-the-art privacy technologies2. "I believe that has not changed," Ienca says. What's more, a 2024 analysis3 of the data policies of 30 consumer neurotech companies by the Neurorights Foundation, a non-profit organization in New York City, showed that nearly all had complete control over the data users provided. That means most firms can use the information as they please, including selling it.
Responding to such concerns, the government of Chile and the legislators of four US states have passed laws that give direct recordings of any form of nerve activity protected status. But Ienca and Nita Farahany, an ethicist at Duke University in Durham, North Carolina, fear that such laws are insufficient because they focus on the raw data and not on the inferences that companies can make by combining neural information with parallel streams of digital data. Inferences about a person's mental health, say, or their political allegiances could still be sold to third parties and used to discriminate against or manipulate a person.
"The data economy, in my view, is already quite privacy-violating and cognitive- liberty-violating," Ienca says. Adding neural data, he says, "is like giving steroids to the existing data economy".
Several key international bodies, including the United Nations cultural organization UNESCO and the Organisation for Economic Co-operation and Development, have issued guidelines on these issues. Furthermore, in September, three US senators introduced an act that would require the Federal Trade Commission to review how data from neurotechnology should be protected.
Heading to the clinicWhile their development advances at pace, so far no implanted BCI has been approved for general clinical use. Synchron's device is closest to the clinic. This relatively simple BCI allows users to select on-screen options by imagining moving their foot. Because it is inserted into a blood vessel on the surface of the motor cortex, it doesn't require neurosurgery. It has proved safe, robust and effective in initial trials4, and Oxley says Synchron is discussing a pivotal trial with the US Food and Drug Administration that could lead to clinical approval.
Elon Musk's neurotech firm Neuralink in Fremont, California, has surgically implanted its more complex device in the motor cortices of at least 13 volunteers who are using it to play computer games, for example, and control robotic hands. Company representatives say that more than 10,000 people have joined waiting lists for its clinical trials.
At least five more BCI companies have tested their devices in humans for the first time over the past two years, making short-term recordings (on timescales ranging from minutes to weeks) in people undergoing neurosurgical procedures. Researchers in the field say the first approvals are likely to be for devices in the motor cortex that restore independence to people who have severe paralysis — including BCIs that enable speech through synthetic voice technology.
As for what's next, Farahany says that moving beyond the motor cortex is a widespread goal among BCI developers. "All of them hope to go back further in time in the brain," she says, "and to get to that subconscious precursor to thought."
(Score: 2) by liar on Wednesday December 03, @03:56AM (4 children)
If you haven't seen it, a movie you might find interesting: The Creator (2023) From Wikipedia:
Plot
In 2055, an artificial intelligence created by the U.S. government detonates a nuclear warhead over Los Angeles, California. In response, most of the Western world pledges to eradicate AI to prevent humanity's extinction. Their efforts are resisted by New Asia, a region comprising East, South and Southeast Asia, whose people continue to embrace AI. The U.S. military aims to assassinate "Nirmata",[a] the chief architect behind New Asia's AI advancements, using the USS NOMAD (North American Orbital Mobile Aerospace Defense), a space station capable of launching destructive attacks from orbit.
Noli nothis permittere te terere.
(Score: 1) by khallow on Wednesday December 03, @09:16PM (3 children)
(Score: 2) by liar on Wednesday December 03, @10:19PM
I have only watched part of this movie so far, but as far as I can tell, the AI is built into multiple bodies ( a synthetic race?)... individuals, viewed as Human in Asia, and machines in the west. And, was the detonation an accident...
Noli nothis permittere te terere.
(Score: 0) by Anonymous Coward on Wednesday December 03, @10:22PM (1 child)
Oh c'mon! Did you ever question how how the Enterprise goes back in time?
(Score: 1) by khallow on Wednesday December 10, @03:12AM