Bloomberg writes that Mitsui & Co., best known among investors as Japan's top oil and iron-ore trader, and its partners have built a wearable suit—a backpack fitted with belts and leg supports—that enhances a user's ability to lift and move heavy objects. The idea is that when worn by farmers, or at nursing homes or construction sites, strength is enhanced.
The device, known as the Assist Suit AWN-03, was developed at ActiveLink, Panasonic Corp.'s robot-development unit. Weighing in at 6 kilograms (13.2 pounds), the suit allows the wearer to lift as much as 15 kilograms without stressing the lower back, according to Mitsui, which demonstrated the outfit to media at its Tokyo headquarters on Aug. 14.
Such a machine is tailor-made for Japan, where labor shortages and a shrinking and aging population are already causing construction delays, says the trading house. But the Assist Suit is just a first step. In two years, Mitsui and its partners aim to release the next iteration, complete with mechanical arms and legs. Further out, future versions could start to take on Aliens-like proportions. "What we have in mind is the Aliens power loader," says Tomoya Tsutsumi, an official at Mitsui's construction and industrial machinery division.
General contractor Kajima Corp. and Yamato Holdings Co., which offers door-to-door parcel delivery services, are among dozens of companies planning to try the technology, according to Tsutsumi. The target is to sell 1,000 units in the initial year after the Assist Suit's release.
"Young workers tend to want to work in a more comfortable environment so businesses are having trouble finding enough workers when labor conditions are harsh," Tsutsumi said.
Article includes a Youtube video demonstrating some pretty nifty stuff.
[More After the Break]
Scientists working at Korea University, Korea, and TU Berlin, Germany have developed a brain-computer control interface for a lower limb exoskeleton by decoding specific signals from within the user's brain.
Using an electroencephalogram (EEG) cap, the system allows users to move forwards, turn left and right, sit and stand simply by staring at one of five flickering light emitting diodes (LEDs)
...
Each of the five LEDs flickers at a different frequency, and when the user focusses their attention on a specific LED this frequency is reflected within the EEG readout. This signal is identified and used to control the exoskeleton.A key problem has been separating these precise brain signals from those associated with other brain activity, and the highly artificial signals generated by the exoskeleton.
"Exoskeletons create lots of electrical 'noise'" explains Klaus Muller, an author on the paper. "The EEG signal gets buried under all this noise -- but our system is able to separate not only the EEG signal, but the frequency of the flickering LED within this signal."
Brain-scanning with EEG caps has been making appearances at Makers Faire for the last couple of years. Has anyone experimented with these kinds of rigs? Are they the right interface for exoskeletons, or is there a better way?
(Score: 2) by wonkey_monkey on Thursday August 20 2015, @08:29AM
Using an electroencephalogram (EEG) cap, the system allows users to move forwards, turn left and right, sit and stand simply by staring at one of five flickering light emitting diodes (LEDs)
If the system relies on the user staring at flickering lights (which at the very least will cause eye fatigue), why not just cut out the middleman and use eye-tracking of some kind? Wouldn't a near-eye camera and eye tracking rig be able to determine which point was being looked at?
Or is there another ultimate aim, such as training the brain to reproduce the right signals without the stimulus?
systemd is Roko's Basilisk